They have finally done it: they've figured out a way to make LLMs even heavier computation-wise
technology
On the road to fully automated luxury gay space communism.
Spreading Linux propaganda since 2020
- Ways to run Microsoft/Adobe and more on Linux
- The Ultimate FOSS Guide For Android
- Great libre software on Windows
- Hey you, the lib still using Chrome. Read this post!
Rules:
- 1. Obviously abide by the sitewide code of conduct. Bigotry will be met with an immediate ban
- 2. This community is about technology. Offtopic is permitted as long as it is kept in the comment sections
- 3. Although this is not /c/libre, FOSS related posting is tolerated, and even welcome in the case of effort posts
- 4. We believe technology should be liberating. As such, avoid promoting proprietary and/or bourgeois technology
- 5. Explanatory posts to correct the potential mistakes a comrade made in a post of their own are allowed, as long as they remain respectful
- 6. No crypto (Bitcoin, NFT, etc.) speculation, unless it is purely informative and not too cringe
- 7. Absolutely no tech bro shit. If you have a good opinion of Silicon Valley billionaires please manifest yourself so we can ban you.
I mean LLMs have gotten orders of magnitude more efficient in just the past year, but also using these types approaches might make it possible to use much smaller models, and iterate on the result.
Damnit, I saw MCTS and thought it'd be something neat, but then it's LLMs because of course every piece of tech news is LLMs.
Making the LLM use an LLM to figure out what to say almost feels like a pretty good tech news shitpost
overthinking how someone might react to what it could say in multiple branches with growing resource usage
This is it, if they can get it to reliably decide to just not say anything at all in the end then I have been fully replaced
Lol same, what I was thinking while reading the features was "wow, they found a way to simulate masking!"
Interesting. I’m not sophisticated enough to judge this particular implementation but the concept of generating entire conversation trees to judge the quality of an output intrigues me for sure and I’d be interested in reading more about it and any research around it. Got any good links for further reading?
I think that's an interesting approach as well. There are a bunch of research papers on using MCTS with LLMs, a few examples here:
https://arxiv.org/abs/2503.19309
https://arxiv.org/abs/2505.23229
https://arxiv.org/abs/2504.02426