this post was submitted on 10 Jun 2025
2 points (100.0% liked)
Programming
21165 readers
118 users here now
Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!
Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.
Hope you enjoy the instance!
Rules
Rules
- Follow the programming.dev instance rules
- Keep content related to programming in some way
- If you're posting long videos try to add in some form of tldr for those who don't want to watch videos
Wormhole
Follow the wormhole through a path of communities [email protected]
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I don't think y'all are disagreeing but maybe this sentence is somewhat confusing:
Maybe the "doesnt" shouldn't be there.
No it is here because that's what they claim.
Nobody yet know how it work, we don't know how LLMs process information.
Anyone who claim it really think, or it isn't thinking, is believing, this is not something the current ML field know.
Well, the neural network is given a prefix (series of tokens) and a token, and it spits out how likely is it that the token follows the prefix. Text is generated by calculating this probability for all known tokens, then picking one random, weighted based on the calculated probabilities.
And the brain is made out of neurons that sends electric signals between them and operate muscles.
That doesnt explain how the brain think.
It allows us to conclude that an LLM doesn't "think" about what it is saying. Based on the mechanics, the LLM doesn't even know it's a participant in the conversation.