this post was submitted on 27 Jun 2024
389 points (96.6% liked)
Comic Strips
12520 readers
2767 users here now
Comic Strips is a community for those who love comic stories.
The rules are simple:
- The post can be a single image, an image gallery, or a link to a specific comic hosted on another site (the author's website, for instance).
- The comic must be a complete story.
- If it is an external link, it must be to a specific story, not to the root of the site.
- You may post comics from others or your own.
- If you are posting a comic of your own, a maximum of one per week is allowed (I know, your comics are great, but this rule helps avoid spam).
- The comic can be in any language, but if it's not in English, OP must include an English translation in the post's 'body' field (note: you don't need to select a specific language when posting a comic).
- Politeness.
- Adult content is not allowed. This community aims to be fun for people of all ages.
Web of links
- [email protected]: "I use Arch btw"
- [email protected]: memes (you don't say!)
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Yeah.
I'm willing to believe that we can have solid AI software authoring, but I am skeptical that it's gonna be via the raw LLM stuff being done with images and audio and such, where what matters is stuff that looks like other stuff.
Maybe you could use LLMs as a component of a larger system that does effective coding. But I'm skeptical that this alone can be a great solution.
Maybe in very limited situations where the system can reliably validate the code correctness itself. Like, say you want to write a quine. That doesn't take input, and the output is trivial to validate.
But for most software, I'd say that it's not easy for a computer to validate that code is correct.
And in some cases, trying to validate code has got to be worse than doing it yourself. Like, think of multithreaded code absent some sort of elaborate type system that permits fully specifying the constraints imposed by the parallelism requirements, and where such constraints are written and available to you. C or C++ doesn't have such a type system.
Or writing security-sensitive code. Same thing -- absent some kind of type system that permits fully-specifying the requirements of the problem, you can't automatically validate it, and trying to review code to understand whether it's secure...ugh.
I can maybe see some kind of "grammar check", having an LLM looking for code that you wrote that has a portion that is unusual compared to existing code that it's seen.
Programming is basically translation from a list of (precise) requirements to code in a programming language. And LLMs can do translation of human language pretty well. But I expect that a major problem for LLM-driven programming is that there's no training corpus for the requirements, the "source language" for translation.