this post was submitted on 26 Oct 2024
1153 points (97.4% liked)
Science Memes
10853 readers
3255 users here now
Welcome to c/science_memes @ Mander.xyz!
A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.
Rules
- Don't throw mud. Behave like an intellectual and remember the human.
- Keep it rooted (on topic).
- No spam.
- Infographics welcome, get schooled.
This is a science community. We use the Dawkins definition of meme.
Research Committee
Other Mander Communities
Science and Research
Biology and Life Sciences
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- !reptiles and [email protected]
Physical Sciences
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
Humanities and Social Sciences
Practical and Applied Sciences
- !exercise-and [email protected]
- [email protected]
- !self [email protected]
- [email protected]
- [email protected]
- [email protected]
Memes
Miscellaneous
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Any use of an LLM in understanding any subject or create any medium, be it papers or artwork, results in intellectual failure, as far as I'm concerned. Imagine if this were a doctor or engineer relying on hallucinated information, people could die.
there is no LLM involved in ryven's comment:
i hope noone is dependent on your reading comprehension mate, or i'll have some bad news
lmao fuck off, why put so much effort into defending the bullshit machines?
EDIT: I honestly didnt even read your comment, too much time wasted arguing with bota and techbros, but if you mean to try to explain the user meant copying the assignment instructions then said user should never have used the word "prompt" in this context to begin with.
Damn, if you're this stupid I understand why you're scared of the machines.
No one in this thread is talking about or "defending" LLMs but you.