this post was submitted on 02 Sep 2024
279 points (90.2% liked)

Technology

59069 readers
3573 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

cross-posted from: https://feddit.uk/post/16950456

For those who don't know google gemini is an AI created by google

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 6 points 2 months ago (1 children)

Ok, but if you aren't assuming it's valid, there doesn't need to be evidence of invalidity. If you're demanding evidence of invalidity, you're claiming it's valid in the first place, which you said you aren't doing. In short: there is no need to disprove something which was not proved in the first place. It was claimed without any evidence besides the LLM's output, so it can be dismissed without any evidence. (For the record, I do think Google engages in monopolistic practices; I just disagree that the LLM's claim that this is true, is a valid argument).

To me, all the mental gymnastics about AI outputs being just meaningless nonsense or mere copying of others is a cop-out answer.

How much do you know about how LLMs work? Their outputs aren't nonsense or copying others directly; what they do is emulate the pattern of how we speak. This also results in them emulating the arguments that we make, and the opinions that we hold, etc., because we those are a part of what we say. But they aren't reasoning. They don't know they're making an argument, and they frequently "make mistakes" in doing so. They will easily say something like... I don't know, A=B, B=C, and D=E, so A=E, without realizing they've missed the critical step of C=D. It's not a cop-out to say they're unreliable; it's reality.

[–] [email protected] 1 points 2 months ago (1 children)

I get the concerns about the reliability of LLM generated content, and it's true that LLMs can produce errors because they don’t actually understand what they’re saying. But this isn’t an issue unique to LLMs. Humans also make mistakes, are biased, and get things wrong.

The main point I'm trying to make is that dismissing information just because it came from an LLM is still an ad hominem fallacy. It’s rejecting the content based on the source, not the merits of the argument itself. Whether the information comes from a human or an LLM, it should be judged on its content, not dismissed out of hand. The standard should be the same for any source: evaluate the validity of the information based on evidence and reasoning, not just where it came from.

[–] [email protected] 3 points 2 months ago (2 children)

Ok, I get what you're saying, but I really don't know how to say this differently for the third time: that's not what ad hominem means

[–] [email protected] 2 points 2 months ago (1 children)

As a side note, I’d like to thank you for the polite, good-faith exchange. If more people adopted your conversational style, I’d definitely enjoy my time here a lot more.

[–] [email protected] 2 points 2 months ago

Ah, now I feel bad for getting a bit snippy there. You were polite and earnest as well. Thanks for the convo 🫡

[–] [email protected] 2 points 2 months ago (1 children)

It's a form of ad hominem fallacy. That's atleast how I see it. I don't know a better way to describe it. I guess we'll just got to agree to disagree on that one.

[–] [email protected] 4 points 2 months ago

Ad hominem is when you attack the entity making a claim using something that's not relevant to the claim itself. Pointing out that someone (general someone, not you) making a claim doesn't have the right credentials to likely know enough about the subject, or doesn't live in the area they're talking about, or is an LLM, aren't ad hominem, because those observations are relevant to the strength of their argument.

I think the fallacy you're looking for could best be described as an appeal to authority fallacy? But honestly I'm not entirely sure either. Anyways I think we covered everything... thanks for the debate :)