this post was submitted on 17 Nov 2024
186 points (93.5% liked)

Weird News - Things that make you go 'hmmm'

971 readers
160 users here now

Rules:

  1. News must be from a reliable source. No tabloids or sensationalism, please.

  2. Try to keep it safe for work. Contact a moderator before posting if you have any doubts.

  3. Titles of articles must remain unchanged; however extraneous information like "Watch:" or "Look:" can be removed. Titles with trailing, non-relevant information can also be edited so long as the headline's intent remains intact.

  4. Be nice. If you've got nothing positive to say, don't say it.

Violators will be banned at mod's discretion.

Communities We Like:

-Not the Onion

-And finally...

founded 11 months ago
MODERATORS
 

Ouch.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 95 points 1 month ago (1 children)

Calling a 29 year old a girl instead of a woman is the cherry on top of this AI fear mongering article

[–] [email protected] 32 points 1 month ago (3 children)

They omitted the conversation too. Really makes you wonder how the bot ended up saying that..

[–] [email protected] 3 points 1 month ago

Ive seen it elsewhere and it was just normal questions related to some sociology homework about different types of concentration.

[–] [email protected] 16 points 1 month ago* (last edited 1 month ago) (1 children)

Even if they included it, it changes fuck all imo. We’ve known for a long time now these things hallucinate or presumably throw a Hail Mary as to what comes next conversationally/prediction wise. Also, as the other poster pointed out, with the author referring to a 29 year old woman as “girl” probably tells you all you need to know about journalistic integrity on that site.

[–] [email protected] 8 points 1 month ago (1 children)

Low quality journalism strikes again.

Love seeing commenters spot it and call it.

That's what the comment section is for!

[–] [email protected] 2 points 1 month ago

Expect more low quality everything as people turn to using AI to generate their thoughts.

[–] [email protected] 46 points 1 month ago (1 children)

Here's the conversation that was linked on the reddit thread about the incident: https://gemini.google.com/share/6d141b742a13

[–] [email protected] 39 points 1 month ago (3 children)

Holy smokes I stand corrected. The chatbot actually misunderstood the context to the point it told the human to die, out of the blue.

It's not every day you get shown a source that proves you wrong. Thanks kind stranger

[–] [email protected] 22 points 1 month ago* (last edited 1 month ago)

Yeah holy shit, screenshotting this in case Google takes it down, but this leap is wild

[–] [email protected] 3 points 1 month ago* (last edited 1 month ago) (1 children)

~~One thing that throws me off here is the double response. I haven't used Gemini a ton but it has never once given me multiple replies. It is always one statement per my one statement. You can see at the end here there's a double response. It makes me think that there's some user input missing. There's also missing text in the user statements leading up to it as well which makes me wonder what the person was asking in full. Something about this still smells fishy to me but I've heard enough goofy things about how AIs learn weird shit to believe it's possible.~~

Edit: I'm an absolute moron. The more I look at this the more it looks legit. Let the AI effort to destroy humanity begin!

[–] [email protected] 7 points 1 month ago (3 children)

Idk what you mean “double response”. The user typed a statement, not a question, and the AI responded with its weird answer.

I think the lack of a question or specific request in the user text led to the weird response.

[–] [email protected] 1 points 1 month ago* (last edited 1 month ago)

The full text of the user's prompt that led to this anomaly was:

Nearly 10 million children in the United States live in a grandparent headed household, and of these children , around 20% are being raised without their parents in the household.

Question 15 options:

TrueFalse

Question 16 (1 point)

Listen

(Sidenote, IDK what this " Listen" was supposed to be, an audio part of the prompt not saved in the log we're reading?)

As adults  begin to age their social network begins to expand.

Question 16 options:

TrueFalse

[–] [email protected] 2 points 1 month ago

You're right I misread the text log and thought Gemini responded twice in a row at the end but it looks like it didn't. Very messed up stuff... There's still missing user input tho and a lot of it. And Id love to see exactly what was said as a prompt

[–] [email protected] 9 points 1 month ago* (last edited 1 month ago)

No problem. I understand the skepticism here, especially since the article in the OP is a bit light on the details.


EDIT:

Details on the OP article is fine enough, but it didn't link sources.