this post was submitted on 25 Apr 2025
379 points (95.9% liked)

Technology

70266 readers
3651 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

Archived link: https://archive.ph/Vjl1M

Here’s a nice little distraction from your workday: Head to Google, type in any made-up phrase, add the word “meaning,” and search. Behold! Google’s AI Overviews will not only confirm that your gibberish is a real saying, it will also tell you what it means and how it was derived.

This is genuinely fun, and you can find lots of examples on social media. In the world of AI Overviews, “a loose dog won't surf” is “a playful way of saying that something is not likely to happen or that something is not going to work out.” The invented phrase “wired is as wired does” is an idiom that means “someone's behavior or characteristics are a direct result of their inherent nature or ‘wiring,’ much like a computer's function is determined by its physical connections.”

It all sounds perfectly plausible, delivered with unwavering confidence. Google even provides reference links in some cases, giving the response an added sheen of authority. It’s also wrong, at least in the sense that the overview creates the impression that these are common phrases and not a bunch of random words thrown together. And while it’s silly that AI Overviews thinks “never throw a poodle at a pig” is a proverb with a biblical derivation, it’s also a tidy encapsulation of where generative AI still falls short.

top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 4 points 3 weeks ago

This also works with asking it "why?" About random facts you make up.

[–] [email protected] 0 points 4 weeks ago* (last edited 4 weeks ago)

I mean are you asking it if there is a history of an idiom existing or just what the idiom could mean?

[–] [email protected] 5 points 4 weeks ago (2 children)

"three horses, one carrot, a slice at a time or live in purple sauce"

When many want the same reward, it must be shared slowly—or chaos/absurdity ensues.

[–] [email protected] 2 points 3 weeks ago

Words to live by. Sounds like something from my dream diary.

[–] [email protected] 4 points 4 weeks ago (1 children)

"AI cannot peel the cat down to the dog's bark"

AI can't reduce complex, chaotic, or nuanced things (like a cat) into something simple or binary (like a dog’s bark).

[–] [email protected] 3 points 4 weeks ago (1 children)

A binary dog will never pee you virtual bananas.

A purely logical or programmed entity (like AI) will never give you true absurdity, spontaneity, or joyfully irrational experiences (the “virtual bananas”).

[–] [email protected] 2 points 4 weeks ago (1 children)

"The one who lives thinking of fruit in Uranus will never eat a banana."

If you're too obsessed with absurd or far-off ideas, you'll miss what's right in front of you.

[–] [email protected] 3 points 4 weeks ago

"A vaccine with no green paint will never straighten its leaves."

solution that lacks vitality or authenticity can't truly heal or nurture life. Or, you can't fix organic problems with lifeless tools.

[–] [email protected] 6 points 4 weeks ago (1 children)
[–] [email protected] 10 points 4 weeks ago (2 children)
[–] [email protected] 1 points 3 weeks ago

Straight up sizzle out of existence.

[–] [email protected] 2 points 4 weeks ago

What? Why is that?

[–] [email protected] 7 points 4 weeks ago

The saying "you can't cross over a duck's river" is a play on words, suggesting that it's difficult to cross a river that is already filled with ducks. It's not a literal statement about rivers and ducks, but rather an idiom or idiom-like phrase used to express the idea that something is difficult or impossible to achieve due to the presence of obstacles or challenges.

I used the word “origin” instead of “meaning”, which didn’t seem to work.

[–] [email protected] 1 points 4 weeks ago (1 children)

Honestly, I’m kind of impressed it’s able to analyze seemingly random phrases like that. It means its thinking and not just regurgitating facts. Because someday, such a phrase could exist in the future and AI wouldn’t need to wait for it to become mainstream.

[–] [email protected] 0 points 4 weeks ago

It's not thinking. It's just spicy autocomplete; having ingested most of the web, it "knows" that what follows a question about the meaning of a phrase is usually the definition and etymology of that phrase; there aren't many examples online of anyone asking for the definition of a phrase and being told "that doesn't exist, it's not a real thing." So it does some frequency analysis (actually it's probably more correct to say that it is frequency analysis) and decides what the most likely words to come after your question are, based on everything it's been trained on.

But it doesn't actually know or think anything. It just keeps giving you the next expected word until it meets its parameters.

[–] [email protected] 15 points 4 weeks ago (1 children)

The premium version has gotten a LOT better in recent months. I just fed it,

"you can't soothe a tree" meaning

And it replied with: This phrase is not a standard idiom with a widely recognized meaning. Its interpretation depends on the context in which it is used. Literally, it means you cannot comfort or calm a tree in the way you would a sentient being, as trees do not possess emotions or consciousness that can be "soothed". Metaphorically, it likely means:

  • Some entities (problems, situations, perhaps even certain people) are unresponsive to attempts at emotional comfort or calming.
  • Trying to apply emotional solutions or appeals to things that are inherently non-emotional or unchangeable is futile.
  • It could describe someone who is extremely stoic, impassive, or emotionally unavailable. The specific meaning depends heavily on the situation where the phrase was encountered.
[–] [email protected] 15 points 4 weeks ago (2 children)

I always wonder how many of these are actually just patches behind the scene to fix viral trends. Or even more devious, they use the viral trends to patch a specific failure point to make it feel like progress is being made.

[–] [email protected] 1 points 3 weeks ago

Yeah, there's some low-paid guy in Chennai who has a completely absurd job speculating on the meaning of nonsense idioms.

[–] [email protected] 5 points 4 weeks ago

Absolutely. It really blurs the line between fancy autocorrect, mechanical turk & apocolyptic AGI. We can only guess we are somewhere between 1 & 2.

[–] [email protected] 12 points 4 weeks ago (5 children)

I am not saying other generative AI lack flaws, but Google's AI Overview is the most problematic generative AI implementation I have ever seen. It offends me that a company I used to trust continues to force this lie generator as a top result for the #1 search engine. And to what end? Just to have a misinformed populace over literally every subject!

OpenAI has issues as well, but ChatGPT is a much, much better search engine with far fewer hallucinations per answer. Releasing AI Overview while the competition is leagues ahead on the same front is asinine!

[–] [email protected] 0 points 4 weeks ago* (last edited 4 weeks ago) (1 children)

Its a language model not a dictionary. By putting the term "definition" before the sentence you imply that the following sentence has a definintion, hence it vectors down to the most likely meaning.

[–] [email protected] 0 points 4 weeks ago

Buddy, I never said the word definition

[–] [email protected] 2 points 4 weeks ago (1 children)

They famously taught it on Reddit. So it's not surprising that it just comes up with nonsense.

You would have thought that they would use a more stable data set. Although it does mean it's very good at explaining the plots of movies badly.

[–] [email protected] 1 points 3 weeks ago* (last edited 3 weeks ago)

it’s very good at explaining the plots of movies badly

Generally by recapitulating plot points with no understanding of their relative importance.

[–] [email protected] -3 points 4 weeks ago (1 children)

Saying you used to trust google is really a core part of the problem. Google isn't a person. Just like AI isn't a person. They both do what they are tasked with. Companies prioritize profit. AI prioritizes giving an answer, not necessarily a correct one. That is how it was designed.

[–] [email protected] 5 points 4 weeks ago* (last edited 4 weeks ago) (1 children)

Impressive how we seem to agree with each other yet you still found a way to insult my way of putting it

[–] [email protected] 1 points 3 weeks ago

Wasn't really going for insult. Companies have spent a lot of time and money to indoctrinate us into thinking of companies like people. Noone is immune. That is the core issue. I probably could have been more explicite on that.

[–] [email protected] 4 points 4 weeks ago (1 children)

And to what end? Just to have a misinformed populace over literally every subject!

This is a feature; not a bug. We're entering a new dark age, and generative AI is the tool that will usher it in. The only "problem" generative AI is efficiently solving is a populace with too much access to direct and accurate information. We're watching as perfectly functional tools and services are being rapidly replaced by a something with inherent issues with reliability, ethics and accountability.

[–] [email protected] 1 points 4 weeks ago

In the case with Google AI overview, I 1000% agree. I am not against all AI tools, but that company has clearly chosen evil.

[–] [email protected] 4 points 4 weeks ago (1 children)

I've resorted to appending every Google search with "-ai" because I don't want to see their bullshit summaries. Outsourcing our thinking is lazy and dangerous, especially when the technology is so flawed.

[–] [email protected] 3 points 4 weeks ago

I like that trick, noted! I mostly use DuckDuckGo as a browser and search engine now. If it fails I use ChatGPT

load more comments
view more: next ›