this post was submitted on 28 Feb 2024
474 points (97.4% liked)

Technology

59137 readers
2342 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
(page 2) 37 comments
sorted by: hot top controversial new old
[–] [email protected] 0 points 8 months ago* (last edited 8 months ago)

The AIrmageddon..

[–] [email protected] 32 points 8 months ago (1 children)

Anecdotally speaking, I've been suspecting this was happening already with code related AI as I've been noticing a pretty steep decline in code quality of the code suggestions various AI tools have been providing.

Some of these tools, like GitHub's AI product, are trained on their own code repositories. As more and more developers use AI to help generate code and especially as more novice level developers rely on AI to help learn new technologies, more of that AI generated code is getting added to the repos (in theory) that are used to train the AI. Not that all AI code is garbage, but there's enough that is garbage in my experience, that I suspect it's going to be a garbage in, garbage out affair sans human correction/oversight. Currently, as far as I can tell, these tools aren't really using much in the way of good metrics to rate whether the code they are training on is quality or not, nor whether it actually even works or not.

More and more often I'm getting ungrounded output (the new term for hallucinations) when it comes to code, rather than the actual helpful and relevant stuff that had me so excited when I first started using these products. And I worry that it's going to get worse. I hope not, of course, but it is a little concerning when the AI tools are more consistently providing useless / broken suggestions.

load more comments (1 replies)
[–] [email protected] 72 points 8 months ago (3 children)

My team has been calling models that use ai generated data "Habsberg models"

[–] [email protected] 20 points 8 months ago (3 children)

I feel there is a good joke here, but I miss the knowledge to understand it. Care to enlighten me?

load more comments (3 replies)
[–] [email protected] 12 points 8 months ago

Lmao that's a perfect name for it

load more comments (1 replies)
[–] [email protected] 32 points 8 months ago (1 children)

Back when i was though concept art as a subject at college my teacher had a name for this.

“Incest” cause every generation of art that references other art becomes more and more strange looking and detached from reality.

If you thought Skyrim weapons look ridiculous you should have seen my classmates Skyrim inspired weapons.

load more comments (1 replies)
[–] [email protected] 3 points 8 months ago

now that the low hanging fruit of internet scraping is exhausted, we're gonna have to start purpose-building datasets. this will be expensive and might be the new bottleneck on AI progress.

[–] [email protected] 15 points 8 months ago
[–] [email protected] 25 points 8 months ago (2 children)

That’s it! I’m starting my own internet, with blackjack. And hookers.

[–] [email protected] 10 points 8 months ago

that's why the fediverse exists

[–] [email protected] 20 points 8 months ago (1 children)

I fail to see how that differs from the current Internet

load more comments (1 replies)
[–] [email protected] 10 points 8 months ago

This article is from June 12, 2023. That's practically stone-aged as far as AI technology has been progressing.

The paper it's based on used a very simplistic approach, training AIs purely on the outputs of its previous "generation." Turns out that's not a realistic real-world scenario, though. In reality AIs can be trained on a mixture of human-generated and AI-generated content and it can actually turn out better than training on human-generated content alone. AI-generated content can be curated and custom-made to be better suited to training, and the human-generated stuff adds back in the edge cases that might disappear when doing repeated training generations.

[–] [email protected] 20 points 8 months ago (2 children)

I mean it makes sense. Machine learning is fantastic at noticing patterns, and the stuff they generate most definitely do have patterns. We might not notice them, but the models will pick up on them and eventually, if you keep training them on that data, they'll skew more and more in that direction.

They've been marketing things like there isn't a limit to how good these things can get, but there is. Nothing is infinite.

[–] [email protected] 16 points 8 months ago (2 children)

I've tried to make this point several times to folks in the industry. I work in AI, and yet every time I approach some people with "you know it ultimately just repeats patterns", I'm met with scoffs and those people telling me I'm just not "seeing the big picture".

But I am, and the truth is that there are limits. This tech is not the digital singularity the marketers and business goons want everyone to think it is.

[–] [email protected] 9 points 8 months ago (2 children)

It repeats things that sort of sound intelligent to try and convince everyone that actual intelligent thought is taking place? It really is just like humans!

[–] [email protected] 2 points 8 months ago (1 children)

They don't really parrot unless they're overfitted.

It's more that they have been trained to produce a certain kind of result. One method you can train them on is by basically assigning a score on how good the output is. Doing this manually takes a lot of time (Google has been doing this for years via captcha), or you could train other models to score text for you.

The obvious problem with the latter solution is that then you need to ensure that that model is scoring roughly in line with how humans would score it; the technical term for this is alignment. There's a pretty funny story about that with GPT-2, presented in a really cute animation format by Robert Miles.

[–] [email protected] 1 points 8 months ago

Here is an alternative Piped link(s):

presented in a really cute animation format by Robert Miles.

Piped is a privacy-respecting open-source alternative frontend to YouTube.

I'm open-source; check me out at GitHub.

load more comments (1 replies)
[–] [email protected] 1 points 8 months ago

It can be both though.

load more comments (1 replies)
[–] [email protected] 36 points 8 months ago (1 children)
[–] [email protected] 3 points 8 months ago (3 children)

Can you explain further more ? Sorry I may understand but not sure

load more comments (3 replies)
[–] [email protected] 67 points 8 months ago (9 children)

Maybe we need to label AI-generated content to, you know, avoid confusion.

[–] [email protected] 4 points 8 months ago (1 children)

Ah, you're suggesting using RFC 3514. Good thinking.

load more comments (1 replies)
[–] [email protected] 8 points 8 months ago (1 children)

Sounds great, how do we enforce it?

[–] [email protected] 14 points 8 months ago (1 children)

If the AIs want to avoid digital incest they'll enforce it for themselves.

load more comments (1 replies)
[–] [email protected] 6 points 8 months ago (1 children)

Sounds like something an advanced language learning model would say....

[–] [email protected] 4 points 8 months ago

It's important to understand that a language modelling AI can only produce responses based on its inputs.

[–] [email protected] 111 points 8 months ago (2 children)

Sorry, best we can do is a race to the bottom fueled by greed and incompetence.

[–] [email protected] 6 points 8 months ago

I'm sure we can compromise on a mandatory database of registered AI-generated content that only the corporations can read from but everyone using AI-generated content is required by law to write to, with hefty fines (but only for regular people).

[–] [email protected] 44 points 8 months ago (1 children)

That will be a refeshing change.

[–] [email protected] 4 points 8 months ago (1 children)

That's what has been happening, and is likely what will continue to happen. Not much change there really...

[–] [email protected] 28 points 8 months ago (1 children)
[–] [email protected] -5 points 8 months ago (1 children)

Wouldn't it have to be funny to be a joke?

[–] [email protected] 3 points 8 months ago (2 children)

I thought I was being funny. Sorry if it didn't tickle you just right.

[–] [email protected] 2 points 8 months ago

Please respect my personal space and refrain from tickling me.

load more comments (1 replies)
load more comments (5 replies)
[–] [email protected] 193 points 8 months ago (3 children)

Collapse faster, please. Sick of ai bullshit clogging up my searches.

[–] [email protected] 42 points 8 months ago* (last edited 8 months ago) (2 children)

The collapse won't stop ai output from spamming the internet though. It will just make it worse and more likely to be incorrect

load more comments (2 replies)
[–] [email protected] 89 points 8 months ago (3 children)

The Internet is fucked now, the only valuable untainted training data is the Internet as it existed prior to this AI bullshit coming online. Confirmed human content is going to be super valuable, so expect our privacy to be fucked as well..

load more comments (3 replies)
load more comments (1 replies)
load more comments
view more: ‹ prev next ›