this post was submitted on 07 Apr 2024
339 points (93.1% liked)

Technology

59378 readers
3246 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] -3 points 7 months ago (2 children)

But the fact that this tech really kicked off just three years ago and is already threatening so many jobs, is pretty telling. Not only will LLMs continue to get better, but they're a big step towards AGI and that's always been an existential crisis we knew was coming. This is the the time to start adapting, quick.

[–] [email protected] 2 points 7 months ago (1 children)

What does it threaten really?

It works for contact centers for bots to answer short simple questions, so that agents' time would be used more efficiently. I'm not sure it saves that much money TBF.

It works for image classification. And still needs checking.

It works for OCR. And still needs checking.

It works for voice recognition and transcription, which is actually cool. Still needs checking.

but they’re a big step towards AGI

What makes you think that? Was the Mechanical Turk a big step towards thinking robots?

They are very good at pretending to be that big step for people who don't know how they work.

[–] [email protected] 1 points 7 months ago* (last edited 7 months ago) (1 children)

You're right that it doesn't save too much money making people more efficient. That's why they will replace employees instead. That's the threat.

Yes they make mistakes. So do people. They just have to make less than an employee does and we're on the right track for that. AI will always make mistakes and this is actually a step in the right direction. Deterministic systems that rely on concrete input and perfectly crafted statistical models can't work in the real world. Once the system it is trying to evaluate (most systems in the real world) is sufficiently complex, you encounter unknown situations where you have to spend infinite time and energy gathering information and computing... or guess.

Our company is small and our customer inquiries increased several fold because our product expanded. We were panicking thinking we needed to train and hire a whole customer support department overnight, where we currently have one person. But instead we implement AI representatives. Our feedback actually became more positive because these agents can connect with you instantly, pull nebulous requests from confusing messages, and alert the appropriate employee of any action needed. Does it make mistakes? Sure, not enough to matter. It's simple for our customer service person to reach out and correct the mistake.

I think people that think this isn't a big deal for AGI don't understand how the human mind works. I find it funny when they try and articulate why they think LLMs are just a trick. "It's not really creating anything, it's just pulling a bunch of relevant material from its training data and using it as a basis for a similar output." And... What is it you think you do?

[–] [email protected] 1 points 7 months ago (1 children)

And… What is it you think you do?

Unlike an LLM, I rebuild myself, for example.

[–] [email protected] 1 points 7 months ago* (last edited 7 months ago)

It's trivial to copy an LLM, but if you mean self improvement: https://arxiv.org/abs/2401.10020

[–] [email protected] 8 points 7 months ago (1 children)

They didn't just appear out of nowhere, they're the result of decades of research and development. You're also making the assumption that additional progress is guaranteed. AI has hit walls and dead ends in the past, there's no reason to assume that we're not hitting a local maximum again right now.

[–] [email protected] -4 points 7 months ago* (last edited 7 months ago) (1 children)

And there's no reason to believe that it is. I know there's been speculation about model collapse and limits of available training data. But there's also been advancements like training data efficiency and autonomous agents. Your response seems to ignore the massive amounts of progress we've seen in the space.

Also the computer, internet, and smart phone were based on decades of research and development. Doesn't mean they didn't take off and change everything.

The fact that you're saying AI hit walls in the past and now we're here, is a pretty good indication that progress is guaranteed.

[–] [email protected] 5 points 7 months ago (2 children)

You said there's no reason and then you list potential reasons right after. Yes, there has been progress and no one is arguing against that, but the two big issues are:

  1. What exists is being overhyped as far more capable than it really is.
  2. How much room there is to grow with current techniques is still unknown.

The computer, internet, and smart phone are all largely deterministic with actions resulting in direct known outcomes. AI as we know it is based on highly complex statistical models and relies heavily on the data it is trained on. It has far more things that can go wrong which makes it unsuitable for critical applications (just look at the disasters when it's used as a customer service representative). That's not even getting into the legal issues that have yet to actually be answered. Just look at the CTO of OpenAI squirming on the question of what Sora was trained on (timestamped).

Being able to overcome walls in the past doesn't guarantee overcoming walls in the present. That's like saying being able to jump over a hurdle is the same as leaping over a skyscraper. There's also the question of timing, it took decades for those previous walls to be overcome. Impact to the workforce is largely overstated and is being used as an excuse for cost cutting. It's just like the articles about automation after the great recession. I'm still waiting on robots that can flip burgers (article from 2012).

[–] [email protected] -4 points 7 months ago* (last edited 7 months ago) (1 children)

I listed reasons people usually cite and why I don't think they're a good reason to assume there won't be progress. I agree it's over-hyped today, because people are excited about the obvious potential tomorrow. I think it's foolish to hide behind that as if it's proof that it doesn't have potential.

Let's say you're right and we hit a wall for 50 years on any progress on AI. There's nothing magical about the human brain's ability to make logical decisions on observations and learning. It's going to happen. And our current system of economy that attributes a person's value to their labor will be in deep shit when it happens. It could take a century to make an appropriate change here. We're already way behind, even with a set back to AI.

I think it's funny when people complain about AI learning from copyright. AI's express goal is to be similar to a human consciousness. Have you ever talked to a human who's never watched a TV show, or a movie, or read a book from this century? An AI that's not aware of those things would be like a useless alien to us.

If people just want to use legal hangups to stop AI, fair play. But that plan is doomed, infinite brainpower is just too valuable. Copyright isn't there to protect the little guy, that was the original 28 year law. Its current form was lobbied by corporations to stifle competition. And they'll dismantle it (or ignore it) in a heartbeat once it suits them.

[–] [email protected] 4 points 7 months ago

The topic at hand is this survey which claims significant impacts to the workforce within five years and this is what I'm speaking towards. As for copyright, these models are straight-up not possible without that data and the link can be clearly demonstrated, they have their training data available which they may have to expose in a court case. Forget about the little guy, the large corporations who own the data will not be happy letting them build this lucrative AI without them getting paid for it. There will be legal fights and it is a potential complication in rolling this stuff out so it should be considered.

[–] [email protected] 1 points 7 months ago

Here is an alternative Piped link(s):

the CTO of OpenAI squirming on the question of what Sora was trained on (timestamped)

Piped is a privacy-respecting open-source alternative frontend to YouTube.

I'm open-source; check me out at GitHub.