this post was submitted on 04 Apr 2024
122 points (87.2% liked)

Technology

59429 readers
3081 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[โ€“] [email protected] 0 points 7 months ago (1 children)

The thought that internet becomes shitty enough that you need a GPU to browse it is really frightening me. If we really reach that point that may be to run an AI which filters out AI generated spam which would really depress me ๐Ÿ˜ญ

[โ€“] [email protected] 4 points 7 months ago* (last edited 7 months ago) (1 children)

The thought that internet becomes shitty enough that you need a GPU to browse it is really frightening me.

I mean, there was a point where an FPU was a separate chip and wasn't the norm; now it's built into the CPU.

I think that it's probably safe to say that, in the future, there will be broader use of parallel processing, as we've fundamental limits on what we know we can do there with existing laws of physics with serial processing. That could wind up being part of the CPU. It could live on a separate piece of hardware -- which may not necessarily be a "GPU" -- parallel processing hardware entered the PC because the most-immediate need was to do 3d graphics rendering, but as you can see from the LLMs that people are running on GPUs today, that's not the only application. The parallel compute accelerator cards that Nvidia is selling today for an arm and a leg on servers aren't aimed at doing 3d graphics.

It may not be 3d graphics rendering or running LLMs that becomes the primary application. But I'd be reasonably comfortable saying that down the line, relative to today, there will be more parallel-processing hardware in computers than is present today.

[โ€“] [email protected] 1 points 7 months ago

That could wind up being part of the CPU

For a lot of newer processors, it already is. Intel, Apple, Samsung, and Qualcomm like to brag about their processors having some sort of neural contraption meant to assist with AI processing.

If it stays around, it might be good enough that you don't need a GPU to do it, since the CPU has an onboard chip that can handle that work instead, since tensor processors like that are a bit more efficient than GPUs, but are also more specialised.