this post was submitted on 27 Jan 2025
882 points (98.1% liked)

Technology

61263 readers
4176 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

cross-posted from: https://lemm.ee/post/53805638

(page 4) 50 comments
sorted by: hot top controversial new old
[–] [email protected] 22 points 3 days ago

I've never been so happy to cancel a subscription.

[–] [email protected] 11 points 4 days ago (4 children)

nvidia falling doesn't make much sense to me, GPUs are still needed to run the model. Unless Nvidia is involved in its own AI model or something?

[–] [email protected] 24 points 4 days ago (1 children)

If you need far less computing power to train the models, far less gpus are needed, and that hurts nvidia

[–] [email protected] 6 points 3 days ago (1 children)

does it really need less power? I'm playing around with it now and I'm pretty impressed so far. it can do math, at least.

[–] [email protected] 18 points 3 days ago (4 children)

That's the claim, it has apparently been trained using a fraction of the compute power of the GPT models and achieves similar results.

[–] [email protected] 4 points 3 days ago (1 children)

But I feel like that will just lead to more training with the same (or more) hardware with a more efficient model. Bitcoin mining didn't slow down only because it got harder. However I don't know enough about the training process. I assume more efficient use of the hardware would allow for larger models to be trained on the same hardware and training data?

load more comments (1 replies)
load more comments (3 replies)
load more comments (3 replies)
load more comments
view more: ‹ prev next ›