this post was submitted on 20 Apr 2024
1 points (100.0% liked)
LocalLLaMA
2249 readers
1 users here now
Community to discuss about LLaMA, the large language model created by Meta AI.
This is intended to be a replacement for r/LocalLLaMA on Reddit.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Why in the world would you need such a large budget? A mac studio can run the 70b variant just fine at $12k
If possible, to run the upcoming llama 400B one. But this is just hypothetical.
Depends on what you're doing with it, but prompt/context processing is a lot faster on Nvidia GPUs than on Apple chips, though if you are using the same prefix all the time it's a bit better.
The time to first token is a lot faster on datacenter GPUs, especially as context length increases, and consumer GPUs don't have enough vram.
So the answer would be "an alibi for the other $88k"
I'll take 'Someone got seed funding and now needs progress to unlock the next part of the package' for $10 please Alex.