this post was submitted on 24 Apr 2025
103 points (100.0% liked)

Privacy

2027 readers
100 users here now

Welcome! This is a community for all those who are interested in protecting their privacy.

Rules

PS: Don't be a smartass and try to game the system, we'll know if you're breaking the rules when we see it!

  1. Be civil and no prejudice
  2. Don't promote big-tech software
  3. No reposting of news that was already posted
  4. No crypto, blockchain, NFTs
  5. No Xitter links (if absolutely necessary, use xcancel)

Related communities:

Some of these are only vaguely related, but great communities.

founded 5 months ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[โ€“] [email protected] 1 points 2 days ago (1 children)

Exactly.

But Ed said you could also use your own models, train it yourself.

[โ€“] [email protected] 2 points 2 days ago* (last edited 2 days ago)

From my own fractured understanding, this is indeed true, but the "DeepSeek" everybody is excited about, which performs as well as OpenAI's best products but faster, is a prebuilt flagship model called R1. (Benchmarks here.)

The training data will never see the light of day. It would be an archive of every ebook under the sun, every scraped website, just copyright infringement as far as the eye can see. That would be the source they would have to release to be open source, and I doubt they would.

But DeepSeek does have the code for "distilling" other companies' more complex models into something smaller and faster (and a bit worse) - but, of course, the input models are themselves not open source, because those models (like Facebook's restrictive Llama model) were also trained on stolen data. (I've downloaded a couple of these distillations just to mess around with them. It feels like having a dumber, slower ChatGPT in a terminal.)

Theoretically, you could train a model using DeepSeek's open source code and ethically sourced input data, but that would be quite the task. Most people just add an extra layer of training data and call it a day. Here's one such example (I hate it.) I can't even imagine how much data you would have to create yourself in order to train one of these things from scratch. George RR Martin himself probably couldn't train an AI to speak in a comprehensible manner by feeding it his life's work.