this post was submitted on 16 Mar 2025
567 points (98.3% liked)

Privacy

35752 readers
327 users here now

A place to discuss privacy and freedom in the digital world.

Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.

In this community everyone is welcome to post links and discuss topics related to privacy.

Some Rules

Related communities

much thanks to @gary_host_laptop for the logo design :)

founded 5 years ago
MODERATORS
(page 2) 47 comments
sorted by: hot top controversial new old
[–] [email protected] 42 points 2 days ago (2 children)

But when China steals all their (arguably not copywrite-able) work...

load more comments (2 replies)
[–] [email protected] 8 points 2 days ago
[–] [email protected] 16 points 2 days ago (1 children)
[–] [email protected] 10 points 2 days ago (1 children)

Of course it is if you copy to monetise which is what they do.

[–] [email protected] 10 points 2 days ago

They monetize it, erase authorship and bastardize the work.

Like if copyright was to protect against anything, it would be this.

[–] [email protected] 61 points 2 days ago* (last edited 2 days ago) (3 children)

Fuck Sam Altmann, the fartsniffer who convinced himself & a few other dumb people that his company really has the leverage to make such demands.

"Oh, but democracy!" - saying that in the US of 2025 is a whole 'nother kind of dumb.
Anyhow, you don't give a single fuck about democracy, you're just scared because a chinese company offers what you offer for a fraction of the price/resources.

Your scared for your government money and basically begging for one more handout "to save democracy".

Yes, I've been listening to Ed Zitron.

[–] [email protected] 9 points 2 days ago (2 children)

gosh Ed Zitron is such an anodyne voice to hear, I felt like I was losing my mind until I listened to some of his stuff

load more comments (2 replies)
load more comments (2 replies)
[–] [email protected] 13 points 2 days ago

Sam Altman is a lying hype-man. He deserves to see his company fail.

[–] [email protected] 9 points 2 days ago* (last edited 2 days ago) (1 children)

This is why they killed that former employee.

[–] [email protected] 11 points 2 days ago (1 children)

Say his name y’all

Suchir Balaji

load more comments (1 replies)
[–] [email protected] 160 points 2 days ago (4 children)

If this passes, piracy websites can rebrand as AI training material websites and we can all run a crappy model locally to train on pirated material.

[–] [email protected] 68 points 2 days ago

That would work if you were rich and friends with government officials. I don’t like your chances otherwise.

[–] [email protected] 14 points 2 days ago

You are a glass half full sort of person!

[–] [email protected] 15 points 2 days ago

Another win for piracy community

load more comments (1 replies)
[–] [email protected] 3 points 2 days ago* (last edited 2 days ago)

Please let it be over, yes.

Nobody even tries to write code from scratch anymore. I think it will have a lot of negative effects on programmers over time.

[–] [email protected] 2 points 2 days ago

I think it would be interesting as hell if they had to cite where the data was from on request. See if it's legitimate sources or just what a reddit user said five years ago

[–] [email protected] 3 points 2 days ago

"Your proposal is acceptable."

[–] [email protected] 2 points 2 days ago
[–] [email protected] 78 points 2 days ago (1 children)

Oh it's "over"? Fine for me

[–] [email protected] 7 points 2 days ago

Ho no, what will we do without degenerate generative AIs?!

[–] [email protected] 8 points 2 days ago (6 children)

Obligatory: I'm anti-AI, mostly anti-technology

That said, I can't say that I mind LLMs using copyrighted materials that it accesses legally/appropriately (lots of copyrighted content may be freely available to some extent, like news articles or song lyrics)

I'm open to arguments correcting me. I'd prefer to have another reason to be against this technology, not arguing on the side of frauds like Sam Altman. Here's my take:

All content created by humans follows consumption of other content. If I read lots of Vonnegut, I should be able to churn out prose that roughly (or precisely) includes his idiosyncrasies as a writer. We read more than one author; we read dozens or hundreds over our lifetimes. Likewise musicians, film directors, etc etc.

If an LLM consumes the same copyrighted content and learns how to copy its various characteristics, how is it meaningfully different from me doing it and becoming a successful writer?

[–] [email protected] 5 points 2 days ago

Except the reason Altman is so upset has nothing to do with this very valid discussion.

As I commented elsewhere:

Fuck Sam Altmann, the fartsniffer who convinced himself & a few other dumb people that his company really has the leverage to make such demands.

He doesn't care about democracy, he's just scared because a chinese company offers what his company offers, but for a fraction of the price/resources.

He's scared for his government money and basically begging for one more handout “to save democracy”.

Yes, I’ve been listening to Ed Zitron.

[–] [email protected] 5 points 2 days ago

If an LLM consumes the same copyrighted content and learns how to copy its various characteristics, how is it meaningfully different from me doing it and becoming a successful writer?

That is the trillion-dollar question, isn’t it?

I’ve got two thoughts to frame the question, but I won’t give an answer.

  1. Laws are just social constructs, to help people get along with each other. They’re not supposed to be grand universal moral frameworks, or coherent/consistent philosophies. They’re always full of contradictions. So… does it even matter if it’s “meaningfully” different or not, if it’s socially useful to treat it as different (or not)?
  2. We’ve seen with digital locks, gig work, algorithmic market manipulation, and playing either side of Section 230 when convenient… that the ethos of big tech is pretty much “define what’s illegal, so I can colonize the precise border of illegality, to a fractal level of granularity”. I’m not super stoked to come with an objective quantitative framework for them to follow, cuz I know they’ll just flow around it like water and continue to find ways to do antisocial shit in ways that technically follow the rules.
[–] [email protected] 6 points 2 days ago

Yup. Violating IP licenses is a great reason to prevent it. According to current law, if they get Alice license for the book they should be able to use it how they want.
I'm not permitted to pirate a book just because I only intend to read it and then give it back. AI shouldn't be able to either if people can't.

Beyond that, we need to accept that might need to come up with new rules for new technology. There's a lot of people, notably artists, who object to art they put on their website being used for training. Under current law if you make it publicly available, people can download it and use it on their computer as long as they don't distribute it. That current law allows something we don't want doesn't mean we need to find a way to interpret current law as not allowing it, it just means we need new laws that say "fair use for people is not the same as fair use for AI training".

[–] [email protected] 9 points 2 days ago (1 children)

Right. The problem is not the fact it consumes the information, the problem is if the user uses it to violate copyright. It’s just a tool after all.

Like, I’m capable of violating copyright in infinitely many ways, but I usually don’t.

[–] [email protected] 4 points 2 days ago* (last edited 2 days ago)

The problem is that the user usually can't tell if the AI output is infringing someone's copyright or not unless they've seen all the training data.

[–] [email protected] -3 points 2 days ago (2 children)

In your example, you could also be sued for ripping off his style.

[–] [email protected] -1 points 2 days ago* (last edited 2 days ago) (1 children)

Edited for clarity: If that were the case then Weird AL would be screwed.

Original: In that case Weird AL would be screwed

[–] [email protected] 3 points 2 days ago (1 children)

No because what he does is already a settled part of the law.

[–] [email protected] 1 points 2 days ago

That's the point. It's established law so OP wouldn't be sued

[–] [email protected] 6 points 2 days ago (2 children)

You can sue for anything in the USA. But it is pretty much impossible to successfully sue for "ripping off someone's style". Where do you even begin to define a writing style?

[–] [email protected] 0 points 2 days ago

There are lots of ways to characterize writing style. Go read Finnegans Wake and tell me James Joyce doesn't have a characteristic style.

[–] [email protected] 0 points 2 days ago

"style", in terms of composition, is actually a component in proving plagiarism.

load more comments (1 replies)
[–] [email protected] 86 points 2 days ago (5 children)
[–] [email protected] 2 points 2 days ago (1 children)

Yeah but I don't sell ripped dvds and copies of other peoples art.

load more comments (1 replies)
[–] [email protected] 20 points 2 days ago (1 children)

What OpenAI is doing is not piracy.

[–] [email protected] 3 points 2 days ago (1 children)

Whatever it is, it isn't theft

[–] [email protected] 15 points 2 days ago (3 children)

Also true. It’s scraping.

In the words of Cory Doctorow:

Web-scraping is good, actually.

Scraping against the wishes of the scraped is good, actually.

Scraping when the scrapee suffers as a result of your scraping is good, actually.

Scraping to train machine-learning models is good, actually.

Scraping to violate the public’s privacy is bad, actually.

Scraping to alienate creative workers’ labor is bad, actually.

We absolutely can have the benefits of scraping without letting AI companies destroy our jobs and our privacy. We just have to stop letting them define the debate.

[–] [email protected] -3 points 2 days ago (1 children)

Our privacy was long gone well before AI companies were even founded, if people cared about their privacy then none of the largest tech companies would exist because they all spy on you wholesale.

The ship has sailed on generating digital assets. This isn't a technology that can be invented. Digital artists will have to adapt.

Technology often disrupts jobs, you can't fix that by fighting the technology. It's already invented. You fight the disruption by ensuring that your country takes care of people who lose their jobs by providing them with support and resources to adapt to the new job landscape.

For example, we didn't stop electronic computers to save the job of Computer (a large field of highly trained humans who did calculations) and CAD destroyed the drafting profession. Digital artists are not the first to experience this and they won't be the last.

[–] [email protected] 4 points 2 days ago (1 children)

Our privacy was long gone well before AI companies were even founded, if people cared about their privacy then none of the largest tech companies would exist because they all spy on you wholesale.

In the US. The EU has proven that you can have perfectly functional privacy laws.

If your reasoning is based o the US not regulating their companies and so that makes it impossible to regulate them, then your reasoning is bad.

load more comments (1 replies)
load more comments (2 replies)
[–] [email protected] 43 points 2 days ago (3 children)

When a corporation does it to get a competitive edge, it is.

[–] [email protected] 2 points 2 days ago

Only if it's illegal to begin with. We need to abolish copyright, as with the internet and digital media in general, the concept has become outdated as scarcity isn't really a thing anymore. This also applies to anything that can be digitized.

The original creator can still sell their work and people can still choose to buy it, and people will if it is convenient enough. If it is inconvenient or too expensive, people will pirate it instead, regardless of the law.

[–] [email protected] 23 points 2 days ago* (last edited 2 days ago) (1 children)

No it's not.

It can be problematic behaviour, you can make it illegal if you want, but at a fundamental level, making a copy of something is not the same thing as stealing something.

[–] [email protected] 8 points 2 days ago* (last edited 2 days ago) (1 children)

it uses the result of your labor without compensation. it's not theft of the copyrighted material. it's theft of the payment.

it's different from piracy in that piracy doesn't equate to lost sales. someone who pirates a song or game probably does so because they wouldn't buy it otherwise. either they can't afford or they don't find it worth doing so. so if they couldn't pirate it, they still wouldn't buy it.

but this is a company using labor without paying you, something that they otherwise definitely have to do. he literally says it would be over if they couldn't get this data. they just don't want to pay for it.

[–] [email protected] -3 points 2 days ago* (last edited 2 days ago) (5 children)

That information is published freely online.

Do companies have to avoid hiring people who read and were influenced by copyrighted material?

I can regurgitate copyrighted works as well, and when someone hires me, places like Stackoverflow get fewer views to the pages that I've already read and trained on.

Are companies committing theft by letting me read the internet to develop my intelligence? Are they committing theft when they hire me so they don't have to do as much research themselves? Are they committing theft when they hire thousands of engineers who have read and trained on copyrighted material to build up internal knowledge bases?

What's actually happening, is that the debates around AI are exposing a deeply and fundamentally flawed copyright system. It should not be based on scarcity and restriction but rewarding use. Information has always been able to flow freely, the mistake was linking payment to restricting it's movement.

load more comments (5 replies)
[–] [email protected] 27 points 2 days ago

It’s only theft if they support laws preventing their competitors from doing it too. Which is kind of what OpenAI did, and now they’re walking that idea back because they’re losing again.

[–] [email protected] 7 points 2 days ago

Piracy is only theft if AI can't be made profitable.

load more comments (1 replies)
load more comments
view more: ‹ prev next ›