this post was submitted on 29 Sep 2024
1170 points (98.2% liked)
People Twitter
5226 readers
2311 users here now
People tweeting stuff. We allow tweets from anyone.
RULES:
- Mark NSFW content.
- No doxxing people.
- Must be a tweet or similar
- No bullying or international politcs
- Be excellent to each other.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Let's have a talk about social media platform censorship. Tiktok and YouTube members who self censor common words like death or rape in legitimate conversations about the topics are learning to temper their language or face consequences. Unimportant consequences.
It may seem small by comparison, but if you condition it at a low level, each step beyond is easy to swallow. Spread it out over an entire population, and you see huge results.
Is this why I see totally unnecessary self censorship for words like “r*cism”? Even here on Lemmy. I assume some of this originated on Twitter, where people abuse the reporting system as a form of retaliation.
Yes. It's so absurd.
You just described Newspeak (Nineteen Eighty-Four novel):
Source
I did pick up what Orwell was putting down. It's definitely helped shaped my view of the world.
Newspeak was an intentional in-universe conlang designed and handed down by Ingsoc based on "how you speak affects how you think" (which is a hypothesis that has... some kind of name). This is a bunch of people trivially avoiding automated filtering like it's been done since the first puritan implemented the first world filter.
One of the main differences is that self-censoring seggs and raep and ahh-es or whatever still leaves it plenty obvious what you mean, it just outs you as a Tiktok user. Conceptually word filters are a blacklist whereas Newspeak was intended to be a whitelist with the restrictiveness that entails.
The Sapir-Whorf hypothesis or linguistic relativity, just in case that was bothering you.
It's a lot more banal, though. Youtube has to sell advertising, and advertisers don't want to be next to discussions of rape or suicide. These restrictions are enforced algorithmically, hence the self-censorship. And in any case, it doesn't achieve the objective of newspeak, as those concepts are still being discussed.
I don't think it's right to divorce the censorship from the result just because the justification is different.
What I mean is that even though that conditioning is taking place for a banal reason it's still true that it's conditioning and will affect the acceptance of moves like this debate fact checking decision that are serious and do have consequences. So therefore it still matters and is still dangerous.
Yet.
But I get what you are saying. I just find the similarities, although banal, kind of funny. In a scary kind of way.