this post was submitted on 25 Jan 2024
3 points (71.4% liked)

Technology

59296 readers
4517 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
top 34 comments
sorted by: hot top controversial new old
[–] [email protected] 3 points 9 months ago

I was a part of a Blursed AI group on Facebook that had been a lot of fun, but suddenly this week it first shifted to Taylor Swift porn, and then to alt-right MAGA shit very quickly. The comments on the Trump and MAGA related images were very on-board with it too. The change was so abrupt that I got the fuck out of there. Seemed orchestrated.

[–] [email protected] 1 points 9 months ago
[–] [email protected] 1 points 9 months ago

!https://i.kym-cdn.com/entries/icons/facebook/000/038/932/tco_-_2021-11-24T113041.108.jpg

[–] [email protected] 2 points 9 months ago (1 children)

She probably has the money to sue each and every poster into oblivion.

[–] [email protected] 0 points 9 months ago* (last edited 9 months ago) (1 children)

For what? She's a public figure who does photo shoots. That's fair game buddy.

[–] [email protected] 0 points 9 months ago (1 children)

The fuck is wrong with you?

[–] [email protected] -1 points 9 months ago* (last edited 9 months ago) (1 children)

I didn't say it's moral or good or just -- but it's not illegal. My usage of fair game is in reference to fair use which is allowed under copyright law.

Try not eating the crayons.

[–] [email protected] 0 points 9 months ago (1 children)

If you meant a different term with a different meaning then you should use that term and even in that case I doubt that the fair use clause would stand as a defense.

[–] [email protected] -1 points 9 months ago* (last edited 9 months ago)

I mean -- no one has taken it to court so you can't say one way or another. Is it fair use if I put a cover of a magazine into stable diffusion and have it change the background? If I put michelle obama on mars did i just break the law? [no] This is a legal rats nest.

[–] [email protected] 1 points 9 months ago

Platforms, including X, where the images were first shared

They also now have a dedicated platform at their disposal. Expect more trash to be posted first on X in the future.

[–] [email protected] 8 points 9 months ago (5 children)

People have been creating and posting realistic looking fake celebrity nudes for quite literally decades now, but now they're using AI and its suddenly a problem?

[–] [email protected] 2 points 9 months ago

It's because a person can crank out a deep fake in 3 hours, and a crappy one in one. It never cropped up because... well lets be real it was a couple of weirdos that were doing it, unless it bubbles up from the dark corners of the internet you risk the Streisand effect by bringing attention to it.

AI can crank out 40 in a minute. 7200 in three hours. That's an entirely different beast. The sheer mass and volume ramps up the odds of any image bubbling up from the dark corners of the web falling into the limelight and now this problem that wasn't big enough to merit thought is rearing up it's ugly head right in front of us.

You can generate unique pictures of Taylor Swift faster than even Taylor swift can generate pictures of Taylor Swift. Within one hour of Taylor swift being seen with a man (and you have enough images of the man) you can create a dozen images of her on a date with that man and attempt to sell them to paparazzi.

The problem is volume. Just like how email made everyone connected and allowed the Nigerian Prince scandal to occur.

[–] [email protected] 0 points 9 months ago (1 children)

I don't think many have gone viral on social media before or took less than 5 minutes to create. My uneducated guess is that previously this stuff would be in some niche forum in the recesses of the internet

[–] [email protected] 1 points 9 months ago

My uneducated guess is that previously this stuff would be in some niche forum in the recesses of the internet

Not really. Back at the time there were public usenet groups specifically dedicated to the [hot actress of the day] fake porn.

[–] [email protected] 1 points 9 months ago

I can see why people are upset, I can agree that distribution of these images can be an issue, but this has the same energy of "I am mad that a certain picture of me is on the internet, I demand that they take it down." Sorry, that's not going to happen any time soon.

[–] [email protected] 1 points 9 months ago (1 children)

It's now massively accessible and realistic. Yes, it's a problem.

[–] [email protected] 1 points 9 months ago (1 children)

Google.com/images

Damn look at that, been accessible for decades

[–] [email protected] 1 points 9 months ago (1 children)

The tools are accessible. I wish this place wasn't full of weirdo ai tech bros sometimes.

[–] [email protected] 1 points 9 months ago (1 children)

Because downloading gimp is really hard.

[–] [email protected] -1 points 9 months ago* (last edited 9 months ago)

I'm sure you do not need me to explain to you the cavernous difference between gimp and deepfake. I trust you are capable of that.

[–] [email protected] 2 points 9 months ago (1 children)

I'm not sure if you noticed, but people who write for a living have suddenly started writing quite a lot about how technology that can write and generate media are bad.

[–] [email protected] 1 points 9 months ago (1 children)

Which is so silly, because AI writing still needs a human editor. I write for a living and there tons of work that involves using AI as a tool to increase productivity rather than to replace writers completely... like photoshop didnt put photographers out of business it just changed the work flow.

[–] [email protected] 1 points 9 months ago* (last edited 9 months ago) (1 children)

I work in a clinical setting where some Doctors are trying an AI program for generating their clinical notes out of the casual conversation between them and the patient. It's way off its mark for what we demand in quality. It requires significant editing from the healthcare provider, and if the note is very robust it quickly becomes more of a chore than modern voice transcription. Our review is not great so far.

[–] [email protected] 0 points 9 months ago (1 children)

That's a terrible way to be using a LLM for generating clinical notes.

Sounds more like trying to use a screwdriver to hammer in screws than an issue with the screwdriver itself.

[–] [email protected] 1 points 9 months ago

Sounds more like trying to use a screwdriver to hammer in screws

This is what I think about AI being forced into many things these days. Feels more like an attempt to justify subscription plans than anything actually productive.

[–] [email protected] 1 points 9 months ago

The grover one was hilarious... Jus sayin

[–] [email protected] 0 points 9 months ago (1 children)

Is it a problem though? I mean, it just makes Rule34 pics that much easier to create. And you wouldn't want to kink-shame anyone, now, would you? Why is it always heterosexual men who are kink-shamed? Why is liking naked women a bad thing?

[–] [email protected] 1 points 9 months ago (1 children)

Alright, I'll bite. To start, this is a real person that we are talking about. A real person who did not consent. Does that mean anything to you? The fact that there is a very real person that exists in very real life that has had this happen to them?

Otherwise, I agree. Nothing wrong with the male libido.

[–] [email protected] 1 points 9 months ago (1 children)

I don't need someone's consent to draw lewd fanart of them.

[–] [email protected] 1 points 9 months ago (2 children)

Do you need their permission to distribute that art?

[–] [email protected] -1 points 9 months ago
[–] [email protected] 0 points 9 months ago (1 children)

100%, which is why the fault lies with the bad actors and the platforms that let this proliferate and not the tool itself.

Can you imagine this headline but with Photoshop instead of AI? It would be utterly silly.

This is orchestrated to create anger against AI. There's a lot of money involved in it and that money triples if consumers aren't allowed to run and distribute models on their own PC.

[–] [email protected] 1 points 9 months ago (1 children)

I can agree with you on principal. Just as long as you aren't distributing things like this I don't really see an issue. Not the tool, the distributor / platform. I also agree that these articles are meant to ensure that those technology can be held behind locked doors. I fully support the idea of making AI something that is self-hostable.

That being said there are people in this thread that see nothing wrong with distributing lewd pictures of real people (drawn, ai generated, or even hacked and stolen). That is the only thing that I was addressing.

[–] [email protected] -1 points 9 months ago

Ya I agree, there's a very ugly side of things and It's a shame certain people are rolling themselves in it.

There's also the fact that lewds of celebs and similar material will grow exponentially but the same is also true for all other media so hopefully it balances itself out.

I think that's the main problem with a lot of these articles, they are missing the forest for the trees so to speak. We are looking at an explosion of culture, the bad stuff is just along for the ride. I'm personally excited for it.