this post was submitted on 30 Sep 2023
6 points (100.0% liked)

Open Source

30364 readers
734 users here now

All about open source! Feel free to ask questions, and share news, and interesting stuff!

Useful Links

Rules

Related Communities

Community icon from opensource.org, but we are not affiliated with them.

founded 5 years ago
MODERATORS
top 26 comments
sorted by: hot top controversial new old
[–] [email protected] 1 points 11 months ago (1 children)

I'll believe it when I see it.

I'm so goddamn tired of "open source" turning into subscription models restricting use cases because the company wants to appease conservative investors.

[–] [email protected] 1 points 11 months ago

Mozilla has a very strong track-record though. They've been around for a very long time, and have stuck to free open-source principles the whole time.

[–] [email protected] 0 points 11 months ago (1 children)

As much as I love Mozilla, I know they're going to censor it (sorry, the word is "alignment" now) the hell out of it to fit their perceived values. Luckily if it's open source then people will be able to train uncensored models

[–] [email protected] 1 points 11 months ago (2 children)

What in the world would an "uncensored" model even imply? And give me a break, private platforms choosing to not platform something/someone isn't "censorship", you don't have a right to another's platform. Mozilla has always been a principled organization and they have never pretended to be apathetic fence-sitters.

[–] [email protected] 0 points 11 months ago (2 children)

Anything that prevents it from my answering my query. If I ask it how to make me a bomb, I don't want it to be censored. It's gathering this from public data they don't own after all. I agree with Mozilla's principles, but also LLMs are tools and should be treated as such.

[–] [email protected] 1 points 11 months ago* (last edited 11 months ago) (1 children)

shit just went from 0 to 100 real fucking quick

for real though, if you ask an LLM how to make a bomb, it's not the LLM that's the problem

[–] [email protected] 0 points 11 months ago* (last edited 11 months ago) (2 children)

If it has the information, why not? Why should you be restricted by what a company deems appropriate. I obviously picked the bomb example as an extreme example, but that's the point.

Just like I can demonize encryption by saying I should be allowed to secretly send illegal content. If I asked you straight up if encryption is a good thing, you'd probably agree. If I mentioned its inevitable bad use in a shocking manner, would you defend the ability to do that, or change your stance that encryption is bad?

To have a strong stance means also defending the potential harmful effects, since they're inevitable. It's hard to keep values consistent, even when there are potential harmful effects of something that's for the greater good. Encryption is a perfect example of that.

[–] [email protected] 1 points 11 months ago

If it has the information, why not?

Naive altruistic reply: To prevent harm.

Cynic reply: To prevent liabilities.

If the restaurant refuses to put your fries into your coffee, because that's not on the menu, then that's their call. Can be for many reasons, but it's literally their business, not yours.

If we replace fries with fuse, and coffee with gun powder, I hope there are more regulations in place. What they sell and to whom and in which form affects more people than just buyer and seller.

Although I find it pretty surprising corporations self-regulate faster than lawmakers can say 'AI' in this case. That's odd.

[–] [email protected] 0 points 11 months ago (1 children)

This is a false equivalence. Encryption only works if nobody can decrypt it. LLMs work even if you censor illegal content from their output.

[–] [email protected] 0 points 11 months ago (1 children)

You miss the point. My point is that if you want to have a consistent view point, you need to acknowledge and defend the harmful sides. Encryption can objectively cause harm, but it should absolutely still be defended.

[–] [email protected] 0 points 11 months ago* (last edited 11 months ago) (1 children)

What the fuck is this "you should defend harm" bullshit, did you hit your head during an entry level philosophy class or something?

The reason we defend encryption even though it can be used for harm is because breaking it means you can't use it for good, and that's far worse. We don't defend the harm it can do in and of itself; why the hell would we? We defend it in spite of the harm because the good greatly outweighs the harm and they cannot be separated. The same isn't true for LLMs.

[–] [email protected] 0 points 11 months ago (1 children)

We don't believe that at all, we believe privacy is a human right. Also you're just objectively wrong about LLMs. Offline uncensored LLMs already exist, and will perpetually exist. We don't defend tools doing harm, we acknowledge it.

[–] [email protected] 1 points 11 months ago

We don't believe that at all, we believe privacy is a human right.

That's just a different way to phrase what I said about defending the good side of encryption.

Offline uncensored LLMs already exist, and will perpetually exist

I didn't say they don't exist, I said that the help and harm aren't inseparable like with encryption.

We don't defend tools doing harm, we acknowledge it.

"My point is that if you want to have a consistent view point, you need to acknowledge and defend the harmful sides."

If you want to walk it back, fine, but don't pretend like you didn't say it.

[–] [email protected] 0 points 11 months ago (1 children)

If you ask how to build a bomb and it tells you, wouldn't Mozilla get in trouble?

[–] [email protected] 0 points 11 months ago* (last edited 11 months ago) (1 children)

Do gun manufacturers get in trouble when someone shoots somebody?

Do car manufacturers get in trouble when someone runs somebody over?

Do search engines get in trouble if they accidentally link to harmful sites?

What about social media sites getting in trouble for users uploading illegal content?

Mozilla doesn't need to host an uncensored model, but their open source AI should be able to be trained to uncensored. So I'm not asking them to host this themselves, which is an important distinction I should have made.

Which uncensored LLMs exist already, so any argument about the damage they can cause is already possible.

[–] [email protected] 0 points 11 months ago

Why are lolbertarians on lemmy?

[–] [email protected] 0 points 11 months ago (1 children)

This is something I think a lot of people don't get about all the current ML hype. Even if you disregard all the other huge ethics issues surrounding sourcing training data, what does anybody think is going to happen if you take the modern web, a huge sea of extremist social media posts, SEO optimized scams and malware, and just general data toxic waste, and then train a model on it without rigorously pushing it away from being deranged? There's a reason all the current AI chatbots have had countless hours of human moderation adjustment to make them remotely acceptable to deploy publicly, and even then there are plenty of infamous examples of them running off the rails and saying deranged things.

Talking about an "uncensored" LLM basically just comes down to saying you'd like the unfiltered experience of a robot that will casually regurgitate all the worst parts of the internet at you, so unless you're actively trying to produce a model to do illegal or unethical things I don't quite see the point of contention or what "censorship" could actually mean in this context.

[–] [email protected] 1 points 11 months ago

It means they can’t make porn images of celebs or anime waifus, usually.

[–] [email protected] 0 points 11 months ago (1 children)

I would really like Mozilla to make the best browser in the world please.

[–] [email protected] 1 points 11 months ago (1 children)
[–] [email protected] 1 points 4 months ago

They failed a while ago. Market share decline continues

[–] [email protected] 1 points 11 months ago

Incredibly welcomed. We need more ethical, non-profit AI researchers in the sea of corporate for-profit AI companies.

[–] [email protected] 2 points 11 months ago* (last edited 11 months ago) (2 children)

My mind immediately went to a horizon zero dawn like dystopia where the Mozilla AI is the only thing left protecting humans from various malevolent AIs bent on consuming the human race

[–] [email protected] 0 points 11 months ago

Imagining the Mozilla AI as a personified Firefox and Thunderbird fighting off Cortana, some BARD (sorry) and a bunch of generic evil corporate AIs just makes me excited that Mozilla would be the one fending everyone off.

[–] [email protected] 0 points 11 months ago (1 children)

Mozilla is Gaia, ChatGPT is hades?

[–] [email protected] 1 points 11 months ago* (last edited 11 months ago)

I think by that point ChatGPT would be more like Apollo, keeping the knowledge of humanity. I feel like one of the more corporate AIs will go full HADES, I'm thinking Bard. It will get a mysterious signal from space that switches it's core protocol from "don't be evil" to "be evil."