this post was submitted on 13 Jun 2024
-178 points (26.2% liked)

Linux

47341 readers
1365 users here now

From Wikipedia, the free encyclopedia

Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).

Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.

Rules

Related Communities

Community icon by Alpár-Etele Méder, licensed under CC BY 3.0

founded 5 years ago
MODERATORS
 

Whenever AI is mentioned lots of people in the Linux space immediately react negatively. Creators like TheLinuxExperiment on YouTube always feel the need to add a disclaimer that "some people think AI is problematic" or something along those lines if an AI topic is discussed. I get that AI has many problems but at the same time the potential it has is immense, especially as an assistant on personal computers (just look at what "Apple Intelligence" seems to be capable of.) Gnome and other desktops need to start working on integrating FOSS AI models so that we don't become obsolete. Using an AI-less desktop may be akin to hand copying books after the printing press revolution. If you think of specific problems it is better to point them out and try think of solutions, not reject the technology as a whole.

TLDR: A lot of ludite sentiments around AI in Linux community.

top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 4 points 3 months ago (1 children)

I've yet to see a need for "AI integration ✨" in to the desktop experience. Copilot, LLM chat bots, TTS, OCR, and translation using machine learning are all interesting but I don't think OS integration is beneficial.

[–] [email protected] -3 points 3 months ago (1 children)

Time 💫 will ✨ prove 💫 you ✨ wrong. 💫

[–] [email protected] 2 points 3 months ago (2 children)

not every high tech product or idea makes it, you don't see a lot of netbooks or wifi connected kitchen appliances these days either; having the ability to make tiny devices or connecting every single device is not justification enough to actually do it. i view ai integration similarly: having an llm in some side bar to change the screen brightness, find some time or switch the keyboard layout isn't really useful. being able to select text in an image viewer or searching through audio and video for spoken words for example would be a useful application for machine learning in the DE, that isn't really what's advertised as "AI" though.

[–] [email protected] 0 points 3 months ago (1 children)

Changing the brightness or WiFi settings can be very useful for many people. Not everyone is a Linux nerd and knows all the ins and outs of basic computing.

[–] [email protected] 2 points 3 months ago

maybe, but these people wouldn't own a pc with a dedicated gpu or neutral network accelerator.

[–] [email protected] 1 points 3 months ago

i don't really think anyone would be against the last two examples to be integrated in dolphin, nautilus, gwenview... either.

[–] [email protected] 0 points 3 months ago

I agree. However, I think it is related to Capitalism and all the sociopathic corporations out there. It's almost impossible to think that anything good will come from the Blue Church controlling even more tech. Capitalism have always used any opportunity to enslave/extort people - that continues with AI under their control.

However, I was also disappointed when I found out how negative 'my' crowd were. I wanted to create an open source lowend AGI to secure poor people a descent life without being attacked by Capitalism every day/hour/second, create abundance, communities, production and and in general help build a social sub society in the midst of the insane blue church and their propagandized believers.

It is perfectly doable to fight the Capitalist religion with homegrown AI based on what we know and have today. But nobody can do it alone, and if there's no-one willing to fight the f*ckers with AI, then it takes time..

I definitely intend to build a revolution-AGI to kill off the Capitalist religion and save exploited poor people. No matter what happens, there will be at least one AGI that are trained on revolution, anti-capitalism and building something much better than this effing blue nightmare. The worlds first aggressive 'Commie-bot' ha! 😍

[–] [email protected] 19 points 3 months ago

As someone whose employer is strongly pushing them to use AI assistants in coding: no. At best, it's like being tied to a shitty intern that copies code off stack overflow and then blows me up on slack when it magically doesn't work. I still don't understand why everyone is so excited about them. The only tasks they can handle competently are tasks I can easily do on my own (and with a lot less re-typing.)

Sure, they'll grow over the years, but Altman et al are complaining that they're running out of training data. And even with an unlimited body of training data for future models, we'll still end up with something about as intelligent as a kid that's been locked in a windowless room with books their whole life and can either parrot opinions they've read or make shit up and hope you believe it. I'll think we'll get a series of incompetent products with increasing ability to make wrong shit up on the fly until C-suite moves on to the next shiny bullshit.

That's not to say we're not capable of creating a generally-intelligent system on par with or exceeding human intelligence, but I really don't think LLMs will allow for that.

tl;dr: a lot of woo in the tech community that the linux community isn't as on board with

[–] [email protected] 2 points 3 months ago (2 children)

You're getting a lot of flack in these comments, but you are absolutely right. All the concerns people have raised about "AI" and the recent wave of machine learning tech are (mostly) valid, but that doesn't mean AI isn't incredibly effective in certain use cases. Rather than hating on the technology or ignoring it, the FOSS community should try to find ways of implementing AI that mitigate the problems, while continuing to educate users about the limitations of LLMs, etc.

[–] [email protected] 3 points 3 months ago

One comment that agrees 🥲

[–] [email protected] 2 points 3 months ago* (last edited 3 months ago) (1 children)

It's spelled flak, not flack. It's from the German word Flugabwehrkanone which literally means aerial defense cannon.

[–] [email protected] 2 points 3 months ago

Oh, that's very interesting. I knew about flak in the military context, but never realized it was the same word used in the idiom. The idiom actually makes a lot more sense now.

[–] [email protected] 5 points 3 months ago (1 children)

That's easy, move over to Windows or Mac and enjoy. I'll stay in my dumb as Linux distros, thank you.

[–] [email protected] 4 points 3 months ago (1 children)

The AI in my head is a bit underpowered but it gets the job done

[–] [email protected] 2 points 3 months ago

Same as mine. But mine also gets confused regularly, and it gets worse with every new version (age) 🤣🤣

[–] [email protected] 15 points 3 months ago

The first problem, as with many things AI, is nailing down just what you mean with AI.

The second problem, as with many things Linux, is the question of shipping these things with the Desktop Environment / OS by default, given that not everybody wants or needs that and for those that don't, it's just useless bloat.

The third problem, as with many things FOSS or AI, is transparency, here particularly training. Would I have to train the models myself? If yes: How would I acquire training data that has quantity, quality and transparent control of sources? If no: What control do I have over the source material the pre-trained model I get uses?

The fourth problem is privacy. The tradeoff for a universal assistant is universal access, which requires universal trust. Even if it can only fetch information (read files, query the web), the automated web searches could expose private data to whatever search engine or websites it uses. Particularly in the wake of Recall, the idea of saying "Oh actually we want to do the same as Microsoft" would harm Linux adoption more than it would help.

The fifth problem is control. The more control you hand to machines, the more control their developers will have. This isn't just about trusting the machines at that point, it's about trusting the developers. To build something the caliber of full AI assistants, you'd need a ridiculous amount of volunteer efforts, particularly due to the splintering that always comes with such projects and the friction that creates. Alternatively, you'd need corporate contributions, and they always come with an expectation of profit. Hence we're back to trust: Do you trust a corporation big enough to make a difference to contribute to such an endeavour without amy avenue of abuse? I don't.


Linux has survived long enough despite not keeping up with every mainstream development. In fact, what drove me to Linux was precisely that it doesn't do everything Microsoft does. The idea of volunteers (by and large unorganised) trying to match the sheer power of a megacorp (with a strict hierarchy for who calls the shots) in development power to produce such an assistant is ridiculous enough, but the suggestion that DEs should come with it already integrated? Hell no

One useful applications of "AI" (machine learning) I could see: Evaluating logs to detect recurring errors and cross-referencing them with other logs to see if there are correlations, which might help with troubleshooting.
That doesn't need to be an integrated desktop assistant, it can just be a regular app.

Really, that applies to every possible AI tool. Make it an app, if you care enough. People can install it for themselves if they want. But for the love of the Machine God, don't let the hype blind you to the issues.

[–] [email protected] 1 points 3 months ago

You're mid right, the when something AI based is announced this is really criticized by some people and there are almost right. When something new pops, like windows recall, it is certain that this "new" feature is really not what AI is capable, and asks really important questions about privacy. But you're right on the fact that Linux should be a bit more interested on AI and tried to made it the right way! But for now there's no really good use cases of AI inside a distro. LLMs are good but do not need to be linked to user activities. Image generators are great but do not need to be linked to user activities... As exemple when Windows tried Recall and failed. Apple iOS 18 wants to implement that, and this should be surely a success inside the Apple minded people. But here where FOSS, privacy and anti Big-Techs guys are the main people that's absolutely sure that every for-profit "new AI" feature would be really hated. I'm not against this mind just giving facts

[–] [email protected] 12 points 3 months ago

Imo you immensely overestimate the capabilities of these models. What they show to the public are always hand picked situations even if they say they dont

[–] [email protected] 13 points 3 months ago

Whenever AI is mentioned lots of people in the Linux space immediately react negatively.

Because whenever AI is mentioned it usually isn't even close to what AI meant.

[–] [email protected] 5 points 3 months ago

This article should be ignored.

[–] [email protected] 39 points 3 months ago (1 children)

Tech Enthusiasts: Everything in my house is wired to the Internet of Things! I control it all from my smartphone! My smart-house is bluetooth enabled and I can give it voice commands via alexa! I love the future!

Programmers / Engineers: The most recent piece of technology I own is a printer from 2004 and I keep a loaded gun ready to shoot it if it ever makes an unexpected noise.

[–] [email protected] 1 points 3 months ago (1 children)

That doesn't describe me or any other programmer I know.

[–] [email protected] 2 points 3 months ago

It doesn't describe me either, but I had nothing meaningful to contribute to the discussion.

[–] [email protected] 17 points 3 months ago (1 children)

Is there no electron wrapper around ChatGPT yet? Jeez we better hurry, imagine having to use your browser like... For pretty much everything else.

[–] [email protected] 2 points 3 months ago

I did not buy these gaming memory sticks for nothing, bring me more electron!

[–] [email protected] 35 points 3 months ago (3 children)

...this looks like it was written by a supervisor who has no idea what AI actually is, but desperately wants it shoehorned into the next project because it's the latest buzzword.

[–] [email protected] 4 points 3 months ago

I see you've met my employer

[–] [email protected] 4 points 3 months ago

"I saw a new toy on tv, and I want it NOW!"

  • Basically how the technobro mind works.
[–] [email protected] 17 points 3 months ago (1 children)

Guys we need AI on our blockchain web3.0 iot. Just imagine the synergy

[–] [email protected] 3 points 3 months ago

Here we have a straight-shooter with upper management written all over him

[–] [email protected] 29 points 3 months ago

There are already a lot of open models and tools out there. I totally disagree that Linux distros or DEs should be looking to bake in AI features. People can run an LLM on their computer just like they run any other application.

[–] [email protected] 1 points 3 months ago

Try to think less about "communities" and maybe you will be happy.

[–] [email protected] 7 points 3 months ago

General Ludd had some good points tho...

[–] [email protected] 23 points 3 months ago

Gnome and other desktops need to start working on integrating FOSS

In addition to everything everyone else has already said, why does this have anything to do with desktop environments at all? Remember, most open-source software comes from one or two individual programmers scratching a personal itch—not all of it is part of your DE, nor should it be. If someone writes an open-source LLM-driven program that does something useful to a significant segment of the Linux community, it will get packaged by at least some distros, accrete various front-ends in different toolkits, and so on.

However, I don't think that day is coming soon. Most of the things "Apple Intelligence" seems to be intended to fuel are either useless or downright offputting to me, and I doubt I'm the only one—for instance, I don't talk to my computer unless I'm cussing it out, and I'd rather it not understand that. My guess is that the first desktop-directed offering we see in Linux is going to be an image generator frontend, which I don't need but can see use cases for even if usage of the generated images is restricted (see below).

Anyway, if this is your particular itch, you can scratch it—by paying someone to write the code for you (or starting a crowdfunding campaign for same), if you don't know how to do it yourself. If this isn't worth money or time to you, why should it be to anyone else? Linux isn't in competition with the proprietary OSs in the way you seem to think.

As for why LLMs are so heavily disliked in the open-source community? There are three reasons:

  1. The fact that they give inaccurate responses, which can be hilarious, dangerous, or tedious depending on the question asked, but a lot of nontechnical people, including management at companies trying to incorporate "AI" into their products, don't realize the answers can be dangerously innacurate.
  2. Disputes over the legality and morality of using scraped data in training sets.
  3. Disputes over who owns the copyright of LLM-generated code (and other materials, but especiallly code).

Item 1 can theoretically be solved by bigger and better AI models, but 2 and 3 can't be. They have to be decided by the courts, and at an international level, too. We might even be talking treaty negotiations. I'd be surprised if that takes less than ten years. In the meanwhile, for instance, it's very, very dangerous for any open-source project to accept a code patch written with the aid of an LLM—depending on the conclusion the courts come to, it might have to be torn out down the line, along with everything built on top of it. The inability to use LLM output for open source or commercial purposes without taking a big legal risk kneecaps the value of the applications. Unlike Apple or Microsoft, the Linux community can't bribe enough judges to make the problems disappear.

[–] [email protected] 15 points 3 months ago (1 children)

Sounds like something an AI would post. Quick, what color are your eyes?

[–] [email protected] 3 points 3 months ago

Potato colored in my case.

[–] [email protected] 16 points 3 months ago

Good.

The Luddites were right.

load more comments
view more: next ›