self

joined 1 year ago
MODERATOR OF
[–] [email protected] 0 points 5 hours ago

you’re identical to the other poster in both tone and content so either you’re using the same LLM to write your posts or you’re otherwise extremely familiar with each other. either way we don’t particularly need you here, but before you go:

Also, from my experience as a student current LLMs are better at explaining than the average teacher is.

Overall, I think our education system is largely outdated and schools of the future won’t look like one teacher explaining and 25 people more or less not listening

one day you’ll hopefully grow old enough to realize how lazy this shit looks to other people

[–] [email protected] 0 points 5 hours ago (1 children)

With readily available video, it is criminal to continue live lecturing and presentation of static information. There is no chance that the live presentation of information is anywhere near the quality of a polished and edited video. There is very little chance that any given lecturer is truly the best at presenting such information.

christ

However, the ability for AI to adapt to any functional thought and help with individualized problem solving is something that no teacher is capable of with more than one student at a time.

it doesn’t do this

It is not always correct, but it is in the same realm of accuracy as an above average teacher. Maybe you too were aware of just how many teachers did not even know the subjects they were tasked with teaching in primary school, I certainly was.

I’m sorry your teachers sucked bad enough you could replace them with a prerecorded video and a statistical language model that’s notorious for generating confident, dangerous lies. I don’t think most kids should have that kind of experience in school though, and if they are currently maybe we should do what it takes (funding, regulation, strikes) to not go in that direction.

[–] [email protected] 0 points 1 day ago

that’s great thanks

[–] [email protected] 0 points 1 day ago

my IDE, notepad.exe, didn’t support capitalizing words until they added copilot to it. so therefore qed editors couldn’t do that without LLMs. computer science is so easy!

[–] [email protected] 0 points 1 day ago (1 children)

Brownlee, who will take us through “mind-blowing demonstrations of AI’s capabilities,” is a curious addition — he’s previously complained of his work being stolen to feed AI. But a big enough check is a pretty convincing counterargument. [Twitter, archive]

I’m disappointed but not at all surprised Marques is involved in this. I think he was caught by surprise by how popular his mildly AI-critical videos were, and this is him doing damage control and trying to get back into the grifters’ good graces.

[–] [email protected] 0 points 1 day ago

it’s always fun when techbros speedrun the narcissist’s prayer like this

[–] [email protected] 0 points 1 day ago (4 children)

oh my god the maitai guy’s actually getting torn apart in the comments

Yeah some of you guys are very good at hacking things. We expected this to get broken eventually, but didn't anticipate how many people would be trying for the bounty, and their persistence. Our logs show over 2000 "saves" before 1 got through. We'll keep trying to get better, and things like this game give us an idea on how to improve.

after it’s pointed out 2000 near-misses before a complete failure is ridiculously awful for anything internet-facing:

Maitai helps LLMs adhere to the expectations given to them. With that said, there are multiple layers to consider when dealing with sensitive data with chatbots, right? First off, you'd probably want to make sure you authenticate the individual on the other end of the convo, then compartmentalize what data the LLM has access to for only that authenticated user. Maitai would be just 1 part of a comprehensive solution.

so uh, what exactly is your product for, then? admit it, this shit just regexed for the secret string on output, that’s why the pirate poem thing worked

[–] [email protected] 0 points 2 days ago

hahaha ok fuck off now

[–] [email protected] 0 points 2 days ago (1 children)

please be very careful with the VSLAM (camera+sensors) ones, and note carefully that iRobot avoided responsibility for this by claiming the impacted people were testers (a claim the alleged testers appear to disagree with)

[–] [email protected] 0 points 2 days ago

and I’m over here like “what if we just included a peltier element… but bigger” and then the satellite comes out covered in noctua fans and RGB light strips

[–] [email protected] 0 points 2 days ago (4 children)

it’s weird how they’re pumping this specific bullshit out now that a common talking point is “well you can’t say you hate AI, because the non-generative bits do actually useful things like protein folding”, as if any of us were the ones who chose to market this shit as AI, and also as if previous AI booms weren’t absolutely fucking turgid with grifts too

[–] [email protected] 0 points 2 days ago

What LLMs actually do, i.e. Trabant tasks

more of a Power Wheels Barbie Jeep whose battery got left out in the sun too long, but I’ll allow it

 

after the predictable failure of the Rabbit R1, it feels like we’ve heard relatively nothing about the Humane AI Pin, which released first but was rapidly overshadowed by the R1’s shittiness. as it turns out, the reason why we haven’t heard much about the Humane AI pin is because it’s fucked:

Between May and August, more AI Pins were returned than purchased, according to internal sales data obtained by The Verge. By June, only around 8,000 units hadn’t been returned, a source with direct knowledge of sales and return data told me. As of today, the number of units still in customer hands had fallen closer to 7,000, a source with direct knowledge said.

it’s fucked in ways you might not have seen coming, but Humane should have:

Once a Humane Pin is returned, the company has no way to refurbish it, sources with knowledge of the return process confirmed. The Pin becomes e-waste, and Humane doesn’t have the opportunity to reclaim the revenue by selling it again. The core issue is that there is a T-Mobile limitation that makes it impossible (for now) for Humane to reassign a Pin to a new user once it’s been assigned to someone.

 

as I was reading through this one, the quotes I wanted to pull kept growing in size until it was just the whole article, so fuck it, this one’s pretty damning

here’s a thin sample of what you can expect, but it gets much worse from here:

Internal conversations at Nvidia viewed by 404 Media show when employees working on the project raised questions about potential legal issues surrounding the use of datasets compiled by academics for research purposes and YouTube videos, managers told them they had clearance to use that content from the highest levels of the company.

A former Nvidia employee, whom 404 Media granted anonymity to speak about internal Nvidia processes, said that employees were asked to scrape videos from Netflix, YouTube, and other sources to train an AI model for Nvidia’s Omniverse 3D world generator, self-driving car systems, and “digital human” products. The project, internally named Cosmos (but different from the company’s existing Cosmos deep learning product), has not yet been released to the public.

 

so Andreessen Horowitz posted another manifesto just over a week ago and it’s the most banal fash shit you can imagine:

Regulatory agencies have been green lit to use brute force investigations, prosecutions, intimidation, and threats to hobble new industries, such as Blockchain.

Regulatory agencies are being green lit in real time to do the same to Artificial Intelligence.

does this shit ever get deeper than Regulation Bad? fuck no it doesn’t. is this Horowitz’s attempt to capitalize on the Supreme Court’s judiciary coup? you fucking bet.

here’s some more banal shit:

We find there are three kinds of politicians:

Those who support Little Tech. We support them.

Those who oppose Little Tech. We oppose them.

Those who are somewhere in the middle – they want to be supportive, but they have concerns. We work with them in good faith.

I find there are three kinds of politicians:

  • those who want hamburger. I give them hamburger.
  • those who abstain from hamburger. I do not give them hamburger.
  • those who have questions about hamburger. I refer them to the shift supervisor in good faith.
 

who could have seen this coming, other than everyone who told the homebrew tree inverter guy this was a bad idea they absolutely shouldn’t do

0
test post (awful.systems)
 

federation may be fucked :(

 

(via https://hachyderm.io/@jbcrawford/112202942593125987, archive: https://archive.is/VnqRZ)

surprise, Amazon’s godawful surveillance grocery stores were just exploiting hidden labor and calling it innovation, and even that was too expensive

even worse, the few times I’ve seen one of these fucking things in the wild, it still had 1-2 employees hovering near the entrance to make sure nobody did the utterly obvious (fuck with the payment system and get free shit), a job that’s also known as a fucking cashier, but with much worse pay, much harder labor (physically stopping shoplifters), and no counter to lean on or opportunity to even sit down

 

(here’s a Verge article about the Waymo car getting burned during a Chinese New Year celebration)

a self-driving car got destroyed (to a round of applause from the crowd) in San Francisco! will the robot car fans on the orange site take this opportunity to explore why the tech seems to be extremely unpopular among the populations of the cities where it’s deployed?

of course the fuck not, time to spin the wheel of racist dog whistles and see which one we land on! a note to the roving orange site fans (hi, fuck off), these replies are either heavily upvoted or have broad agreement in the thread (or I’m posting them here cause I want to laugh at some stupid shit, you don’t dictate the terms of my enjoyment)

This isn't a revolt against AI. SF attracts anarchist mobs and they'll vandalize buses, trains, police cars, bikes, whatever is around.

we’re off to a strong start with some bullshit straight from musk’s twitter (which he stole from the fever dreams of the conservatives on his platform)

Alternatively: this is San Francisco where on a good day the locals don’t need much excuse to set fire to a car (although I usually associate it with the Giants winning a World Series) and this poor dumb stupid driverless Waymo drove into a celebratory and by the looks of it somewhat drunken crowd on the Streets of Chinatown during the Chinese New Year where in following its prime directive to do no harm, it got itself stuck up the creek without a paddle so to speak. Waymo probably should have accounted for that ahead of time and told their cars not to go near Chinatown this evening.

remember that no matter what, the robot car is the victim here. there’s no chance Waymo was doing anything dangerous or assholeish in the area; much like robocop, the car is an innocent victim of its fucking prime directives??? and you wouldn’t set fire to robocop, would you?

This is a hilarious take. A few youths went bonkers and defaced private property. Has nothing to do with philosophical beliefs or a Big Tech agenda. You should debate the finer points of the Big Tech agenda with them while they run up to you in a maddened rage.

yeah! I can’t wait until these angry mobs set fire to your robot car body! then you’ll see!

Arguments about driverless cars aside, the youth in this country are seriously lost. It only takes one generation of poor parenting and poor civic policies to ruin a culture.

this one is downvoted, but this reply isn’t:

Sounds like they were right. The youth at that point was lost, and are now raising people who will literally burn down a waymo for fun, or because of some horrifically ignorant idea about fairness.

oh you poor woke kids don’t like when shitty dangerous robot cars are on the streets? are you gonna start crying about how it’s “unfair” they’re covering up pedestrian injuries and traffic accidents now? your grandpa would never stand for this

 

(via mastodon)

 

remember, regardless of how outspoken you are in life, nothing will stop the capitalists from reanimating your defiled corpse into a shitheaded centrist zombie if there’s a buck in it:

“I'd just like to say that as much as I think billionaires are destroying the fabric of society with unchecked greed and blatant self-interest at the expense of basic human rights for everyone else, it is a little strange to me that people get mad at them. People are the ones who gave them the money in the first place," the AI Carlin said.

(editor’s note: the above is supposed to be a joke from the comedy special these fucking assholes hijacked Carlin’s corpse to promote. I can’t find the punchline, but it’s supposed to be a joke)

 

from the linked github thread:

Your project is in violation of the AGPL, and you have stated this is intentional and you have no plans to open source it. This is breaking the law, and as such I've began to help you with the first steps of re-open sourcing the plugin.

the project author (who gets paid for violating the AGPL via patreon) responds like a mediocre crypto grifter and insists their violation of the law be debated on the discord they control (where their shitty community can shout down the reporter):

While keeping code private doesn't guarantee security, it does make it harder for bad actors to keep up with changes. You are welcome to debate this matter in the MakePlace discord: https://discord.com/invite/YuvcPzCuhq If you are able to convince the MakePlace community that keeping the code open-source is better, I will respect the wishes of the community.

aaaand the smackdown:

Respectfully, I won't attempt to "debate" or "convince" anyone; I'm leaving this pull request and my fork here for others to see and use. It is not a matter of "better"; you are violating a software license and the law. It does not "make it harder" for anyone; Harmony hooking exists, IL modification exists, you can modify plugins from other plugins.

 

(via Timnit Gebru)

Although the board members didn’t use the language of abuse to describe Altman’s behavior, these complaints echoed some of their interactions with Altman over the years, and they had already been debating the board’s ability to hold the CEO accountable. Several board members thought Altman had lied to them, for example, as part of a campaign to remove board member Helen Toner after she published a paper criticizing OpenAI, the people said.

The complaints about Altman’s alleged behavior, which have not previously been reported, were a major factor in the board’s abrupt decision to fire Altman on Nov. 17, according to the people. Initially cast as a clash over the safe development of artificial intelligence, Altman’s firing was at least partially motivated by the sense that his behavior would make it impossible for the board to oversee the CEO.

For longtime employees, there was added incentive to sign: Altman’s departure jeopardized an investment deal that would allow them to sell their stock back to OpenAI, cashing out equity without waiting for the company to go public. The deal — led by Joshua Kushner’s Thrive Capital — values the company at almost $90 billion, according to a report in the Wall Street Journal, more than triple its $28 billion valuation in April, and it could have been threatened by tanking value triggered by the CEO’s departure.

huh, I think this shady AI startup whose product is based on theft that cloaks all its actions in fake concern for humanity might have a systemic ethics problem

view more: next ›