froztbyte

joined 2 years ago
[–] [email protected] 0 points 4 days ago (2 children)

meme-consumerism is so fucking diseased, ugh

[–] [email protected] 0 points 4 days ago* (last edited 4 days ago) (1 children)

lol, yeah

"perverse incentives rule everything around me" is a big thing (observable) in "startup"[0] world because everything[1] is about speed/iteration. for example: why bother spending a few weeks working out a way to generate better training data for a niche kind of puzzle test if you can just code in "personality" and make the autoplag casinobot go "hah, I saw a puzzle almost like this just last week, let's see if the same solution works...."

i.e. when faced with a choice of hard vs quick, cynically I'll guess the latter in almost all cases. there are occasional exceptions, but none of the promptfondlers and modelfarmers are in that set imo

[0] - look, we may wish to argue about what having billions in vc funding categorizes a business as. but apparently "immature shitderpery" is still squarely "startup"

[1] - in the bayfucker playbook. I disagree.

[–] [email protected] 0 points 5 days ago* (last edited 5 days ago) (3 children)

(excuse possible incoherence it’s 01:20 and I’m entirely in filmbrain (I’ll revise/edit/answer questions in morning))

re (1): while that is a possibility, keep in mind that all this shit also operates/exists in a metrics-as-targets obsessed space. they might not present end user with hit% but the number exists, and I have no reason to believe that isn’t being tracked. combine that with social effects (public humiliation of their Shiny New Model, monitoring usage in public, etc etc) - that’s where my thesis of directed prompt-improvement is grounded

re (2): while they could do something like that (synthetic derivation, etc), I dunno if that’d be happening for this. this is outright a guess on my part, a reach based on character based on what I’ve seen from some the field, but just…..I don’t think they’d try that hard. I think they might try some limited form of it, but only so much as can be backed up in relatively little time and thought. “only as far as you can stretch 3 sprints” type long

(the other big input in my guesstimation re (2) is an awareness of the fucked interplay of incentives and glorycoders and startup culture)

[–] [email protected] 0 points 5 days ago (5 children)

I would be 0% surprised to learn that the modelfarmers "iterated" to "hmm, people are doing a lot of logic tests, let's handle those better" and that that's what gets here

(I have no evidence for this, but to me it seems a completely obvious/evident way for them to try keep the party going)

[–] [email protected] 0 points 5 days ago (1 children)
[–] [email protected] 0 points 5 days ago (3 children)

kiiiiiiind of a pretty weird post to be making with a thread necro-reply tbh

[–] [email protected] 0 points 5 days ago

thought up a new simile for agi noises:

all the “super-intelligent agi” stuff is them rattling a sheet of metal and calling it thunder

[–] [email protected] 0 points 5 days ago* (last edited 5 days ago)

this has been happening for a while, just getting coverage again now. first coverage was months ago. morphed/evolved pretty quickly out of the typosquatting shit

((a lot of people in the) security space absolutely fucking loves "giving names" to things that have been (known to be) happening before, and acting like suddenly they're the ones who first saw the thing. see this nonsense for another good example of that happening)

[–] [email protected] 0 points 1 week ago (3 children)

for the same reasons as this, not really a thing I'd post, even in jest

the existence of these callcenters (often in india, but hardly only there - much of africa is beset by the same problem) is an outright fucking feature of the years-long capitalist market optimisation hell-loop. and as annoying as their "output" (for lack of a better word) may be in one's daily life, at the end of the day it's still a bunch of people at the bottom rung getting fucked

[–] [email protected] 0 points 1 week ago (1 children)

+1! and there's so many better options anyway!

right off the top of my head (in addition to inside/instead): incognito, insecurely

 

Unfortunately I can’t snip from mobile easily now, but maybe someone else can archive it and comment with archive link?

 

Too tired to sneer at the book myself right now but the article doesn’t pull punches either

Figured it’s worth posting since the book author has featured here more than once recently and has definitely been an enabler to The Shit

1
submitted 1 year ago* (last edited 1 year ago) by [email protected] to c/[email protected]
 

Not entirely the usual fare, but i figured some here would appreciate it

I often rag on the js/node/npm ecosystem for being utter garbage, and this post is a quite a full demonstration of many of the shortcomings and outright total design failures present in that space

 

starting out[0] with "I was surprised by the following results" and it just goes further down almost-but-not-quite Getting It Avenue

close, but certainly no cigar

choice quotes:

Why is it impressive that a model trained on internet text full of random facts happens to have a lot of random facts memorized? … why does that in any way indicate intelligence or creativity?

That’s a good point.

you don't fucking say

I have a website (TrackingAI.org) that already administers a political survey to AIs every day. So I could easily give the AIs a real intelligence test, and track that over time, too.

really, how?

As I started manually giving AIs IQ tests

oh.

Then it proceeds to mis-identify every single one of the 6 answer options, leading it to pick the wrong answer. There seems to be little rhyme or reason to its misidentifications

if this fuckwit had even the slightest fucking understanding of how these things work, it would be glaringly obvious

there's plenty more, so remember to practice stretching before you start your eyerolls

 

[open scene]

background: a brightly lit airy Social Gathering space with multicoloured furniture (meeting rooms are so 2010). people have been arriving in clumps of 2~5 for over 30 minutes, and the presentation can start soon

sundar: I want to thank you all for coming. this one should be quick today.

* sundar briefly sweeps his eyes across the room before continuing *

sundar: guys! GUYS! we made the prompt VIDEO CAPABLE! it can follow A STREAMING SEQUENCE OF IMAGES!! you can immediately start testing this from your corporate account (whispers if you're in the right orgs). for the public scoff, we'll start with Ask Us pricing in a few months, and we'll force it on the usual product avenues. the office and mail suites stand ready to roll out the integration updates before anyone can ask. you know how the riffraff gets....

* some motion and noise in the back *

sundar: ... sorry melanie, what's that? speak up melanie I can't hear your question. you know how much that mask muffles your voice...

* a game of broken telephone for moving a handheld microphone to the back of the room ensues *

melanie: hi sundar, congratulations to the team for their achievement. I wanted to ask: how does gemini pro solve the issues other models have faced? what new innovations have been accomplished? how is it dealing with the usual issues of correctness, energy consumption, cultural contexts? how is it trained on areas where no datasets exist? were any results sourced from cooperation with the AI ethics and responsibility workgroups that have found so many holes in our previous models?

sundar: * smiles brightly, stares directly into middle of crowd. moves hand to the electronic shutter control, and starts pressing the increase button multiple times until shutter is entirely opaque *

[sundar walks off into the fake sunset, breaks open the boardroom whiskey]

[inside the private exec room]

sundar: FUCK! that was too close. didn't we fire those types already in the last layoffs...? someone get me HR, we need to do something

[end scene]

 

archive

"There's absolutely no probability that you're going to see this so-called AGI, where computers are more powerful than people, in the next 12 months. It's going to take years, if not many decades, but I still think the time to focus on safety is now," he said.

just days after poor lil sammyboi and co went out and ran their mouths! the horror!

Sources told Reuters that the warning to OpenAI's board was one factor among a longer list of grievances that led to Altman's firing, as well as concerns over commercializing advances before assessing their risks.

Asked if such a discovery contributed..., but it wasn't fundamentally about a concern like that.

god I want to see the boardroom leaks so bad. STOP TEASING!

“What we really need are safety brakes. Just like you have a safety break in an elevator, a circuit breaker for electricity, an emergency brake for a bus – there ought to be safety breaks in AI systems that control critical infrastructure, so that they always remain under human control,” Smith added.

this appears to be a vaguely good statement, but I'm gonna (cynically) guess that it's more steered by the fact that MS now repeatedly burned their fingers on human-interaction AI shit, and is reaaaaal reticent about the impending exposure

wonder if they'll release a business policy update about usage suitability for *GPT and friends

 

archive (e: twitter [archive] too, archive for nitter seems a bit funky)

it'd be nice if these dipshits, like, came off a factory line somewhere. then you could bin them right at the QC failure

 

Mr. Altman’s departure follows a deliberative review process [by the board]

"god, he's really cost us... how much can we get back?"

which concluded that he was not consistently candid in his communications with the board

not only with the board, kids

hindering its ability to exercise its responsibilities

you and me both, brother

 

nitter archive

just in case you haven't done your daily eye stretches yet, here's a workout challenge! remember to count your reps, and to take a break between paragraphs! duet your score!

oh and, uh.. you may want to hide any loose keyboards before you read this. because you may find yourself wanting to throw something.

 

will this sure is gonna go well :sarcmark:

it almost feels like when Google+ got shoved into every google product because someone had a bee in their bonnet

flipside, I guess, is that we'll soon (at scale!) get to start seeing just how far those ideas can and can't scale

view more: ‹ prev next ›