Update on the state of AI drug discovery companies: AI Does Not Make It Easy by Derek Lowe
TechTakes
Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.
This is not debate club. Unless it’s amusing debate.
For actually-good tech, you want our NotAwfulTech community
The Bookseller: Penguin Random House underscores copyright protection in AI rebuff
Penguin Random House (PRH) has amended its copyright wording across all imprints globally, confirming it will appear “in imprint pages across our markets”. The new wording states: “No part of this book may be used or reproduced in any manner for the purpose of training artificial intelligence technologies or systems”, and will be included in all new titles and any backlist titles that are reprinted.
Now that the content mafia has realized GenAI isn't gonna let them get rid of all the expensive and troublesome human talent. it's time to give Big AI a wedgie.
Now that the content mafia has realized GenAI isn’t gonna let them get rid of all the expensive and troublesome human talent. it’s time to give Big AI a wedgie.
Considering the massive(ly inflated) valuations running around Big AI and the massive amounts of stolen work that powers the likes of CrAIyon, ChatGPT, DALL-E and others, I suspect the content mafia is likely gonna try and squeeze every last red cent they can out of the AI industry.
Forget Gladwell
All nonfiction writers can end up writing incorrect or controversial things, but why does every Gladwell book push half-formed and inaccurate theories? For years, my loose feeling about Gladwell was that he writes like someone who doesn’t care about being correct, which is not a way I would describe any other author I've encountered. There is something uniquely odd about his work.
has the era of active sabotage of the autoplag inputs begun? let's hope so
They added sleeps to training jobs? Sounds like they deserve a raise for improving energy efficiency instead…
I thought they were gonna do that themselves by feeding on their own outputs littered all over the www. Maybe they can use some help.
that's also happening, but yeah it's going to have to be a team effort
Considering Glaze and Nightshade have been around for a while, and I talked about sabotaging scrapers back in July, arguably, it already has.
Hell, I ran across a much smaller scale case of this a couple days ago:
Not sure how effective it is, but if Elon's stealing your data for his autoplag no matter what, you might as well try to force-feed it as much poison as you can.
It's almost completely ineffective, sorry. It's certainly not as effective as exfiltrating weights via neighborly means.
On Glaze and Nightshade, my prior rant hasn't yet been invalidated and there's no upcoming mathematics which tilt the scales in favor of anti-training techniques. In general, scrapers for training sets are now augmented with alignment models, which test inputs to see how well the tags line up; your example might be rejected as insufficiently normal-cat-like.
I think that "force-feeding" is probably not the right metaphor. At scale, more effort goes into cleaning and tagging than into scraping; most of that "forced" input is destined to be discarded or retagged.
I saw people say they would add 10% opaque layers of the musk with Epstein's accomplice (whos name i forgot for a second and too lazy to look her up) photo. Would be nice if there was a tool to do so automatically. (Not that i post on twitter anymore).
tbh that sounds like a pretty easy script to write! Too bad I am not near a computer rn
Wouldn't really need a script, though. Just open up photoshop or GIMP and add a layer after everything is finished.
It would be funny if someone was literally beating up servers with a wooden shoe.
"percussive maintenance"
momty python style giant sabot descends on Microsoft data centre