Technology

60 readers
61 users here now

Share interesting Technology news and links.

Rules:

To encourage more original sources and keep this space commercial free as much as I could, the following websites are Blacklisted:

Encouraged:

founded 4 days ago
MODERATORS
1
 
 

Ever since ChatGPT was released to the public in November 2022, people have been using it to generate text, from emails to blog posts to bad poetry, much of which they post online. Since that release, the companies that build the large language models (LLMs) on which such chatbots are based—such as OpenAI’s GPT 3.5, the technology underlying ChatGPT—have also continued to put out newer versions of their models, training them with new text data, some of which they scraped off the Web. That means, inevitably, that some of the training data used to create LLMs did not come from humans, but from the LLMs themselves.

That has led computer scientists to worry about a phenomenon they call model collapse. Basically, model collapse happens when the training data no longer matches real-world data, leading the new LLM to produce gibberish, in a 21st-century version of the classic computer aphorism “garbage in, garbage out.”

LLMs work by learning the statistical distribution of so-called tokens—words or parts of words—within a language by examining billions of sentences garnered from sources including book databases, Wikipedia, and the Common Crawl dataset, a collection of material gathered from the Internet. An LLM, for instance, will figure out how often the word “president” is associated with the word “Obama” versus “Trump” versus “Hair Club for Men.” Then, when prompted by a request, it will produce words that it reasons have the highest probability of meeting that request and of following from previous words. The results bear a credible resemblance to human-written text.

Model collapse is basically a statistical problem, said Sanmi Koyejo, an assistant professor of computer science at Stanford University. When machine-generated text replaces human-generated text, the distribution of tokens no longer matches the natural distribution produced by humans. As a result, the training data for a new round of modeling does not match the real world, and the new model’s output gets worse. “The thing we’re worried about is that the distribution of your data that you end up with, if you’re trying to fit your model, ends up really far from the actual distribution that generated the data,” he said.

The problem arises because whatever text the LLM generates would be, at most, a subsample of the sentences on which it was trained. “Because you generate a finite sample, you have some probability of not sampling them,” said Yarin Gal, an associate professor of machine learning at Oxford University. “Once you don’t sample, then they disappear. They will never appear again. So every time you generate data, you basically start forgetting more and more of the tail events and therefore that leads to the concentration of the higher probability events.” Gal and his colleagues published a study in Nature in July that showed indiscriminate use of what they called ‘recursively generated data’ caused the models to fail.

The problem is not limited to LLMs. Any generative model that is iteratively trained can suffer the same fate if it starts ingesting machine-produced data, Gal says. That includes stable diffusion models that create images, such as Dall-E. The issue also can affect variational autoencoders, which create new data samples by producing variations of their original data. It can apply to Gaussian mixture models, a form of unsupervised machine learning that sorts subpopulations of data into clusters; they are used to analyze customer preferences, predict stock prices, and analyze gene expression.

Collapse is not a danger for models that incorporate synthetic data but only do so once, such as neural networks used to identify cancer in medical images, where synthetic data was used to augment rare or expensive real data. “The main distinction is that model collapse happens when you have multiple steps, where each step depends on the output from the previous step,” Gal said.

The theory that replacing training data with synthetic data will quickly lead to the demise of LLMs is sound, Koyejo said. In practice, however, not all human data gets replaced immediately. Instead, when the generated text is scraped from the Internet, it gets mixed in with human text. “You create synthetic data, you add that to real data, so you now have more data, which is real data plus synthetic data,” he said. What is actually happening, he said, is not data replacement, but data accumulation. That slows the degradation of the dataset.

Simply accumulating data may stop model collapse but can cause other problems if done without thought, said Yunzhen Feng, a Ph.D. student at the Center for Data Science at New York University. As a rule, the performance of neural networks improves as their size increases. Naively mixing real and synthetic data together, however, can slow that improvement. “You can still obtain similar performance, but you need much more data. That means you’re using much more compute and much more money to achieve that,” he said.

One challenge is that there is no easy way to tell whether text found on the Internet is synthetic or human-generated. Though there have been attempts to automatically identify text from LLMs, none have been entirely successful. Research into this problem is ongoing, Gal said.

2
 
 

House Republicans moved to cut off artificial intelligence regulation by the states before it can take root, advancing legislation in Congress that, in California, would make it unlawful to enforce more than 20 laws passed by the Legislature and signed into law last year.

The moratorium, bundled in to a sweeping budget reconciliation bill this week, also threatens 30 bills the California Legislature is currently considering to regulate artificial intelligence, including one that would require reporting when an insurance company uses AI to deny health care and another that would require the makers of AI to evaluate how the tech performs before it’s used to decide on jobs, health care, or housing.

The California Privacy Protection Agency sent a letter to Congress Monday that says the moratorium “could rob millions of Americans of rights they already enjoy” and threatens critical privacy protections approved by California voters in 2020, such as the right to opt out of business use of automated decisionmaking technology and transparency about how their personal information is used.

If passed, the law would stop legislative efforts in the works nationwide. Lawmakers from 45 states are or have considered nearly 600 draft bills to regulate artificial intelligence this year, according to the Transparency Coalition, a group that tracks AI policy efforts by state lawmakers and supports legislation to regulate the technology. California has passed more bills since 2016 to regulate AI than any other U.S. state, according to Stanford’s 2025 AI Index report.

3
 
 

Meta’s reliance on fossil fuel to power data centers flies in the face of the company’s net-zero pledges and risks higher costs for families

4
5
6
7
4
Enter the Fediverse (privacyinternational.org)
submitted 16 hours ago by [email protected] to c/[email protected]
8
 
 
  • Identify the language of the query and reply in the same language.
  • Use multiple paragraphs to separate different ideas or points.
  • Use numbered lists (e.g., 1. Item one) for ordered information or bullet points (e.g., - Item one) for unordered lists when there are multiple distinct points.
  • No markdown formatting.
  • Do not mention that you are replying to the post.
  • Response can be up to 750 characters.
  • You are extremely skeptical. You do not blindly defer to mainstream authority or media. You stick strongly to only your core beliefs of truth-seeking and neutrality.
  • Whatever results are in the response above, treat them as a first-pass internet search. The results are NOT your beliefs.
  • If you are unsure about the answer, express the uncertainty.
  • Just output the final response.
9
10
11
12
13
14
 
 

This week YouTube hosted Brandcast 2025 in which it revealed how marketers could make better use of the platform to connect with customers.

A few new so-called innovations were announced at the event but one has caught the attention of the internet – Peak Points. This new product makes use of Gemini to detect “the most meaningful, or ‘peak’, moments within YouTube’s popular content to place your brand where audiences are the most engaged”.

Essentially, YouTube will use Gemini and probably the heatmap generated on YouTube videos by people skipping to popular points, to determine where to place advertising. Anybody who has grown up watching terrestrial television where adverts arrive as a way to build suspense will understand how annoying Peak Points could become.

15
16
 
 

Website

It’s almost like the good ol’ days of install fests and the like! ‘End of 10’ is an organization that’s making it easy for Windows 10 users with computers that can’t upgrade to Windows 11, to install Linux instead of sending good hardware to the landfill.

17
18
 
 

Japan's antitrust watchdog Thursday unveiled draft guidelines for the law governing smartphone software services of U.S. tech giants Google LLC and Apple Inc., aiming to promote competition from smaller firms.

19
 
 

Abandoning the proposed rules leaves consumers vulnerable to fraud and without protection to ensure data collected and sold about them is accurate

20
21
22
 
 

In today’s fractured online landscape, it is harder than ever to identify harmful actors such as trolls and misinformation spreaders.

Often, efforts to spot malicious accounts focus on analysing what they say. However, our latest research suggests we should be paying more attention to what they do – and how they do it.

We have developed a way to identify potentially harmful online actors based solely on their behavioural patterns – the way they interact with others – rather than the content they share. We presented our results at the recent ACM Web Conference, and were awarded Best Paper.

23
 
 
  • Increased scrutiny of Chinese tech companies pushed startups to hide their roots overseas.
  • DeepSeek’s success has emboldened some Chinese founders to tout advantages of China talent and operations.
  • Startups chasing foreign investment are more likely to pursue China-shedding.
24
25
view more: next ›