Search Articles

Search Results: LLMTraining

Scaling Evolution Strategies to Billion‑Parameter Models with Low‑Rank Perturbations

Scaling Evolution Strategies to Billion‑Parameter Models with Low‑Rank Perturbations

A new paper introduces EGGROLL, a low‑rank perturbation technique that turns the traditionally expensive Evolution Strategies into a viable tool for training billion‑parameter neural networks. By replacing full‑rank noise with a pair of small matrices, the method cuts memory and compute costs while matching the performance of classic ES on reinforcement learning and large‑language‑model benchmarks.
When Large Language Models Get ‘Brain Rot’: Inside AI’s Junk-Data Problem

When Large Language Models Get ‘Brain Rot’: Inside AI’s Junk-Data Problem

New research from UT Austin, Texas A&M, and Purdue suggests that large language models can exhibit a digital form of 'brain rot' when overexposed to junk data from social platforms. For developers and AI leaders, it’s a warning shot: data quality isn’t just a tuning detail—it’s the difference between a reliable co-pilot and a manipulative, brittle system.

Web Admin Declares War on Generic User-Agents, Citing LLM Scraping Epidemic

A prominent technical blogger has implemented aggressive blocking against HTTP requests with generic User-Agent strings, citing an unsustainable flood of crawlers harvesting data for LLM training. This move highlights the escalating tension between website operators and the opaque, resource-intensive scraping fueling AI models.
Anthropic's $1.5B Settlement Sets Landmark Precedent for AI Copyright Battles

Anthropic's $1.5B Settlement Sets Landmark Precedent for AI Copyright Battles

Anthropic agrees to pay $1.5 billion to settle a class-action lawsuit alleging it trained its Claude AI on pirated books, marking the largest copyright recovery in history. The ruling establishes critical boundaries for AI development while forcing the industry to confront ethical data sourcing.
How Moonshot AI Cracked Qualitative LLM Training with Bill James-Inspired Rubrics

How Moonshot AI Cracked Qualitative LLM Training with Bill James-Inspired Rubrics

Facing the challenge of improving language model writing without verifiable metrics, Moonshot AI adopted an unconventional approach inspired by baseball statistician Bill James. By developing targeted, imperfect rubrics to guide reinforcement learning, Kimi K2 achieved top-tier qualitative performance while avoiding reward hacking pitfalls.