Anthropic Accuses Three Chinese AI Labs of Copying Claude at Scale
February 24, 2026
D.A.D. today covers 13 stories from 3 sources. What's New, What's Innovative, What's Controversial, What's in the Lab, and What's in Academe.
D.A.D. Joke of the Day: My AI writes emails so polished my coworkers think I've finally matured. Little do they know I've just outsourced my personality.
What's New
AI developments from the last 24 hours
Debate Rages Over Unintended Risks of Age-Verification Laws
An online post has raised an underexplored paradox about age verification requirements for social media: proving users are old enough to use a platform requires collecting exactly the kind of sensitive data—government IDs, facial scans, behavioral tracking—that privacy laws aim to protect. Major platforms handle this differently: Meta uses video selfies for facial age estimation on Instagram, TikTok scans public videos to infer ages, and Google combines behavioral signals with ID checks. A Wired report notes Roblox's age-verification system is allegedly being circumvented by users selling child-aged accounts to adults seeking access to restricted areas. The post attracted over 1,000 comments on Hacker News, reflecting how deeply the tension between child safety and data privacy resonates across the tech community.
Why it matters: As regulators push stricter child-safety rules while simultaneously tightening data protection laws, platforms face an increasingly impossible compliance position—and users face growing pressure to hand over biometric and identity data just to prove they're allowed online.
Discuss on Hacker News · Source: spectrum.ieee.org
AI Tools Help Developers Rewrite 25,000 Lines of Browser Code in Two Weeks
The Ladybird browser project is switching from C++ to Rust for memory safety, starting with its JavaScript engine. Founder Andreas Kling used Claude Code and Codex to help translate roughly 25,000 lines of code in about two weeks—a task he estimated would have taken months by hand. The port passed all 52,898 conformance tests and 12,461 regression tests with zero failures and no performance loss. On Hacker News, some noted the original submission title omitted the AI assistance.
Why it matters: This is a concrete case study in AI-assisted code migration at scale—the kind of tedious, high-stakes translation work that typically consumes months of developer time, compressed dramatically with AI tooling.
Discuss on Hacker News · Source: ladybird.org
Only Five AI Models Ace a Deceptively Simple Reasoning Test
A simple reasoning test stumped most major AI models: 'The car wash is 50 meters away. Should I walk or drive?' The correct answer is drive—you need the car at the car wash. But 42 of 53 models said walk. Developer Felix Wunderlich ran the test 10 times per model for consistency. Only five achieved 100% accuracy: Claude Opus 4.6, Gemini 2.0 Flash Lite, Gemini 3 Flash, Gemini 3 Pro, and Grok-4. GPT-5 managed just 7 of 10. Entire model families—all Llama, all Mistral—scored zero.
Why it matters: Simple tests like this expose how models can excel at complex tasks while missing obvious real-world logic—a reminder that benchmark performance doesn't always translate to reliable common-sense reasoning in everyday prompts.
Discuss on Hacker News · Source: opper.ai
Wolfram Pitches Its Math Engine as Essential AI Infrastructure
Stephen Wolfram is positioning Wolfram Language and Wolfram|Alpha as essential infrastructure for AI systems, arguing that LLMs need external tools for precise computation and verified knowledge—capabilities they struggle with natively. The pitch builds on their ChatGPT plugin from early 2023. Wolfram frames his 40-year-old computational language not just as a calculator for AI, but as a medium through which AI systems could 'think' computationally. The announcement is strategic positioning rather than a new product launch.
Why it matters: This signals Wolfram's bet that the AI stack will include specialized reasoning layers, positioning the company as picks-and-shovels infrastructure rather than competing with foundation model makers directly.
Discuss on Hacker News · Source: writings.stephenwolfram.com
What's Innovative
Clever new use cases for AI
Interactive Timeline Tracks 171 AI Models From 2017 to 2026
An open-source project called AI Timeline offers an interactive visualization of 171 large language models from the original Transformer paper (2017) through GPT-5.3 (projected for 2026). The tool lets users filter by open vs. closed models, key milestones, and providers. Early feedback on Hacker News flagged missing models like GPT-J and GPT-NeoX, and users requested features like architecture evolution views and trend predictions.
Why it matters: As the model landscape fragments across dozens of providers, a reference timeline helps teams understand which models are derivatives of what—context that matters when evaluating vendors or explaining AI history to stakeholders.
Discuss on Hacker News · Source: llm-timeline.com
What's Controversial
Stories sparking genuine backlash, policy fights, or heated disagreement in the AI community
Anthropic Accuses Three Chinese AI Labs of Copying Claude at Scale
Anthropic claims three Chinese AI companies—MiniMax, DeepSeek, and Moonshot—distilled Claude's capabilities at scale using fraudulent accounts. The company alleges the firms created over 24,000 fake accounts and generated more than 16 million exchanges with Claude to extract its outputs for training their own models. Anthropic also said it noticed attempts to modify outputs to conform with Chinese censorship rules. Community reaction has been skeptical, with observers noting the irony given ongoing criticism of Western AI labs for training on copyrighted materials without permission. Others questioned whether such distillation practices might be more widespread across the industry.
Why it matters: This escalates tensions between US and Chinese AI labs and raises uncomfortable questions about industry-wide training practices—if true, it suggests model distillation may be a quiet norm that's only controversial when competitors do it.
Discuss on Hacker News · Source: anthropic.com
What's in the Lab
New announcements from major AI labs
OpenAI Abandons Key Coding Benchmark, Cites Data Contamination
OpenAI announced it will stop evaluating its models on SWE-bench Verified, a widely-used benchmark for measuring AI coding ability on real GitHub issues. The company claims the benchmark has become unreliable due to contamination—flawed test cases and training data leakage that inflate scores artificially. SWE-bench has been the go-to metric for comparing coding agents from Anthropic, Google, and others. OpenAI did not specify which benchmark it would use instead.
Why it matters: When a major lab abandons a benchmark, it signals either genuine measurement problems or strategic repositioning—either way, expect coding agent comparisons to get murkier before they get clearer.
OpenAI Launches Enterprise Program for Production AI Deployments
OpenAI launched Frontier Alliance Partners, an enterprise program aimed at helping companies move AI projects from pilot phase to production deployment. The program focuses on secure, scalable agent implementations—the autonomous AI systems that can take actions on users' behalf. No details on pricing, participating partners, or specific capabilities were announced.
Why it matters: Signals OpenAI is prioritizing the 'last mile' problem many companies face: getting AI out of experiments and into actual business operations, particularly as agentic AI becomes the next competitive battleground.
What's in Academe
New papers on AI and its effects from researchers
Massive Video Dataset Shows AI Can Learn Reasoning Skills It Wasn't Taught
Researchers released the Very Big Video Reasoning (VBVR) Dataset, a resource for testing AI systems' ability to reason about video content. At over one million video clips and 200 reasoning tasks, it's roughly 1,000 times larger than existing datasets. The scale enabled one of the first large-scale studies of how video reasoning improves with more training data—researchers report early signs that models can generalize to reasoning tasks they weren't explicitly trained on, a capability that typically emerges only at scale.
Why it matters: Video understanding remains a frontier for AI—this benchmark could accelerate development of systems that analyze surveillance footage, meeting recordings, or product demos without manual tagging.
Plain-English Commands Can Now Control Optical Network Hardware
Researchers developed AgentOptics, an AI framework that lets natural language commands control complex optical network equipment—fiber links, wavelength multiplexers, and similar infrastructure. In testing across 410 tasks, the system achieved 87-99% success rates, compared to just 50% for approaches where AI generates code to run on devices. The framework uses a standardized tool layer to translate plain-English instructions into device actions across eight types of optical hardware.
Why it matters: This is telecom infrastructure research, but it signals how AI agents may eventually manage network operations that underpin enterprise connectivity—potentially reducing the specialized expertise needed to provision and troubleshoot high-speed optical links.
Algorithm Claims Order-of-Magnitude Speedup for Economic Simulations
Researchers developed Recurrent Structural Policy Gradient (RSPG) for solving Mean Field Games—mathematical models used to simulate how large populations of agents make decisions under uncertainty. The team claims their approach converges an order of magnitude faster than existing methods and can handle scenarios where agents must account for shared historical information. They also released MFAX, an open-source framework for building these simulations.
Why it matters: This is specialized research infrastructure—potentially relevant for economists modeling market behavior or researchers in logistics and crowd dynamics, but unlikely to affect most business workflows near-term.
LLM-Based Optimizer Claims to Solve Complex Problems Without Traditional Math
Researchers developed AdaEvolve, a framework that uses LLMs to solve optimization problems without requiring traditional mathematical gradients. The system works by having the AI adaptively adjust its search strategy at three levels—fine-tuning its exploration intensity, routing computational resources to promising approaches, and generating entirely new tactics when progress stalls. The researchers claim it outperforms existing open-source methods across 185 different optimization challenges, including logistics-style combinatorial problems and algorithm design tasks, though specific performance numbers weren't provided.
Why it matters: If validated, this suggests LLMs could handle complex business optimization problems—scheduling, resource allocation, supply chain routing—that currently require specialized software or operations research expertise.
Workshop Challenges Researchers to Train AI on Child-Sized Data
The BabyLM Workshop announced its 4th edition for 2026, seeking papers and competition participants focused on training language models with far less data than typical approaches—closer to what a human child learns from. The competition challenges researchers to build effective models using limited, developmentally plausible datasets. This year adds a Multilingual track alongside its general track.
Why it matters: This is academic research infrastructure—interesting if you follow the debate over whether AI needs trillion-token datasets or whether more efficient learning is possible, but not something that will affect enterprise AI tools anytime soon.
What's Happening on Capitol Hill
Upcoming AI-related committee hearings
Tuesday, February 24 — Building an AI-Ready America: Teaching in the AI Age House · House Education and the Workforce Subcommittee on Early Childhood, Elementary, and Secondary Education (Hearing) 2175, Rayburn House Office Building
Tuesday, February 24 — Powering America's AI Future: Assessing Policy Options to Increase Data Center Infrastructure House · House Science, Space, and Technology Subcommittee on Investigations and Oversight (Hearing) 2318, Rayburn House Office Building
What's On The Pod
Some new podcast episodes
The Cognitive Revolution — Intelligence with Everyone: RL @ MiniMax, with Olive Song, from AIE NYC & Inference by Turing Post