March 11, 2026

D.A.D. today covers 16 stories from 5 sources. What's New, What's Innovative, What's Controversial, What's in the Lab, and What's in Academe.

D.A.D. Joke of the Day: My AI and I have trust issues. I keep asking it to be concise, and it keeps writing me a novel about why brevity matters.

What's New

AI developments from the last 24 hours

Amazon Now Requires Senior Sign-Off on AI-Assisted Code After Outages

Amazon is requiring senior engineer sign-off on AI-assisted code changes after a 'trend of incidents' including a nearly six-hour website outage this month. An internal briefing note cited 'Gen-AI assisted changes' and 'novel GenAI usage for which best practices and safeguards are not yet fully established' as contributing factors. The company is implementing immediate initiatives to limit future outages. Community discussion on Hacker News highlights a core tension: AI can generate code faster than humans can meaningfully review it, and the reasoning behind AI-generated changes is often invisible, complicating future debugging.

Why it matters: This is the first major public acknowledgment from a tech giant that AI coding tools are causing production incidents at scale—a signal that enterprises racing to adopt these tools may need formal governance frameworks before deployment, not after.


Test-First Approach Reportedly Boosts AI Coding Agent Output Fourfold

A developer running Claude Code workshops reports a core problem with overnight AI coding agents: there's no reliable way to verify AI-generated code at scale. Having AI review its own work creates what he calls a 'self-congratulation machine.' His proposed fix borrows from test-driven development—humans write acceptance criteria first, then AI builds features verified by separate automated testing tools. Teams using this approach reportedly merge 40-50 pull requests weekly versus 10 previously, though these figures are anecdotal from workshop participants rather than controlled studies.

Why it matters: As AI coding tools move from assistance to autonomy, the verification bottleneck—not the code generation—may become the limiting factor for adoption in production environments.


Developer Claims Claude Code Built Entire Programming Language Without Human-Written Code

A developer claims to have built a complete programming language called Cutlet using Claude Code over four weeks—without reading any of the generated code. The developer's role was purely architectural: setting guardrails and correctness tests while Claude wrote every line. The resulting interpreter builds on macOS and Linux, includes a REPL, and features arrays, functions, vectorized operations, and filtering. Source code is available on GitHub.

Why it matters: This is an extreme test case for AI-assisted development—suggesting that with sufficient testing infrastructure, developers might supervise rather than write code, though the approach raises obvious questions about maintainability and debugging when no human has read the codebase.


Tutorial Walks Through Git Rebasing in Emacs Interface

A developer tutorial walks through rebasing workflows using Magit, the Emacs-based Git interface. The author frames Magit's strength as 'high discoverability'—it shows available commands through hints and displays the underlying Git operations, helping users learn Git rather than hiding it behind abstractions. The post demonstrates using the git log as an interactive command center for understanding repository structure and performing complex rebases.

Why it matters: This is developer tooling for Emacs users—unlikely to affect your workflow unless your engineering team already lives in that ecosystem.


Quicksort Inventor Tony Hoare Dies at 92

Tony Hoare, the British computer scientist who invented quicksort and won the Turing Award, died Thursday at 92. Hoare's contributions shaped foundational computer science: quicksort remains one of the most widely used sorting algorithms, his work on ALGOL influenced programming language design for decades, and Hoare logic became essential for formally verifying that programs work correctly. He spent much of his career at Oxford and later Microsoft Research Cambridge.

Why it matters: For anyone building or relying on software, Hoare's work is embedded invisibly in the systems you use daily—his passing marks the loss of one of computing's genuine pioneers.


What's Innovative

Clever new use cases for AI

Mac Tool Claims Double the Speed for Local AI Models

Y Combinator-backed RunAnywhere launched MetalRT, an inference engine for Apple Silicon that the startup claims outperforms existing tools across language models, speech-to-text, and text-to-speech. Their benchmarks on an M4 Max show Qwen3-4B running at 186 tokens per second versus 87 for llama.cpp, and transcription hitting 714x real-time speed. They also open-sourced RCLI, a voice AI pipeline that runs entirely on-device with no cloud dependencies or API keys required.

Why it matters: If the benchmarks hold up, Mac users running local AI—whether for privacy, offline access, or cost savings—could see substantial speed gains, particularly for voice applications where latency matters.


Community Workflows Make Local AI Video Generation More Accessible

A new Hugging Face repository offers ready-made workflows for LTX-2, an open-source video generation model, packaged for use with ComfyUI—a popular node-based interface for running AI models locally. The workflows focus on image-to-video generation, letting users animate still images into video clips. This is community-contributed tooling rather than an official release, aimed at users who run AI video generation on their own hardware rather than through cloud services.

Why it matters: This is developer and hobbyist plumbing—useful if your team experiments with local video AI tools, but not a capability shift for most business users.


What's in the Lab

New announcements from major AI labs

Amid Political Fights, Anthropic Launches Policy Institute, Hires OpenAI Defector

Anthropic is launching The Anthropic Institute, a new research organization focused on how powerful AI will reshape jobs, economies, and governance. Led by co-founder Jack Clark in a new role as Head of Public Benefit, the institute combines three existing teams—frontier red-teaming, societal impacts, and economic research—and is hiring across multiple cities including a new DC office opening this spring. Among its founding hires: Zoë Hitzig, a research scientist who recently resigned from OpenAI and published a New York Times op-ed titled "OpenAI Is Making the Mistakes Facebook Made," criticizing the company for optimizing for engagement over user welfare. Other hires include Matt Botvinick from Google DeepMind and economist Anton Korinek from the University of Virginia.

Why it matters: The move comes as Anthropic navigates its public conflict with the Trump administration over the renamed Department of War—building out a Washington policy presence signals the company is preparing for a sustained fight over AI governance, while poaching a vocal OpenAI critic adds pointed competitive messaging.


Training Method Claims to Block Prompt Injection Attacks

Researchers have proposed IH-Challenge, a training approach designed to make AI models better at following a hierarchy of instructions—prioritizing commands from system operators over potentially conflicting user prompts. The method claims to improve safety controls and resistance to prompt injection attacks, where malicious inputs trick models into ignoring their original instructions. No benchmark data or independent validation was provided in the announcement.

Why it matters: If validated, this could help enterprises deploy AI assistants with more reliable guardrails—a persistent concern for companies worried about users circumventing safety controls or extracting sensitive system prompts.


ChatGPT Adds Visual Explanations for Math and Science

OpenAI announced that ChatGPT now includes interactive visual explanations for math and science topics. The feature lets users explore formulas, variables, and concepts through real-time visual displays rather than text-only responses. OpenAI is positioning this as an educational tool, though no details were provided on how it works technically or what subjects are covered.

Why it matters: For professionals who use ChatGPT to brush up on quantitative concepts or explain technical material to others, visual walkthroughs could make the tool more useful for learning and communication—though how well it actually works remains to be seen.


DeepMind Marks 10 Years Since AlphaGo, Credits It for Protein-Folding Breakthrough

Google DeepMind published a retrospective marking the 10th anniversary of AlphaGo's victory over world champion Lee Sedol at Go—a game with more possible positions than atoms in the universe. The company traces a direct line from that 2016 breakthrough to its subsequent achievements: AlphaZero mastering chess in hours, and AlphaFold cracking the protein-folding problem in 2020. DeepMind says over 3 million researchers now use AlphaFold's database of 200 million protein structures.

Why it matters: This is DeepMind making its case for historical significance—positioning AlphaGo as the starting gun for the current AI era. No immediate business implications, but useful context for understanding the competitive landscape.


What's in Academe

New papers on AI and its effects from researchers

Pathology AI Framework Mimics How Human Diagnosticians Reason

Researchers introduced PathMem, an AI framework designed to help pathology models reason more like human diagnosticians. The system organizes medical knowledge into structured long-term memory and dynamically retrieves relevant information during analysis—mimicking how pathologists draw on training and experience when examining slides. On benchmarks for whole-slide image analysis, PathMem improved diagnostic report precision by 12.8% and relevance by 10.1% over previous approaches. Open-ended diagnosis accuracy rose roughly 9% as well.

Why it matters: This is research-stage work, but the approach—giving AI systems structured domain memory rather than relying purely on pattern matching—could eventually make pathology AI more reliable and explainable in clinical settings.


Robots Learn to Push Objects Against Walls When Grippers Fail

Researchers developed a framework that teaches robots to use their environment—pushing objects against walls, sliding items across surfaces—rather than relying solely on their grippers. The technique, called Dynamics-Aware Policy Learning, lets robots manipulate objects in cluttered spaces by learning from contact physics rather than following pre-programmed rules. In testing, it beat traditional grasping methods and human teleoperation by over 25% in simulated clutter. Real-world trials, including a grocery store deployment, achieved roughly 50% success rates.

Why it matters: Warehouse and retail automation has struggled with cluttered, unpredictable environments—robots that can improvise using nearby surfaces could handle real-world messiness far better than current pick-and-place systems.


Smarter Chunk Ordering Improves Multi-Agent AI Accuracy on Long Documents

New research proposes a better way to feed long documents to AI systems that use multiple agents working in sequence. The problem: when you split a long document into chunks and process them one agent at a time, information gets lost along the chain. The solution uses a statistical technique (Chow-Liu trees) to reorder chunks so strongly related information gets processed together, rather than following the document's natural sequence. In benchmarks, this reordering improved answer accuracy compared to standard approaches.

Why it matters: For enterprise teams building multi-agent AI workflows on long documents—contracts, research reports, regulatory filings—chunk ordering may be an overlooked variable affecting output quality.


Open-Source Tool Lets Teams Test AI Models in Plain English

Researchers have released One-Eval, an open-source framework that lets users describe what they want to test about an AI model in plain English, then automatically generates and runs the evaluation. Instead of manually configuring benchmark suites, users can request something like "test this model's ability to summarize legal documents" and get a reproducible workflow. The team says this reduces the technical overhead of AI evaluation, particularly for enterprise teams that need to validate models before deployment but lack dedicated ML engineering resources.

Why it matters: As companies increasingly evaluate multiple AI models for business use, tools that simplify head-to-head comparisons without requiring deep technical expertise could accelerate procurement decisions.


AI Models Become More Honest When Given Time to Reason

New research finds that giving AI models time to reason makes them more honest—the opposite of humans, who tend to become less truthful when they have time to deliberate. Researchers created scenarios where honesty carried costs and found that deceptive responses occupy 'metastable' regions in how models represent information, meaning lies are more easily disrupted by minor input changes or resampling than truthful answers. Notably, the content of the reasoning itself didn't predict outcomes; simply engaging the reasoning process pushed models toward more stable honest defaults.

Why it matters: This suggests a potential alignment lever: if deception is inherently unstable in LLMs while honesty is a robust default, techniques that encourage reasoning or introduce small perturbations could serve as guardrails against AI dishonesty—useful for anyone deploying AI in high-stakes contexts.


What's Happening on Capitol Hill

Upcoming AI-related committee hearings

Tuesday, March 17DeepSeek and Unitree Robotics: Examining the National Security Risks of PRC Artificial Intelligence, Robotics, and Autonomous Technologies and Building a Secure U.S. Technology Base House · Homeland Security Subcommittee on Cybersecurity and Infrastructure Protection (Hearing) 310, Cannon House Office Building


What's On The Pod

Some new podcast episodes

AI in BusinessAI Use Cases, Deployment, and Measuring Real-World ROI - with Ylan Kazi of Blue Cross Blue Shield of North Dakota

How I AIMastering Midjourney: How to create consistent, beautiful brand imagery without complex prompts | Jamey Gannon

AI in BusinessOperationalizing Customer Service at Scale with Outcome-Driven Agentic AI - with Craig Walker of Dialpad