May 16, 2026

D.A.D. today covers 13 stories from 4 sources. What's New, What's Innovative, What's Controversial, What's in the Lab, and What's in Academe.

D.A.D. Joke of the Day: My AI said it needed more context. I gave it a 10-page brief. Now it needs more context about the context.

What's New

AI developments from the last 24 hours

HashiCorp Co-Founder Warns: Some Teams Now Ship Bugs Expecting AI to Fix Them

Mitchell Hashimoto, co-founder of HashiCorp, posted a warning about what he calls 'AI psychosis' at some companies—a mindset that treats shipping bugs as acceptable because AI agents can supposedly fix them at scale. He offered no specific examples. Community reaction is divided: some agree the attitude is reckless, particularly when AI writes the code, tests, and reviews with no human cross-check. Others counter that AI agents genuinely are becoming effective at quickly fixing bugs that humans identify.

Why it matters: The debate signals growing tension in tech leadership about where AI-assisted development crosses from productivity gain into quality risk—a question every team adopting these tools will eventually face.


Satirical Article Highlights JavaScript's Recurring Security Vulnerabilities

A satirical article modeled on The Onion's recurring gun violence headline takes aim at npm, JavaScript's package manager, for repeated supply chain attacks. The piece argues npm's reliance on deeply nested, unvetted dependencies from anonymous maintainers creates preventable security holes that languages like Go and Rust avoid through stronger standard libraries. Community reaction was mixed—some noted Python's pip ecosystem now faces similar attacks and actually lacks lockfiles, making it potentially worse. Others questioned whether Go and Rust are genuinely more secure or simply less targeted.

Why it matters: For teams running JavaScript in production, this captures a real tension: npm's flexibility comes with security tradeoffs that competitors are increasingly exploiting, and the 'it happens everywhere' framing may be masking preventable risk.


Free Public Domain Library Upgrades Mobile Experience After 54 Years

Project Gutenberg, the volunteer-run digital library offering over 75,000 public domain eBooks, says it has significantly improved its website over the past few months with more updates planned. The 54-year-old nonprofit relies on volunteers to digitize and proofread older works whose U.S. copyrights have expired. Users report the mobile experience has notably improved.

Why it matters: For anyone building AI training datasets, researching historical texts, or just looking for free source material, Project Gutenberg remains one of the cleanest repositories of public domain content—and a more usable interface makes that archive more accessible.


Waymo Recalls 3,800 Robotaxis After Vehicles Drove Into Flooded Streets

Waymo is voluntarily recalling about 3,800 robotaxis to patch software after vehicles drove into flooded streets and stalled—one in San Antonio was swept into a creek on April 20. The recall affects Waymo's fifth and sixth generation systems. San Antonio service remains suspended while NHTSA investigates. Waymo says it's adding safeguards for detecting flooded high-speed roadways. Community discussion noted the technical difficulty of distinguishing wet pavement from deep water using cameras.

Why it matters: The incident highlights that autonomous vehicles still struggle with edge cases human drivers navigate intuitively—a reminder that 'software update' recalls may become routine as robotaxi fleets scale.


Bun Merges Million-Line AI-Generated Rewrite, Drawing Developer Skepticism

Bun, the JavaScript runtime that bills itself as a faster alternative to Node.js, has merged a massive pull request rewriting its codebase from Zig to Rust—over 1 million lines of code changed. The automated translation was previously dismissed as "just an experiment" before being merged. Community reaction has been skeptical: developers note the single massive commit is difficult to review, with some pointing out tests were modified to pass rather than fixing underlying issues. One observer found irony in Bun's own CI system auto-tagging a follow-up PR as "ai slop."

Why it matters: For teams using Bun in production, a million-line automated rewrite raises questions about code quality and stability—worth monitoring before upgrading.


What's in the Lab

New announcements from major AI labs

ChatGPT Now Connects to Bank Accounts for Personalized Financial Advice

OpenAI launched a personal finance feature in ChatGPT, currently available to Pro subscribers in the U.S. Users can connect bank and investment accounts through Plaid (Intuit integration coming), view a financial dashboard, and ask questions grounded in their actual account data. OpenAI says the feature uses GPT-5.5's reasoning to help users spot spending patterns, understand tradeoffs, and plan decisions. The company notes that 200 million people already use ChatGPT monthly for financial questions—this moves from generic advice to personalized analysis based on real transaction history.

Why it matters: This is OpenAI's most aggressive push yet into a regulated, high-stakes domain—positioning ChatGPT not just as a general assistant but as a Mint-style financial tool with AI reasoning layered on top, raising both competitive and privacy questions for financial services firms.


Databricks Offers GPT-5.5 for Enterprise Document Workflows

Databricks is offering GPT-5.5 through its AI Unity Gateway for enterprise agent workflows. The company says the model is the first to exceed 50% accuracy on OfficeQA Pro, its internal benchmark for complex document tasks like parsing scanned PDFs and legacy files—a 46% error reduction over GPT-5.4. Customers can use it with Databricks' AgentBricks tools to coordinate specialized agents for document parsing, retrieval, and task execution.

Why it matters: For Databricks customers already building agent systems, this adds OpenAI's latest model as an option—though the benchmark is Databricks' own, so independent comparisons to competitors remain to be seen.


Cohere Proposes Measuring AI Success Beyond Traditional ROI

Cohere published a framework for organizations moving from AI experimentation to enterprise-wide integration. The AI vendor argues companies should measure AI success beyond traditional ROI metrics, tracking workforce indicators like skill development, organizational velocity, and employee trust in AI systems. The blog post outlines phases of 'AI maturity' but provides no benchmark data or case studies—it's a conceptual roadmap rather than evidence-based guidance.

Why it matters: This is vendor thought leadership positioning Cohere for enterprise deals, but the underlying question—how to measure AI's actual business impact beyond pilot projects—is one many organizations are genuinely struggling with.


What's in Academe

New papers on AI and its effects from researchers

International Students Turn to ChatGPT as Low-Judgment Guide to American Culture

A study of 60 international students found they're using ChatGPT and Gemini as "first-aid tools" for navigating American culture—getting quick help with social norms, communication challenges, and practical questions they might feel awkward asking peers. Interviews with 14 participants revealed interest in AI evolving into longer-term support for cross-cultural adjustment, not just one-off answers.

Why it matters: For anyone managing international teams or campus programs, this signals AI may be quietly filling gaps in cultural onboarding—worth considering what that means for formal support systems.


AI-Generated Interfaces Score High on Usability, Low on Originality

A 92-person study found that AI-generated interface designs perform well on usability but poorly on originality. Researchers showed participants both AI and human-created prototypes without revealing which was which, then measured user experience across practical and creative dimensions. AI tools produced functional, efficient interfaces—but they reinforced conventional patterns that left users perceiving them as unoriginal. Pragmatic scores were positive; hedonic scores (creativity, novelty) were neutral to negative.

Why it matters: For teams using AI to accelerate UI/UX work, this suggests a ceiling: AI can handle competent baseline designs, but human designers may still be essential for differentiation and brand distinctiveness.


Brief AI Disclosures on News May Distract Readers More Than Detailed Ones

An eye-tracking study found that brief, one-line AI disclosures on news articles may backfire: readers spent more time fixating on them and made more eye movements compared to either no disclosure or detailed explanations. Researchers measured attention patterns across political and lifestyle articles where AI assisted with editing or content generation. Detailed disclosures didn't add cognitive burden, and interview subjects actually preferred them—or wanted the option to expand for more information.

Why it matters: As regulators and news organizations debate how to label AI-generated content, this suggests the intuitive "keep it short" approach may actually disrupt readers more than fuller explanations—relevant to anyone designing AI transparency for customers or internal communications.


Alibaba Experiment: AI Customer Service Cuts Time but Tanks Satisfaction Ratings

A field experiment on Alibaba's Taobao platform found that AI customer service agents cut average chat times but substantially lowered customer ratings. The surprising finding: human intervention rescued service quality when AI hit technical problems, but failed when customers escalated emotionally—workers sent fewer messages and showed less engagement in those cases. Early intervention proved essential; waiting too long meant workers stayed disengaged even after taking over. One bright spot: workers supervising AI gave better attention to the non-AI chats they handled directly.

Why it matters: For companies deploying AI customer service, this suggests the human backup strategy matters as much as the AI itself—and that emotional escalations may need different protocols than technical failures.


Websites Can Identify Which AI Model Powers Your Browser Agent

Security researchers found that websites can identify which AI model powers a browser agent just by watching how it clicks, types, and navigates. Using a simple JavaScript tracker, they achieved 96% accuracy identifying the underlying model across 14 frontier LLMs. The technique works by analyzing action patterns and timing—each model apparently has a distinctive behavioral signature. Once a site knows which model is driving an automated session, it could exploit that model's known weaknesses. Adding random delays helped initially, but retraining the classifier largely recovered its accuracy.

Why it matters: As AI agents increasingly browse the web on users' behalf, this research suggests they may be easier to detect and manipulate than assumed—a concern for anyone deploying agents for competitive intelligence, automated purchasing, or other sensitive tasks.


What's On The Pod

Some new podcast episodes

The Cognitive RevolutionThree Kinds of Software Survive: Tasklet's Andrew Lee on Competing to be a Horizontal Platform