Sources

The Agent Infrastructure Rewrite & The ARC-AGI-3 Reality Check — 2026-03-30#

Highlights#

The industry is waking up to the reality that legacy infrastructure built for human speed is becoming the primary bottleneck for agentic workflows. Simultaneously, the AGI hype narrative took a massive hit today: François Chollet’s newly released ARC-AGI-3 benchmark completely stumped the latest frontier models, underscoring the stark difference between scaling compute and achieving genuine novel reasoning.

Top Stories#

  • ARC-AGI-3 Destroys Frontier Models: François Chollet launched ARC-AGI-3, a grueling benchmark of 135 novel game environments where untrained humans score 100%, but top models like GPT 5.4, Gemini 3.1 Pro, and Opus 4.6 score below 1%. The scoring system heavily punishes brute-force compute, resetting the industry scoreboard to near zero and offering a $2M Kaggle prize for open-sourced solutions. (Source)
  • Re-engineering for Agent Speed: Aaron Levie and Jeff Dean highlighted that our software tools—from file systems to search indexes and payment rails—must be rebuilt to support agents running 24/7 in parallel. Because current tools were designed for human latency, Amdahl’s law dictates that even infinitely fast AI models will only yield a 2-3x overall improvement without a massive infrastructure overhaul. (Source)
  • US Enterprise Generative AI Adoption Stalls: New time-series data on the labor market effects of generative AI indicates that LLM adoption at work in the US actually dropped over the past quarter, despite remaining higher than previous years. (Source)
  • Public Rejection of AI Data Centers: A newly released Quinnipiac poll shows severe NIMBYism against AI infrastructure, with Americans opposing the construction of AI data centers in their communities by a staggering 65-24 margin. (Source)
  • Sycamore Raises $65M for Enterprise Agent OS: Sri Viswanath announced Sycamore Labs, a trusted agent operating system for the enterprise, backed by a massive $65M seed round led by Coatue Management and Lightspeed Venture Partners. (Source)

Articles Worth Reading#

Why Local Coding Agents Still Struggle (Source) Georgi Gerganov offers a highly technical breakdown of why getting reliable performance from local coding agents remains incredibly difficult. He notes that the friction isn’t just a matter of model intelligence, but rather brittle harnesses, chat template parsing errors, and prompt construction intricacies across the stack. To evaluate local setups properly, he recommends developers write custom harnesses rather than blindly plugging local models into clients like Claude Code, and he highlights Qwen3.5 as a step-change model worth deploying.

Claire Vo on “Blameful” Incident Reports (Source) Pushing back against mealy-mouthed corporate apologies, Claire Vo argues that Sev-0 post-mortems sent to customers need to be aggressively clear and slightly more “blameful”. She critiques the passive voice often used when data is exposed—as if “incident fairies” caused the breach—and demands plain-language explanations of exactly how data was compromised alongside clear remediation paths. It is a sharp reminder for tech leadership that transparent incident management is the ultimate barometer of client trust.

The US Brain Drain and Regulatory Capture (Source) In a stark warning about American scientific leadership, commentators note that an estimated 95,000 scientists have left federal agencies following funding cuts and office closures. Meanwhile, the FTC has reportedly been ordered to stop “burdening” AI companies, allowing actors like Clarifai to legally retain models trained on millions of stolen dating profiles. As the US bleeds academic talent and cancels science grants, nations like China are aggressively funding their research infrastructure, threatening to surpass the US as a scientific superpower.