Sources
- AI Engineer
- All-In Podcast
- Andrej Karpathy
- Anthropic
- Apple
- Apple Developer
- AWS Events
- ByteByteGo
- Computerphile
- Cursor
- Dwarkesh Patel
- EO
- Fireship
- GitHub
- Google Cloud Tech
- Google DeepMind
- Google for Developers
- Hung-yi Lee
- Lenny's Podcast
- Lex Clips
- Lex Fridman
- Life at Google
- Marques Brownlee
- Microsoft
- No Priors: AI, Machine Learning, Tech, & Startups
- Numberphile
- NVIDIA
- OpenAI
- Perplexity
- Quanta Magazine
- Slack
- The Pragmatic Engineer
- Visual Studio Code
Tech Videos — 2026-04-12#
Watch First#
Building Towards Self-Driving Codebases with Long-Running, Asynchronous Agents offers a highly credible look into the mechanics of long-running coding agents from Cursor’s founder, cutting through the hype to explain the concrete architectural hurdles of scaling AI from autocomplete to massive, unsupervised pull requests.
Highlights by Theme#
Developer Tools & Platforms#
On the NVIDIA Developer channel, Cursor’s founder details their move to cloud-based “async agents,” noting a notable internal benchmark where an agent successfully managed an 8-hour, 10,000-line PR that migrated a video renderer from React to Rust. For a deeper dive into the execution environments wrapping these tools, Harness Engineering from Hung-yi Lee demonstrates that even a tiny 2B parameter model (Gemma 4) can successfully debug code if provided a proper “harness” that grants bash and Python execution capabilities alongside explicit step-by-step operating rules.
AI & Machine Learning#
The NVIDIA Developer talk drops a crucial technical insight regarding long-running agents: models trained via reinforcement learning suffer severe performance degradation when task lengths exceed their training distribution of hundreds of thousands of tokens. To mitigate this train-test mismatch, Cursor utilizes a multi-agent planner-worker architecture to fan out tasks and compress context lengths back into safe bounds. Similarly, Hung-yi Lee’s lecture explores how injecting environmental feedback—like compiler errors—into a continuous multi-turn loop essentially acts as a “textual gradient,” allowing models to self-correct and update their behavior without actually updating model weights.
Everything Else#
In management and culture, Keith Rabois argues on Lenny’s Podcast that AI tools are merging engineering, product, and design into a singular, business-driven function where the hardest remaining skill is simply knowing what to build. He also drops a pragmatic take for engineering leaders dealing with rapid scale—where 70% of a leader’s time is spent fighting “success disasters”—asserting that top-performing startup teams must prioritize relentless momentum and winning over psychological safety. A brief segment from No Priors echoes this macro-shift, noting that the accelerating diffusion of AI will force entirely new economic and organizational structures. Finally, for a non-tech detour, clips from Lex Clips and Dwarkesh Patel explore historical governance, contrasting the short-lived Viking expansion with the thousand-year stability of the Byzantine Empire, and examining how a 29-year-old Machiavelli managed diplomatic bureaucracy during massive geopolitical instability.