<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Llms on MacWorks</title><link>https://macworks.dev/tags/llms/</link><description>Recent content in Llms on MacWorks</description><generator>Hugo</generator><language>en</language><atom:link href="https://macworks.dev/tags/llms/index.xml" rel="self" type="application/rss+xml"/><item><title>2026-04-13</title><link>https://macworks.dev/docs/week/simonwillison/simonwillison-2026-04-13/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/week/simonwillison/simonwillison-2026-04-13/</guid><description>&lt;h1 id="simon-willison--2026-04-13"&gt;Simon Willison — 2026-04-13&lt;a class="anchor" href="#simon-willison--2026-04-13"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="highlight"&gt;Highlight&lt;a class="anchor" href="#highlight"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;Today&amp;rsquo;s standout is Simon&amp;rsquo;s hands-on research into the newly released &lt;code&gt;servo&lt;/code&gt; crate using Claude Code. It perfectly captures his classic approach to AI-assisted exploration, demonstrating how quickly you can prototype a Rust CLI tool and evaluate WebAssembly compatibility with an LLM sidekick.&lt;/p&gt;
&lt;h2 id="posts"&gt;Posts&lt;a class="anchor" href="#posts"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;[Exploring the new servo crate]&lt;/strong&gt; · &lt;a href="https://simonwillison.net/2026/Apr/13/servo-crate-exploration/#atom-everything"&gt;Source&lt;/a&gt;
Following the initial release of the embeddable &lt;code&gt;servo&lt;/code&gt; browser engine on crates.io, Simon tasked Claude Code for web with exploring its capabilities. The AI successfully generated a working Rust CLI tool called &lt;code&gt;servo-shot&lt;/code&gt; for taking web screenshots. While compiling Servo itself to WebAssembly proved unfeasible due to its heavy use of threads and SpiderMonkey dependencies, Claude instead built a playground page utilizing a WebAssembly build of the &lt;code&gt;html5ever&lt;/code&gt; and &lt;code&gt;markup5ever_rcdom&lt;/code&gt; crates to parse HTML fragments.&lt;/p&gt;</description></item><item><title>Engineer Reads</title><link>https://macworks.dev/docs/today/engineer-blogs-2026-04-14/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/today/engineer-blogs-2026-04-14/</guid><description>&lt;h1 id="engineering-reads--2026-04-14"&gt;Engineering Reads — 2026-04-14&lt;a class="anchor" href="#engineering-reads--2026-04-14"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="the-big-idea"&gt;The Big Idea&lt;a class="anchor" href="#the-big-idea"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;The defining characteristic of good software engineering isn&amp;rsquo;t output volume, but the human constraints—specifically &amp;ldquo;laziness&amp;rdquo; and &amp;ldquo;doubt&amp;rdquo;—that force us to distill complexity into crisp abstractions and exercise restraint. As AI effortlessly generates code and acts on probabilistic certainty, our primary architectural challenge is deliberately designing simplicity and deferral into these systems.&lt;/p&gt;
&lt;h2 id="deep-reads"&gt;Deep Reads&lt;a class="anchor" href="#deep-reads"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;[Fragments: April 14]&lt;/strong&gt; · Martin Fowler · &lt;a href="https://martinfowler.com/fragments/2026-04-14.html"&gt;Martin Fowler&amp;rsquo;s Blog&lt;/a&gt;
Fowler synthesizes recent reflections on how AI-native development challenges our classical engineering virtues. He draws on Bryan Cantrill to argue that human &amp;ldquo;laziness&amp;rdquo;—our finite time and cognitive limits—is the forcing function for elegant abstractions, whereas LLMs inherently lack this constraint and will happily generate endless layers of garbage to solve a problem. Through a personal anecdote about simplifying a playlist generator via YAGNI rather than throwing an AI coding agent at it, he highlights the severe risk of LLM-induced over-complication. The piece then shifts to adapting our practices, touching on Jessitron&amp;rsquo;s application of Test-Driven Development to multi-agent workflows and Mark Little&amp;rsquo;s advocacy for AI architectures that value epistemological &amp;ldquo;doubt&amp;rdquo; over decisive certainty. Engineers navigating the integration of LLMs into their daily workflows should read this to re-calibrate their mental models around the enduring value of human constraints and system restraint.&lt;/p&gt;</description></item><item><title>Engineer Reads</title><link>https://macworks.dev/docs/week/blogs/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/week/blogs/</guid><description>&lt;h1 id="engineering-reads--week-of-2026-04-02-to-2026-04-10"&gt;Engineering Reads — Week of 2026-04-02 to 2026-04-10&lt;a class="anchor" href="#engineering-reads--week-of-2026-04-02-to-2026-04-10"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="week-in-review"&gt;Week in Review&lt;a class="anchor" href="#week-in-review"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;This week&amp;rsquo;s reading reflects a fundamental inflection point: raw LLM intelligence is no longer the bottleneck in software development. Instead, the industry is pivoting toward the hard systems engineering required to constrain probabilistic models—whether through strict data ledgers, living specifications, or formal verification harnesses. The dominant debate centers on how we preserve architectural taste, mechanical sympathy, and system ethics as the mechanical act of writing code becomes increasingly commoditized.&lt;/p&gt;</description></item><item><title>2026-04-10</title><link>https://macworks.dev/docs/week/simonwillison/simonwillison-2026-04-10/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/week/simonwillison/simonwillison-2026-04-10/</guid><description>&lt;h1 id="simon-willison--2026-04-10"&gt;Simon Willison — 2026-04-10&lt;a class="anchor" href="#simon-willison--2026-04-10"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="highlight"&gt;Highlight&lt;a class="anchor" href="#highlight"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;Simon points out the non-obvious reality that ChatGPT&amp;rsquo;s Advanced Voice Mode is actually running on an older, weaker model compared to their flagship developer tools. Drawing on insights from Andrej Karpathy, he highlights the widening capability gap between consumer-facing voice interfaces and B2B-focused reasoning models that benefit from verifiable reinforcement learning.&lt;/p&gt;
&lt;h2 id="posts"&gt;Posts&lt;a class="anchor" href="#posts"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href="https://simonwillison.net/2026/Apr/10/voice-mode-is-weaker/#atom-everything"&gt;ChatGPT voice mode is a weaker model&lt;/a&gt;&lt;/strong&gt;
Simon reflects on the counterintuitive fact that OpenAI&amp;rsquo;s Advanced Voice Mode runs on a GPT-4o era model with an April 2024 knowledge cutoff. Prompted by a tweet from Andrej Karpathy, he contrasts this consumer feature with top-tier coding models capable of coherently restructuring entire codebases or finding system vulnerabilities. Karpathy notes this divergence in capabilities exists because coding tasks offer explicit, verifiable reward functions ideal for reinforcement learning and hold significantly more B2B value.&lt;/p&gt;</description></item><item><title>2026-04-08</title><link>https://macworks.dev/docs/week/simonwillison/simonwillison-2026-04-08/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/week/simonwillison/simonwillison-2026-04-08/</guid><description>&lt;h1 id="simon-willison--2026-04-08"&gt;Simon Willison — 2026-04-08&lt;a class="anchor" href="#simon-willison--2026-04-08"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="highlight"&gt;Highlight&lt;a class="anchor" href="#highlight"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;The most substantial piece today is a deep-dive into Meta&amp;rsquo;s new Muse Spark model and its chat harness, where Simon successfully extracts the platform&amp;rsquo;s system tool definitions via direct prompting. His exploration of Meta&amp;rsquo;s built-in Python Code Interpreter and &lt;code&gt;visual_grounding&lt;/code&gt; capabilities highlights a powerful, sandbox-driven approach to combining generative AI with programmatic image analysis and exact object localization.&lt;/p&gt;
&lt;h2 id="posts"&gt;Posts&lt;a class="anchor" href="#posts"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href="https://simonwillison.net/2026/Apr/8/muse-spark/#atom-everything"&gt;Meta’s new model is Muse Spark, and meta.ai chat has some interesting tools&lt;/a&gt;&lt;/strong&gt;
Meta has launched Muse Spark, a new hosted model currently accessible as a private API preview and directly via the meta.ai chat interface. By simply asking the chat harness to list its internal tools and their exact parameters, Simon documented 16 different built-in tools. Standouts include a Python Code Interpreter (&lt;code&gt;container.python_execution&lt;/code&gt;) running Python 3.9 and SQLite 3.34.1, mechanisms for creating web artifacts, and a highly capable &lt;code&gt;container.visual_grounding&lt;/code&gt; tool. He ran hands-on experiments generating images of a raccoon wearing trash, then used the platform&amp;rsquo;s Python sandbox and grounding tools to extract precise, nested bounding boxes and perform object counts (like counting whiskers or his classic pelicans). Although the model is closed for now, infrastructure scaling and comments from Alexandr Wang suggest future versions could be open-sourced.&lt;/p&gt;</description></item><item><title>2026-04-04</title><link>https://macworks.dev/docs/week/blogs/engineer-blogs-2026-04-04/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/week/blogs/engineer-blogs-2026-04-04/</guid><description>&lt;h1 id="engineering-reads--2026-04-04"&gt;Engineering Reads — 2026-04-04&lt;a class="anchor" href="#engineering-reads--2026-04-04"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="the-big-idea"&gt;The Big Idea&lt;a class="anchor" href="#the-big-idea"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;Raw LLM intelligence is no longer the primary bottleneck for AI-assisted development; the real engineering challenge is building the system scaffolding—memory, tool execution, and repository context—that turns a stateless model into an effective, autonomous coding agent.&lt;/p&gt;
&lt;h2 id="deep-reads"&gt;Deep Reads&lt;a class="anchor" href="#deep-reads"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;[Components of A Coding Agent]&lt;/strong&gt; · Sebastian Raschka · &lt;a href="https://magazine.sebastianraschka.com/p/components-of-a-coding-agent"&gt;Sebastian Raschka Magazine&lt;/a&gt;
The core insight of this piece is that an LLM alone is just a stateless text generator; to do useful software engineering, it needs a surrounding agentic architecture. Raschka details the necessary scaffolding: equipping the model with tool use, stateful memory, and deep repository context. The technical mechanism relies on building an environment where the model can fetch file structures, execute commands, and persist state across conversational turns rather than just blindly emitting isolated code snippets. The tradeoff here is a steep increase in system complexity—managing context windows, handling tool execution failures, and maintaining state transitions is often much harder than prompting the model itself. Systems engineers and developers building AI integrations should read this to understand the practical anatomy of modern autonomous developer tools.&lt;/p&gt;</description></item><item><title>2026-04-07</title><link>https://macworks.dev/docs/week/simonwillison/simonwillison-2026-04-07/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/week/simonwillison/simonwillison-2026-04-07/</guid><description>&lt;h1 id="simon-willison--2026-04-07"&gt;Simon Willison — 2026-04-07&lt;a class="anchor" href="#simon-willison--2026-04-07"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="highlight"&gt;Highlight&lt;a class="anchor" href="#highlight"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;Anthropic&amp;rsquo;s decision to restrict access to their new Claude Mythos model underscores a massive, sudden shift in AI capabilities. It is a fascinating look at an industry-wide reckoning as open-source maintainers transition from dealing with &amp;ldquo;AI slop&amp;rdquo; to facing a tsunami of highly accurate, sophisticated vulnerability reports.&lt;/p&gt;
&lt;h2 id="posts"&gt;Posts&lt;a class="anchor" href="#posts"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;[Anthropic’s Project Glasswing - restricting Claude Mythos to security researchers - sounds necessary to me]&lt;/strong&gt; · &lt;a href="https://simonwillison.net/2026/Apr/7/project-glasswing/#atom-everything"&gt;Source&lt;/a&gt;
Anthropic has delayed the general release of Claude Mythos, a general-purpose model similar to Claude Opus 4.6, opting instead to limit access to trusted partners under &amp;ldquo;Project Glasswing&amp;rdquo; so they can patch foundational internet systems. Simon digs into the context, tracking how credible security professionals are warning about the ability of frontier LLMs to chain multiple minor vulnerabilities into sophisticated exploits. He even uses &lt;code&gt;git blame&lt;/code&gt; to independently verify a 27-year-old OpenBSD kernel bug discovered by the model. He concludes that delaying the release until new safeguards are built, while providing $100M in credits to defenders, is a highly reasonable trade-off.&lt;/p&gt;</description></item><item><title>2026-04-05</title><link>https://macworks.dev/docs/archives/simonwillison/simonwillison-2026-04-05/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/archives/simonwillison/simonwillison-2026-04-05/</guid><description>&lt;h1 id="simon-willison--2026-04-05"&gt;Simon Willison — 2026-04-05&lt;a class="anchor" href="#simon-willison--2026-04-05"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="highlight"&gt;Highlight&lt;a class="anchor" href="#highlight"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;Simon highlights a deep-dive post by Lalit Maganti on the realities of &amp;ldquo;agentic engineering&amp;rdquo; when building a robust SQLite parser. The piece beautifully articulates a crucial lesson for our space: while AI is incredible at plowing through tedious low-level implementation details, it struggles significantly with high-level design and architectural decisions where there isn&amp;rsquo;t an objectively right answer.&lt;/p&gt;
&lt;h2 id="posts"&gt;Posts&lt;a class="anchor" href="#posts"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href="https://simonwillison.net/2026/Apr/5/building-with-ai/#atom-everything"&gt;Eight years of wanting, three months of building with AI&lt;/a&gt;&lt;/strong&gt;
Simon shares a standout piece of long-form writing by Lalit Maganti on the process of building &lt;code&gt;syntaqlite&lt;/code&gt;, a parser and formatter for SQLite. Claude Code was instrumental in overcoming the initial hurdle of implementing 400+ tedious grammar rules, allowing Lalit to rapidly vibe-code a working prototype. However, the post cautions that relying on AI for architectural design led to deferred decisions and a confusing codebase, ultimately requiring a complete rewrite with more human-in-the-loop decision making. The core takeaway is that while AI excels at tasks with objectively checkable answers, it remains weak at subjective design and system architecture.&lt;/p&gt;</description></item><item><title>Simon Willison</title><link>https://macworks.dev/docs/week/simonwillison/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/week/simonwillison/</guid><description>&lt;h1 id="simon-willison--week-of-2026-04-04-to-2026-04-10"&gt;Simon Willison — Week of 2026-04-04 to 2026-04-10&lt;a class="anchor" href="#simon-willison--week-of-2026-04-04-to-2026-04-10"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="highlight-of-the-week"&gt;Highlight of the Week&lt;a class="anchor" href="#highlight-of-the-week"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;Anthropic&amp;rsquo;s decision to delay the general release of their highly capable Claude Mythos model under &amp;ldquo;Project Glasswing&amp;rdquo; marks a significant turning point in the AI industry. The move underscores a massive shift in frontier model capabilities, as models evolve from generating text to autonomously chaining multiple minor vulnerabilities into sophisticated exploits, requiring a new level of security safeguards before release.&lt;/p&gt;</description></item></channel></rss>