<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Tech@Youtube on MacWorks</title><link>https://macworks.dev/docs/month/tech-videos/</link><description>Recent content in Tech@Youtube on MacWorks</description><generator>Hugo</generator><language>en</language><atom:link href="https://macworks.dev/docs/month/tech-videos/index.xml" rel="self" type="application/rss+xml"/><item><title>Week 13 Summary</title><link>https://macworks.dev/docs/month/tech-videos/weekly-2026-W13/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/month/tech-videos/weekly-2026-W13/</guid><description>&lt;h1 id="tech-videos--week-of-2026-03-20-to-2026-03-26"&gt;Tech Videos — Week of 2026-03-20 to 2026-03-26&lt;a class="anchor" href="#tech-videos--week-of-2026-03-20-to-2026-03-26"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="watch-first"&gt;Watch First&lt;a class="anchor" href="#watch-first"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;&lt;a href="#"&gt;加快語言模型生成速度 (2/2)：KV Cache&lt;/a&gt; by Hung-yi Lee is the single highest-signal video this week for anyone dealing with AI infrastructure. It breaks down the brutal VRAM bottlenecks of LLM inference and the math behind architectural optimizations like Multi-Head Latent Attention far better than any high-level blog post.&lt;/p&gt;
&lt;h2 id="week-in-review"&gt;Week in Review&lt;a class="anchor" href="#week-in-review"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;The industry is aggressively moving away from monolithic, synchronous LLM chat wrappers toward decoupled, multi-agent swarms constrained by deterministic protocols like MCP and A2A. Simultaneously, the hyper-acceleration of the coding &amp;ldquo;inner loop&amp;rdquo; is exposing massive friction downstream, with machine-generated output completely overwhelming traditional CI/CD and human review pipelines. Infrastructure is shifting closer to the metal, with a clear focus on owning the compute layer and managing sub-millisecond latencies to make agentic workflows economically viable.&lt;/p&gt;</description></item><item><title>Week 14 Summary</title><link>https://macworks.dev/docs/month/tech-videos/weekly-2026-W14/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/month/tech-videos/weekly-2026-W14/</guid><description>&lt;h1 id="tech-videos--week-of-2026-03-28-to-2026-04-03"&gt;Tech Videos — Week of 2026-03-28 to 2026-04-03&lt;a class="anchor" href="#tech-videos--week-of-2026-03-28-to-2026-04-03"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="watch-first"&gt;Watch First&lt;a class="anchor" href="#watch-first"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;For the most impactful video, the Syntax channel&amp;rsquo;s &lt;a href="#"&gt;37,000 Lines of Slop&lt;/a&gt; is the single best watch this week because it provides a brutal, necessary teardown of AI coding hype. It vividly demonstrates why blindly shipping massive LLM output without rigorous human review results in catastrophic production payloads, cutting through the marketing noise of effortless AI development.&lt;/p&gt;
&lt;h2 id="week-in-review"&gt;Week in Review&lt;a class="anchor" href="#week-in-review"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;The dominant theme this week is the awkward transition from isolated LLM chat interfaces to orchestrated, tool-using agents, exposing massive friction in both security and developer workflows. We are also seeing a definitive industry shift toward inference-bound hardware architectures, as scaling laws collide with concrete power, memory, and cooling bottlenecks.&lt;/p&gt;</description></item></channel></rss>