<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Reverse Engineering on MacWorks</title><link>https://macworks.dev/tags/reverse-engineering/</link><description>Recent content in Reverse Engineering on MacWorks</description><generator>Hugo</generator><language>en</language><atom:link href="https://macworks.dev/tags/reverse-engineering/index.xml" rel="self" type="application/rss+xml"/><item><title>Hacker News</title><link>https://macworks.dev/docs/today/hackernews-2026-04-19/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://macworks.dev/docs/today/hackernews-2026-04-19/</guid><description>&lt;h1 id="hacker-news--2026-04-19"&gt;Hacker News — 2026-04-19&lt;a class="anchor" href="#hacker-news--2026-04-19"&gt;#&lt;/a&gt;&lt;/h1&gt;
&lt;h2 id="top-story"&gt;Top Story&lt;a class="anchor" href="#top-story"&gt;#&lt;/a&gt;&lt;/h2&gt;
&lt;p&gt;&lt;a href="https://abacusnoir.com/2026/04/18/zero-copy-gpu-inference-from-webassembly-on-apple-silicon/"&gt;Zero-Copy GPU Inference from WebAssembly on Apple Silicon&lt;/a&gt;
On Apple Silicon, you can share a WebAssembly module&amp;rsquo;s linear memory directly with the GPU—meaning zero copies, no serialization, and no intermediate buffers. By composing &lt;code&gt;mmap&lt;/code&gt;, Metal buffers, and Wasmtime&amp;rsquo;s custom memory allocator, the author ran a 1B parameter Llama model entirely from a Wasm guest with zero-copy overhead. This is pure, hardware-sympathetic engineering, proving that sandboxed runtimes don&amp;rsquo;t have to ruin performance if you just leverage the underlying physics of the chip.&lt;/p&gt;</description></item></channel></rss>