<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: rahen</title><link>https://news.ycombinator.com/user?id=rahen</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 16 Apr 2026 20:21:36 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=rahen" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by rahen in "Want to Write a Compiler? Just Read These Two Papers (2008)"]]></title><description><![CDATA[
<p>I'm also writing a compiler and CS6120 from Cornell has helped me a lot:
<a href="https://www.cs.cornell.edu/courses/cs6120/2025fa/self-guided/" rel="nofollow">https://www.cs.cornell.edu/courses/cs6120/2025fa/self-guided...</a></p>
]]></description><pubDate>Wed, 15 Apr 2026 18:32:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=47783226</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47783226</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47783226</guid></item><item><title><![CDATA[New comment by rahen in "Show HN: I built a tiny LLM to demystify how language models work"]]></title><description><![CDATA[
<p>I don't mean to be 'that guy', but after a quick review, this really feels like low-effort AI slop to me.<p>There is nothing wrong using AI tools to write code, but nothing here seems to have taken more than a generic 'write me a small LLM in PyTorch' prompt, or any specific human understanding.<p>The bar for what constitutes an engineering feat on HN seems to have shifted significantly.</p>
]]></description><pubDate>Mon, 06 Apr 2026 13:36:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=47660754</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47660754</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47660754</guid></item><item><title><![CDATA[New comment by rahen in "Circuit-level PDP-11/34 emulator"]]></title><description><![CDATA[
<p>Out of curiosity, what were those wonderful things you were hearing about the 11/34 back then?</p>
]]></description><pubDate>Sun, 29 Mar 2026 00:55:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=47559488</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47559488</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47559488</guid></item><item><title><![CDATA[New comment by rahen in "Paper Tape Is All You Need – Training a Transformer on a 1976 Minicomputer"]]></title><description><![CDATA[
<p>Yes. The Cray supercomputers from the 80s were crazy good matmul machines in particular. The quad-CPU Cray X-MP (1984) could sustain 800 MFLOPS to 1 GFLOPS, and with a 1 GB SSD, had enough computer power and bandwidth to train a 7-10M-parameter language model in about six months, and infer at 18-25 tok/sec.<p>A mid-90s Cray T3E could have handled GPT-2 124M, 24 years before OpenAI.<p>I also had a punch-card computer from 1965 learn XOR with backpropagation.<p>The hardware was never the bottleneck, the ideas were.</p>
]]></description><pubDate>Sat, 28 Mar 2026 21:35:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47558308</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47558308</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47558308</guid></item><item><title><![CDATA[New comment by rahen in "Paper Tape Is All You Need – Training a Transformer on a 1976 Minicomputer"]]></title><description><![CDATA[
<p>I also have a working design for a small Transformer on the original Game Boy. It has around 4000 parameters fitting in the 8 KB cartridge SRAM, where the "saved game" is the trained model. A TI-82 with its 32 KB of RAM would be even more comfortable.</p>
]]></description><pubDate>Sat, 28 Mar 2026 18:26:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47557092</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47557092</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47557092</guid></item><item><title><![CDATA[New comment by rahen in "Paper Tape Is All You Need – Training a Transformer on a 1976 Minicomputer"]]></title><description><![CDATA[
<p>I encouter two main failure modes. First, the bipolar PROMs degrade at the atomic level, the metal ions in the fuses tend to migrate or 'regrow' over decades, causing bit rot.
Second, the backplanes suffer from mechanical fatigue. After forty years of thermal expansion and structural flexing, especially when inserting boards, the traces and solder joints develop stress cracks. Both are a pain to repair.<p><a href="https://retrocmp.com/articles/trying-to-fix-a-dec-pdp-1134-backplane/149-trying-to-fix-a-dec-pdp-1134-backplane-therapies" rel="nofollow">https://retrocmp.com/articles/trying-to-fix-a-dec-pdp-1134-b...</a></p>
]]></description><pubDate>Sat, 28 Mar 2026 18:22:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47557062</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47557062</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47557062</guid></item><item><title><![CDATA[New comment by rahen in "Paper Tape Is All You Need – Training a Transformer on a 1976 Minicomputer"]]></title><description><![CDATA[
<p>The WASM GUI is probably the easiest way to see the Transformer in action on this machine: <a href="https://dbrll.github.io/ll-34/" rel="nofollow">https://dbrll.github.io/ll-34/</a><p>There's also the original Tetris from 1984 to play.</p>
]]></description><pubDate>Sat, 28 Mar 2026 18:16:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47557015</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47557015</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47557015</guid></item><item><title><![CDATA[New comment by rahen in "Paper Tape Is All You Need – Training a Transformer on a 1976 Minicomputer"]]></title><description><![CDATA[
<p>That thing is a Tamagochi though, it constantly needs attention, pardon the pun. I did most of the development and tuning on ll-34 for that reason.</p>
]]></description><pubDate>Sat, 28 Mar 2026 13:57:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=47554682</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47554682</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47554682</guid></item><item><title><![CDATA[New comment by rahen in "Paper Tape Is All You Need – Training a Transformer on a 1976 Minicomputer"]]></title><description><![CDATA[
<p>Thanks for reposting! I'm the author of ATTN-11. Happy to answer any questions about the fixed-point arithmetic, the PDP-11 hardware, or the training process.</p>
]]></description><pubDate>Sat, 28 Mar 2026 13:49:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=47554584</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47554584</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47554584</guid></item><item><title><![CDATA[Paper Tape Is All You Need – Training a Transformer on a 1976 Minicomputer]]></title><description><![CDATA[
<p>Article URL: <a href="https://github.com/dbrll/ATTN-11">https://github.com/dbrll/ATTN-11</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47518568">https://news.ycombinator.com/item?id=47518568</a></p>
<p>Points: 145</p>
<p># Comments: 26</p>
]]></description><pubDate>Wed, 25 Mar 2026 15:20:08 +0000</pubDate><link>https://github.com/dbrll/ATTN-11</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=47518568</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47518568</guid></item><item><title><![CDATA[New comment by rahen in "The Connection Machine CM-1 "Feynman" T-shirt"]]></title><description><![CDATA[
<p>Around the same time (1984), there was also another very cool piece of technology that often gets overlooked: the CMU WARP. It wasn’t as flashy as the Crays and the Connection Machine, but it was the first systolic array accelerator (what we’d now call TPUs). It packed as much MFLOPS as a Cray 1.<p>It's also the computer that powered the Chevrolet Navlab self-driving car in 1986.</p>
]]></description><pubDate>Tue, 03 Feb 2026 08:54:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=46868384</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46868384</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46868384</guid></item><item><title><![CDATA[New comment by rahen in "Sheaf – A Functional Language for Differentiable Programs"]]></title><description><![CDATA[
<p>I've been building a functional language for differentiable programming that compiles to JAX. The core idea is homoiconicity applied to ML, models are data structures that can inspect and transform themselves.</p>
]]></description><pubDate>Wed, 28 Jan 2026 14:59:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=46796190</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46796190</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46796190</guid></item><item><title><![CDATA[Sheaf – A Functional Language for Differentiable Programs]]></title><description><![CDATA[
<p>Article URL: <a href="https://sheaf-lang.org/">https://sheaf-lang.org/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46796189">https://news.ycombinator.com/item?id=46796189</a></p>
<p>Points: 3</p>
<p># Comments: 1</p>
]]></description><pubDate>Wed, 28 Jan 2026 14:59:58 +0000</pubDate><link>https://sheaf-lang.org/</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46796189</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46796189</guid></item><item><title><![CDATA[New comment by rahen in "The Emacs Widget Library: A Critique and Case Study"]]></title><description><![CDATA[
<p>For those interested, this guy is revamping the Emacs widget library with something more modern and platform agnostic, based on SDL: <a href="https://appetrosyan.github.io/posts/" rel="nofollow">https://appetrosyan.github.io/posts/</a><p>His posts are very insightful.</p>
]]></description><pubDate>Wed, 14 Jan 2026 09:59:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=46614237</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46614237</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46614237</guid></item><item><title><![CDATA[New comment by rahen in "I switched from VSCode to Zed"]]></title><description><![CDATA[
<p>My only complaint regarding the Zed editor is the inability to display two panes of the sidebar one below the other. Not only is it impossible to display them together, but switching between them requires clicking a tiny button in the status bar. To make matters worse, performing a search hides the symbols and the tree view.<p>So right now I'm sticking to Emacs.</p>
]]></description><pubDate>Tue, 06 Jan 2026 12:26:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=46511414</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46511414</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46511414</guid></item><item><title><![CDATA[New comment by rahen in "Go away Python"]]></title><description><![CDATA[
<p>Lisp isn't missing anything, it's a natural fit for AI/ML. It’s the ecosystem's tooling that needs catching up.<p>The code hasn't reached RC yet, but I'll definitely post a Show HN once it's ready for a preview.</p>
]]></description><pubDate>Tue, 30 Dec 2025 19:51:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=46437224</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46437224</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46437224</guid></item><item><title><![CDATA[New comment by rahen in "Go away Python"]]></title><description><![CDATA[
<p>I doubt it considering there are massive Clojure codebases with large teams collaborating on them every day. The lack of Lisp tooling and the prevalence of Python are more a result of inertia, low barrier to entry and ecosystem lock-in.</p>
]]></description><pubDate>Tue, 30 Dec 2025 17:51:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=46435936</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46435936</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46435936</guid></item><item><title><![CDATA[New comment by rahen in "Go away Python"]]></title><description><![CDATA[
<p>"> I think all of ML being in Python is a colossal mistake that we'll pay for for years.<p>Market pressure. Early ML frameworks were in Lisp, then eventually Lua with Torch, but demand dictated the choice of Python because "it's simple" even if the result is cobbled together.<p>Lisp is arguably still the most suitable language for neural networks for a lot of reasons beyond the scope of this post, but the tooling is missing. I’m developing such a framework right now, though I have no illusions that many will adopt it. Python may not be elegant or efficient, but it's simple, and that's what people want.</p>
]]></description><pubDate>Tue, 30 Dec 2025 17:35:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=46435755</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46435755</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46435755</guid></item><item><title><![CDATA[New comment by rahen in "Show HN: Z80-μLM, a 'Conversational AI' That Fits in 40KB"]]></title><description><![CDATA[
<p>I love it, instant Github star.
I wrote an MLP in Fortran IV for a punched card machine from the sixties (<a href="https://github.com/dbrll/Xortran" rel="nofollow">https://github.com/dbrll/Xortran</a>), so this really speaks to me.<p>The interaction is surprisingly good despite the lack of attention mechanism and the limitation of the "context" to trigrams from the last sentence.<p>This could have worked on 60s-era hardware and would have completely changed the world (and science fiction) back then. Great job.</p>
]]></description><pubDate>Mon, 29 Dec 2025 10:09:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=46419112</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46419112</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46419112</guid></item><item><title><![CDATA[New comment by rahen in "Learning Fortran (2024)"]]></title><description><![CDATA[
<p>The reverse is true though, and I find that fascinating with Fortran.<p>I recently learned Fortran IV to build a backpropagated neural network for the IBM 1130 (1965) and was amazed to see it compile with no warning on both the punched card compiler from IBM and on a modern Fortran compiler (gfortran).<p>Some Fortran II conventions, like the arithmetic IF, are now deprecated, but --std=legacy is all it takes to make it work.</p>
]]></description><pubDate>Wed, 17 Dec 2025 23:44:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=46307190</link><dc:creator>rahen</dc:creator><comments>https://news.ycombinator.com/item?id=46307190</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46307190</guid></item></channel></rss>