<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: am17an</title><link>https://news.ycombinator.com/user?id=am17an</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 17 Apr 2026 12:48:21 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=am17an" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by am17an in "Apple's accidental moat: How the "AI Loser" may end up winning"]]></title><description><![CDATA[
<p>Don’t underestimate the march of technology. Just look at your phone, it has more FLOPS than there were in the entire world 40 years ago.</p>
]]></description><pubDate>Mon, 13 Apr 2026 05:46:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47748076</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47748076</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47748076</guid></item><item><title><![CDATA[New comment by am17an in "How the AI Bubble Bursts"]]></title><description><![CDATA[
<p>Thank you, there are two things I would like to point out:<p>1) Google releasing something probably means they don't see it as important. 4-bit KV-cache quantization has been known for a long time. The fact there is almost a mass hysteria about this paper makes me think there is a lack of skepticism in this AI mania, even in relatively tech-savvy crowd.<p>2) But prices for memory companies are crashing! look around, the whole market is crashing.</p>
]]></description><pubDate>Mon, 30 Mar 2026 15:48:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=47575793</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47575793</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47575793</guid></item><item><title><![CDATA[New comment by am17an in "What if AI doesn't need more RAM but better math?"]]></title><description><![CDATA[
<p>There are techniques which already achieve great compression of the cache at 4 bit, eg using hadamard transforms. Going from
4 bit to 3 bit isn’t the great leap people expect this to be. It’s actually slower to run and is generally worse in practice.</p>
]]></description><pubDate>Sun, 29 Mar 2026 16:58:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47564911</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47564911</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47564911</guid></item><item><title><![CDATA[New comment by am17an in "Astral to Join OpenAI"]]></title><description><![CDATA[
<p>Welp, back to pip</p>
]]></description><pubDate>Thu, 19 Mar 2026 14:32:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47440213</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47440213</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47440213</guid></item><item><title><![CDATA[New comment by am17an in "Allow me to get to know you, mistakes and all"]]></title><description><![CDATA[
<p>Working in open source, I've now heard a wide variety of disabilities that people have and they <i>have</i> to be aided by an LLM for writing even descriptions of their PRs.</p>
]]></description><pubDate>Sun, 15 Mar 2026 11:02:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=47386216</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47386216</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47386216</guid></item><item><title><![CDATA[New comment by am17an in "Can I run AI locally?"]]></title><description><![CDATA[
<p>You can still run larger MoE models using expert weight off-loading to the CPU for token generation. They are by and large useable, I get ~50 toks/second on a kimi linear 48B (3B active) model on a potato PC + a 3090</p>
]]></description><pubDate>Fri, 13 Mar 2026 17:44:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47367387</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47367387</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47367387</guid></item><item><title><![CDATA[New comment by am17an in "Intelligence is a commodity. Context is the real AI Moat"]]></title><description><![CDATA[
<p>Sure. “Tell me a joke”</p>
]]></description><pubDate>Thu, 05 Mar 2026 17:39:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47264696</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47264696</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47264696</guid></item><item><title><![CDATA[New comment by am17an in "OpenAI – How to delete your account"]]></title><description><![CDATA[
<p>I was referring to the 35B version. It is surprisingly good for its size. You can use it for implementation tasks without it going off the rails</p>
]]></description><pubDate>Sun, 01 Mar 2026 18:03:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=47209063</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47209063</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47209063</guid></item><item><title><![CDATA[New comment by am17an in "Ghostty – Terminal Emulator"]]></title><description><![CDATA[
<p>Damn I’m jealous that they figured out how to pay their contributors. I’ve been toiling away for free</p>
]]></description><pubDate>Sun, 01 Mar 2026 17:34:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=47208777</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47208777</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47208777</guid></item><item><title><![CDATA[New comment by am17an in "OpenAI – How to delete your account"]]></title><description><![CDATA[
<p>They already have with qwen3.5</p>
]]></description><pubDate>Sat, 28 Feb 2026 15:02:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47196114</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47196114</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47196114</guid></item><item><title><![CDATA[New comment by am17an in "Unsloth Dynamic 2.0 GGUFs"]]></title><description><![CDATA[
<p>What do you use for sub-50ms inference?</p>
]]></description><pubDate>Sat, 28 Feb 2026 11:09:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=47193695</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47193695</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47193695</guid></item><item><title><![CDATA[New comment by am17an in "How do I cancel my ChatGPT subscription?"]]></title><description><![CDATA[
<p>Honestly you can run this on a 16GB VRAM GPU with llama.cpp. Just try it!</p>
]]></description><pubDate>Sat, 28 Feb 2026 08:36:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=47192339</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47192339</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47192339</guid></item><item><title><![CDATA[New comment by am17an in "Ggml.ai joins Hugging Face to ensure the long-term progress of Local AI"]]></title><description><![CDATA[
<p>One often overlooked after that is ggml, the tensor library that runs llama.cpp is not based on pytorch, rather just plain cpp. In a world where pytorch dominates, it shows that alternatives are possible and are worthy to be pursued.</p>
]]></description><pubDate>Sat, 21 Feb 2026 11:05:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47099625</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47099625</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47099625</guid></item><item><title><![CDATA[New comment by am17an in "Do Not Outsource Judgement"]]></title><description><![CDATA[
<p>Holy smokes we're cooked.</p>
]]></description><pubDate>Sat, 14 Feb 2026 07:11:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=47012370</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47012370</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47012370</guid></item><item><title><![CDATA[New comment by am17an in "An AI agent published a hit piece on me"]]></title><description><![CDATA[
<p>Maintainers time is a more scarce resource than free tokens. I would much rather get my time back after reading those PRs</p>
]]></description><pubDate>Fri, 13 Feb 2026 10:50:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=47001298</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=47001298</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47001298</guid></item><item><title><![CDATA[New comment by am17an in "Anthropic's original take home assignment open sourced"]]></title><description><![CDATA[
<p>1) Python is unreadable."
Would you prefer C or C++?<p>> Unironically, yes. Unless I never plan to look at that code again</p>
]]></description><pubDate>Wed, 21 Jan 2026 13:18:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=46705315</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=46705315</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46705315</guid></item><item><title><![CDATA[Every LLM hallucinates that std:vector deletes elements in LIFO order]]></title><description><![CDATA[
<p>Article URL: <a href="https://am17an.bearblog.dev/every-llm-hallucinates-stdvector-deletes-elements-in-a-lifo-order/">https://am17an.bearblog.dev/every-llm-hallucinates-stdvector-deletes-elements-in-a-lifo-order/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46455487">https://news.ycombinator.com/item?id=46455487</a></p>
<p>Points: 6</p>
<p># Comments: 1</p>
]]></description><pubDate>Thu, 01 Jan 2026 16:43:04 +0000</pubDate><link>https://am17an.bearblog.dev/every-llm-hallucinates-stdvector-deletes-elements-in-a-lifo-order/</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=46455487</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46455487</guid></item><item><title><![CDATA[New comment by am17an in "A guide to local coding models"]]></title><description><![CDATA[
<p>Use llama.cpp? I get 250 toks/sec on gpt-oss using a 4090, not sure about the mac speeds</p>
]]></description><pubDate>Mon, 22 Dec 2025 04:56:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=46351436</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=46351436</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46351436</guid></item><item><title><![CDATA[New comment by am17an in "Your job is to deliver code you have proven to work"]]></title><description><![CDATA[
<p>Well a 1000 line PR is still not welcome. It puts too much of a burden on the maintainers. Small PRs are the way to go, tests are great too. If you have to submit a big PR, get buy in from a maintainer first that they will review your code.</p>
]]></description><pubDate>Thu, 18 Dec 2025 16:20:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=46314696</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=46314696</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46314696</guid></item><item><title><![CDATA[New comment by am17an in "How well do you know C++ auto type deduction?"]]></title><description><![CDATA[
<p>I agree, this would be in the same vein as "STL returns a verbose type, it's okay to use auto here because no-one cares"</p>
]]></description><pubDate>Mon, 15 Dec 2025 16:24:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=46276567</link><dc:creator>am17an</dc:creator><comments>https://news.ycombinator.com/item?id=46276567</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46276567</guid></item></channel></rss>