<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: vparseval</title><link>https://news.ycombinator.com/user?id=vparseval</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 06 May 2026 16:56:22 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=vparseval" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by vparseval in "Accelerating Gemma 4: faster inference with multi-token prediction drafters"]]></title><description><![CDATA[
<p>qwen-3.6 is really interesting. The dense 27B model is pretty slow for me whereas the sparse 31B is blazingly fast but it also needs to be since it's so chatty. It produces pages and pages of stream of consciousness stuff. 27B does this to a lesser extent but slow enough that I can actually read it whereas 31B just blasts by.<p>I haven't yet compared either to Gemma 4. I tried that out the day after it came out with the patched llama.cpp that added support for it but I couldn't make tool calling work and so it was kind of useless. I should try again to see if things have changed but judging by what people say, qwen-3.6 seems stronger for coding anyway.</p>
]]></description><pubDate>Wed, 06 May 2026 01:29:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=48031054</link><dc:creator>vparseval</dc:creator><comments>https://news.ycombinator.com/item?id=48031054</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48031054</guid></item><item><title><![CDATA[New comment by vparseval in "Show HN: A (marginally) useful x86-64 ELF executable in 301 bytes"]]></title><description><![CDATA[
<p>Love it! It's entirely inapplicable and useless to me but it embodies the spirit of Show HN and what the spirit of programming in the 80s and 90s was.</p>
]]></description><pubDate>Thu, 09 Apr 2026 04:23:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=47699250</link><dc:creator>vparseval</dc:creator><comments>https://news.ycombinator.com/item?id=47699250</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47699250</guid></item><item><title><![CDATA[New comment by vparseval in "Google releases Gemma 4 open models"]]></title><description><![CDATA[
<p>I found that you can run models locally pretty well that exceed your VRAM by a bit. At least ollama will hand excess off to your system RAM. Maybe performance suffers but I've never actually seen it crap out and I can wait a few minutes for a response.</p>
]]></description><pubDate>Fri, 03 Apr 2026 02:28:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47622607</link><dc:creator>vparseval</dc:creator><comments>https://news.ycombinator.com/item?id=47622607</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47622607</guid></item><item><title><![CDATA[New comment by vparseval in "Is anybody else bored of talking about AI?"]]></title><description><![CDATA[
<p>I don't remember exactly in which book's introduction Hannah Arendt mentioned this, but she pointed out that every time humanity learned a new skill that improved its efficiency in some capacity, that skill as well as adjacent skills diminished irrevocably.<p>AI is the thing that for the first time can think better than us (or so at least some people believe) and is seen as an efficiency booster in the world of cognition and ideas. I'd think Hannah Arendt would be worried with what we are currently seeing and where we might be headed.</p>
]]></description><pubDate>Wed, 25 Mar 2026 01:45:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=47512146</link><dc:creator>vparseval</dc:creator><comments>https://news.ycombinator.com/item?id=47512146</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47512146</guid></item></channel></rss>