<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: deepsquirrelnet</title><link>https://news.ycombinator.com/user?id=deepsquirrelnet</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 17 Apr 2026 10:27:58 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=deepsquirrelnet" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by deepsquirrelnet in "Claude Opus 4.7"]]></title><description><![CDATA[
<p>My tinfoil hat theory, which may not be that crazy, is that providers are sandbagging their models in the days leading up to a new release, so that the next model "feels" like a bigger improvement than it is.<p>An important aspect of AI is that it needs to be seen as moving forward all the time. Plateaus are the death of the hype cycle, and would tether people's expectations closer to reality.</p>
]]></description><pubDate>Thu, 16 Apr 2026 16:01:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47795353</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47795353</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47795353</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "I went to America's worst national parks so you don't have to"]]></title><description><![CDATA[
<p>> Cuyahoga Valley: There is nothing wrong with Cuyahoga Valley. Statistically, you’re from Ohio, so why not?<p>In college, I took an interim elective course on geology of the national parks. On the first day of class, the professor asked an icebreaker for students to say which national park they lived closest to. I said Ohio - Cuyahoga Valley.<p>Well some snot nosed boy scout confidently piped up that there were mostly certainly no national parks in Ohio, and the professor agreed. This is a deep personal grudge that I still hold to this day.</p>
]]></description><pubDate>Mon, 13 Apr 2026 13:55:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47752035</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47752035</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47752035</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "Ask HN: What Are You Working On? (April 2026)"]]></title><description><![CDATA[
<p>CnakeCharmer - 
<a href="https://github.com/dleemiller/CnakeCharmer" rel="nofollow">https://github.com/dleemiller/CnakeCharmer</a><p><a href="https://huggingface.co/datasets/CnakeCharmer/CnakeCharmer" rel="nofollow">https://huggingface.co/datasets/CnakeCharmer/CnakeCharmer</a><p>This project started from a belief that llms should be better at doing python to cython code translations than they are. So we started setting a large set of parallel implementations.<p>Then I realized that Claude code was much better at working on the data using tools (mcp) to check and iterate. The scope transformed into an platform for creating the SFT agentic trace dataset using sandboxed tools for compilation, testing, linting, address sanitizing and benchmarking.<p>We still need to bulk up the GRPO dataset with a large number of good unmatched python examples. But early results using SFT only on gpt-oss 20b are quite good.</p>
]]></description><pubDate>Mon, 13 Apr 2026 13:19:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=47751555</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47751555</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47751555</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "Why do we tell ourselves scary stories about AI?"]]></title><description><![CDATA[
<p>There are so many reason if you look at how it's being sold.<p>* We need to completely deregulate <i>these US companies</i> so China doesn't win and take us over<p>* We need to heavily regulate anybody who is not following the rules that make us the de-facto winner<p>* This is so powerful it will take all the jobs (and therefore if you lead a company that isn't using AI, you will soon be obsolete)<p>* If you don't use AI, you will not be able to function in a future job<p>* We need to lineup an excuse to call our friends in government and turn off the open source spigot when the time is right<p>They have chosen fear as a motivator, and it is clearly working very well. It's easier to use fear now, while it's new and then flip the narrative once people are more familiar with it than to go the other direction. Companies are not just telling a story to hype their product, but why they alone are the ones that should be entrusted to build it.</p>
]]></description><pubDate>Fri, 10 Apr 2026 15:13:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47719383</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47719383</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47719383</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "GLM-5.1: Towards Long-Horizon Tasks"]]></title><description><![CDATA[
<p>I am working on a large scale dataset for producing agent traces for Python <> cython conversion with tooling, and it is second only to gemini pro 3.1 in acceptance rates (16% vs 26%).<p>Mid-sized models like gpt-oss minimax and qwen3.5 122b are around 6%, and gemma4 31b around 7% (but much slower).<p>I haven’t tried Opus or ChatGPT due to high costs on openrouter for this application.</p>
]]></description><pubDate>Tue, 07 Apr 2026 18:46:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47679639</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47679639</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47679639</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "If DSPy is so great, why isn't anyone using it?"]]></title><description><![CDATA[
<p>Good article, and I think the "evolution of every AI system" is spot on.<p>In my opinion, the reason people don't use DSPy is because DSPy aims to be a machine learning platform. And like the article says -- this feels different or hard to people who are not used to engineering with probabilistic outputs. But these days, many more people are programming with probability machines than ever before.<p>The absolute biggest time sink and 'here be dragons' of using LLMs is poke and hope prompt "engineering" without proper evaluation metrics.<p>> You don’t have to use DSPy. But you should build like someone who understands why it exists.<p>And this is the salient point, and I think it's very well stated. It's not about the framework per se, but about the methodology.</p>
]]></description><pubDate>Mon, 23 Mar 2026 16:13:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=47491486</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47491486</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47491486</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "Super Micro Shares Plunge 25% After Co-Founder Charged in $2.5B Smuggling Plot"]]></title><description><![CDATA[
<p>This is even after the Hindenburg research report that found numerous screaming red flags a few years ago.<p><a href="https://hindenburgresearch.com/smci/" rel="nofollow">https://hindenburgresearch.com/smci/</a></p>
]]></description><pubDate>Fri, 20 Mar 2026 18:21:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=47458567</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47458567</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47458567</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "What makes Intel Optane stand out (2023)"]]></title><description><![CDATA[
<p>I worked at Micron in the SSD division when Optane (originally called crosspoint “Xpoint”) was being made. In my mind, there was never a real serious push to productize it. But it’s not clear to me whether that was due to unattractive terms of the joint venture or lack of clear product fit.<p>There was certainly a time when it seemed they were shopping for engineers opinions of what to do with it, but I think they quickly determined it would be a much smaller market anyway from ssds and didn’t end up pushing on it too hard. I could be wrong though, it’s a big company and my corner was manufacturing and not product development.</p>
]]></description><pubDate>Sun, 15 Mar 2026 16:57:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47389251</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47389251</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47389251</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "“This is not the computer for you”"]]></title><description><![CDATA[
<p>Interesting read! I love to see this spirit. I grew up with a different - but similar experience. Only, as an 80s and 90s kid, computers were nothing but limitations. Even when my dad built a machine with a 133MHz Cyrix chip, already a year later, it was outdated by computers with literally double the computing power.<p>That Cyrix machine was already miles ahead of the 386 that was handed down to me to play text based games on and learn dos through hard knocks. I remember leafing through old hard drives that had 10mb of capacity and realizing they had no value despite not being that old.<p>Later in college, I had the confidence to build my own first desktop with parts cobbled together from sketchy resellers. Athlon A1 single core 1ghz. Man that thing could fly.</p>
]]></description><pubDate>Fri, 13 Mar 2026 16:04:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47366275</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47366275</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47366275</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "How much of HN is AI?"]]></title><description><![CDATA[
<p>> I tapped into Pangram. Pangram is a remarkably good, conservative model for detecting LLM-generated text. These detectors have a bad rep among techies, but the objections are often based on outdated assumptions<p>Turing test is really in the rearview, huh?<p>Humans need machines to detect if a machine wrote the text, because humans aren’t sure.</p>
]]></description><pubDate>Thu, 12 Mar 2026 02:19:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47345531</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47345531</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47345531</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "BitNet: 100B Param 1-Bit model for local CPUs"]]></title><description><![CDATA[
<p>The title being misleading is important as well, because this has landed on the front page, and the only thing that would be the only notable part of this submission.<p>The "new" on huggingface banner has weights that were uploaded 11 months ago, and it's 2B params. Work on this in the repo is 2 years old.<p>The amount of publicity compared to the anemic delivery for BitNet is impressive.</p>
]]></description><pubDate>Wed, 11 Mar 2026 17:09:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=47338273</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47338273</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47338273</guid></item><item><title><![CDATA[Show HN: FizzBuzz Forever – Agent Edition]]></title><description><![CDATA[
<p>Step up your game and land your next job when you bust out FizzBuzz Forever Agent Edition.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47331058">https://news.ycombinator.com/item?id=47331058</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 11 Mar 2026 02:07:54 +0000</pubDate><link>https://github.com/dleemiller/fizzbuzz-forever</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47331058</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47331058</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "GLiNER2: Unified Schema-Based Information Extraction"]]></title><description><![CDATA[
<p>Zero-shot encoder models are so cool. I'll definitely be checking this out.<p>If you're looking for a zero-shot classifier, tasksource is in a similar vein.<p><a href="https://huggingface.co/tasksource/ModernBERT-large-nli" rel="nofollow">https://huggingface.co/tasksource/ModernBERT-large-nli</a></p>
]]></description><pubDate>Thu, 05 Mar 2026 22:45:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=47268299</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47268299</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47268299</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "Show HN: Timber – Ollama for classical ML models, 336x faster than Python"]]></title><description><![CDATA[
<p>Does this use something like xnnpack under the hood?</p>
]]></description><pubDate>Mon, 02 Mar 2026 14:07:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=47218145</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47218145</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47218145</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "Qwen3.5 122B and 35B models offer Sonnet 4.5 performance on local computers"]]></title><description><![CDATA[
<p>4-bit quantization on newer nvidia hardware is being supported in training as well these days. I believe the gpt-oss models were trained natively in MXFP4, which is a 4-bit floating point / e2m1 (2-exponent, 1 bit mantissa, 1 bit sign).<p>It doesn't seem terribly common yet though. I think it is challenging to keep it stable.<p>[1] <a href="https://www.opencompute.org/blog/amd-arm-intel-meta-microsoft-nvidia-and-qualcomm-standardize-next-generation-narrow-precision-data-formats-for-ai" rel="nofollow">https://www.opencompute.org/blog/amd-arm-intel-meta-microsof...</a><p>[2] <a href="https://www.opencompute.org/documents/ocp-microscaling-formats-mx-v1-0-spec-final-pdf" rel="nofollow">https://www.opencompute.org/documents/ocp-microscaling-forma...</a></p>
]]></description><pubDate>Sun, 01 Mar 2026 02:22:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=47202953</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47202953</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47202953</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "We do not think Anthropic should be designated as a supply chain risk"]]></title><description><![CDATA[
<p>That’s why I unsubbed today! Otherwise I might forget.</p>
]]></description><pubDate>Sun, 01 Mar 2026 00:13:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=47202083</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47202083</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47202083</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "Unsloth Dynamic 2.0 GGUFs"]]></title><description><![CDATA[
<p>I love the work unsloth is doing. I only wish gguf format had better vllm support. It’s sometimes hard to find trustworthy quants that work well with vllm.</p>
]]></description><pubDate>Sat, 28 Feb 2026 14:10:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47195593</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47195593</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47195593</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "We Will Not Be Divided"]]></title><description><![CDATA[
<p>Isn’t there some kind of term for when the government controls the means of production. I’ll think about it. It’s one of those terms that’s been thrown around so loosely by this regime you knew they were going there.</p>
]]></description><pubDate>Sat, 28 Feb 2026 13:53:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=47195404</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47195404</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47195404</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "I asked Claude for 37,500 random names, and it can't stop saying Marcus"]]></title><description><![CDATA[
<p>Ask an llm to pick a random number from 1-10. My money is on 7.<p>This is known to be a form of collapse from RL training, because base models do not exhibit it [1].<p>1. <a href="https://arxiv.org/abs/2505.00047" rel="nofollow">https://arxiv.org/abs/2505.00047</a></p>
]]></description><pubDate>Wed, 25 Feb 2026 19:47:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=47156818</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47156818</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47156818</guid></item><item><title><![CDATA[New comment by deepsquirrelnet in "US orders diplomats to fight data sovereignty initiatives"]]></title><description><![CDATA[
<p>The market reflects reality. The new reality is that the people who are invested apparently don't need liquidity, and bad news doesn't really matter.</p>
]]></description><pubDate>Wed, 25 Feb 2026 15:28:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=47152837</link><dc:creator>deepsquirrelnet</dc:creator><comments>https://news.ycombinator.com/item?id=47152837</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47152837</guid></item></channel></rss>