<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: IceWreck</title><link>https://news.ycombinator.com/user?id=IceWreck</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 06 Apr 2026 05:38:13 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=IceWreck" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by IceWreck in "Running Gemma 4 locally with LM Studio's new headless CLI and Claude Code"]]></title><description><![CDATA[
<p>> So while feasible it's only great for batch jobs not interactive usage.<p>I mean yeah true but depends on how big the model is. The example I gave (Qwen 3.5 35BA3B) was fitting a 35B Q4 K_M (say 20 GB in size) model in 12 GB VRAM. With a 4070Ti + high speed 32 GB DDR5 ram you can easily get 700 token/sec prompt processing and 55-60 token/sec generation which is quite fast.<p>On the other hand if I try to fit a 120B model in 96 GB of DDR5 + the same 12 GB VRAM I get 2-5 token/sec generation.</p>
]]></description><pubDate>Sun, 05 Apr 2026 22:45:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=47654715</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=47654715</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47654715</guid></item><item><title><![CDATA[New comment by IceWreck in "Running Gemma 4 locally with LM Studio's new headless CLI and Claude Code"]]></title><description><![CDATA[
<p>It does if you use an inference engine where you can offload some of the experts from VRAM to CPU RAM.
That means I can fit a 35 billion param MoE in let's say 12 GB VRAM GPU + 16 gigs of memory.</p>
]]></description><pubDate>Sun, 05 Apr 2026 20:03:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=47653312</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=47653312</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47653312</guid></item><item><title><![CDATA[New comment by IceWreck in "Components of a Coding Agent"]]></title><description><![CDATA[
<p>> This is speculative, but I suspect that if we dropped one of the latest, most capable open-weight LLMs, such as GLM-5, into a similar harness, it could likely perform on par with GPT-5.4 in Codex or Claude Opus 4.6 in Claude Code.<p>People have been doing that for over a year already? GLM officially recommends plugging into Claude Code <a href="https://docs.z.ai/devpack/tool/claude" rel="nofollow">https://docs.z.ai/devpack/tool/claude</a> and any model can be plugged into Codex CLI (it's open source and can be set via config file).</p>
]]></description><pubDate>Sat, 04 Apr 2026 21:22:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=47643558</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=47643558</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47643558</guid></item><item><title><![CDATA[New comment by IceWreck in "Claude Code's source code has been leaked via a map file in their NPM registry"]]></title><description><![CDATA[
<p>> What Google and OpenAi have open sourced is their Agents SDK, a toolkit, not the secret sauce of how their flagship agents are wired under the hood<p>And how is that any different? Claude Code is a harness, similar to open source ones like Codex, Gemini CLI, OpenCode etc. Their prompts were already public because you could connect it to your own LLM gateway and see everything. The code was transpiled javascript which is trivial to read with LLMs anyways.</p>
]]></description><pubDate>Tue, 31 Mar 2026 18:08:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=47591275</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=47591275</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47591275</guid></item><item><title><![CDATA[New comment by IceWreck in "Astral to Join OpenAI"]]></title><description><![CDATA[
<p>basedpyright has existed for years and now we have pyrefly from meta too. I think ty is also working on one.</p>
]]></description><pubDate>Sat, 21 Mar 2026 16:57:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=47468794</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=47468794</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47468794</guid></item><item><title><![CDATA[New comment by IceWreck in "Just-bash: Bash for Agents"]]></title><description><![CDATA[
<p>At this point why not make the agents use a restricted subset of python, typescript or lua or something.<p>Bash has been unchanged for decades but its not a very nice language.<p>I know pydantic has been experimenting with <a href="https://github.com/pydantic/monty" rel="nofollow">https://github.com/pydantic/monty</a> (restricted python) and I think Cloudflare and co were experimenting with giving typescript to agents.</p>
]]></description><pubDate>Thu, 26 Feb 2026 15:18:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=47167218</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=47167218</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47167218</guid></item><item><title><![CDATA[New comment by IceWreck in "zclaw: personal AI assistant in under 888 KB, running on an ESP32"]]></title><description><![CDATA[
<p>I've been using <a href="https://github.com/sipeed/picoclaw" rel="nofollow">https://github.com/sipeed/picoclaw</a></p>
]]></description><pubDate>Sun, 22 Feb 2026 18:48:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=47113546</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=47113546</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47113546</guid></item><item><title><![CDATA[New comment by IceWreck in "QNX Self-Hosted Developer Desktop"]]></title><description><![CDATA[
<p>Blackberry OS 10 was also running QNX under the hook afaik.</p>
]]></description><pubDate>Sat, 27 Dec 2025 10:33:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=46400736</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=46400736</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46400736</guid></item><item><title><![CDATA[New comment by IceWreck in "Nvidia to buy assets from Groq for $20B cash"]]></title><description><![CDATA[
<p>This is exactly what Google did with Windsurf and similar to what Meta did with Scale AI. Seems like a rising trend,</p>
]]></description><pubDate>Fri, 26 Dec 2025 01:02:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=46388210</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=46388210</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46388210</guid></item><item><title><![CDATA[New comment by IceWreck in "NIST was 5 μs off UTC after last week's power cut"]]></title><description><![CDATA[
<p>We need nanosecond precision for trading - basically timestamping exchange/own/other events and to measure latency.</p>
]]></description><pubDate>Mon, 22 Dec 2025 19:24:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=46357799</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=46357799</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46357799</guid></item><item><title><![CDATA[New comment by IceWreck in "Go is portable, until it isn't"]]></title><description><![CDATA[
<p>You're linking to a different version - this is the one that most people use <a href="https://github.com/modernc-org/sqlite" rel="nofollow">https://github.com/modernc-org/sqlite</a></p>
]]></description><pubDate>Sat, 13 Dec 2025 09:12:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=46253226</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=46253226</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46253226</guid></item><item><title><![CDATA[New comment by IceWreck in "Tongyi DeepResearch – open-source 30B MoE Model that rivals OpenAI DeepResearch"]]></title><description><![CDATA[
<p>LlamaCPP supports offloading some experts in a MoE model to CPU. The results are very good and even weaker GPUs can run larger models at reasonable speeds.<p>n-cpu-moe in <a href="https://github.com/ggml-org/llama.cpp/blob/master/tools/server/README.md" rel="nofollow">https://github.com/ggml-org/llama.cpp/blob/master/tools/serv...</a></p>
]]></description><pubDate>Sun, 02 Nov 2025 18:54:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=45792491</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=45792491</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45792491</guid></item><item><title><![CDATA[New comment by IceWreck in "Claude Code 2.0"]]></title><description><![CDATA[
<p>I was using aider quite a lot from ~ 7 months ago to ~ 3 months ago.
I had to stop because they refuse to implement MCPs and Claude/Codex style agentic workflow just yields better results.</p>
]]></description><pubDate>Mon, 29 Sep 2025 19:14:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=45417611</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=45417611</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45417611</guid></item><item><title><![CDATA[(Ab)using Agentic Coding CLIs for Data Cleaning and Standardisation]]></title><description><![CDATA[
<p>Article URL: <a href="https://abifog.com/blog/data-standardisation-with-agentic-clis/">https://abifog.com/blog/data-standardisation-with-agentic-clis/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45399003">https://news.ycombinator.com/item?id=45399003</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 27 Sep 2025 20:18:10 +0000</pubDate><link>https://abifog.com/blog/data-standardisation-with-agentic-clis/</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=45399003</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45399003</guid></item><item><title><![CDATA[New comment by IceWreck in "Seedbox Lite: A lightweight torrent streaming app with instant playback"]]></title><description><![CDATA[
<p>Does it download torrents on your server or web torrent on your browser? - the readme really doesn't say.<p>Imo downloading on the server is more useful. Web torrent is great but I don't think it's very practical in many places.</p>
]]></description><pubDate>Fri, 29 Aug 2025 16:31:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=45066193</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=45066193</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45066193</guid></item><item><title><![CDATA[New comment by IceWreck in "F-Droid build servers can't build modern Android apps due to outdated CPUs"]]></title><description><![CDATA[
<p>Huawei and Honor are seperate app stores?<p>And Oppo and Vivo too?<p>In both instances one company owns the other - why have competing app stores?</p>
]]></description><pubDate>Wed, 13 Aug 2025 08:26:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=44885893</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=44885893</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44885893</guid></item><item><title><![CDATA[New comment by IceWreck in "Ollama's new app"]]></title><description><![CDATA[
<p>Why not Linux? The UI looks to be some kind chrome based thingy - probably electron - should be easy to port to Linux.<p>Also is there a link to the source?</p>
]]></description><pubDate>Wed, 30 Jul 2025 22:23:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=44740213</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=44740213</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44740213</guid></item><item><title><![CDATA[New comment by IceWreck in "Qwen3-30B-A3B"]]></title><description><![CDATA[
<p>You can already use in it ollama by using the unsloth quants:<p>```
ollama run hf.co/unsloth/Qwen3-30B-A3B-Instruct-2507-GGUF:Q4_K_M
```<p>> And what is the best offline model for coding?<p>That would depend on your hardware.</p>
]]></description><pubDate>Tue, 29 Jul 2025 23:20:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=44729376</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=44729376</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44729376</guid></item><item><title><![CDATA[New comment by IceWreck in "Harper – an open-source alternative to Grammarly"]]></title><description><![CDATA[
<p>Slightly controversial compared to other comments here but I haven't used Grammerly at all since LLMs came out. Even a 4B local LLM is good enough to rephrase all forms of text and fix most grammer mistakes.</p>
]]></description><pubDate>Fri, 20 Jun 2025 23:10:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=44332963</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=44332963</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44332963</guid></item><item><title><![CDATA[New comment by IceWreck in "Google Duo will be replaced by Google Meet in Sept 2025"]]></title><description><![CDATA[
<p>I think Allo and YouTube Chat were also around for the same time as Duo.</p>
]]></description><pubDate>Sat, 31 May 2025 08:23:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=44142838</link><dc:creator>IceWreck</dc:creator><comments>https://news.ycombinator.com/item?id=44142838</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44142838</guid></item></channel></rss>