<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: tarruda</title><link>https://news.ycombinator.com/user?id=tarruda</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 05 Apr 2026 23:59:24 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=tarruda" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by tarruda in "What changes when you turn a Linux box into a router"]]></title><description><![CDATA[
<p>I currently do something similar.<p>My router is a 16GB n150 mini PC with dual NICs. The actual router OS is within openwrt VM managed by Incus (VM/Container hypervisor) that has both NICs passed through.<p>One of the NICs is connected to another OpenWrt wifi access point, and the other is connected to the ISP modem.<p>The n150 also has a wifi card that I setup as an additional AP I can connect to if something goes wrong with the virtualization setup.<p>Been running this for at least 6 months and has been working pretty well.</p>
]]></description><pubDate>Fri, 03 Apr 2026 22:25:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=47633148</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47633148</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47633148</guid></item><item><title><![CDATA[New comment by tarruda in "StepFun 3.5 Flash is #1 cost-effective model for OpenClaw tasks (300 battles)"]]></title><description><![CDATA[
<p>Benchmarks don't tell the whole story. For one-shot coding tasks, I found Step 3.5 Flash to be stronger even than Qwen 3.5 397B.</p>
]]></description><pubDate>Thu, 02 Apr 2026 11:23:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=47612928</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47612928</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47612928</guid></item><item><title><![CDATA[New comment by tarruda in "StepFun 3.5 Flash is #1 cost-effective model for OpenClaw tasks (300 battles)"]]></title><description><![CDATA[
<p>Since that discussion, they released the base model and a midtrain checkpoint:<p>- <a href="https://huggingface.co/stepfun-ai/Step-3.5-Flash-Base" rel="nofollow">https://huggingface.co/stepfun-ai/Step-3.5-Flash-Base</a><p>- <a href="https://huggingface.co/stepfun-ai/Step-3.5-Flash-Base-Midtrain" rel="nofollow">https://huggingface.co/stepfun-ai/Step-3.5-Flash-Base-Midtra...</a><p>I'm not aware of other AI labs that released base checkpoint for models in this size class. Qwen released some base models for 3.5, but the biggest one is the 35B checkpoint.<p>They also released the entire training pipeline:<p>- <a href="https://huggingface.co/datasets/stepfun-ai/Step-3.5-Flash-SFT" rel="nofollow">https://huggingface.co/datasets/stepfun-ai/Step-3.5-Flash-SF...</a><p>- <a href="https://github.com/stepfun-ai/SteptronOss" rel="nofollow">https://github.com/stepfun-ai/SteptronOss</a></p>
]]></description><pubDate>Wed, 01 Apr 2026 17:09:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=47603626</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47603626</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47603626</guid></item><item><title><![CDATA[New comment by tarruda in "Flash-MoE: Running a 397B Parameter Model on a Laptop"]]></title><description><![CDATA[
<p>> Have you compared against MLX?<p>I don't think MLX supports similar 2-bit quants, so I never tried 397B with MLX.<p>However I did try 4-bit MLX with other Qwen 3.5 models and yes it is significantly faster. I still prefer llama.cpp due to it being a one in all package:<p>- SOTA dynamic quants (especially ik_llama.cpp)
- amazing web ui with MCP support
- anthropic/openai compatible endpoints (means it can be used with virtually any harness)
- JSON constrained output which basically ensures tool call correctness.
- routing mode</p>
]]></description><pubDate>Sun, 22 Mar 2026 17:35:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=47479961</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47479961</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47479961</guid></item><item><title><![CDATA[New comment by tarruda in "Flash-MoE: Running a 397B Parameter Model on a Laptop"]]></title><description><![CDATA[
<p>>  in neovim and I feel so grateful that this tool is still receiving love and getting new updates.<p>@justinmk deserves the credit for this!</p>
]]></description><pubDate>Sun, 22 Mar 2026 17:30:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=47479908</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47479908</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47479908</guid></item><item><title><![CDATA[New comment by tarruda in "Flash-MoE: Running a 397B Parameter Model on a Laptop"]]></title><description><![CDATA[
<p>In my case it the 2.46BPW has been working flawless for tool calling, so I don't think 2-bit was the culprit for JSON failing.<p>They did reduce the number of experts, so maybe that was it?</p>
]]></description><pubDate>Sun, 22 Mar 2026 17:29:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=47479897</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47479897</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47479897</guid></item><item><title><![CDATA[New comment by tarruda in "Flash-MoE: Running a 397B Parameter Model on a Laptop"]]></title><description><![CDATA[
<p>Yes. Note that the only reason I acquired this device was to run LLMs, so I can dedicate its whole RAM to it. Probably not viable for a 128G device where you are actively using for other things.</p>
]]></description><pubDate>Sun, 22 Mar 2026 17:25:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=47479844</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47479844</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47479844</guid></item><item><title><![CDATA[New comment by tarruda in "Flash-MoE: Running a 397B Parameter Model on a Laptop"]]></title><description><![CDATA[
<p>> What's the tok/s you get these days?<p>I ran llama-bench a couple of weeks ago when there was a big speed improvement on llama.cpp (<a href="https://github.com/ggml-org/llama.cpp/pull/20361#issuecomment-4039467718" rel="nofollow">https://github.com/ggml-org/llama.cpp/pull/20361#issuecommen...</a>):<p><pre><code>    % llama-bench -m ~/ml-models/huggingface/ubergarm/Qwen3.5-397B-A17B-GGUF/smol-IQ2_XS/Qwen3.5-397B-A17B-smol-IQ2_XS-00001-of-00004.gguf -fa 1 -t 1 -ngl 99 -b 2048 -ub 2048 -d 0,10000,20000,30000,40000,50000,60000,70000,80000,90000,100000,150000,200000,250000
    ggml_metal_device_init: tensor API disabled for pre-M5 and pre-A19 devices
    ggml_metal_library_init: using embedded metal library
    ggml_metal_library_init: loaded in 0.008 sec
    ggml_metal_rsets_init: creating a residency set collection (keep_alive = 180 s)
    ggml_metal_device_init: GPU name:   MTL0
    ggml_metal_device_init: GPU family: MTLGPUFamilyApple7  (1007)
    ggml_metal_device_init: GPU family: MTLGPUFamilyCommon3 (3003)
    ggml_metal_device_init: GPU family: MTLGPUFamilyMetal3  (5001)
    ggml_metal_device_init: simdgroup reduction   = true
    ggml_metal_device_init: simdgroup matrix mul. = true
    ggml_metal_device_init: has unified memory    = true
    ggml_metal_device_init: has bfloat            = true
    ggml_metal_device_init: has tensor            = false
    ggml_metal_device_init: use residency sets    = true
    ggml_metal_device_init: use shared buffers    = true
    ggml_metal_device_init: recommendedMaxWorkingSetSize  = 134217.73 MB
    | ------------------------------ | ---------: | ---------: | ---------- | ------: | -------: | -: | --------------: | -------------------: |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |           pp512 |        189.67 ± 1.98 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |           tg128 |         19.98 ± 0.01 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  pp512 @ d10000 |        168.92 ± 0.55 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  tg128 @ d10000 |         18.93 ± 0.02 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  pp512 @ d20000 |        152.42 ± 0.22 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  tg128 @ d20000 |         17.87 ± 0.01 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  pp512 @ d30000 |        139.37 ± 0.28 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  tg128 @ d30000 |         17.12 ± 0.01 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  pp512 @ d40000 |        128.38 ± 0.33 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  tg128 @ d40000 |         16.38 ± 0.00 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  pp512 @ d50000 |        118.07 ± 0.55 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  tg128 @ d50000 |         15.66 ± 0.00 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  pp512 @ d60000 |        108.44 ± 0.38 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  tg128 @ d60000 |         14.98 ± 0.01 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  pp512 @ d70000 |         98.85 ± 0.18 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  tg128 @ d70000 |         14.36 ± 0.00 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  pp512 @ d80000 |         91.39 ± 0.49 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  tg128 @ d80000 |         13.84 ± 0.00 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  pp512 @ d90000 |         85.76 ± 0.24 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 |  tg128 @ d90000 |         13.30 ± 0.00 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 | pp512 @ d100000 |         80.19 ± 0.83 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 | tg128 @ d100000 |         12.82 ± 0.00 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 | pp512 @ d150000 |         54.46 ± 0.33 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 | tg128 @ d150000 |         10.17 ± 0.09 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 | pp512 @ d200000 |         47.05 ± 0.15 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 | tg128 @ d200000 |          9.04 ± 0.02 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 | pp512 @ d250000 |         40.71 ± 0.26 |
    | qwen35moe 397B.A17B Q8_0       | 113.41 GiB |   396.35 B | MTL,BLAS   |       1 |     2048 |  1 | tg128 @ d250000 |          8.01 ± 0.02 |

    build: d28961d81 (8299)
</code></pre>
So it starts at 20 tps tg and 190 tps pp with empty context and ends at 8 tps tg and 40 tps pp with 250k prefill.<p>I suspect that there are still a lot of optimizations to be implemented for Qwen 3.5 on llama.cpp, wouldn't be surprised to reach 25 tps in a few months.<p>> You're the guy who launched Neovim!<p>That's me ;D<p>> I use it every day.<p>So do I for the past 12 years! Though I admit in the past year I greatly reduced the amount of code I write by hand :/</p>
]]></description><pubDate>Sun, 22 Mar 2026 16:39:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=47479299</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47479299</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47479299</guid></item><item><title><![CDATA[New comment by tarruda in "Flash-MoE: Running a 397B Parameter Model on a Laptop"]]></title><description><![CDATA[
<p>I don't think I've ever seen the M1 ultra GPU exceed 80w in asitop.<p>Update: I just did a quick asitop test while inferencing and the GPU power was averaging at 53.55</p>
]]></description><pubDate>Sun, 22 Mar 2026 16:03:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47478898</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47478898</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47478898</guid></item><item><title><![CDATA[New comment by tarruda in "Flash-MoE: Running a 397B Parameter Model on a Laptop"]]></title><description><![CDATA[
<p>I can't say anything about the OP method, but I already tested the smol-IQ2_XS quant (which has 2.46 BPW) with the pi harness. I did not do a very long session because token generation and prompt processing gets very slow, but I think I worked for up to ~70k context and it maintained a lot of coherence in the session. IIRC the GPQA diamond is supposed to exercise long chains of thought and it scored exceptionally well with 82% (the original BF16 official number is 88%: <a href="https://huggingface.co/Qwen/Qwen3.5-397B-A17B" rel="nofollow">https://huggingface.co/Qwen/Qwen3.5-397B-A17B</a>).<p>Note that not all quants are the same at a certain BPW. The smol-IQ2_XS quant I linked is pretty dynamic, with some tensors having q8_0 type, some q6_k and some q4_k (while the majority is iq2_xs). In my testing, this smol-IQ2_XS quant is the best available at this BPW range.<p>Eventually I might try a more practical eval such as terminal bench.</p>
]]></description><pubDate>Sun, 22 Mar 2026 14:42:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=47478015</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47478015</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47478015</guid></item><item><title><![CDATA[New comment by tarruda in "Flash-MoE: Running a 397B Parameter Model on a Laptop"]]></title><description><![CDATA[
<p>Note that this is not the only way to run Qwen 3.5 397B on consumer devices, there are excellent ~2.5 BPW quants available that make it viable for 128G devices.<p>I've had great success (~20 t/s) running it on a M1 Ultra with room for 256k context. Here are some lm-evaluation-harness results I ran against it:<p><pre><code>    mmlu: 87.86%

    gpqa diamond: 82.32%

    gsm8k: 86.43%

    ifeval: 75.90%
</code></pre>
More details of my experience:<p>- <a href="https://huggingface.co/ubergarm/Qwen3.5-397B-A17B-GGUF/discussions/8" rel="nofollow">https://huggingface.co/ubergarm/Qwen3.5-397B-A17B-GGUF/discu...</a><p>- <a href="https://huggingface.co/ubergarm/Qwen3.5-397B-A17B-GGUF/discussions/2" rel="nofollow">https://huggingface.co/ubergarm/Qwen3.5-397B-A17B-GGUF/discu...</a><p>- <a href="https://gist.github.com/simonw/67c754bbc0bc609a6caedee16fef89e8?permalink_comment_id=5991165#gistcomment-5991165" rel="nofollow">https://gist.github.com/simonw/67c754bbc0bc609a6caedee16fef8...</a><p>Overall an excellent model to have for offline inference.</p>
]]></description><pubDate>Sun, 22 Mar 2026 13:50:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=47477552</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47477552</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47477552</guid></item><item><title><![CDATA[New comment by tarruda in "Motorola GrapheneOS devices will be bootloader unlockable/relockable"]]></title><description><![CDATA[
<p>Thanks for the info!</p>
]]></description><pubDate>Thu, 05 Mar 2026 08:00:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=47258906</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47258906</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47258906</guid></item><item><title><![CDATA[New comment by tarruda in "Motorola GrapheneOS devices will be bootloader unlockable/relockable"]]></title><description><![CDATA[
<p>I would rather have a phone that doesn't let my carrier show random messages whenever they feel like it.</p>
]]></description><pubDate>Wed, 04 Mar 2026 17:06:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=47250537</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47250537</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47250537</guid></item><item><title><![CDATA[New comment by tarruda in "Motorola GrapheneOS devices will be bootloader unlockable/relockable"]]></title><description><![CDATA[
<p>Just a message popup, a window with dark background and some text ad on it.<p>I did not buy this phone from a carrier, just added the SIM card later.<p>Really surprised to learn this doesn't happen to others. Always assumed that the SIM card had some special privilege given by Android.</p>
]]></description><pubDate>Wed, 04 Mar 2026 17:06:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=47250517</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47250517</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47250517</guid></item><item><title><![CDATA[New comment by tarruda in "Motorola GrapheneOS devices will be bootloader unlockable/relockable"]]></title><description><![CDATA[
<p>Just checked, and only "Phone" and "Google" have this permission.<p>There are no preinstalled apps, I bought this phone clean on Germany and then added a Brazil's SIM card when I got back.<p>Could it be that the SIM card has some control over the Phone app?</p>
]]></description><pubDate>Wed, 04 Mar 2026 17:03:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47250476</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47250476</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47250476</guid></item><item><title><![CDATA[New comment by tarruda in "Motorola GrapheneOS devices will be bootloader unlockable/relockable"]]></title><description><![CDATA[
<p>I have a pixel 8a with a TIM SIM card  and every once in a while I see an ad popup on my phone.</p>
]]></description><pubDate>Wed, 04 Mar 2026 12:42:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47246658</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47246658</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47246658</guid></item><item><title><![CDATA[New comment by tarruda in "Motorola GrapheneOS devices will be bootloader unlockable/relockable"]]></title><description><![CDATA[
<p>One thing that annoys me is the ability that my mobile carrier has to just throw ad popups.<p>Is that something that GrapheneOS fixes?</p>
]]></description><pubDate>Wed, 04 Mar 2026 12:30:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=47246541</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47246541</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47246541</guid></item><item><title><![CDATA[New comment by tarruda in "Step 3.5 Flash – Open-source foundation model, supports deep reasoning at speed"]]></title><description><![CDATA[
<p>> so llama.cpp just doesn't handle it correctly.<p>It is a bug in the model weights and reproducible in their official chat UI. More details here: <a href="https://github.com/ggml-org/llama.cpp/pull/19283#issuecomment-3870270263" rel="nofollow">https://github.com/ggml-org/llama.cpp/pull/19283#issuecommen...</a></p>
]]></description><pubDate>Thu, 19 Feb 2026 19:26:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=47077967</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47077967</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47077967</guid></item><item><title><![CDATA[New comment by tarruda in "Step 3.5 Flash – Open-source foundation model, supports deep reasoning at speed"]]></title><description><![CDATA[
<p>I did play with Qwen3 Coder Next a bit, but didn't try it in a coding harness. Will give it a shot later.</p>
]]></description><pubDate>Thu, 19 Feb 2026 19:01:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47077651</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47077651</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47077651</guid></item><item><title><![CDATA[New comment by tarruda in "Step 3.5 Flash – Open-source foundation model, supports deep reasoning at speed"]]></title><description><![CDATA[
<p>No, it is not cheaper. An M3 ultra with 512GB costs $10k which would give you 50 months of Claude or Codex pro plans.<p>However, if you check the prices on Chinese models (which are the only ones you would be able to run on a Mac), they are much cheaper than the US plans. It would take you forever to get to the $10k<p>And of course this is not even considering energy costs or running inference on your own hardware (though Macs should be quite efficient there).</p>
]]></description><pubDate>Thu, 19 Feb 2026 19:00:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47077637</link><dc:creator>tarruda</dc:creator><comments>https://news.ycombinator.com/item?id=47077637</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47077637</guid></item></channel></rss>