<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: htsh</title><link>https://news.ycombinator.com/user?id=htsh</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 30 Apr 2026 04:23:32 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=htsh" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by htsh in "Phi-4-reasoning-vision and the lessons of training a multimodal reasoning model"]]></title><description><![CDATA[
<p>yes! especially b/c i want to process a lot of email and directories full of old, personal documents</p>
]]></description><pubDate>Sun, 08 Mar 2026 15:42:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=47298223</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=47298223</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47298223</guid></item><item><title><![CDATA[New comment by htsh in "Apple's 512GB Mac Studio vanishes, a quiet acknowledgment of the RAM shortage"]]></title><description><![CDATA[
<p>are we sure the RAM market will stop being insane in a year or two or could this be the new norm?</p>
]]></description><pubDate>Sun, 08 Mar 2026 12:34:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47296809</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=47296809</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47296809</guid></item><item><title><![CDATA[New comment by htsh in "Claude Code: connect to a local model when your quota runs out"]]></title><description><![CDATA[
<p>thanks! came in here to ask this.<p>we can do much better with a cheap model on openrouter (glm 4.7, kimi, etc.) than anything that I can run on my lowly 3090 :)</p>
]]></description><pubDate>Wed, 04 Feb 2026 23:15:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=46893290</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=46893290</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46893290</guid></item><item><title><![CDATA[New comment by htsh in "The unreasonable effectiveness of an LLM agent loop with tool use"]]></title><description><![CDATA[
<p>I have been doing this with claude code and openai codex and/or cline. One of the three takes the first pass (usually claude code, sometimes codex), then I will have cline / gemini 2.5 do a "code review" and offer suggestions for fixes before it applies them.</p>
]]></description><pubDate>Fri, 16 May 2025 15:29:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=44006655</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=44006655</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44006655</guid></item><item><title><![CDATA[New comment by htsh in "Qwen3: Think deeper, act faster"]]></title><description><![CDATA[
<p>curious, why the 30b MoE over the 32b dense for local coding?<p>I do not know much about the benchmarks but the two coding ones look similar.</p>
]]></description><pubDate>Mon, 28 Apr 2025 22:03:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=43826622</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=43826622</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43826622</guid></item><item><title><![CDATA[New comment by htsh in "OpenVINO AI effects [denoising and transcription] for Audacity"]]></title><description><![CDATA[
<p>A lot of us have ryzen / nvidia combos... hopefully, soon, though.</p>
]]></description><pubDate>Sun, 16 Feb 2025 16:20:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=43069186</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=43069186</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43069186</guid></item><item><title><![CDATA[New comment by htsh in "Official DeepSeek R1 Now on Ollama"]]></title><description><![CDATA[
<p>assuming you want to run entirely in GPU, with 12gb vram, your sweet spot is likely the distill 14b qwen at a 4bit quant. so just run:<p>ollama run deepseek-r1:14b<p>generally, if the model file size < your vram, it is gonna run well. this file is 9gb.<p>if you don't mind slower generation, you can run models that fit within your vram + ram, and ollama will handle that offloading of layers for you.<p>so the 32b should run on your system, but it is gonna be much slower as it will be using GPU + CPU.<p>prob of interest: 
<a href="https://simonwillison.net/2025/Jan/20/deepseek-r1/" rel="nofollow">https://simonwillison.net/2025/Jan/20/deepseek-r1/</a><p>-h</p>
]]></description><pubDate>Tue, 21 Jan 2025 11:52:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=42779133</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=42779133</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42779133</guid></item><item><title><![CDATA[New comment by htsh in "I turned my open-source project into a full-time business"]]></title><description><![CDATA[
<p>As a longtime user of nodemailer, thank you.<p>I am gonna check out emailengine for future work.</p>
]]></description><pubDate>Tue, 27 Feb 2024 13:19:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=39523721</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=39523721</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39523721</guid></item><item><title><![CDATA[New comment by htsh in "MobileDiffusion: Rapid text-to-image generation on-device"]]></title><description><![CDATA[
<p>Dreambooth was kinda great?<p>That said, I agree that I wish there were more done post-research towards products with some of this stuff.</p>
]]></description><pubDate>Thu, 01 Feb 2024 12:46:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=39215372</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=39215372</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39215372</guid></item><item><title><![CDATA[New comment by htsh in "Brave Leo now uses Mixtral 8x7B as default"]]></title><description><![CDATA[
<p>Yes, offloading some layers to the GPU and VRAM should still help. And 11gb isn't bad.<p>If you're on linux or wsl2, I would run oobabooga with --verbose. Load a GGUF, start with a small number of GPU layers and creep up, keeping an eye on VRAM usage.<p>If you're on windows, you can try out LM Studio and fiddle with layers while you monitor VRAM usage, though windows may be doing some weird stuff sharing ram.<p>Would be curious to see the diffs. Specifically if there's a complexity tax in offloading that makes the CPU-alone faster but in my experience with a 3060 and a mobile 3080, offloading what I can makes a big diff.</p>
]]></description><pubDate>Sat, 27 Jan 2024 13:35:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=39155415</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=39155415</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39155415</guid></item><item><title><![CDATA[New comment by htsh in "Brave Leo now uses Mixtral 8x7B as default"]]></title><description><![CDATA[
<p>openrouter, fireworks, together.<p>we use openrouter but have had some inconsistency with speed. i hear fireworks is faster, swapping it out soon.</p>
]]></description><pubDate>Sat, 27 Jan 2024 11:27:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=39154666</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=39154666</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39154666</guid></item><item><title><![CDATA[New comment by htsh in "Show HN: Voxos.ai – An Open-Source Desktop Voice Assistant"]]></title><description><![CDATA[
<p>Can one enter their own opeanai URL and api-key? (so we can use openai-compatible things like openrouter or lm-studio)?</p>
]]></description><pubDate>Fri, 19 Jan 2024 17:18:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=39058049</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=39058049</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39058049</guid></item><item><title><![CDATA[New comment by htsh in "Show HN: Open source alternative to ChatGPT and ChatPDF-like AI tools"]]></title><description><![CDATA[
<p>That is what the RAG system does. The PDF is chunked and thrown into a vector store. And then when prompted, only the relevant bits are retrieved and stuffed into the context and sent to the LLM.<p>So yeah it's kinda smoke and mirrors. In some cases, for some long PDFs, it works really well. If it's a 500 page PDF with many disparate topics, it may do fine.</p>
]]></description><pubDate>Sun, 10 Dec 2023 13:14:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=38591320</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=38591320</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38591320</guid></item><item><title><![CDATA[New comment by htsh in "Show HN: CopilotKit- Build in-app AI chatbots and AI-powered textareas"]]></title><description><![CDATA[
<p>Cool! Any plans for Svelte?</p>
]]></description><pubDate>Wed, 06 Dec 2023 16:43:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=38546233</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=38546233</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38546233</guid></item><item><title><![CDATA[New comment by htsh in "Ask HN: SaaS pricing pages with high prices and not “contact sales”"]]></title><description><![CDATA[
<p>supabase just added their $599 tier for their soc2/hippa compliant product. really appreciated that.</p>
]]></description><pubDate>Sat, 30 Sep 2023 18:52:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=37718541</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=37718541</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37718541</guid></item><item><title><![CDATA[New comment by htsh in "Exllamav2: Inference library for running LLMs locally on consumer-class GPUs"]]></title><description><![CDATA[
<p>This subreddit remained open. Unfortunately, however, the oobabooga one went closed for a while and lost a lot of momentum. It is also back, however.<p>Are there good lemmy spaces for LLMs?</p>
]]></description><pubDate>Wed, 13 Sep 2023 17:19:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=37499126</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=37499126</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37499126</guid></item><item><title><![CDATA[New comment by htsh in "Chief executives cannot shut up about AI"]]></title><description><![CDATA[
<p>I know how that works. And my point was not that they should or will be replaced, but rather that they are no less expendable than developers (not very much).<p>But the decisions they make are one of the things that can be automated. I do not know if you have been inside one of these places but the executives are not doing a great job deciding (at mine they decided opensearch was a better bet than elastic and switched existing installations).<p>A new regime came in and then bad decision after bad decision drove our best talent away. Consultants, everywhere.<p>Also, that number is much lower. Full time devs are down, contractors and consultants are up. As a full time dev at one of these places, it felt like the number of executives was growing as everything else shrank.<p>Perhaps you are right about the highest levels, but think about all of the middlemen executives and what they do.<p>And even that -- I think an AI could choose to not spend millions on Deloitte or Accenture on software that inevitably failed.</p>
]]></description><pubDate>Fri, 02 Jun 2023 10:00:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=36163200</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=36163200</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36163200</guid></item><item><title><![CDATA[New comment by htsh in "Chief executives cannot shut up about AI"]]></title><description><![CDATA[
<p>Having just left a large enterprise, it certainly feels like executive jobs are replaceable soonest with the AI tech available to us now.<p>Not sure why those of us that live in code editors or even Excel should worry about our jobs more than those that live in Powerpoint.</p>
]]></description><pubDate>Fri, 02 Jun 2023 01:47:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=36159876</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=36159876</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36159876</guid></item><item><title><![CDATA[New comment by htsh in "Ask HN: Where have you found community outside of work?"]]></title><description><![CDATA[
<p>I moved back to NYC after a long time away, and to a different part of town than where many of my old friends live, and getting a dog considerably improved my connection with the folks around me.<p>And of course, as others have said, volunteering.</p>
]]></description><pubDate>Tue, 30 May 2023 19:02:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=36129114</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=36129114</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36129114</guid></item><item><title><![CDATA[New comment by htsh in "Making friends as an adult is hard (2021)"]]></title><description><![CDATA[
<p>1. go to a place where folks you want to hang out with live<p>2. get a dog</p>
]]></description><pubDate>Tue, 18 Apr 2023 11:11:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=35612421</link><dc:creator>htsh</dc:creator><comments>https://news.ycombinator.com/item?id=35612421</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=35612421</guid></item></channel></rss>