<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: deoxykev</title><link>https://news.ycombinator.com/user?id=deoxykev</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 08 Apr 2026 00:18:51 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=deoxykev" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[Solving Automata Cam Profiling with Grasshopper]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.youtube.com/watch?v=grgIhw1YbHw">https://www.youtube.com/watch?v=grgIhw1YbHw</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46969632">https://news.ycombinator.com/item?id=46969632</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 11 Feb 2026 01:31:27 +0000</pubDate><link>https://www.youtube.com/watch?v=grgIhw1YbHw</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=46969632</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46969632</guid></item><item><title><![CDATA[New comment by deoxykev in "Iowa City made its buses free. Traffic cleared, and so did the air"]]></title><description><![CDATA[
<p>I live there in that city. There are hardly any homeless at all here. Not like other cities at least. I could see it being a major problem in other places.</p>
]]></description><pubDate>Mon, 24 Nov 2025 00:25:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=46028887</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=46028887</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46028887</guid></item><item><title><![CDATA[New comment by deoxykev in "Images over DNS"]]></title><description><![CDATA[
<p>How about LLM chat over DNS? <a href="https://github.com/accupham/llm-dns-proxy" rel="nofollow">https://github.com/accupham/llm-dns-proxy</a></p>
]]></description><pubDate>Sat, 20 Sep 2025 17:02:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=45315093</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=45315093</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45315093</guid></item><item><title><![CDATA[New comment by deoxykev in "Images over DNS"]]></title><description><![CDATA[
<p>And it typically works on captive portals too before payment.</p>
]]></description><pubDate>Sat, 20 Sep 2025 16:59:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=45315070</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=45315070</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45315070</guid></item><item><title><![CDATA[New comment by deoxykev in "Perfume reviews"]]></title><description><![CDATA[
<p>Meta-commentary always leans nerdier.</p>
]]></description><pubDate>Fri, 18 Jul 2025 18:52:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=44608419</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=44608419</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44608419</guid></item><item><title><![CDATA[New comment by deoxykev in "It's time to become an ML engineer (2022)"]]></title><description><![CDATA[
<p>Curious to hear what kind of work you do. Because there are definitely fields where productivity as 10x'd because of AI tools.</p>
]]></description><pubDate>Thu, 20 Feb 2025 14:47:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=43115308</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=43115308</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43115308</guid></item><item><title><![CDATA[New comment by deoxykev in "Web awesome: "Shoelace 3.0" open source web components"]]></title><description><![CDATA[
<p>HTMX and shoelace is an awesome combo. Super fast to prototype things and tweak as needed. Being able to copy paste snippets and directly inject data in a straightforward way is a nice way of working. It limits cognitive overhead so you can focus on the domain logic rather than fight javascript dependencies.</p>
]]></description><pubDate>Tue, 18 Feb 2025 18:35:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=43093373</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=43093373</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43093373</guid></item><item><title><![CDATA[New comment by deoxykev in "DeepRAG: Thinking to retrieval step by step for large language models"]]></title><description><![CDATA[
<p>Don't forget to finetune the reranker too if you end up doing the embedding model. That tends to have outsized effects on performance for out of distribution content.</p>
]]></description><pubDate>Tue, 04 Feb 2025 18:55:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=42936906</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=42936906</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42936906</guid></item><item><title><![CDATA[New comment by deoxykev in "Show HN: Klarity – Open-source tool to analyze uncertainty/entropy in LLM output"]]></title><description><![CDATA[
<p>Interesting, I had never heard about min-p until now. From what I understand, it's like a low-pass filter for the token sampling pool which boosts semantic coherence. Like removing static from the radio.<p>Do you have any benchmarks of min-p sampling with the new reasoning models, such as QwQ and R1?</p>
]]></description><pubDate>Mon, 03 Feb 2025 20:01:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=42922200</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=42922200</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42922200</guid></item><item><title><![CDATA[New comment by deoxykev in "How to Run DeepSeek R1 671B Locally on a $2000 EPYC Server"]]></title><description><![CDATA[
<p>Yeah, there is a clear bottleneck somewhere in llama.cpp. Even high end hardware is struggling to get good numbers. The theoretical limit should be higher, but it's not yet.<p>Benchmarks:
<a href="https://github.com/ggerganov/llama.cpp/issues/11474#issuecomment-2629590835">https://github.com/ggerganov/llama.cpp/issues/11474#issuecom...</a></p>
]]></description><pubDate>Mon, 03 Feb 2025 17:45:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=42920777</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=42920777</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42920777</guid></item><item><title><![CDATA[New comment by deoxykev in "Efficient Reasoning with Hidden Thinking"]]></title><description><![CDATA[
<p>I don't think autoregressive models have a fundemental difference in terms of reasoning capability in latent space vs token space. Latent space enables abstract reasoning and pattern recognition, while token space acts as both the discrete interface for communication, and as a interaction medium to extend, refine and synthesize high order reasoning over latent space.<p>Intuively speaking, most people think of writing as a communication tool. But actually it's also a thinking tool that helps create deeper connections over discrete thoughts which can only occupy a fixed slice of our attention at any given time. Attentional capacity the primary limitation-- for humans and LLMs. So use the token space as extended working memory. Besides, even the Coconut paper got mediocre results. I don't think this is the way.</p>
]]></description><pubDate>Mon, 03 Feb 2025 17:43:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=42920748</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=42920748</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42920748</guid></item><item><title><![CDATA[New comment by deoxykev in "Show HN: Klarity – Open-source tool to analyze uncertainty/entropy in LLM output"]]></title><description><![CDATA[
<p>The fundemental challenge of using log probabilities to measure LLM certainty is the mismatch between how language models process information and how semantic meaning actually works. The current models analyze text token by token-- fragments that don't necessarily align with complete words, let alone complex concepts or ideas.<p>This creates a gap between the mechanical measurement of certainty and true understanding, much like mistaking the map for the territory or confusing the finger pointing at the moon with the moon itself.<p>I've done some work before in this space, trying to come up with different useful measures from the logprobs, such as measuring shannon entropy over a sliding window, or even bzip compression ratio as a proxy for information density. But I didn't find anything semantically useful or reliable to exploit.<p>The best approach I found was just multiple choice questions. "Does X entail Y? Please output [A] True or [B] False. Then measure the linprobs of the next token, which should be `[A` (90%) or `[B` (10%). Then we might make a statement like: The LLM thinks there is a 90% probability that X entails Y.</p>
]]></description><pubDate>Mon, 03 Feb 2025 17:34:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=42920656</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=42920656</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42920656</guid></item><item><title><![CDATA[New comment by deoxykev in "DeepSeek gives Europe's tech firms a chance to catch up"]]></title><description><![CDATA[
<p>My take: the distills under 32B aren’t worth running. Quants seem to impact quality much more than other models. 32B and 70B unquantized are very good. 671B is SOTA.</p>
]]></description><pubDate>Mon, 03 Feb 2025 11:38:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=42917208</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=42917208</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42917208</guid></item><item><title><![CDATA[New comment by deoxykev in "How to Run DeepSeek R1 671B Locally on a $2000 EPYC Server"]]></title><description><![CDATA[
<p>8x 3090 will net you around 10-12tok/s</p>
]]></description><pubDate>Sat, 01 Feb 2025 16:16:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=42899469</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=42899469</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42899469</guid></item><item><title><![CDATA[New comment by deoxykev in "Show HN: Flow – A dynamic task engine for building AI agents"]]></title><description><![CDATA[
<p>Have you hit any non-determinism errors keeping workflow state outside temporal?</p>
]]></description><pubDate>Tue, 03 Dec 2024 13:40:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=42305919</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=42305919</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42305919</guid></item><item><title><![CDATA[New comment by deoxykev in "Show HN: Flow – A dynamic task engine for building AI agents"]]></title><description><![CDATA[
<p>Hey, I’m building agents on top of temporal as well. One of the main limitations is child workflows can not spawn other child workflows. Are you doing an activity for every prompt execution and passing those through other activities? Or something more framework-y?</p>
]]></description><pubDate>Tue, 03 Dec 2024 02:10:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=42302391</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=42302391</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42302391</guid></item><item><title><![CDATA[New comment by deoxykev in "Capstone Disassembler Framework"]]></title><description><![CDATA[
<p>Imhex is a really great frontend for Capstone.
<a href="https://github.com/WerWolv/ImHex">https://github.com/WerWolv/ImHex</a></p>
]]></description><pubDate>Wed, 25 Sep 2024 17:22:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=41649687</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=41649687</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41649687</guid></item><item><title><![CDATA[New comment by deoxykev in "Serving AI from the Basement – 192GB of VRAM Setup"]]></title><description><![CDATA[
<p>Are you able to run 405B? 4Bit quant vram requirements are just shy of 192GB.</p>
]]></description><pubDate>Mon, 09 Sep 2024 14:17:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=41488792</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=41488792</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41488792</guid></item><item><title><![CDATA[New comment by deoxykev in "Mistral AI Launches New 8x22B MOE Model"]]></title><description><![CDATA[
<p>4 bit quants should require 85GB VRAM, so this will fit nicely on 4x 24G consumer GPUs, plus some leftover for KV cache optimization.</p>
]]></description><pubDate>Wed, 10 Apr 2024 03:09:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=39986596</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=39986596</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39986596</guid></item><item><title><![CDATA[New comment by deoxykev in "Show HN: Beyond text splitting – improved file parsing for LLMs"]]></title><description><![CDATA[
<p>How does this compare to LayoutLMv3? Was it trained on forms at all?</p>
]]></description><pubDate>Mon, 08 Apr 2024 14:22:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=39970025</link><dc:creator>deoxykev</dc:creator><comments>https://news.ycombinator.com/item?id=39970025</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39970025</guid></item></channel></rss>