<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: krackers</title><link>https://news.ycombinator.com/user?id=krackers</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 19 Apr 2026 12:37:52 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=krackers" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by krackers in "Anonymous request-token comparisons from Opus 4.6 and Opus 4.7"]]></title><description><![CDATA[
<p>If I have a conversation with claude then come back 30 minutes later to resume the conversation, the KV values for that prefill prefix are going to be exactly the same. That's the whole point of this caching in the first place.<p>If you're willing to incur a latency penalty on a "cold resume" (which is fine for most use-cases), why couldn't they just move it to disk. The size of the KV cache should scale on the order of something like (context_length * n_layers * residual_length). I think for a standard V3-MoE model at 1M token length, this should be on the order of 100G at FP16? And you can surely play tricks with KV compression (e.g. the recent TurboQuant paper). It doesn't seem like an outrageous amount of data to put onto cheap scratch HDD (and it doesn't grow indefinitely since really old conversations can be discarded).</p>
]]></description><pubDate>Sun, 19 Apr 2026 03:02:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=47821479</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47821479</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47821479</guid></item><item><title><![CDATA[New comment by krackers in "Anonymous request-token comparisons from Opus 4.6 and Opus 4.7"]]></title><description><![CDATA[
<p>>pay for reinitializing the cache<p>Why can't they save the kv cache to disk then later reload it to memory?</p>
]]></description><pubDate>Sat, 18 Apr 2026 22:03:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=47819914</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47819914</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47819914</guid></item><item><title><![CDATA[New comment by krackers in "America Lost the Mandate of Heaven"]]></title><description><![CDATA[
<p>>manufacturing alone isn’t going to grow their economy any further.<p>But why does the economy need to grow? If you can manufacture everything you need, and you have access to the raw resources, what else do you need as a country. In what sense is growing your economy with VC scams like Juicero better than actually having industrial output?</p>
]]></description><pubDate>Sat, 18 Apr 2026 20:53:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=47819449</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47819449</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47819449</guid></item><item><title><![CDATA[New comment by krackers in "Average is all you need"]]></title><description><![CDATA[
<p>The harder part is understanding the nature of the data you're working with. There's always some catch ("oh that field `foo` was never backfilled, so for queries before 2020 you have to recompute it by joining with legacyBar instead")</p>
]]></description><pubDate>Sat, 18 Apr 2026 00:42:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47812157</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47812157</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47812157</guid></item><item><title><![CDATA[New comment by krackers in "Slop Cop"]]></title><description><![CDATA[
<p>Isn't this called the tricolon? Ironically the names of the patterns all seem AI generated.</p>
]]></description><pubDate>Sat, 18 Apr 2026 00:38:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=47812133</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47812133</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47812133</guid></item><item><title><![CDATA[New comment by krackers in "Google Gemma 4 Runs Natively on iPhone with Full Offline AI Inference"]]></title><description><![CDATA[
<p>More info on specific design choices needed to run models here [1]. I mean it is possible given that apple themselves did it in [2], but it's also not as general purpose or flexible as a GPU.<p>[1] <a href="https://news.ycombinator.com/item?id=43881692">https://news.ycombinator.com/item?id=43881692</a>
[2] <a href="https://machinelearning.apple.com/research/neural-engine-transformers" rel="nofollow">https://machinelearning.apple.com/research/neural-engine-tra...</a></p>
]]></description><pubDate>Fri, 17 Apr 2026 23:44:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=47811820</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47811820</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47811820</guid></item><item><title><![CDATA[New comment by krackers in "Codex for almost everything"]]></title><description><![CDATA[
<p>Neat, good to know! And it does seem my mental model of event loop was broken. Accessibility related interactions don't have any related NSEvent.<p>They are handled as part of the "conceptual" run loop, but they seem to be dispatched internally by AXRuntime library from a callback off some mach port. And because of this, the call to nextEventMatchingEventMask in the main -[NSApplication run] loop never even sees any such NSEvent.<p><pre><code>    -[NSApplication(NSEvent) _nextEventMatchingEventMask:untilDate:inMode:dequeue:]  (in AppKit)
        _DPSNextEvent  (in AppKit)
          _BlockUntilNextEventMatchingListInModeWithFilter  (in HIToolbox)
            ReceiveNextEventCommon  (in HIToolbox)
              RunCurrentEventLoopInMode  (in HIToolbox)
                CFRunLoopRunSpecific  (in CoreFoundation)
                  __CFRunLoopRun  (in CoreFoundation)
                    __CFRunLoopDoSource1  (in CoreFoundation)
                      __CFRUNLOOP_IS_CALLING_OUT_TO_A_SOURCE1_PERFORM_FUNCTION__  (in CoreFoundation)
                        mshMIGPerform  (in HIServices)
                          _XPerformAction  (in HIServices)
                            _AXXMIGPerformAction  (in HIServices)

</code></pre>
In some sense this is sort of similar to apple events, which are also "hidden" from the caller of nextEventMatchingEventMask. From what I can see those are handled by DPSNextEvent, which sorts based on the raw carbon EventRef. aevt types have `AEProcessAppleEvent` called on them, then the event is just consumed silently. Others get converted to a CGEvent and returned back to caller for it to handle. But of course accessibility events didn't exist in Classic mac, so they can't be handled at this layer so they were pushed further down. You can almost see the historical legacy here..<p>[1] <a href="https://www.cocoawithlove.com/2009/01/demystifying-nsapplication-by.html" rel="nofollow">https://www.cocoawithlove.com/2009/01/demystifying-nsapplica...</a></p>
]]></description><pubDate>Fri, 17 Apr 2026 23:02:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47811515</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47811515</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47811515</guid></item><item><title><![CDATA[New comment by krackers in "Codex for almost everything"]]></title><description><![CDATA[
<p>Could you elaborate on what you mean? My understanding of the Cocoa event loop was that ultimately everything is received as an NSEvent at the application layer (maybe that's wrong though).<p>Do you mean that you can just AXUIElementPerformAction once you have a reference to it and the OS will internally synthesize the right type of event, even if it's not in the foreground?</p>
]]></description><pubDate>Fri, 17 Apr 2026 18:19:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=47808941</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47808941</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47808941</guid></item><item><title><![CDATA[New comment by krackers in "Codex for almost everything"]]></title><description><![CDATA[
<p>Which specific ones though allow you to send input to a window without raising it? People have been trying to do "focus follows mouse [without auto raise]" for a long time on mac, and the synthetic event equivalent to command+click is the only discovered method I'm aware of, e.g. used in <a href="https://github.com/sbmpost/AutoRaise" rel="nofollow">https://github.com/sbmpost/AutoRaise</a><p>There is also this old blog post by Yegge [1] which mentions `AXUIElementPostKeyboardEvent` but there were plenty of bugs with that, and I haven't seen anyone else build on it. I guess the modern equivalent is `CGEventPostToPSN`/`CGEventPostToPid`. I guess it's a good candidate though, perhaps the Sky team they acquired knows the right private APIs to use to get this working.<p>Edit: The thread at [2] also has some interesting tidbits, such as Automator.app having "Watch Me Do" which can also do this, and a CLI tool that claims to use the CGEventPostToPid API [3]. Maybe there's more ways to do it than I realized.<p>[1] <a href="https://steve-yegge.blogspot.com/2008/04/settling-osx-focus-follows-mouse-debate.html" rel="nofollow">https://steve-yegge.blogspot.com/2008/04/settling-osx-focus-...</a>
[2] <a href="https://www.macscripter.net/t/keystroke-to-background-app-as-vs-automator/77570" rel="nofollow">https://www.macscripter.net/t/keystroke-to-background-app-as...</a>
[3] <a href="https://github.com/socsieng/sendkeys" rel="nofollow">https://github.com/socsieng/sendkeys</a></p>
]]></description><pubDate>Thu, 16 Apr 2026 20:56:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=47799403</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47799403</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47799403</guid></item><item><title><![CDATA[New comment by krackers in "Codex for almost everything"]]></title><description><![CDATA[
<p>>background computer use<p>How does that even work technically? macOS doesn't support multiple cursors. On native Cocoa apps you can pass input to a window without raising via command+click so possibly they synthesized those events, but fewer and fewer apps support that these days. And AppleScript is basically dead, so they can't be using that either.<p>I also read they acquired the Sky team (who I think were former Apple employees). No wonder they were able to pull of something so slick.</p>
]]></description><pubDate>Thu, 16 Apr 2026 20:32:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47799128</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47799128</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47799128</guid></item><item><title><![CDATA[New comment by krackers in "Tax Wrapped 2025"]]></title><description><![CDATA[
<p>>See what the federal government spent with your tax dollars.<p>Is thinking of it in this sense actually accurate? I always assumed since every government has embraced MMT they can spend whatever they want simply by printing it out of thin air. Then taxation could be understood as the only crude knob to "destroy money", and also has the effect of forcing USD to be the primary national currency (e.g. owning bitcoin won't do you any good if you ultimately need to pay taxes in USD).</p>
]]></description><pubDate>Tue, 14 Apr 2026 01:39:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47760240</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47760240</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47760240</guid></item><item><title><![CDATA[New comment by krackers in "The economics of software teams: Why most engineering orgs are flying blind"]]></title><description><![CDATA[
<p>If good writing was easy then "LLM slop writing" wouldn't be a thing.</p>
]]></description><pubDate>Mon, 13 Apr 2026 23:39:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=47759378</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47759378</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47759378</guid></item><item><title><![CDATA[New comment by krackers in "Google has the same AI adoption curve as John Deere"]]></title><description><![CDATA[
<p>> just cancelled IntelliJ for a thousand engineers<p>IntelliJ can't cost more than the AI provider subscriptions, and it will actually handle large refactors without breaking your codebase.</p>
]]></description><pubDate>Mon, 13 Apr 2026 20:24:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=47757366</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47757366</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47757366</guid></item><item><title><![CDATA[New comment by krackers in "The hottest college major [Computer Science] hit a wall. What happened?"]]></title><description><![CDATA[
<p>>just people seem more willing to take the trap door ideas<p>It's mainly due to pressure from above. People who want to do a good job and are allowed the time to will be fastidious with or without AI. But now AI provides a shortcut and band-aid where things can be papered over or products can be launched quickly. Ship fast and then iterate" doesn't work when you're building on shaky foundations, but good luck convincing people of that.</p>
]]></description><pubDate>Mon, 13 Apr 2026 18:46:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=47756282</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47756282</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47756282</guid></item><item><title><![CDATA[New comment by krackers in "Six (and a half) intuitions for KL divergence"]]></title><description><![CDATA[
<p>I think you probably meant this, but when used with RL it's usually KL(π || π_ref), which has high loss when the in-training policy π produces output that's unlikely in the reference. But yeah as you noted, I guess this also means that there is no penalty if π _does not_ produce output in π_ref, which leads to a form of mode-collapse.<p>This collapse in variety matches with what I've seen some studies show that "sloppification" is not present in the base model, and is only introduced during the RL phase.</p>
]]></description><pubDate>Sun, 12 Apr 2026 05:24:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47736378</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47736378</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47736378</guid></item><item><title><![CDATA[New comment by krackers in "I still prefer MCP over skills"]]></title><description><![CDATA[
<p>>MCP exposes capabilities and Skills may shape how capabilities are used.<p>This is my understanding as well. What most people seem to ultimately be debating is "dedicated tool calls" (which is what MCP boils down to) versus a stateful environment that admits a single uber-tool (bash) that can compose things via scripting.<p>I guess this is what riles people up, like emacs vs vim. Some people see perfectly good CLI tools lying around and don't see why they need to basically reimplement a client against API. Others closer to the API provider side imagine it cleaner to expose a tailored slim-down surface. Devs that just use claude code on a laptop think anything other than CLI orchestration is overcomplicating it, while others on the enterprise side need a more fine-grain permission model and don't want to spin up an entire sandbox env just to run bash.<p>It's also not either or. You can can "compose" regular tool calls as well, even without something as heavy weight as an entire linux env. For instance you could have all tools exposed as FFI in QuickJS or something. The agent can invoke and compose tools by writing and executing JS programs. How well this works depends on the post-training of the model though, if agents are RL'd to emit individual tool calls via<p><pre><code>    <tool>{"myTool": {"arg1": 1}}</tool>
    <tool>{"myTool": {"arg1": 2}}</tool>
</code></pre>
tokens, then they're probably not going to be as successful shoving entire JS scripts in there like<p><pre><code>   <tool>
      const resp1 = myTool(1);
      const resp2 = myTool(2);
      console.log(resp1, resp2);
   </tool></code></pre></p>
]]></description><pubDate>Sun, 12 Apr 2026 02:09:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47735584</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47735584</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47735584</guid></item><item><title><![CDATA[New comment by krackers in "Six (and a half) intuitions for KL divergence"]]></title><description><![CDATA[
<p>See this video, beautiful explanation that doesn't already assume familiarity with entropy <a href="https://www.youtube.com/watch?v=ErfnhcEV1O8" rel="nofollow">https://www.youtube.com/watch?v=ErfnhcEV1O8</a></p>
]]></description><pubDate>Sun, 12 Apr 2026 01:08:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=47735350</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47735350</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47735350</guid></item><item><title><![CDATA[Apple Silicon and Virtual Machines: Beating the 2 VM Limit (2023)]]></title><description><![CDATA[
<p>Article URL: <a href="https://khronokernel.com/macos/2023/08/08/AS-VM.html">https://khronokernel.com/macos/2023/08/08/AS-VM.html</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47733971">https://news.ycombinator.com/item?id=47733971</a></p>
<p>Points: 236</p>
<p># Comments: 177</p>
]]></description><pubDate>Sat, 11 Apr 2026 20:58:48 +0000</pubDate><link>https://khronokernel.com/macos/2023/08/08/AS-VM.html</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47733971</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47733971</guid></item><item><title><![CDATA[New comment by krackers in "Filing the corners off my MacBooks"]]></title><description><![CDATA[
<p>There's a more thorough version of this at <a href="https://www.youtube.com/watch?v=RSaJAAqSAMw" rel="nofollow">https://www.youtube.com/watch?v=RSaJAAqSAMw</a> and the end-result doesn't look as tacky</p>
]]></description><pubDate>Fri, 10 Apr 2026 22:59:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=47724846</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47724846</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47724846</guid></item><item><title><![CDATA[New comment by krackers in "ChatGPT Pro now starts at $100/month"]]></title><description><![CDATA[
<p>Is this related to the paper on Recursive Language Models? I remember it mentioned something similar about "symbolic recursion", but the way you describe it makes it sound too simple, why is there an entire paper about it?</p>
]]></description><pubDate>Fri, 10 Apr 2026 07:27:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47714761</link><dc:creator>krackers</dc:creator><comments>https://news.ycombinator.com/item?id=47714761</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47714761</guid></item></channel></rss>