<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: versteegen</title><link>https://news.ycombinator.com/user?id=versteegen</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 15 Apr 2026 02:38:23 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=versteegen" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by versteegen in "Day 1 of ARC-AGI-3"]]></title><description><![CDATA[
<p>Where do you see that? I only skimmed the prompts but don't see any aspects of any of the games explained in there. There are a few hints which are legitimate prior knowledge about games in general, though some looks too inflexible to me. Prior knowledge ("Core priors") is a critical requirement of the ARC series, read the reports.</p>
]]></description><pubDate>Fri, 27 Mar 2026 12:26:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47541876</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47541876</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47541876</guid></item><item><title><![CDATA[New comment by versteegen in "Day 1 of ARC-AGI-3"]]></title><description><![CDATA[
<p>The dataset miscomparison is a big problem. The prompt is super specific to ARC-AGI-3, which is perfectly fine to do, but skimming it I saw nothing that appears specific to the 25 games in the dataset. Especially considering they've only had one day for overfitting. Could be quite subtle leakage though.</p>
]]></description><pubDate>Fri, 27 Mar 2026 12:24:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=47541852</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47541852</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47541852</guid></item><item><title><![CDATA[New comment by versteegen in "Day 1 of ARC-AGI-3"]]></title><description><![CDATA[
<p>...Their agent is called "Agentica ARC-AGI-3 agent for Opus 4.6 (120k) High".<p>Yes, it's unfair to compare results for the 25 (easier) public games against scores for the 55 semi-private games (scores for which are taken from <a href="https://arcprize.org/leaderboard">https://arcprize.org/leaderboard</a>).<p>But you're wrong to say that a custom harness invalidates the result. Yes, the official "ARC verified" scoreboard <i>for frontier LLMs</i> requires (<a href="https://arcprize.org/policy">https://arcprize.org/policy</a>):<p>>  using extremely generic and miminal LLM testing prompts, no client-side "harnesses", no hand-crafted tools, and no tailored model configuration<p>but these are limitations placed in order to compare LLMs from frontier labs on equal footing, <i>not</i> limitations that apply to submissions in general. It's not as if a solution to ARC-AGI-3 must involve training a custom LLM! This Agentica harness is completely legitimate approach to ARC-AGI-3, similar to J. Berman's for ARC-AGI-1/2, for example.</p>
]]></description><pubDate>Fri, 27 Mar 2026 12:03:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=47541681</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47541681</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47541681</guid></item><item><title><![CDATA[New comment by versteegen in "ARC-AGI-3"]]></title><description><![CDATA[
<p>> An AI that can only perform at the average human level is useless unless it can be trained for the job like humans can.<p>Yes, if you want skilled labour. But that's not at all what ARC-AGI attempts to test for: it's testing for general intelligence as possessed by anyone without a mental incapacity.</p>
]]></description><pubDate>Thu, 26 Mar 2026 00:58:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47525470</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47525470</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47525470</guid></item><item><title><![CDATA[New comment by versteegen in "No, it doesn't cost Anthropic $5k per Claude Code user"]]></title><description><![CDATA[
<p>> Aren't they losing money on the retail API pricing, too?<p>No, they aren't, and probably neither is anyone else offering API pricing. And Anthropic's API margins may be higher than anyone else.<p>For example, DeepSeek released numbers showing that R1 was served at approximately "a cost profit margin of 545%" (meaning 82% of revenue is profit), see my comment <a href="https://news.ycombinator.com/item?id=46663852">https://news.ycombinator.com/item?id=46663852</a></p>
]]></description><pubDate>Tue, 10 Mar 2026 06:07:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=47319564</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47319564</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47319564</guid></item><item><title><![CDATA[New comment by versteegen in "No, it doesn't cost Anthropic $5k per Claude Code user"]]></title><description><![CDATA[
<p>Ed Zitron made that claim (in particular here: [1]). In the same article he admits he not a programmer, and had to ask someone else to try out Claude Code and ccusage for him. He doesn't have any understanding of how LLMs or caching works. But he's prominent because he's received leaked financial details for Anthropic and OpenAI, eg [2]<p>[1] <a href="https://www.wheresyoured.at/anthropic-is-bleeding-out/" rel="nofollow">https://www.wheresyoured.at/anthropic-is-bleeding-out/</a>
[2] <a href="https://www.wheresyoured.at/costs/" rel="nofollow">https://www.wheresyoured.at/costs/</a></p>
]]></description><pubDate>Tue, 10 Mar 2026 06:04:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47319551</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47319551</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47319551</guid></item><item><title><![CDATA[New comment by versteegen in "10% of Firefox crashes are caused by bitflips"]]></title><description><![CDATA[
<p>I'm surprised "faulty PSU" is not on GP's list of common problems. Almost every unstable computer I've ever experienced has been due to either a dying PSU (not an under-specced one) or dying power conversion capacitors on the motherboard.</p>
]]></description><pubDate>Fri, 06 Mar 2026 09:56:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=47273002</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47273002</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47273002</guid></item><item><title><![CDATA[New comment by versteegen in "Claude's Cycles [pdf]"]]></title><description><![CDATA[
<p>AFAICT, Claude was not asked to prove its algorithm works for all odd n, but was instead told to move on to even n.</p>
]]></description><pubDate>Wed, 04 Mar 2026 13:02:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=47246828</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47246828</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47246828</guid></item><item><title><![CDATA[New comment by versteegen in "GPT‑5.3 Instant"]]></title><description><![CDATA[
<p>> Gemini 2.5 Pro's reasoning traces (before they nerfed them) were a good example. The deep technical analysis, and then the human-friendly version in the final output. But I found their reasoning more readable than the final output!<p>They were also sometimes more useful: you could see whether it reasoned its way to an answer, or used faulty reasoning, or if it was just contextual recall. Huge shame they replaced them with garbage (though a bit better now).<p>> the language is surprisingly offputting. I don't know if it got worse<p>I'm pretty sure it did.</p>
]]></description><pubDate>Wed, 04 Mar 2026 10:45:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=47245688</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47245688</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47245688</guid></item><item><title><![CDATA[New comment by versteegen in "Show HN: I built a site where you hire yourself instead of applying for jobs"]]></title><description><![CDATA[
<p>Yes, .wow also.</p>
]]></description><pubDate>Sat, 28 Feb 2026 00:28:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=47188213</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47188213</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47188213</guid></item><item><title><![CDATA[New comment by versteegen in "Statement from Dario Amodei on our discussions with the Department of War"]]></title><description><![CDATA[
<p>That misses my point: the evidence is the extensive argumentation provided for why it reduces risk. To quote Karnofsky:<p>> I wish people simply evaluated whether the changes seem good on the merits, without starting from a strong presumption that the mere fact of changes is either a bad thing or a fine thing. It should be hard to change good policies for bad reasons, not hard to change all policies for any reason.</p>
]]></description><pubDate>Fri, 27 Feb 2026 23:51:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=47187750</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47187750</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47187750</guid></item><item><title><![CDATA[New comment by versteegen in "Julia: Performance Tips"]]></title><description><![CDATA[
<p>Yes. And I did port my GUI layer to CimGui.jl. The rest of it is pretty intertwined with Makie, didn't do that yet. The Makie version does look better than ImGui though.</p>
]]></description><pubDate>Fri, 27 Feb 2026 23:47:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47187716</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47187716</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47187716</guid></item><item><title><![CDATA[New comment by versteegen in "Julia: Performance Tips"]]></title><description><![CDATA[
<p>I recently used Makie to create an interactive tool for inspecting nodes of a search graph (dragging, hiding, expanding edges, custom graph layout), with floating windows of data and buttons. Yes, it's great for interactive plots (you can keep using the REPL to manipulate the plot, no freezing), yes Observables and GridLayout are great, and I was very impressed with Makie's plotting abilities from making the  basics easy to the extremely advanced, but no, it was the wrong tool. Makie doesn't really do floating windows (subplots), and I had to jump through hoops to create my own float system which uses GridLayout for the GUI widgets inside them. I did get it to all work nearly flawlessly in the end, but I should probably have used a Julia imGUI wrapper instead: near instant start time!</p>
]]></description><pubDate>Fri, 27 Feb 2026 11:38:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=47179360</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47179360</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47179360</guid></item><item><title><![CDATA[New comment by versteegen in "Statement from Dario Amodei on our discussions with the Department of War"]]></title><description><![CDATA[
<p>> They pragmatically changed their views of safety just recently, so those values for which they would burn at the stake are very fluid.<p>Yes it was a pragmatic change, no it was not a change in their values. The commentary here on HN about Anthropic's RSP change was completely off the mark. They "think these changes are the right thing for reducing AI risk, both from Anthropic and from other companies if they make similar changes", as stated in this detailed discussion by Holden Karnofsky, who takes "significant responsibility for this change":<p><a href="https://www.lesswrong.com/posts/HzKuzrKfaDJvQqmjh/responsible-scaling-policy-v3" rel="nofollow">https://www.lesswrong.com/posts/HzKuzrKfaDJvQqmjh/responsibl...</a><p>> I strongly think today’s environment does not fit the “prisoner’s dilemma” model. In today’s environment, I think there are companies not terribly far behind the frontier that would see any unilateral pause or slowdown as an opportunity rather than a warning.<p>> What I didn’t expect was that RSPs (at least in Anthropic’s case) would come to be seen as hard unilateral commitments (“escape clauses” notwithstanding) that would be very difficult to iterate on.</p>
]]></description><pubDate>Fri, 27 Feb 2026 08:09:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=47177936</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47177936</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47177936</guid></item><item><title><![CDATA[New comment by versteegen in "Claude Sonnet 4.6"]]></title><description><![CDATA[
<p>Sure, depending on the particular product, having control and direct local access to the data would be desirable or deal breaker. But for this Clickup integration that's not so important to us (we can duplicate where necessary), while still using Clickup lets us use all the other features we can get via the web app.<p>(The emacs mode includes an MCP server)</p>
]]></description><pubDate>Thu, 19 Feb 2026 14:54:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=47074429</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47074429</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47074429</guid></item><item><title><![CDATA[New comment by versteegen in "Claude Sonnet 4.6"]]></title><description><![CDATA[
<p>I'd agree that this effect is probably mainly due to architectural parameters such as the number and dimensions of heads, and hidden dimension. But not so much the model size (number of parameters) or less training.<p>Saw something about Sonnet 4.6 having had a greatly increased amount of RL training over 4.5.</p>
]]></description><pubDate>Wed, 18 Feb 2026 15:30:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=47062087</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47062087</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47062087</guid></item><item><title><![CDATA[New comment by versteegen in "Claude Sonnet 4.6"]]></title><description><![CDATA[
<p>Agreed, and here's a real example from a tiny startup: Clickup's web app is too damn slow and bloated with features and UI, so we created emacs modes to access and edit Clickup workspaces (lists, kanban boards, docs, etc) via the API. Just some limited parts we care about. I was initially skeptical that it would work well or at all, but wow, it really has significantly improved the usefulness of Clickup by removing barriers.</p>
]]></description><pubDate>Wed, 18 Feb 2026 15:13:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=47061839</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=47061839</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47061839</guid></item><item><title><![CDATA[New comment by versteegen in "GPT-5.3-Codex"]]></title><description><![CDATA[
<p>> The only valid ARC AGI results are from tests done by the ARC AGI non-profit using an unreleased private set. I believe lab-conducted ARC AGI tests must be on public sets and taken on a 'scout's honor' basis that the lab self-administered the test correctly<p>Not very accurate. For each of ARC-AGI-1 and ARC-AGI-2 there is training set and three eval sets: public, semi-private, and private. The ARC foundation runs frontier LLMs on the semi-private set, and the labs give them pre-release API access so they can report release-day evals. They mostly don't allow anyone else to access the semi-private set (except for live Kaggle leaderboards which use it), so you see independent researchers report on the public eval set instead, often very dubious. The private is for Kaggle competitions only, no frontier LLMs evals are possible.<p>(ARC-AGI-1 results are now largely useless because most of its eval tasks became the ARC-2 training set. However some labs have said they don't train LLMs on the training sets anyway.)</p>
]]></description><pubDate>Sun, 08 Feb 2026 10:30:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=46933061</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=46933061</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46933061</guid></item><item><title><![CDATA[New comment by versteegen in "OpenAI could reportedly run out of cash by mid-2027"]]></title><description><![CDATA[
<p>Paid/API LLM inference is profitable, though. For example, DeepSeek R1 had "a cost profit margin of 545%" [1] (ignoring free users and using a placeholder $2/hour figure H800 GPU, which seems ballpark of real to me due to Chinese electricity subsidies). Dario has said each Anthropic model is profitable over its lifetime. (And looking at ccusage stats and thinking Anthropic is losing thousands per Claude Code user is nonsense, API prices aren't their real costs. That's why opencode gives free access to GLM 4.7 and other models: it was far cheaper than they expected due to the excellent cache hit rates.) If anyone ran out of money they would stop spending on experiments/research and training runs and be profitable... until their models were obsolete. But it's impossible for everyone to go bankrupt.<p>[1] <a href="https://github.com/deepseek-ai/open-infra-index/blob/main/202502OpenSourceWeek/day_6_one_more_thing_deepseekV3R1_inference_system_overview.md" rel="nofollow">https://github.com/deepseek-ai/open-infra-index/blob/main/20...</a></p>
]]></description><pubDate>Sun, 18 Jan 2026 01:13:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=46663852</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=46663852</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46663852</guid></item><item><title><![CDATA[New comment by versteegen in "Ask HN: How can we solve the loneliness epidemic?"]]></title><description><![CDATA[
<p>It's excellent that you're working on loneliness! Somehow. What is it your startup actually does?</p>
]]></description><pubDate>Fri, 16 Jan 2026 11:00:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=46645184</link><dc:creator>versteegen</dc:creator><comments>https://news.ycombinator.com/item?id=46645184</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46645184</guid></item></channel></rss>