<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: throw5</title><link>https://news.ycombinator.com/user?id=throw5</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 06 Apr 2026 02:05:10 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=throw5" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by throw5 in "Eight years of wanting, three months of building with AI"]]></title><description><![CDATA[
<p>Yes, how dare someone take an idea, develop it, and publish it outside the algorithm-driven rage pit. Truly terrible behavior! /s<p>Expanding a thought beyond 280 characters and publishing it somewhere other than the X outrage machine is something we should be encouraging.</p>
]]></description><pubDate>Sun, 05 Apr 2026 16:16:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=47650903</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47650903</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47650903</guid></item><item><title><![CDATA[New comment by throw5 in "Eight years of wanting, three months of building with AI"]]></title><description><![CDATA[
<p>> This article is describing a problem that is still two steps removed from where AI code becomes actually useful.<p>But it does a good job of countering the narrative you often see on LinkedIn, and to some extent on HN as well, where AI is portrayed as all-capable of developing enterprise software. If you spend any time in discussions hyping AI, you will have seen plenty of confident claims that traditional coding is dead and that AI will replace it soon. Posts like this is useful because it shows a more grounded reality.<p>> 90 percent of the things users want either A) dont exist or B) are impossible to find, install and run without being deeply technical. These things dont need to scale, they dont need to be well designed. They are for the most part targeted, single user, single purpose, artifacts.<p>Yes, that is a particular niche where AI can be applied effectively. But many AI proponents go much further and argue that AI is already capable of delivering complex, production-grade systems. They say, you don't need engineers anymore. They say, you only need product owners who can write down the spec. From what I have seen, that claim does not hold up and this article supports that view.<p>Many users may not be interested in scalability and maintainability... But for a number of us, including the OP and myself, the real question is whether AI can handle situations where scalability, maintainability and sound design DO actually matter. The OP does a good job of understanding this.</p>
]]></description><pubDate>Sun, 05 Apr 2026 16:06:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47650801</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47650801</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47650801</guid></item><item><title><![CDATA[New comment by throw5 in "Claude Code runs Git reset –hard origin/main against project repo every 10 mins"]]></title><description><![CDATA[
<p>Does any one of this help me if Claude runs `git reset --hard`?<p>If I am working in a sandbox, I have uncommitted changes in a sandbox and if Claude runs `git reset --hard` on those uncommitted changes in the sandbox, I've got the same problem?<p>> Care about the data in that workspace? Push it first.<p>But you're changing the problem. If I push everything, then yeah I've got no problem. But between pushing one change and the next, you're gonna have uncommitted changes, won't you? and if Claude runs `git reset --hard` at that time, same problem, isn't it?</p>
]]></description><pubDate>Mon, 30 Mar 2026 03:22:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=47570074</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47570074</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47570074</guid></item><item><title><![CDATA[New comment by throw5 in "Claude Code runs Git reset –hard origin/main against project repo every 10 mins"]]></title><description><![CDATA[
<p>This! The safeguards need to be outside LLM and they need to be deterministic.<p>Now I wish I could reject `git reset --hard` on my local system somehow.</p>
]]></description><pubDate>Mon, 30 Mar 2026 01:08:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=47569289</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47569289</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47569289</guid></item><item><title><![CDATA[New comment by throw5 in "Claude Code runs Git reset –hard origin/main against project repo every 10 mins"]]></title><description><![CDATA[
<p>Yes, if something is reproducible and undesirable, it is a bug and RLHF can reduce it. I'm not disupting that. "reduce" is the keyword here. You can't eliminate them entirely.<p>My point is that fixing one bug does not eliminate the <i>class</i> of bugs. Heck, it does not even fix that one bug deterministically. You only reduce its probability like you rightly said.<p>With git commands, there is not like a system like Lean that can formally reject invalid proofs. Really I think the mathematicians have got it easier with LLMs because a proof is either valid or invalid. It's not so clear cut with git commands. Almost any command can be valid in some narrow context, which makes it much harder to reject undesirable outputs entirely.<p>Until the underlying probabilities of undesirable output become negligible so much that they become practically impossible, these kinds of issues will keep surfacing even if you address individual bugs. Will the probabilities become so low someday that these issues are practically impossible? Maybe. But we are not there yet. Until then, we should recalibrate our expectations and rely on deterministic safeguards outside the LLM.</p>
]]></description><pubDate>Mon, 30 Mar 2026 00:53:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47569173</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47569173</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47569173</guid></item><item><title><![CDATA[New comment by throw5 in "Claude Code runs Git reset –hard origin/main against project repo every 10 mins"]]></title><description><![CDATA[
<p>Yes, exactly. People often overlook that, even with guardrails, it is still probabilities all the way down.<p>You can reduce the risk, but not drive it to zero, and at scale even very small failure rates will surface.</p>
]]></description><pubDate>Mon, 30 Mar 2026 00:31:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=47569015</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47569015</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47569015</guid></item><item><title><![CDATA[New comment by throw5 in "Claude Code runs Git reset –hard origin/main against project repo every 10 mins"]]></title><description><![CDATA[
<p>> It is meaningless to say that because the author was able to reproduce it multiple times.<p>I don't know how that refutes what I'm saying.<p>The behaviour was reproduced multiple times, so it is clearly an observable outcome, not a one-off. It just shows that the probability of `git reset --hard` is > 0 even with RLHF and post-training.</p>
]]></description><pubDate>Mon, 30 Mar 2026 00:16:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47568918</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47568918</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47568918</guid></item><item><title><![CDATA[New comment by throw5 in "Claude Code runs Git reset –hard origin/main against project repo every 10 mins"]]></title><description><![CDATA[
<p>> Just by a thing being common in training data doesn't mean it will be produced.<p>That's not what I said at all. I never said it <i>will</i> be produced. I said there is <i>some probability</i> of it being produced.<p>> False, it goes against the RL/HF and other post training goals.<p>It is correct that frequency in training data alone does not determine outputs, and that post-training (RLHF, policies, etc.) is meant to steer the model away from undesirable behavior.<p>But those mechanisms do not make such outputs <i>impossible</i>. They just make them less likely. The underlying system is still <i>probabilistic</i> and operating with incomplete context.<p>I am not sure how you can be so confident that a probabilistic model would never produce `git reset --hard`. There is nothing inherent in how LLMs work that makes that sequence impossible to generate.</p>
]]></description><pubDate>Mon, 30 Mar 2026 00:05:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=47568833</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47568833</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47568833</guid></item><item><title><![CDATA[New comment by throw5 in "Claude Code runs Git reset –hard origin/main against project repo every 10 mins"]]></title><description><![CDATA[
<p>Isn't this a natural consequence of how these systems work?<p>The model is probabilistic and sequences like `git reset --hard` are very common in training data, so they have some probability to appear in outputs.<p>Whether such a command is appropriate depends on context that is not fully observable to the system, like whether a repository or changes are disposable or not. Because of that, the system cannot rely purely on fixed rules and has to figure intent from incomplete information, which is also probabilistic.<p>With so many layers of probabilities, it seems expected that sometimes commands like this will be produced even if they are not appropriate in that specific situation.<p>Even a 0.01% failure rate due to context corruption, misinterpretation of intent, or guardrail errors would show up regularly at scale, that is like 1 in 10000 queries.</p>
]]></description><pubDate>Mon, 30 Mar 2026 00:00:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=47568793</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47568793</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47568793</guid></item><item><title><![CDATA[New comment by throw5 in "90% of Claude-linked output going to GitHub repos w <2 stars"]]></title><description><![CDATA[
<p>> I'm with you on all points except for it being bought.<p>Stars get bought all the time. I've been around startup scene and this is basically part of the playbook now for open core model. You throw your code up on GitHub, call it open source, then buy your stars early so it looks like people care. Then charge for hosted or premium features.<p>There's a whole market for it too. You can literally pay for stars, forks, even fake activity. Big star count makes a project look legit at a glance, especially to investors or people who don't dig too deep. It feeds itself. More people check it out, more people star it just because others already did.</p>
]]></description><pubDate>Thu, 26 Mar 2026 03:21:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=47526305</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47526305</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47526305</guid></item><item><title><![CDATA[New comment by throw5 in "The “small web” is bigger than you might think"]]></title><description><![CDATA[
<p>> But what I think is an even better solution is to do it at the content level: sign the content, like a GPG signature<p>How would this work in reality? With the current state of browsers this is not possible because the ISP can still insert their content into the page and the browser will still load it with the modified content that does not match the signature. Nothing forces the GPG signature verification with current tech.<p>If you mean that browsers need to be updated to verify GPG signature, I'm not sure how realistic that is. Browsers cannot verify the GPG signature and vouch for it until you solve the problem of key revocation and key expiry. If you try to solve key revocation and key expiry, you are back to the same problems that certificates have.</p>
]]></description><pubDate>Mon, 16 Mar 2026 19:38:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47403748</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47403748</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47403748</guid></item><item><title><![CDATA[New comment by throw5 in "Hazardous substances found in all headphones tested by ToxFREE project"]]></title><description><![CDATA[
<p>How about in-ear earphones? They use silicone tips, right? Are there any known harmful effects of those?<p>The study names brands like Bose, Panasonic, Samsung, and Sennheiser. What about Apple airpods? Anyone knows what's that made of and if they've got any harmful effects?</p>
]]></description><pubDate>Sun, 15 Mar 2026 01:55:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=47383474</link><dc:creator>throw5</dc:creator><comments>https://news.ycombinator.com/item?id=47383474</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47383474</guid></item></channel></rss>