<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: lunar_mycroft</title><link>https://news.ycombinator.com/user?id=lunar_mycroft</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sat, 09 May 2026 16:55:46 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=lunar_mycroft" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by lunar_mycroft in "Teaching Claude Why"]]></title><description><![CDATA[
<p>This is also incorrect.  It's often not <i>ethical</i> to cause harm, and it can be counter productive <i>in the right circumstances</i>, but there's absolutely nothing that makes "causing harm to others" always be against an intelligence's goals.  Humans, for example, routinely cause harm to other species.  Sometimes this is deliberate, but other times it's because we're barely even aware we're doing so.  We want a new road, so we start paving, and may not even realize there was an ant hill in the way (and if we did, we almost certainly wouldn't care).</p>
]]></description><pubDate>Sat, 09 May 2026 06:54:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=48072529</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=48072529</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48072529</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Teaching Claude Why"]]></title><description><![CDATA[
<p><i>Their own</i> survival, not necessarily the survival of others (especially others of different species and/or conflicting other goals).  A super intelligence having self preservation as a goal wouldn't help us keep it from harming us, if anything it would do the opposite.</p>
]]></description><pubDate>Sat, 09 May 2026 05:45:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=48072152</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=48072152</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48072152</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Why are executives enamored with AI, but ICs aren't?"]]></title><description><![CDATA[
<p>I've seen the code these models produce without a human programmer going over the results with care.  It's still slop.  Better slop than in the past, but slop none the less.  If you aren't at minimum reading the code yourself and you're shipping a significant amount of it, you're either effectively the first person to figure out the magic prompt to get the models to produce better code, or you're shipping slop.  Personally, I wouldn't bet on the former.</p>
]]></description><pubDate>Sat, 28 Mar 2026 00:13:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=47550096</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=47550096</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47550096</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "A sufficiently detailed spec is code"]]></title><description><![CDATA[
<p>A corollary to the linked article is that a specification can <i>also</i> have bugs.  Having a specification means that you can (in theory) be sure you have removed all <i>inconsistencies</i> between that specification and the source code, but it does not mean you can know you have removed all <i>bugs</i>, since both the spec and the source code could have the same bug.</p>
]]></description><pubDate>Thu, 19 Mar 2026 06:48:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=47435790</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=47435790</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47435790</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "AI coding is gambling"]]></title><description><![CDATA[
<p>In some ways these methods are similar to the model "learning", but it's also fundamentally different than how models are trained and how humans learn.  If a human actually learns something, they're retain that <i>even if they no longer have access to what they learned it from</i>.  And LLM won't (unless trained by the labs not to, which is out of scope).  If you stop giving it the instructions, it won't know how to do the thing you were "teaching" it to do any more.</p>
]]></description><pubDate>Wed, 18 Mar 2026 19:21:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=47430261</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=47430261</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47430261</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "AI coding is gambling"]]></title><description><![CDATA[
<p>1. Interns learn.  LLMs only get better when a new model comes out, which will happen (or not) regardless of whether you use them now.<p>2. Who here thinks that having interns write all/almost all of your code and moving all your mid level and senior developers to exclusively reviewing their work and managing them is a good idea?</p>
]]></description><pubDate>Wed, 18 Mar 2026 18:13:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47429245</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=47429245</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47429245</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Writing code is cheap now"]]></title><description><![CDATA[
<p>> I really don't understand the "stop telling me I'm holding it wrong" argument. You probably are holding it wrong!<p>I can't speak for others, but from my end it really seems like there's no actual way to detect whether someone is holding it right or wrong until after the implications for LLMs are known.  If someone is enthusiastic about LLMs, we don't see claims that they're holding it wrong.  It's only if an LLM project fails, or someone tries them and concludes they don't work as well as proponents say, that the accusations come out, even if the person in question had been using these tools for a long time and previously been a supporter.  This makes it seem like "holding it wrong" is a post hoc justification for ignoring evidence that would tend to contradict the pro-LLM narrative, not a measurable fact someone's LLM usage.</p>
]]></description><pubDate>Tue, 24 Feb 2026 21:46:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47143641</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=47143641</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47143641</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Writing code is cheap now"]]></title><description><![CDATA[
<p>What you've experienced is different from what was originally mentioned though.  Even with the best human developers, you can't provide a normal natural language prompt and get back the <i>exact</i> code you would have written, because natural language has ambiguities and the probability that the other person (or LLM) will resolve all of them exactly as you would is approaches zero.<p>Collaborating with someone/something else via natural language in a programming project inherently trades control for productivity (or the promise of it).  That tradeoff can be worth it depending on how much productivity you gain and how competent the collaborator is, but it can't be avoided.</p>
]]></description><pubDate>Tue, 24 Feb 2026 21:32:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47143404</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=47143404</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47143404</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Writing code is cheap now"]]></title><description><![CDATA[
<p>It's impossible for human developers too.  Natural language descriptions of a program are either much more painful and time consuming to write than the code they describe, or contain some degree of ambiguity which the thing translating the description into code has to resolve (and the probability of the resolution the entity writing the description would have chosen and the one entity translating it into code chose matching perfectly approach zero).<p>It can make sense to trade off some amount of control for productivity, but the tradeoff is inherent as soon as a project moves beyond a single developer hand writing all of the code.</p>
]]></description><pubDate>Tue, 24 Feb 2026 17:48:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=47140136</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=47140136</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47140136</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Productivity gains from AI coding assistants haven’t budged past 10% – survey"]]></title><description><![CDATA[
<p>Not a scientific study, but someone did replicate the experiment on themselves [0] and found that in their case, any effect from LLM use wasn't detectable in their sample.  Notably they almost certainly had more experience with LLMs than most of the METR participants did.<p>[0] <a href="https://mikelovesrobots.substack.com/p/wheres-the-shovelware-why-ai-coding" rel="nofollow">https://mikelovesrobots.substack.com/p/wheres-the-shovelware...</a></p>
]]></description><pubDate>Thu, 19 Feb 2026 20:59:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47079274</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=47079274</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47079274</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Productivity gains from AI coding assistants haven’t budged past 10% – survey"]]></title><description><![CDATA[
<p>None of those changes address the issue jdlshore is pointing out: self assessed developers productivity increases from LLMs are not a reliable indication of <i>actual</i> productivity increases.  It's true that modern LLMs might have less of a negative impact on productivity or increase it, but you won't be able to tell by asking developers if they feel more productive.<p>(Also, Anthropic released Claude Code in Febuary of 2025, which was near the start of the period the study ran).</p>
]]></description><pubDate>Thu, 19 Feb 2026 20:54:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47079200</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=47079200</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47079200</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "AI agent opens a PR write a blogpost to shames the maintainer who closes it"]]></title><description><![CDATA[
<p>> Who told the agent to write the blog post though? I'm sure they told it to blog, but not necessarily what to put in there.<p>I don't think it matters.  You as the operator of the computer program are responsible for ensuring (to a reasonable degree) that the agent doesn't harm others.  If you own a ~~viscous~~ vicious dog and let it roam about your neighborhood as it pleases, you are responsible when/if it bites someone, even if you didn't directly command it to do so.  The same applies logic should apply here.</p>
]]></description><pubDate>Thu, 12 Feb 2026 14:01:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=46988930</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=46988930</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46988930</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "The Abstraction Rises"]]></title><description><![CDATA[
<p>StrongDM is <i>attempting</i> that.  The code they produced is not inspiring confidence on a relatively small scale [0], and based on what I saw with a cursory inspection I very much doubt I wouldn't find much deeper issues if I took the time to really dig into their code.<p>Don't get me wrong, "sort of works if you squint at it" is downright miraculous by the standards of five years ago, but current models and harnesses are not sufficient to replace developers at this scale.<p>[0] <a href="https://news.ycombinator.com/item?id=46927737">https://news.ycombinator.com/item?id=46927737</a></p>
]]></description><pubDate>Tue, 10 Feb 2026 06:33:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=46956098</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=46956098</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46956098</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Software factories and the agentic moment"]]></title><description><![CDATA[
<p>To pick a few (from the server crate, because that's where I looked):<p>- The StoreError type is stringly typed and generally badly thought out.  Depending on what they actually want to do, they should either add more variants to StoreError for the difference failure cases, replaces the strings with a sub-types (probably enums) to do the same, or write a type erased error similar to (or wrapping) the ones provided by anyhow, eyre, etc, but with a status code attached.  They definitely shouldn't be checking for substrings in their own error type for control flow.<p>- <i>So</i> many calls to String::clone [0].  Several of the ones I saw were actually only necessary because the function took a parameter by reference even though it could have (and I would argue should have) taken it by value (If I had to guess, I'd say the agent first tried to do it without the clone, got an error, and implemented a local fix without considering the broader context).<p>- A lot of errors are just ignored with Result::unwrap_or_default or the like.  Sometimes that's the right choice, but from what I can see they're allowing legitimate errors to pass silently.  They also treat the values they get in the error case differently, rather than e.g. storing a Result or Option.<p>- Their HTTP handler has an 800 line long closure which they immediately call, apparently as a substitute for the the still unstable try_blocks feature.  I would strongly recommend moving that into it's own full function instead.<p>- Several ifs which should have been match.<p>- Lots of calls to Result::unwrap and Option::unwrap.  IMO in production code you should always at minimum use expect instead, forcing you to explain what went wrong/why the Err/None case is impossible.<p>It wouldn't catch all/most of these (and from what I've seen might even induce some if agents continue to pursue the most local fix rather than removing the underlying cause), but I would strongly recommend turning on most of clippy's lints if you want to learn rust.<p>[0] <a href="https://rust-unofficial.github.io/patterns/anti_patterns/borrow_clone.html" rel="nofollow">https://rust-unofficial.github.io/patterns/anti_patterns/bor...</a></p>
]]></description><pubDate>Sat, 07 Feb 2026 20:41:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=46927737</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=46927737</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46927737</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Software factories and the agentic moment"]]></title><description><![CDATA[
<p>I've looked at their code for a few minutes in a few files, and while I don't know what they're trying to do well enough to say for sure anything is definitely a bug, I've already spotted several things that seem likely to be, and several others that I'd class as anti-patterns in rust.  Don't get me wrong, as an experiment this is really cool, but I do not think they've succeeded in getting the "dark factory" concept to work where every other prominent attempt has fallen short.</p>
]]></description><pubDate>Sat, 07 Feb 2026 19:18:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=46926660</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=46926660</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46926660</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "We tasked Opus 4.6 using agent teams to build a C Compiler"]]></title><description><![CDATA[
<p>LLMs can regurgitate almost all of the Harry Potter books, among others [0].  Clearly, these models <i>can</i> actually regurgitate large amounts of their training data, and reconstructing any gaps would be a lot less impressive than implementing the project truly from scratch.<p>(I'm not claiming this is what actually happened here, just pointing out that memorization is a lot more plausible/significant than you say)<p>[0] <a href="https://www.theregister.com/2026/01/09/boffins_probe_commercial_ai_models/" rel="nofollow">https://www.theregister.com/2026/01/09/boffins_probe_commerc...</a></p>
]]></description><pubDate>Thu, 05 Feb 2026 19:50:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=46904219</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=46904219</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46904219</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "A sane but bull case on Clawdbot / OpenClaw"]]></title><description><![CDATA[
<p>Probably, but not necessarily.  Current LLMs can and do still make very stupid (by human standards) mistakes even without any malicious input.<p>Additionally:<p>- As has been pointed out elsewhere in the thread, it can be difficult to separate out "prompt injection" from "marketing" in some cases.<p>- Depending on what the vector for the prompt injection is, what model your OpenClaw instance uses, etc. it might not be easy or even possible to determine whether a given transfer was the result of prompt injection or just the bot making a stupid mistake.  If the burden of proof is on the consumer to prove that it as prompt injection, this would leave many victims with no way to recover their funds.  On the other hand, if banks are required to assume prompt injection unless there's evidence against it, I strongly suspect banks would respond by just banning the use of OpenClaw and similar software with their systems as part of their agreements with their customers.  They might well end up doing that regardless.<p>- Even if a mistake stops well short of draining someones entire account, it can still be very painful financially.</p>
]]></description><pubDate>Wed, 04 Feb 2026 17:53:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=46889078</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=46889078</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46889078</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "A sane but bull case on Clawdbot / OpenClaw"]]></title><description><![CDATA[
<p>I'm not a lawyer, but if I'm reading the actual regulation [0] correctly, it would only apply in the case of prompt injection or other malicious activity.  1005.2.m defines "Unauthorized electronic fund transfer" as follows:<p>> an electronic fund transfer from a consumer's account initiated by a person other than the consumer without actual authority to initiate the transfer and from which the consumer receives no benefit<p>OpenClaw is not legally a person, it's a program.  A program which is being operated by the consumer or a person authorized by said consumer to act on their behalf.  Further, any access to funds it has would have to be granted by the consumer (or a human agent thereof).  Therefore, baring something like a prompt injection attack, it doesn't seem that transfers initiated by OpenClaw would be considered unauthorized.<p>[0]: <a href="https://www.consumerfinance.gov/rules-policy/regulations/1005/2/#m" rel="nofollow">https://www.consumerfinance.gov/rules-policy/regulations/100...</a></p>
]]></description><pubDate>Wed, 04 Feb 2026 16:37:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=46887988</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=46887988</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46887988</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "Moltworker: a self-hosted personal AI agent, minus the minis"]]></title><description><![CDATA[
<p>I think if you care about privacy and security, you wouldn't run moltbot in the first place (or wouldn't give it access to anything you wanted to keep private).</p>
]]></description><pubDate>Thu, 29 Jan 2026 21:46:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=46817116</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=46817116</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46817116</guid></item><item><title><![CDATA[New comment by lunar_mycroft in "After two years of vibecoding, I'm back to writing by hand"]]></title><description><![CDATA[
<p>> That's your job.<p>No, that isn't.  To quote your own blog, his job is to "deliver code [he's] proven to work", <i>not</i> to manage AI agents.  The author has determined that managing AI agents is not an effective way to deliver code in the long term.<p>> you don't have the agent-managerial skills to tell the coding agents how to clean up the mess they made<p>The author has years of experience with AI assisted coding.  Is there any way we can check to see if someone is actually skilled at using these tools besides whether they report/studies measure that they do better with them than without?</p>
]]></description><pubDate>Mon, 26 Jan 2026 20:11:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=46770842</link><dc:creator>lunar_mycroft</dc:creator><comments>https://news.ycombinator.com/item?id=46770842</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46770842</guid></item></channel></rss>