<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: agentultra</title><link>https://news.ycombinator.com/user?id=agentultra</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 12 Apr 2026 09:17:47 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=agentultra" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by agentultra in "AI assistance when contributing to the Linux kernel"]]></title><description><![CDATA[
<p>A piece of wood is either cut to spec or not. You don’t have to try and convince the table saw with a prompt that it is a table saw.<p>These tools are nothing alike and the reductionism of this metaphor isn’t helpful.</p>
]]></description><pubDate>Sat, 11 Apr 2026 20:53:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47733945</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47733945</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47733945</guid></item><item><title><![CDATA[New comment by agentultra in "AI assistance when contributing to the Linux kernel"]]></title><description><![CDATA[
<p><i>Also I, a programmer, can immediately see whether the "probabilistic device" generated code that looks like it should.</i><p>I highly doubt that.<p>Empirical studies show that humans have very little effect on error rates when reviewing code. That effect disappears quickly the more code you read.<p>Most programmers are bad at detecting UB and memory ownership and lifetime errors.<p>A piece of wood comes off the table it’s cut or it’s not.<p>Code is far more complex.</p>
]]></description><pubDate>Sat, 11 Apr 2026 20:16:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47733641</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47733641</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47733641</guid></item><item><title><![CDATA[New comment by agentultra in "AI assistance when contributing to the Linux kernel"]]></title><description><![CDATA[
<p>How do the reviewers feel about this? Hopefully it won't result in them being overwhelmed with PRs. There used to be a kind of "natural limit" to error rates in our code given how much we could produce at once and our risk tolerance for approving changes. Given empirical studies on informal code review which demonstrate how ineffective it is at preventing errors... it seems like we're gearing up to aim a fire-hose of code at people who are ill-prepared to review code at these new volumes.<p>How long until people get exhausted with the new volume of code review and start "trusting" the LLMs more without sufficient review, I wonder?<p>I don't envy Linus in his position... hopefully this approach will work out well for the team.</p>
]]></description><pubDate>Sat, 11 Apr 2026 17:47:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=47732530</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47732530</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47732530</guid></item><item><title><![CDATA[New comment by agentultra in "AI assistance when contributing to the Linux kernel"]]></title><description><![CDATA[
<p>There's a stark difference between a table saw and an LLM that weakens this argument.<p>A table saw isn't a probabilistic device.</p>
]]></description><pubDate>Sat, 11 Apr 2026 17:35:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=47732412</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47732412</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47732412</guid></item><item><title><![CDATA[New comment by agentultra in "The future of everything is lies, I guess – Part 5: Annoyances"]]></title><description><![CDATA[
<p>An important distinction to make, and I whole heartedly agree.<p>It’s not LLMs replacing workers, it’s people. People who have a lot of money and don’t sell their labour for a paycheque. And the systems that compel them to such actions.</p>
]]></description><pubDate>Sat, 11 Apr 2026 16:55:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=47732080</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47732080</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47732080</guid></item><item><title><![CDATA[New comment by agentultra in "The future of everything is lies, I guess – Part 5: Annoyances"]]></title><description><![CDATA[
<p>To lie requires recognition of the truth and an intention to deceive. LLM’s don’t have such abilities. They are systems that generate plausible sequences of symbols based on training inputs, alignments, reinforcement, and inference. These systems don’t know or care what truth is and therefore cannot lie.<p>It’s already bad. I’m not looking forward to the future. These systems are terrible. It’s a future without people that they want for some reason. I’d rather deal with people incompetent, tired, annoyed people than an LLM.</p>
]]></description><pubDate>Sat, 11 Apr 2026 16:38:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47731952</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47731952</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47731952</guid></item><item><title><![CDATA[New comment by agentultra in "I Quit. The Clankers Won"]]></title><description><![CDATA[
<p>Best of luck in your journey!<p>To those reading this thread though, be wary of the answers LLMs generate: they're plausible sounding and the LLM's are designed to be sycophants. Be wary, double check their answers to your queries against credible sources.<p>And read the source!</p>
]]></description><pubDate>Wed, 01 Apr 2026 16:02:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=47602676</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47602676</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47602676</guid></item><item><title><![CDATA[New comment by agentultra in "I quit. The clankers won"]]></title><description><![CDATA[
<p>> Anyone not learning things via LLM coding right now either doesn't care at all about the underlying code/systems<p>How many bytes is a pointer in C? How many bytes is a shared pointer in C++? What does sysctl do? What about fsync?<p>What is a mutex lock? How is it different from a spin lock?<p>You want to find the n nearest points to a given point on a 2-D Cartesian plane.  Could you write the code to solve that on your own?<p>Can you answer any of these questions without searching for the answer?<p>I don't use LLMs and I learn things fine. Always have. For several decades. I care deeply about the underlying code and systems. It annoys me when people say they do and they cannot even understand how the computer works. I'm fine with people having domain-specific knowledge of programming: maybe you've only been interested in web development and scripting DOM elements. But don't pretend that your expertise in that area means you understand how to write an operating system.<p>Or worse: that it prevents you from learning how to write an operating system.<p>You can do that without an LLM. There's no royal road. You have to understand the theory, read the books, read the code, write the code, make mistakes, fix mistakes, read papers, talk to other people with more experience than you... and just write code. And rewrite it. And do it all again.<p>I find the opposite is true: those who use LLM coding <i>exclusively</i> never enjoyed programming to begin with, only learned as much as they needed to, and want the end results.</p>
]]></description><pubDate>Wed, 01 Apr 2026 14:13:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=47601197</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47601197</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47601197</guid></item><item><title><![CDATA[New comment by agentultra in "Slop is not necessarily the future"]]></title><description><![CDATA[
<p>True. I’ve never bought a piece of commercial software and wanted to inspect the source first before making that decision.<p>But I will demand my money back or sue you if your crappy code leaks my personal information, destroys my property, performs worse than advertised, or otherwise harms me in some way.<p>There was sloppy code before LLMs. It’s what they were trained on. And it’s why they generate slop.<p>All that code that was rushed out to meet an arbitrary deadline made up by the sales team, written by junior and lazy senior developers, pushed by the, “code doesn’t matter,” folks. Code written by the enterprise architecture astronauts with a class hierarchy deeper than the Mariana Trench. A few years down the line you get bloated, slow, hard to maintain spaghetti piles of dung. Windows rendering text that stutter when you scroll them. Virtual keyboards that take seconds to pop up. Browser tabs that take more available memory than was available to send astronauts to the moon and back.<p>When humans write it you generally have a few people on a team who are concerned with these things. They try to reign in the slop generating, “always be shipping,” people. You need a mix of both. Because each line of code is a liability as much as it’s a new feature.</p>
]]></description><pubDate>Wed, 01 Apr 2026 13:08:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=47600354</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47600354</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47600354</guid></item><item><title><![CDATA[New comment by agentultra in "How the AI Bubble Bursts"]]></title><description><![CDATA[
<p>It sounds like most of the data centers promised in 2025 and 2026 are not even built yet and most of the GPUs bought haven't even been installed.<p>If it does all go down in flames, even floor value is not going to be that valuable.<p>I can't predict the future but it's smelling a lot like a recession already under way that is bigger than the sub-prime crash.</p>
]]></description><pubDate>Mon, 30 Mar 2026 13:22:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=47573966</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47573966</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47573966</guid></item><item><title><![CDATA[New comment by agentultra in "Coding agents could make free software matter again"]]></title><description><![CDATA[
<p>I think it will wall people off from software.<p>I don’t know what SaaS has to do with FOSS. The point of FOSS was to allow me to modify the software I run on my system. If the device drivers for some hardware I depend on are no longer supported by the company I bought it from, if it’s open source, I can modify and extend the software myself.<p>The Copy Left licenses ensure that I share my modifications back if I distribute them. It’s a thing for the public good.<p>Agent-based software development walls people off from that. Mostly by ensuring that the provenance of the code it generates is not known and by deskilling people so that they don’t know what to prompt or how to fix their code.</p>
]]></description><pubDate>Sun, 29 Mar 2026 23:29:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=47568572</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47568572</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47568572</guid></item><item><title><![CDATA[New comment by agentultra in "Reports of code's death are greatly exaggerated"]]></title><description><![CDATA[
<p>I wouldn’t call what LLMs are doing an <i>abstraction</i>. They generate code. You just don’t have to write it. It can feel like it’s hiding details behind a new, precise semantic layer… but you’ll find out once the project gets to a certain size that is not the case: the details absolutely matter and you’ll be untangling a large knot of code (or prompting the AI to fix it for the seventh time).<p>It’s a good thought and I tend to think that this is the way I would feel more productive: better languages that give us the ability to write better abstractions. Abstractions should provide us with new semantic layers that lose no precision and encapsulate lots of detail.<p>They shouldn’t require us to follow patterns in our code and religiously generate boilerplate and configuration. That’s indirection and slop. It’s wasted code, wasted effort, and is why I find frameworks like React to be… not pleasant to use. I would rather generate the code that adds a button. It should be a single expression but for many reasons, in React, it isn’t.</p>
]]></description><pubDate>Wed, 25 Mar 2026 02:36:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=47512522</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47512522</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47512522</guid></item><item><title><![CDATA[New comment by agentultra in "Reports of code's death are greatly exaggerated"]]></title><description><![CDATA[
<p>The LLM didn’t make a compiler. It generated code that could plausibly implement one. Humans made the compilers it was trained on. It took many such examples and examples of other compilers and thousands of books and articles and blog posts to train the model. It took years of tweaking, fitting, aligning and other tricks to make the model respond to queries with better, more plausible output.  It never made, invented, or reasoned about compilers. It’s an algorithm and system running on a bunch of computers.<p>The C compiler Anthropic got excited about was not a “working” compiler in the sense that you could replace GCC with it and compile the Linux kernel for all of the target platforms it supports. Their definition of, “works,” was that it passed some very basic tests.<p>Same with SQLite translation from C to Rust. Gaping, poorly specified English prose is insufficient. Even with a human in the loop iterating on it. The Rust version is orders of magnitude slower and uses tons more memory. It’s not a drop in Rust-native replacement for SQLite. It’s something else if you want to try that.<p>What mechanism in these systems is responsible for guessing the requirements and constraints missing in the prompts? If we improve that mechanism will we get it to generate a slightly more plausible C compiler or will it tell us that our specifications are insufficient and that we should learn more about compilers first?<p>I’m sure its possible that there are cases where these tools can be useful. I’m not sure this is it though. AGI is purely hypothetical. We don’t simulate a black hole inside a computer and expect gravity to come out of it. We don’t simulate the weather systems on Earth and expect hurricanes to manifest from the computer.  Whatever bar the people selling AI system have for AGI is a moving goalpost, a gimmick, a dream of potential to keep us hooked on what they’re selling right now.<p>It’s unfortunate that the author nearly hits on why but just misses it. The quotes they chose to use nail it. The blog post they reference <i>nearly</i> gets it too. But they both end up giving AI too much credit.<p>Generating a whole React application is probably a breath of fresh air. I don’t doubt anyone would enjoy that and marvel at it. Writing React code is very tedious. There’s just no reason to believe that it is anything more than it is or that we will see anything more than incremental and small improvements from here. If we see any more at all. It’s possible we’re near the limits of what we can do with LLMs.</p>
]]></description><pubDate>Mon, 23 Mar 2026 03:53:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=47485336</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47485336</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47485336</guid></item><item><title><![CDATA[New comment by agentultra in "I'm OK being left behind, thanks"]]></title><description><![CDATA[
<p>One area where it may end up leaving you behind is if you’re looking for a job right now. There are <i>a lot</i> of companies putting vibe coding in their job requirements. The more companies that do this the harder it will be to find employment if you’re not adopting this tool/workflow.</p>
]]></description><pubDate>Fri, 20 Mar 2026 16:17:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=47456769</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47456769</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47456769</guid></item><item><title><![CDATA[New comment by agentultra in "A sufficiently detailed spec is code"]]></title><description><![CDATA[
<p>Unfortunately even formal specifications have this problem. Nothing can replace thinking. But sycophancy, I agree, is a problem. These tools are designed to be pleasing, to generate plausible output; but they cannot think critically about the tasks they're given.<p>Nothing will save you from a bad specification. And there's no royal road to knowing how to write good ones.</p>
]]></description><pubDate>Thu, 19 Mar 2026 15:42:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=47441336</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47441336</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47441336</guid></item><item><title><![CDATA[New comment by agentultra in "A sufficiently detailed spec is code"]]></title><description><![CDATA[
<p>We do have such detailed specifications. But they are written in a language with a narrow interface. It’s a technique called, “program synthesis,” and you can find an example of such a language called, <i>Synquid</i>.<p>It might be illuminating to see what a mathematically precise specification can and cannot do when it comes to generating programs. A major challenge in formal methods is proving that the program implements the specification faithfully, known as the <i>specification gap</i>. If you have a very high level and flexible specification language, such as TLA+, there is a lot of work to do to verify that the program you write meets the specification you wrote. For something like Synquid that is closer to the code there are more constraints on expressivity.<p>The point is that spoken language is not sufficiently precise to define a program.<p>Just because an LLM can fill in plausible details where sufficient detail is lacking doesn’t indicate that it’s solving the specification gap. If the program happens to implement the specification faithfully you got lucky. You still don’t actually know that’s true until you verify it.<p>It’s different with a narrow interface though: you can be very precise and very abstract with a good mathematical system for expressing specifications. It’s a lot more work and requires more training to do than filling in a markdown file and trying to coax the algorithm into outputting what you want through prose and fiction.</p>
]]></description><pubDate>Thu, 19 Mar 2026 14:11:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47439823</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47439823</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47439823</guid></item><item><title><![CDATA[New comment by agentultra in "Toward automated verification of unreviewed AI-generated code"]]></title><description><![CDATA[
<p>This might work on small, self contained projects.<p>No side effects is a hefty constraint.<p>Systems tend to have multiple processes all using side effects. There are global properties of the system that need specification and tests are hard to write for these situations. Especially when they are temporal properties that you care about (eg: if we enter the A state then eventually we must enter the B state).<p>When such guarantees involve multiple processes, even property tests aren’t going to cover you sufficiently.<p>Worse, when it falls over at 3am and you’ve never read the code… is the plan to vibe code a big fix right there? Will you also remember to modify the specifications first?<p>Good on the author for trying. Correctness is hard.</p>
]]></description><pubDate>Tue, 17 Mar 2026 20:22:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=47417792</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47417792</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47417792</guid></item><item><title><![CDATA[New comment by agentultra in "Leanstral: Open-source agent for trustworthy coding and formal proof engineering"]]></title><description><![CDATA[
<p>Very cool but I haven’t been able to convince software developers in industry to write property based tests. I sometimes joke that we will start writing formal proofs until the tests improve. Just so that they will appreciate the difference a little more.<p>I can’t even convince most developers to use model checkers. Far more informal than a full proof in Lean. Still highly useful in many engineering tasks. People prefer boxes and arrows and waving their hands.<p>Anyway, I don’t know that I’d want to have a system vibe code a proof. These types of proofs, I suspect, aren’t going to be generated to be readable, elegant, and be well understood by people. Like programs they generate it will look plausible.<p>And besides, you will still need a human to review the proof and make sure it’s specifying the right things. This doesn’t solve that requirement.<p>Although I have thought that it would be useful to have a system that could prove trivial lemmas in the proof. That would be very neat.</p>
]]></description><pubDate>Tue, 17 Mar 2026 14:29:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=47413212</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47413212</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47413212</guid></item><item><title><![CDATA[New comment by agentultra in "The "are you sure?" Problem: Why AI keeps changing its mind"]]></title><description><![CDATA[
<p>There isn’t a mind to change. Unfortunately the article is slop. Too bad, won’t read the rest.<p>I wish there was a tag or something we could put on headlines to avoid giving views to slop.</p>
]]></description><pubDate>Mon, 16 Mar 2026 13:07:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=47398498</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47398498</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47398498</guid></item><item><title><![CDATA[New comment by agentultra in "AI didn't simplify software engineering: It just made bad engineering easier"]]></title><description><![CDATA[
<p>The end user of a bridge doesn’t care about most things the engineer who designed it does. They care that the bridge spans the gap and is safe to use. The company building the bridge and the group financing its construction care about a few more things like how much it will cost to provide those things to the end user, how long it will last. The engineer cares about a few more things: will the tolerances in the materials used in the struts account for the shearing forces of the most extreme weather in the region?<p>So it is with software.<p>You might not need a blueprint if you’re building a shed in your back yard. This is the kind of software that and user might write or script themselves. If it’s kind of off nobody is going to get hurt.<p>In many cities in North America you can’t construct a dwelling with plumbing connections to a sewer and a permanent foundation without an engineer. And you need an engineer and a blueprint to get the license to go ahead with your construction.<p>Because if you get it wrong you can make people in the dwelling sick and damage the surrounding environment and infrastructure.<p>Software-wise this is where you’re handling other people’s sensitive data. You probably have more than one component that needs to interact with others. If you get it wrong people could lose money, assets could get damaged, etc.<p>This is where I think the software industry needs to figure out liability and maybe professionalize a bit. Right now the liability is with the company and the deterrents are basically no worse than a speeding ticket in most cases. It’s more profitable to keep speeding and pay off the ticket than to prevent harm from throwing out sloppy code and seeing what sticks.<p>Then if you are building a sky scraper… well yeah, that’s the large scale stuff you build at big tech.<p>There are different degrees of software with different requirements. While not <i>engineers</i> by professional accreditation, in practice I would say most software developers are doing engineering… or trying.<p>What I agree with in the article is that AI tools make <i>bad engineering easier.</i> That is for people building houses and skyscrapers who should be thinking about blueprints: they are working under assumptions that the AI is “smart” and will build sky-scrapers for them. They’re not thinking about the things they ought to be and less about things that will pass on to the customers: cost and a product that isn’t fit for use.<p>A bridge that falls down if you drive too fast over it isn’t a useful bridge even though it looks like a bridge.</p>
]]></description><pubDate>Sat, 14 Mar 2026 22:23:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=47381920</link><dc:creator>agentultra</dc:creator><comments>https://news.ycombinator.com/item?id=47381920</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47381920</guid></item></channel></rss>