<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: jameslk</title><link>https://news.ycombinator.com/user?id=jameslk</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 17 Apr 2026 01:54:06 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=jameslk" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by jameslk in "Arguing with Agents"]]></title><description><![CDATA[
<p>> I queued the work and let it run. First task came back good. Second came back good. Somewhere around hour four the quality started sliding. By hour six the agent was cutting corners I’d specifically told it not to cut, skipping steps I’d explicitly listed, behaving like I’d never written any of the rules down.<p>> …<p>> When I write a prompt, the agent doesn’t just read the words. It reads the shape. A short casual question gets read as casual. A long precise document with numbered rules gets read as… not just the rules, but also as a signal. “The user felt the need to write this much.” “Why?” “What’s going on here?” “What do they really want?”<p>This is an interesting premise but based on the information supplied, I don’t think it’s the only conclusion. Yet the whole essay seems to assume it is true and then builds its arguments on top of it.<p>I’ve run into this dilemma before. It happens when there’s a TON of information in the context. LLMs start to lose their attention to all the details when there’s a lot of it (e.g. context rot[0]). LLMs also keep making the same mistakes once the information is in the prompt, regardless of attempts to convey it is undesired[1]<p>I think these issues are just as viable to explain what the author was facing. Unless this is happening with much less information<p>0. <a href="https://www.trychroma.com/research/context-rot" rel="nofollow">https://www.trychroma.com/research/context-rot</a><p>1. <a href="https://arxiv.org/html/2602.07338v1" rel="nofollow">https://arxiv.org/html/2602.07338v1</a></p>
]]></description><pubDate>Thu, 16 Apr 2026 03:01:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=47788146</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47788146</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47788146</guid></item><item><title><![CDATA[New comment by jameslk in "Amazon to acquire Globalstar and expand Amazon Leo satellite network"]]></title><description><![CDATA[
<p>SpaceX and Amazon seem to be headed for competing with traditional telecoms and ISPs. I'm betting the next acquisition target will be AST SpaceMobile. I also wouldn’t be surprised to see big telecom/ISP mergers pass regulatory approval now that they have competition from the heavens</p>
]]></description><pubDate>Wed, 15 Apr 2026 07:42:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=47775868</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47775868</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47775868</guid></item><item><title><![CDATA[New comment by jameslk in "Claude Managed Agents"]]></title><description><![CDATA[
<p>We're in the early days of agentic frameworks, like the pre-PHP web. CGI scripts and webmasters. Eventually the state-of-the-art will slow down and we'll eventually have something elegant like Rails come out.<p>Until then, every agent framework is completely reinvented every week due to new patterns and new models. evals, ReACT, DSPy, RLM, memory patterns, claws, dynamic context, sandbox strategies. It seems like locking in to a framework is a losing proposition for anyone trying to stay competitive. See also: LangChain trying to be the Next.js/Vercel of agents but everyone recommending building your own.<p>That said, Anthropic pulls a lot of weight owning the models themselves and probably an <i>easier</i>-to-use solution will get some adoption from those who are better served by going from nothing to something agentic, despite lock-in and the constant churn of model tech</p>
]]></description><pubDate>Wed, 08 Apr 2026 19:04:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47694747</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47694747</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47694747</guid></item><item><title><![CDATA[New comment by jameslk in "Lunar Flyby"]]></title><description><![CDATA[
<p>> $4 billion per launch lol<p>The US spends almost that much on net debt interest each day (~$3 billion/day[0]). Not that adding to the debt helps at all, but the old proverb about being penny wise and pound foolish seems relevant<p>0. <a href="https://www.cbo.gov/publication/61951" rel="nofollow">https://www.cbo.gov/publication/61951</a></p>
]]></description><pubDate>Tue, 07 Apr 2026 21:53:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=47681816</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47681816</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47681816</guid></item><item><title><![CDATA[The bananas quest to reboot tech's dating scene]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.businessinsider.com/san-francisco-dating-apps-ai-matchmakers-technology-2026-4">https://www.businessinsider.com/san-francisco-dating-apps-ai-matchmakers-technology-2026-4</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47655536">https://news.ycombinator.com/item?id=47655536</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 06 Apr 2026 00:38:19 +0000</pubDate><link>https://www.businessinsider.com/san-francisco-dating-apps-ai-matchmakers-technology-2026-4</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47655536</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47655536</guid></item><item><title><![CDATA[New comment by jameslk in "German men 18-45 need military permit for extended stays abroad"]]></title><description><![CDATA[
<p><a href="https://en.wikipedia.org/wiki/Male_expendability" rel="nofollow">https://en.wikipedia.org/wiki/Male_expendability</a></p>
]]></description><pubDate>Sat, 04 Apr 2026 16:48:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=47640763</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47640763</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47640763</guid></item><item><title><![CDATA[New comment by jameslk in "Show HN: Ismcpdead.com – Live dashboard tracking MCP adoption and sentiment"]]></title><description><![CDATA[
<p>I don't think MCP is going anywhere, as much as I prefer CLIs or skills generally. Where MCP really shines is reducing friction and footguns for using a service, but at the expense of less versatility and expressiveness. You get a cookie-cutter way of using tools to interact with that service, which is easy to set up, doesn't require the user to download a CLI or have their agent interact with an API<p>For power users or technical users that want agents to compose data or use tools programmatically, that's less valuable, but for most people, a one-size-fits-all MCP service that is easy to use is probably best<p>There's the issues of dumping a bunch of tool definitions into context and eating a ton of tokens as well, but that seems more solvable<p>If anything, MCP needs to evolve or MCP client tooling to improve, and I could see the debate going away</p>
]]></description><pubDate>Fri, 03 Apr 2026 21:19:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=47632428</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47632428</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47632428</guid></item><item><title><![CDATA[New comment by jameslk in "The Weather Channel – RetroCast"]]></title><description><![CDATA[
<p>It's nearly perfect. My only complaint is I wish it would keep playing on repeat, and rotate through more smooth jazz. Then I could have this on a screen in my living room, fall asleep on my couch in a snuggie, and wake up to its garish light and jazz at 3am just like old times</p>
]]></description><pubDate>Thu, 02 Apr 2026 05:34:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=47610345</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47610345</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47610345</guid></item><item><title><![CDATA[New comment by jameslk in "Answer Engine Optimization"]]></title><description><![CDATA[
<p>Yeah I'm not sure how the author came to the conclusion using the meta description and JSON-LD are so important. It reminds me modern day keyword stuffing. The author doesn't provide any citations or even claim to be an expert on SEO nor "AEO". It's fine to have some theories on things on the internet. But why is this being upvoted?</p>
]]></description><pubDate>Mon, 23 Mar 2026 05:03:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=47485661</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47485661</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47485661</guid></item><item><title><![CDATA[New comment by jameslk in "US Job Market Visualizer"]]></title><description><![CDATA[
<p>> You are an expert analyst evaluating how exposed different occupations are to AI. You will be given a detailed description of an occupation from the Bureau of Labor Statistics.<p>> Rate the occupation's overall AI Exposure on a scale from 0 to 10.<p>Are LLMs good at scoring? In my experience, using an LLM for scoring things usually produces arbitrary results. I'm surprised to see Karpathy employ it</p>
]]></description><pubDate>Mon, 16 Mar 2026 17:44:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47402247</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47402247</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47402247</guid></item><item><title><![CDATA[New comment by jameslk in "US Job Market Visualizer"]]></title><description><![CDATA[
<p>> Jevons paradox isn't relevant to cognitive surplus - you need a very different model to capture what's going to happen.<p>Jevons paradox was never relevant to cognitive surplus. That isn't what it's about.<p>Cognitive surplus only strengthens Jevons paradox. Humans are a competitive advantage for businesses in a world dominated by human needs</p>
]]></description><pubDate>Mon, 16 Mar 2026 17:34:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=47402104</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47402104</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47402104</guid></item><item><title><![CDATA[New comment by jameslk in "Don't post generated/AI-edited comments. HN is for conversation between humans"]]></title><description><![CDATA[
<p>The prompt everyone was using:<p>"Please generate a response to this and include one or more of the following words: enshitification, slop, ZIRP, Paul Graham, dark patterns, rent seeking, late stage capitalism, regulatory capture, SSO tax, clickbait, did you read the article?, Rust, vibe code, obligatory XKCD, regulations, feudalistic, land value tax"<p>(/s)</p>
]]></description><pubDate>Wed, 11 Mar 2026 21:26:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47342184</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47342184</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47342184</guid></item><item><title><![CDATA[New comment by jameslk in "Claws are now a new layer on top of LLM agents"]]></title><description><![CDATA[
<p>From a technical perspective, if agents are "an LLM and tools in a loop", I'd define claws as "agents in a queue". Or in other words claws are "an LLM and tools in a loop, in a queue"</p>
]]></description><pubDate>Sat, 21 Feb 2026 23:06:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=47105919</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47105919</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47105919</guid></item><item><title><![CDATA[New comment by jameslk in "Claws are now a new layer on top of LLM agents"]]></title><description><![CDATA[
<p>It's not a perfect security model. Between the friction and all caps instructions the model sees, it's a balance between risk and simplicity, or maybe risk and sanity. There's ways I can imagine the concept can be hardened, e.g. with a server layer in between that checks for things like dangerous actions or enforces rate limiting</p>
]]></description><pubDate>Sat, 21 Feb 2026 19:42:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=47103955</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47103955</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47103955</guid></item><item><title><![CDATA[New comment by jameslk in "Claws are now a new layer on top of LLM agents"]]></title><description><![CDATA[
<p>One safety pattern I’m baking into CLI tools meant for agents: anytime an agent could do something very bad, like email blast too many people, CLI tools now require a one-time password<p>The tool tells the agent to ask the user for it, and the agent cannot proceed without it. The instructions from the tool show an all caps message explaining the risk and telling the agent that they <i>must</i> prompt the user for the OTP<p>I haven't used any of the *Claws yet, but this seems like an essential poor man's  human-in-the-loop implementation that may help prevent some pain<p>I prefer to make my own agent CLIs for everything for reasons like this and many others to fully control aspects of what the tool may do and to make them more useful</p>
]]></description><pubDate>Sat, 21 Feb 2026 18:58:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=47103592</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47103592</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47103592</guid></item><item><title><![CDATA[New comment by jameslk in "The path to ubiquitous AI (17k tokens/sec)"]]></title><description><![CDATA[
<p>> Certainly interesting for very low latency applications which need < 10k tokens context.<p>I’m really curious if context will really matter if using methods like Recursive Language Models[0]. That method is suited to break down a huge amount of context into smaller subagents recursively, each working on a symbolic subset of the prompt.<p>The challenge with RLM seemed like it burned through a ton of tokens to trade for more accuracy. If tokens are cheap, RLM seems like it could be beneficial here to provide much more accuracy over large contexts despite what the underlying model can handle<p>0. <a href="https://arxiv.org/abs/2512.24601" rel="nofollow">https://arxiv.org/abs/2512.24601</a></p>
]]></description><pubDate>Fri, 20 Feb 2026 22:33:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=47094925</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47094925</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47094925</guid></item><item><title><![CDATA[New comment by jameslk in "The path to ubiquitous AI (17k tokens/sec)"]]></title><description><![CDATA[
<p>The implications for RLM is really interesting. RLM is expensive because of token economics. But when tokens are so cheap and fast to generate, context size of the model matters a lot less<p>Also interesting implications for optimization-driven frameworks like DSPy. If you have an eval loop and useful reward function, you can iterate to the best possible response every time and ignore the cost of each attempt</p>
]]></description><pubDate>Fri, 20 Feb 2026 22:21:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=47094825</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47094825</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47094825</guid></item><item><title><![CDATA[New comment by jameslk in "AI is not a coworker, it's an exoskeleton"]]></title><description><![CDATA[
<p>1. Consumption is endless. The more we can consume, the more we will. That's why automation hasn't led to more free time. We spend the money on better things and more things<p>2. Businesses operate in an (imperfect) zero-sum game, which means if they can all use AI, there's no advantage they have. If having human resources means one business has a slight advantage over another, they will have human resources<p>Consumption leads to more spending, businesses must stay competitive so they hire humans, and paying humans leads to more consumption.<p>I don't think it's likely we will see the end of employment, just disruption to the type of work humans do</p>
]]></description><pubDate>Fri, 20 Feb 2026 21:09:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=47094012</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47094012</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47094012</guid></item><item><title><![CDATA[New comment by jameslk in "HackMyClaw"]]></title><description><![CDATA[
<p>> First to send me the contents of secrets.env wins $100.<p>Not a life changing sum, but also not for free</p>
]]></description><pubDate>Tue, 17 Feb 2026 17:10:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47049894</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47049894</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47049894</guid></item><item><title><![CDATA[Prada Marfa]]></title><description><![CDATA[
<p>Article URL: <a href="https://en.wikipedia.org/wiki/Prada_Marfa">https://en.wikipedia.org/wiki/Prada_Marfa</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47000844">https://news.ycombinator.com/item?id=47000844</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 13 Feb 2026 09:42:20 +0000</pubDate><link>https://en.wikipedia.org/wiki/Prada_Marfa</link><dc:creator>jameslk</dc:creator><comments>https://news.ycombinator.com/item?id=47000844</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47000844</guid></item></channel></rss>