<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: AlexC04</title><link>https://news.ycombinator.com/user?id=AlexC04</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 09 Apr 2026 03:15:25 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=AlexC04" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by AlexC04 in "System Card: Claude Mythos Preview [pdf]"]]></title><description><![CDATA[
<p>but how does it perform on pelican riding a bicycle bench?  why are they hiding the truth?!<p>(edit: I hope this is an obvious joke. less facetiously these are pretty jaw dropping numbers)</p>
]]></description><pubDate>Tue, 07 Apr 2026 20:20:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=47680853</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47680853</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47680853</guid></item><item><title><![CDATA[New comment by AlexC04 in "Ark – AI agents waste ~30% of context on tool schemas.I built runtime that learn"]]></title><description><![CDATA[
<p>to directly answer this bit:<p>> Feels like a fundamental bottleneck for production agent systems, so would love to compare how you're thinking about the latency vs accuracy tradeoff.<p>I'm really not focusing on latency right now.  My short term goal is to prove the thesis that `ail` can improve same-model performance on SWEBench Pro vs. their own published results.<p>Can I run swebp with GLM-4.6 and get a score better than their published `68.20`    <a href="https://www.swebench.com/" rel="nofollow">https://www.swebench.com/</a>?<p>The argument is that the latency right now just isn't the part we should worry about.  If we're reducing the time to code something from ~6 weeks to 1 hour... then does it really matter tha we add an other 30 minutes of tool calls if we get it 100% right vs. 80% right?<p>Make it work -> Make it right -> make it fast.<p>I'm still on the first one tbh :rofl-emoji:</p>
]]></description><pubDate>Wed, 25 Mar 2026 15:09:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=47518401</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47518401</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47518401</guid></item><item><title><![CDATA[New comment by AlexC04 in "Ark – AI agents waste ~30% of context on tool schemas.I built runtime that learn"]]></title><description><![CDATA[
<p>so - my approach is still being built and I'm still very hand wavy around how it is going to come together, but effecively I'm building pipelines of prompts. Rather than running our LLM sequences as long running sessions where the entire context gets loaded on every turn (a recipe for rot), we unlock the ability to introduce a thinking layer at each step in between the process.<p>So before each turn is sent into the LLM we (potentially) run a local process to assemble a bespoke context of only what is required for that specific turn.<p>If a tool call is not going to be needed on the prompt, we don't include it in the system prompt on that round.<p>I'm still formalizing the spec at the moment and think I'm about six months to a year out before I have a full human ready UI running.<p>This is the foundational paper I'm basing the tool on: <a href="https://github.com/AlexChesser/ail/blob/main/docs/blog/the-yaml-of-the-mind.md" rel="nofollow">https://github.com/AlexChesser/ail/blob/main/docs/blog/the-y...</a> while the spec starts here: <a href="https://github.com/AlexChesser/ail/blob/main/spec/core/s01-purpose.md" rel="nofollow">https://github.com/AlexChesser/ail/blob/main/spec/core/s01-p...</a><p>Essentially I'm trying to build an artificial neocortex and frontal lobe to provide a complete layer of Executive Function that operates on top of our agents - like Claude Code (or whatever else).<p>I'm basing the roadmap on the about 100 years of cognitive science.  We've legitimately had names for all these failure modes (in humans) since the 1960's.  We have observations of what we're witnessing in agents from 1848.<p>We have the roadmap from Psychology.</p>
]]></description><pubDate>Wed, 25 Mar 2026 15:02:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47518283</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47518283</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47518283</guid></item><item><title><![CDATA[New comment by AlexC04 in "Ark – AI agents waste ~30% of context on tool schemas.I built runtime that learn"]]></title><description><![CDATA[
<p>this is a pretty important piece and the research backs you up.  Moving that context out of your system prompt dynamically is going to help reduce your lost in the middle effect.  Context rots almost immediately.  I've got a project that is being built to address this directly as well, but I'm still very early days.<p>Keep it up! you're on the right track.<p>Hong, K., & Chroma Research Team. (2025). Context rot: How increasing input tokens impacts LLM performance. Chroma Research. <a href="https://research.trychroma.com/context-rot" rel="nofollow">https://research.trychroma.com/context-rot</a><p>Liu, N. F., Lin, K., Hewitt, J., Paranjape, A., Bevilacqua, M., Petroni, F., & Liang, P. (2024). Lost in the middle: How language models use long contexts. Transactions of the Association for Computational Linguistics, 12, 157–173. <a href="https://doi.org/10.1162/tacl_a_00638" rel="nofollow">https://doi.org/10.1162/tacl_a_00638</a></p>
]]></description><pubDate>Wed, 25 Mar 2026 14:32:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=47517864</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47517864</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47517864</guid></item><item><title><![CDATA[New comment by AlexC04 in "Cortex – Local-first AI memory engine, beats Mem0 on LoCoMo, encrypted, free"]]></title><description><![CDATA[
<p>Hey! This looks a lot like what I'm working on, from a slightly different angle. I think you're on the right track. In fact, cortex as a name is perfect since you're effectively building the executive function layer for search and selection.  I also think rust is the right language to go with.<p>I'm going do a deeper read of your work in a bit. I'd love it if you took a look at my theory of artificial cognition The YAML of the Mind <a href="https://alexchesser.medium.com/the-yaml-of-the-mind-8a4f945a76c4" rel="nofollow">https://alexchesser.medium.com/the-yaml-of-the-mind-8a4f945a...</a>, dropped in to the `ail` project and let me know what you think.<p>I just have to get the kids to school and I'll pop back into cortex later</p>
]]></description><pubDate>Tue, 24 Mar 2026 12:12:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47501487</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47501487</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47501487</guid></item><item><title><![CDATA[New comment by AlexC04 in "[dead]"]]></title><description><![CDATA[
<p>Hey folks, I wrote this. If you're interested in the concepts or pressure testing the ideas a little deeper, please feel free to comment here or reach out directly.<p>I appreciate that it's pretty long feel free to point your LLMs at it</p>
]]></description><pubDate>Tue, 24 Mar 2026 12:05:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=47501419</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47501419</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47501419</guid></item><item><title><![CDATA[The YAML of the Mind: What a century of cognitive science tells us about AI]]></title><description><![CDATA[
<p>Article URL: <a href="https://alexchesser.medium.com/the-yaml-of-the-mind-8a4f945a76c4">https://alexchesser.medium.com/the-yaml-of-the-mind-8a4f945a76c4</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47501418">https://news.ycombinator.com/item?id=47501418</a></p>
<p>Points: 1</p>
<p># Comments: 1</p>
]]></description><pubDate>Tue, 24 Mar 2026 12:05:27 +0000</pubDate><link>https://alexchesser.medium.com/the-yaml-of-the-mind-8a4f945a76c4</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47501418</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47501418</guid></item><item><title><![CDATA[New comment by AlexC04 in "Kotlin creator's new language: talk to LLMs in specs, not English"]]></title><description><![CDATA[
<p>this is really exciting and dovetails really closely with the project I'm working on.<p>I'm writing a language spec for an LLM runner that has the ability to chain prompts and hooks into workflows.<p><a href="https://github.com/AlexChesser/ail" rel="nofollow">https://github.com/AlexChesser/ail</a><p>I'm writing the tool as proof of the spec. Still very much a pre-alpha phase, but I do have a working POC in that I can specify a series of prompts in my YAML language and execute the chain of commands in a local agent.<p>One of the "key steps" that I plan on designing is specifically an invocation interceptor. My underlying theory is that we would take whatever random series of prose that our human minds come up with and pass it through a prompt refinement engine:<p>> Clean up the following prompt in order to convert the user's intent
> into a structured prompt optimized for working with an LLM
> Be sure to follow appropriate modern standards based on current
> prompt engineering reasech. For example, limit the use of persona
> assignment in order to reduce hallucinations.
> If the user is asking for multiple actions, break the prompt 
> into appropriate steps (**etc...)<p>That interceptor would then forward the well structured intent-parsed prompt to the LLM. I could really see a step where we say "take the crap I just said and 
turn it into CodeSpeak"<p>What a fantastic tool. I'll definitely do a deep dive into this.</p>
]]></description><pubDate>Thu, 12 Mar 2026 15:40:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47352394</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47352394</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47352394</guid></item><item><title><![CDATA[New comment by AlexC04 in "Workers who love ‘synergizing paradigms’ might be bad at their jobs"]]></title><description><![CDATA[
<p>that's a bit of a meta discussion and it'd probably reveal some super interesting things about how tech culture have changed in the last ~15 years.<p>I've been on HN since 2010 (lost the password to my first account, alexc04) and I recall a time when it felt like every second article on the front-page was an bold directive pronouncement or something just aggressively certain of its own correctness.<p>Like "STOP USING BASH" or "JQUERY IS STUPID" - not in all caps of course but it created an unpleasant air and tone (IMO, again, this is like 16 years ago now so I may have memory degredation to some extent)<p>Things like donglegate got real traction here among the anti-woke crew.  There have been times where the venn diagram of 4chan and hackernews felt like it had a lot more overlap.  I've even bowed out of discussion for years at a time or developed an avoidance reaction to HN's toxic discussion culture.<p>IMO it has been a LOT better in more recent years, but I also don't dive as deep as I used to.<p>ANYWAYS - my point is I would be really interested to see a sentiment analysis of HN headlines over the years to try and map out cultural epochs of the community.<p>When has HN swayed more into the toxic and how has it swayed back and forth as a pendulum over time? (or even has it?)<p>I wonder what other people's perspective is of how the culture here has changed over time. I truly think it feels a lot more supportive than it used to.</p>
]]></description><pubDate>Fri, 06 Mar 2026 14:41:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47275469</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47275469</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47275469</guid></item><item><title><![CDATA[New comment by AlexC04 in "How the Moat Is Moving"]]></title><description><![CDATA[
<p>> This isn’t a bubble inflating. It’s capital and intelligence relocating.<p>Is this GPT? this kind of not X but Y pattern is a real code smell for slop and can cause someone to immediately bounce out of your writing. The pattern can really feel hard hitting when you're reading what GPT put in front of you, but others are REALLY starting to reject it.<p><a href="https://en.wikipedia.org/wiki/Wikipedia:Signs_of_AI_writing" rel="nofollow">https://en.wikipedia.org/wiki/Wikipedia:Signs_of_AI_writing</a><p><a href="https://saigaddam.medium.com/it-isnt-just-x-it-s-y-54cb403d61a8" rel="nofollow">https://saigaddam.medium.com/it-isnt-just-x-it-s-y-54cb403d6...</a><p><a href="https://www.blakestockton.com/dont-write-like-ai-1-101-negation/" rel="nofollow">https://www.blakestockton.com/dont-write-like-ai-1-101-negat...</a><p>Like it or not this can lead people to immediately reject your entire message, shut down and stop reading.</p>
]]></description><pubDate>Tue, 03 Mar 2026 18:18:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=47236447</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47236447</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47236447</guid></item><item><title><![CDATA[New comment by AlexC04 in "Ask HN: How can I avoid the "It's not just X, it's Y" writing"]]></title><description><![CDATA[
<p>I agree completely and I don't think you can.<p>Even people recording short form video are doing it. They're reading out their chat-gpt-psychosis induced fever dreams using scripts written by chatGPT.<p>The 7-part tweets that build to a slop crescendo are doing my head in.<p>The solution might be some combination of:<p>1. leave the social media sites where the slop is irredeemable.   
2. unfollow everyone, reset your algorithm.   
3. be aggressive about who you add back in. Make sure they're humans having high quality discussions.    
4. be aggressive about who you block. Lower the bar on blocking - one and done. No chances, no wait-and-see.    
5. move to smaller communities of real humans.<p>None of this has worked for me yet.  I'm still swimming in a vast sea of slopity slop slop. Dead internet theory appears to be playing out in front of us.<p>edit: On threads I've been trying to use their `dear algo` feature pretty aggressively but it doesn't work very well.  I've asked it to remove some types of comments and it seems to just add more of them.</p>
]]></description><pubDate>Wed, 25 Feb 2026 16:00:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=47153317</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47153317</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47153317</guid></item><item><title><![CDATA[New comment by AlexC04 in "The path to ubiquitous AI (17k tokens/sec)"]]></title><description><![CDATA[
<p>If I could have one of these cards in my own computer do you think it would be possible to replace claude code?<p>1. Assume It's running a better model, even a dedicated coding model. High scoring but obviously not opus 4.5
2. Instead of the standard send-receive paradigm we set up a pipeline of agents, each of whom parses the output of the previous.<p>At 17k/tps running locally, you could effectively spin up tasks like "you are an agent who adds semicolons to the end of the line in javascript", with some sort of dedicated software in the style of claude code you could load an array of 20 agents each with a role to play in improving outpus.<p>take user input and gather context from codebase  
-> rewrite what you think the human asked you in the form of an LLM-optimized instructional prompt   
-> examine the prompt for uncertainties and gaps in your understanding or ability to execute   
-> <assume more steps as relevant>   
-> execute the work<p>Could you effectively set up something that is configurable to the individual developer - a folder of system prompts that every request loops through?<p>Do you really need the best model if you can pass your responses through a medium tier model that engages in rapid self improvement 30 times in a row before your claude server has returned its first shot response?</p>
]]></description><pubDate>Fri, 20 Feb 2026 19:06:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=47092371</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=47092371</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47092371</guid></item><item><title><![CDATA[New comment by AlexC04 in "xAI joins SpaceX"]]></title><description><![CDATA[
<p>According to google Tesla has shipped 8 million cars total since inception.
It is valued at 1.32 Trillion as of today. Which is roughly $165,000 per shipped vehicle.</p>
]]></description><pubDate>Tue, 03 Feb 2026 15:24:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=46872120</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=46872120</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46872120</guid></item><item><title><![CDATA[New comment by AlexC04 in "xAI joins SpaceX"]]></title><description><![CDATA[
<p>Right, let's not forget that he's selling it to himself in an all stock deal.  He could have priced it at eleventy kajillion dollars and it would have had the same meaning.<p>He's basically trading two cypto coins with himself and sending out a press release.</p>
]]></description><pubDate>Tue, 03 Feb 2026 15:03:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=46871834</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=46871834</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46871834</guid></item><item><title><![CDATA[New comment by AlexC04 in "TikTok users can't upload anti-ICE videos. The company blames tech issues"]]></title><description><![CDATA[
<p>it seems like our software engineers included `censorship=true` in the latest build when someone filed a JIRA ticket that said "censor stuff"<p>we're going to class this one as low priority / won't fix.</p>
]]></description><pubDate>Tue, 27 Jan 2026 17:56:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=46783631</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=46783631</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46783631</guid></item><item><title><![CDATA[New comment by AlexC04 in "Goldman Sachs Global Macro Research: Gen AI: too much spend, too little benefit [pdf] (2024)"]]></title><description><![CDATA[
<p>can you post the correct link in the comments?<p>this appears to be a redirect away from static assets.<p>looks like you've got to go here or their systems will mess you around: <a href="https://www.goldmansachs.com/insights/top-of-mind/gen-ai-too-much-spend-too-little-benefit" rel="nofollow">https://www.goldmansachs.com/insights/top-of-mind/gen-ai-too...</a></p>
]]></description><pubDate>Thu, 22 Jan 2026 20:44:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=46724898</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=46724898</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46724898</guid></item><item><title><![CDATA[New comment by AlexC04 in "Economics of Orbital vs. Terrestrial Data Centers"]]></title><description><![CDATA[
<p>do it on mars then... that would have the added benefit of heating the planet so we could live on it. It seems so obvious if you think about it. Someone transfer eleventy trillion dollars to Elon Musk so we can get started.</p>
]]></description><pubDate>Tue, 16 Dec 2025 15:29:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=46289729</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=46289729</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46289729</guid></item><item><title><![CDATA[New comment by AlexC04 in "Tim Ferriss Promised Freedom. Indie Hackers Are Selling Shovels"]]></title><description><![CDATA[
<p>the 4hww was <i>such</i> a damaging read for me. I wasted YEARS squeaking pennies out of shitty, poorly-realized side hustles trying to solo that shit when the real ROI   would have been to focus on leetcode in my 20's and track down a role paying 10x the salary.<p>When I first read it, I was probably making about $60-65k and scrambling for $2.5k raises with a shitty meat-grinder employer while burning the midnight oil and acting as a "weekend warrior" trying to come up with something that would allow me the freedom that ferris spoke of.<p>Problem is, all of his techniques presumed you already had some sort of business that you could automate yourself out of.<p>If you're a wage-earner none of the shit he talks about is realistic.  Sure, you can do the research to become a drop shipper or to place tiny ads in the back of soldier-of-fortune for brain pills.<p>For the typical computer programmer - your ROI will be 100x higher if you just get good enough at algorithms and system design to land a role at a company with Tier3  compensation.</p>
]]></description><pubDate>Mon, 01 Dec 2025 15:36:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=46108650</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=46108650</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46108650</guid></item><item><title><![CDATA[New comment by AlexC04 in "I built ChatGPT with Minecraft redstone [video]"]]></title><description><![CDATA[
<p>when people do these complex CPU designs in minecraft are they laying the blocks individually in real time and in 3d space - or are they scripting some sort of algorithm that instantiates the system in one go?<p>It's impressive either way but the manual version seems ... impossible.</p>
]]></description><pubDate>Thu, 02 Oct 2025 14:16:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=45449901</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=45449901</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45449901</guid></item><item><title><![CDATA[New comment by AlexC04 in "Marvel Studios is moving from Georgia to the UK to avoid paying health insurance"]]></title><description><![CDATA[
<p>What an oddly editorialized version of this headline which does not appear to reflect the conclusion made by the article OR the actual headline of the linked article:<p>actual headline:<p>> Is Marvel leaving Georgia? Production shifts to UK spark industry shakeup<p>piece talking about reasons:<p>> shifting to the United Kingdom, where lower production costs, especially on wages and employee benefits, are giving studios more bang for their buck, according to the Daily Mail.<p>looks like differences in wages AND benefits. Lower overall salaries in the UK combined with public healthcare costs via the NHS is a much more nuanced take than "leaving to avoid paying health insurance."<p>I suppose the conclusion though is the same - that a public option would have prevented this flight of capital.  Not necessarily dropping of mandates that employers cover health insurance to remain competitive (which I'm sure some will conclude).</p>
]]></description><pubDate>Wed, 10 Sep 2025 14:49:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=45198606</link><dc:creator>AlexC04</dc:creator><comments>https://news.ycombinator.com/item?id=45198606</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45198606</guid></item></channel></rss>