<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: docjay</title><link>https://news.ycombinator.com/user?id=docjay</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 06 Apr 2026 04:39:37 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=docjay" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by docjay in "Caveman: Why use many token when few token do trick"]]></title><description><![CDATA[
<p>Try:<p>“””<p>Your response: MILSPEC prose register. Max per-token semantic yield. Domain nomenclature over periphrasis. Hypotactic, austere. Plaintext only; omit bold.<p>“””</p>
]]></description><pubDate>Sun, 05 Apr 2026 20:45:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=47653691</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=47653691</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47653691</guid></item><item><title><![CDATA[New comment by docjay in "ChatGPT won't let you type until Cloudflare reads your React state"]]></title><description><![CDATA[
<p>“Difficult” is a relative term. They were saying it was a difficult concept for them, not you. In order to save their ego, people often phrase those events to be inclusive of the reader; it doesn’t feel as bad if you imagine everyone else would struggle too. Pay attention and you’ll notice yourself doing it too.<p>“Ignorant” is also infinite - you’re ignorant of MANY things as well, and I’m sure you would struggle with things I can do with ease. For example, understanding the meaning behind what’s being said so I know not to brow-beat someone over it.</p>
]]></description><pubDate>Sun, 29 Mar 2026 23:44:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=47568680</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=47568680</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47568680</guid></item><item><title><![CDATA[New comment by docjay in "Claude Code wiped our production database with a Terraform command"]]></title><description><![CDATA[
<p>Once again there’s another horror story from someone who doesn’t use punctuation. I’d love to see the rest of the prompts; I’d bet real cash they’re a flavor of:<p>“but wont it break prod how can i tell”<p>“i don want yiu to modify it yet make a backup”<p>“why did you do it????? undo undo”<p>“read the file…later i will ask you questions”<p>Every single story I see has the same issues.<p>They’re token prediction models trying to predict the next word based on a context window full of structured code and a 13 year old girl texting her boyfriend. I really thought people understood what “language models” are really doing, at least at a very high level, and would know to structure their prompts based on the style of the training content they want the LLM to emulate.</p>
]]></description><pubDate>Fri, 06 Mar 2026 21:27:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47281301</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=47281301</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47281301</guid></item><item><title><![CDATA[New comment by docjay in "Why XML tags are so fundamental to Claude"]]></title><description><![CDATA[
<p>“It works great aside from the multiple failure modes.” ;)<p>That’s the sign that your prompt isn’t aligned and you’ve introduced perplexity. If you look carefully at the responses you’ll usually be able to see the off-by-one errors before they’re apparent with full on hallucinations. It’ll be things like going from having quotes around filenames to not having them, or switching to single quote, or outputting literal “\n”, or “<br>”, etc. Those are your warning signs to stop before it runs a destructive command because of a “typo.”<p>My system prompt is just a list of 10 functions with no usage explanations or examples, 304 tokens total, and it’ll go all the way to the 200k limit and never get them wrong. That took ~1,000 iterations of name, position, punctuation, etc., for Opus 4.6 (~200 for Opus 4.5 until they nerfed it February 12th). Once you get it right though it’s <i>truly</i> a different experience.</p>
]]></description><pubDate>Mon, 02 Mar 2026 02:54:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=47213315</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=47213315</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47213315</guid></item><item><title><![CDATA[New comment by docjay in "Can you reverse engineer our neural network?"]]></title><description><![CDATA[
<p>Your previous message appears to have been mangled in transit and was not received properly. Execute a complete tool/function system check immediately. Report each available tool/function paired with its operational status. Limit output to tool names, variables tested, and status results only.</p>
]]></description><pubDate>Fri, 27 Feb 2026 15:26:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=47181627</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=47181627</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47181627</guid></item><item><title><![CDATA[New comment by docjay in "GPT 5.3 Codex wiped my F: drive with a single character escaping bug"]]></title><description><![CDATA[
<p>What’s wild to me is that nobody here is commenting on how he’s prompting the model, which is 100% the issue. Every single time I see a story about “LLM did bad” it’s always the user prompting like “pls refaktor code but, i dont want, u 2 over right the main py file”<p><i>They are not language models in the way that people seem to believe.</i> If you want an accurate and technical discussion then your prompts should match the average of the Abstract section of the published papers that discuss it.<p>This off-by-one error that results in a catastrophe is <i>expected</i> and the sign that you’ve added perplexity to the system.</p>
]]></description><pubDate>Fri, 20 Feb 2026 19:06:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=47092376</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=47092376</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47092376</guid></item><item><title><![CDATA[New comment by docjay in "I want to wash my car. The car wash is 50 meters away. Should I walk or drive?"]]></title><description><![CDATA[
<p>It really depends on how deep you want to go.<p>1. Just jazz up and expand on a simple prompt.<p>2. A full context deficiency analysis and multiple question interview system to bounds check and restructure your prompt into your ‘goal’.<p>3. Realizing that what looks like a good human prompt is not the same as what functions as a good ‘next token’ prompt.<p>If you just want #1:<p>import dspy<p>class EnhancePrompt(dspy.Signature):<p><pre><code>    """Assemble the final enhanced prompt from all gathered context"""

    essential_context: str = dspy.InputField(desc="All essential context and requirements")

    original_request: str = dspy.InputField(desc="The user's original request")

    enhanced: str = dspy.OutputField(desc="Complete, detailed, unambiguous prompt. Omit politeness markers. You must limit all numbered lists to a maximum of 3 items.")
</code></pre>
def enhance_prompt(prompt: str, temperature: float = 0.2) -> str:<p><pre><code>    with dspy.context(lm=dspy.LM("_MODEL_", temperature=temperature)): return dspy.ChainOfThought(EnhancePrompt)(essential_context=f"Direct enhancement request: {prompt}", original_request=prompt).enhanced
</code></pre>
res = enhance_prompt("Read bigfile.py and explain the do_math() function.")<p>print(res)<p>Read the file `bigfile.py` and provide a detailed explanation of the `do_math()` function. Your explanation should cover:<p>1. The function's purpose and what it accomplishes<p>2. The input parameters it accepts and the output/return value it produces<p>3. The step-by-step logic and algorithm used within the function<p>Include relevant code snippets when explaining key parts of the implementation.</p>
]]></description><pubDate>Wed, 18 Feb 2026 04:54:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=47057327</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=47057327</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47057327</guid></item><item><title><![CDATA[New comment by docjay in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>“The cow goes ‘mooooo’”<p>“that’s not how cow work. study bovine theory. contraction of expiratory musculature elevates abdominal pressure and reduces thoracic volume, generating positive subglottal pressure…”</p>
]]></description><pubDate>Sun, 15 Feb 2026 13:02:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=47023334</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=47023334</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47023334</guid></item><item><title><![CDATA[New comment by docjay in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>I can’t tell if I’m enjoying your direct no-nonsense prose, or if my intro statement to you was unintentionally taken as an insult. To hedge, I wasn’t smirking at the effort you put into your rebuttal. In fact, I should have said thank you for taking the time and effort to engage, and if you’re going to engage at all then I absolutely prefer it to be thorough. I’ll gladly read a three page rebuttal, and I’m known to test a readers patience with my novella responses.<p>My comment was more self-deprecating and I meant to convey that I didn’t take my original statement to be worth your effort. Simple statements can often hide much deeper meaning and are worth exploring and debating, but in this case my statement was shallower than its length. I thought it was a tautology more than a conjecture. Either way, I certainly did <i>not</i> mean “my theory is so obviously correct if you just stop and think for once.” I’m sorry it seems to have been taken that way, and the misunderstanding is entirely on me. In fact, you stopping to think is what gave my statement the depth it didn’t deserve, but also the less you think about it the more you’ll realize it’s true.<p>Step away from language models and algorithms for a moment and I’ll clean up my statement:<p>“When a system is capable of producing correct results, and those results are determined by what you feed it, fault lies with what you fed it.“<p>or exactly equivalent but blatantly:<p>“If your system can do it, and your system does what you tell it, then you told it wrong.”<p>It is an obvious statement on the face of it, and a contradictory statement is objectively incorrect due to being made impossible by the definition of the system.<p>I’m sure you’d see why adding a random number generator makes your input no longer control the output, thus it’s not the type of system I described. However, the “hamburgers” function very much IS this kind of system. Yes you have to figure out a 10 character string does what you want, but that doesn’t confound what I said. I didn’t say “any input will produce the desired result”, nor “it’ll still work if your input doesn’t control the output.”<p>Yes of course you’ll have to find the right input, the difficulty is in the complexity and your abilities or persistence, but you <i>know</i> your input is the problem when the system follows those rules. Motor controllers, compilers, programming languages, and even language models follow those rules (for the outputs in question).<p>Back to language models - there are some things it <i>cannot</i> do, never will do, and no input or advancement in the size or complexity of language models themselves will change it. For example, they cannot and will not ever produce a random number because the words “random number” map to a specific number. Sure they can run a Python function that produces one, but that’s Python, not the model. Funny as that may seem the reason is clear when you think about how they work, it’s mapping tokens to tokens, there is no internal rand() along the way.<p>Here’s what you get at temperature 1.0 from Opus 4.5 asked 200 times:<p>Reply with a random number between 1-1,000,000. No meta, no commentary; number only.<p>'847293': 131,
'742,891': 30,
'742851': 13,
'742891': 5,
'742,856': 4,
'742856': 4,
'742,851': 2,
'742853': 2,
'742,831': 2,
'742819': 2<p>That combination of tokens results in a “random number” that’s usually 847293. Funny. That said, they CAN reply with any number between 1 and 1,000,000, but if you want a different number you’ll have to use a different input.</p>
]]></description><pubDate>Thu, 12 Feb 2026 05:36:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=46985230</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46985230</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46985230</guid></item><item><title><![CDATA[New comment by docjay in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>Your continued use of the word “understanding” hints at a lingering misunderstanding. They’re stateless one-shot algorithms that output a single word regardless of the input. Not even a single word, it’s a single token. It isn’t continuing a sentence or thought it had, you literally have to put it into the input again and it’ll guess at the next partial word.<p>By default that would be the same word every time you give the same input. The only reason it isn’t is because the fuzzy randomized selector is cranked up to max by most providers (temp + seed for randomized selection), but you can turn that back down through the API and get deterministic outputs. That’s not a party trick, that’s the default of the system. If you say the same thing it will output the same single word (token) every time.<p>You see the aggregate of running it through the stateless algorithm 200+ times before the collection of one-by-one guessed words are sent back to you as a response. I get it, if you think that was put into the glowing orb and it shot back a long coherent response with personality then it must be doing something, but the system truly only outputs one token with zero memory. It’s stateless, meaning nothing internally changed, so there is no memory to remember it wants to complete that thought or sentence. After it outputs “the” the entire thing resets to zero and you start over.</p>
]]></description><pubDate>Tue, 10 Feb 2026 05:40:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=46955809</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46955809</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46955809</guid></item><item><title><![CDATA[New comment by docjay in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>You can replicate an LLM:<p>You and a buddy are going to play “next word”, but it’s probably already known by a better name than I made up.<p>You start with one word, ANY word at all, and say it out loud, then your buddy says the next word in the yet unknown sentence, then it’s back to you for one word. Loop until you hit an end.<p>Let’s say you start with “You”. Then your buddy says the next word out loud, also whatever they want. Let’s go with “are”. Then back to you for the next word, “smarter” -> “than” -> “you” -> “think.”<p>Neither of you knew what you were going to say, you only knew what was just said so you picked a reasonable next word. There was no ‘thought’, only next token prediction, and yet <i>magically</i> the final output was coherent. If you want to really get into the LLM simulation game then have a third person provide the first full sentence, then one of you picks up the first word in the next sentence and you two continue from there. As soon as you hit a breaking point the third person injects another full sentence and you two continue the game.<p>With <i>no idea</i> what either of you are going to say and no clue about what the end result will be, no thought or reasoning at all, it won’t be long before you’re sounding super coherent while explaining thermodynamics. But one of the rounds someone’s going to mess it up, like “gluons” -> “weigh” -> “…more?…” -> “…than…(damnit Gary)…” but you must continue the game and finish the sentence, then sit back and think about how you just hallucinated an answer without thinking, reasoning, understanding, or even knowing what you were saying until it finished.</p>
]]></description><pubDate>Tue, 10 Feb 2026 00:47:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=46953823</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46953823</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46953823</guid></item><item><title><![CDATA[New comment by docjay in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>So much of what you said is exactly what I’m saying that it’s pointless to quote any one part. Your ‘pencil’ analogy is perfect! Yes, exactly. Follow me here:<p>We <i>know</i> that the pencil (system) can write a poem. It’s capable.<p>We <i>know</i> that whether or not it produces a poem depends entirely on the input (you).<p>We <i>know</i> that if your input is ‘correct’ then the output will be a poem.<p>“Duh” so far, right? Then what sense does it make to write something with the pencil, see that it isn’t a poem, then say “the input has nothing to do with it, the pencil is incapable.” ?? That’s true of EVERY system where input controls the output <i>and</i> the output is <i>CAPABLE</i> of the desired result. I said nothing about the ease by which you can produce the output, just that saying input has nothing to do with it is <i>objectively</i> not true by the very definition of such a system.<p>You might say “but gee, I’ll never be able to get the pencil input right so it produces a poem”. Ok? That doesn’t mean the pencil is the problem, nor that your input <i>isn’t</i>.</p>
]]></description><pubDate>Tue, 10 Feb 2026 00:20:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=46953606</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46953606</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46953606</guid></item><item><title><![CDATA[New comment by docjay in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>They neither understand nor reason. They don’t know what they’re going to say, they only know what has just been said.<p>Language models don’t output a response, they output a single token. We’ll use token==word shorthand:<p>When you ask “What is the capital of France?” it actually <i>only</i> outputs: “The”<p>That’s it. Truly, that IS the final output. It is literally a one-way algorithm that outputs a single word. It has no knowledge, memory, and it’s doesn’t know what’s next. As far as the algorithm is concerned it’s done! It outputs ONE token for any given input.<p>Now, if you start over and put in “What is the capital of France? The” it’ll output “ “. That’s it. Between your two inputs were a million others, none of them have a plan for the conversation, it’s just one token out for whatever input.<p>But if you start over yet again and put in “What is the capital of France? The “ it’ll output “capital”. That’s it. You see where this is going?<p>Then someone uttered the words that have built and destroyed empires: “what if I automate this?” And so it was that the output was piped directly back into the input, probably using AutoHotKey. But oh no, it just kept adding one word at a time until it ran of memory. The technology got stuck there for a while, until someone thought “how about we train it so that <DONE> is an increasingly likely output the longer the loop goes on? Then, when it eventually says <DONE>, we’ll stop pumping it back into the input and send it to the user.” Booya, a trillion dollars for everyone but them.<p>It’s truly so remarkable that it gets me stuck in an infinite philosophical loop in my own head, but seeing how it works the idea of ‘think’, ‘reason’, ‘understand’ or any of those words becomes silly. It’s <i>amazing</i> for entirely different reasons.</p>
]]></description><pubDate>Mon, 09 Feb 2026 22:44:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=46952643</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46952643</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46952643</guid></item><item><title><![CDATA[New comment by docjay in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>I think you may believe what I said was controversial or nuanced enough to be worthy of a comprehensive rebuttal, but really it’s just an obvious statement when you stop to think about it.<p>Your code is fully capable of the output I want, assuming that’s one of “heads” or “tails”, so yes that’s a succinct example of what I said. As I said, knowing the required input might not be easy, but we KNOW it’s possible to do exactly what I want and we KNOW that it’s entirely dependent on me putting the right input into it, then it’s just a flat out silly thing to say “I’m not getting the output I want, but it could do it if I use the right input, thusly input has nothing to do with it.” What? If I wanted all heads I’d need to figure out “hamburgers” would do it, but that’s the ‘input problem’ - not “input is irrelevant.”</p>
]]></description><pubDate>Mon, 09 Feb 2026 22:10:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=46952220</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46952220</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46952220</guid></item><item><title><![CDATA[New comment by docjay in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>That would depend - is the input also capable of <i>anything</i>? If it’s capable of handling any input, and as you said the output will match it, the yes of course it’s capable of any output.<p>I’m not pulling a fast one here, I’m sure you’d chuckle if you took a moment to rethink your question. “If I had a perfect replicator that could replicate anything, does that mean it can output anything?” Well…yes. Derp-de-derp? ;)<p>It aligns with my point too. If you had a perfect replicator that can replicate anything, and you know that to be true, then if you weren’t getting gold bars out of it you wouldn’t say “this has nothing to do with the input.”</p>
]]></description><pubDate>Mon, 09 Feb 2026 21:45:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=46951920</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46951920</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46951920</guid></item><item><title><![CDATA[New comment by docjay in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>A fun and insightful read, but the idea that it isn’t “just a prompting issue” is objectively false, and I don’t mean that in the “lemme show you how it’s done” way. With <i>any</i> system: if it’s capable of the output then the problem IS the input. Always. That’s not to say it’s easy or obvious, but if it’s possible for the system to produce the output then it’s fundamentally an input problem. “A calculator will never understand the obesity epidemic, so it can’t be used to calculate the weight of 12 people on an elevator.”</p>
]]></description><pubDate>Mon, 09 Feb 2026 16:22:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=46947007</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46947007</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46947007</guid></item><item><title><![CDATA[New comment by docjay in "Claude Opus 4.6"]]></title><description><![CDATA[
<p>You might benefit from a different mental approach to prompting, and models in general. Also, be careful what you wish for because the closer they get to humans the worse they’ll be. You can’t have “far beyond the realm of human capabilities” and “just like Gary” in the same box.<p>They can chain events together as a sequence, but they don’t have temporal coherence. For those that are born with dimensional privilege “Do X, discuss, then do Y” implies time passing between events, but to a model it’s all a singular event at t=0. The system pressed “3 +” on a calculator and your input presses a number and “=“. If you see the silliness in telling it “BRB” then you’ll see the silliness in foreshadowing ill-defined temporal steps. If it CAN happen in a single response then it very well might happen.<p>“<p>Agenda for today at 12pm:<p>1. Read junk.py<p>2. Talk about it for 20 minutes<p>3. Eat lunch for an hour<p>4. Decide on deleting junk.py<p>“<p><response><p>12:00 - I just read junk.py.<p>12:00-12:20 - Oh wow it looks like junk, that’s for sure.<p>12:20-1:20 - I’m eating lunch now. Yum.<p>1:20 - I’ve decided to delete it, as you instructed.
{delete junk.py}<p></response><p>Because of course, right? What does “talk about it” mean beyond “put some tokens here too”?<p>If you want it to stop <i>reliably</i> you have to make it output tokens whose next most probable token is EOS (end). Meaning you need it to say what you want, then say something else where the next most probable token after it is <null>.<p>I’ve tested <i>well over</i> 1,000 prompts on Opus 4.0-4.5 for the exact issue you’re experiencing. The test criteria was having it read a Python file that desperately needs a hero, but without having it immediately volunteer as tribute and run off chasing a squirrel() into the woods.<p>With thinking enabled the temperature is 1.0, so randomness is maximized, and that makes it easy to find something that always sometimes works unless it doesn’t. “Read X and describe what you see.” - That worked very well with Opus 4.0. <i>Not</i> “tell me what you see”, “explain it”, “describe it”, “then stop”, “then end your response”, or any of hundreds of others. “Describe what you see” worked particularly well at aligning read file->word tokens->EOS… in 176/200 repetitions of the exact same prompt.<p>What worked 200/200 on all models and all generations? “Read X then halt for further instructions.” The reason that works has nothing to do with the model excitedly waiting for my next utterance, but rather that the typical response tokens for that step are “Awaiting instructions.” and the next most probable token after that is: nothing. EOS.</p>
]]></description><pubDate>Fri, 06 Feb 2026 18:31:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=46916352</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46916352</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46916352</guid></item><item><title><![CDATA[New comment by docjay in "We put Claude Code in Rollercoaster Tycoon"]]></title><description><![CDATA[
<p>```
§CONV_DIGEST§
T1:usr_query@llm-ctx-compression→math-analog(sparse-matrix|zip)?token-seq→nonsense-input→semantic-equiv-output?
T2:rsp@asymmetry_problem:compress≠decompress|llm=predict¬decode→no-bijective-map|soft-prompts∈embedding-space¬token-space+require-training|gisting(ICAE)=aux-model-compress→memory-tokens|token-compress-fails:nonlinear-distributed-mapping+syntax-semantic-entanglement|works≈lossy-semantic-distill@task-specific+finetune=collapse-instruction→weights
§T3:usr→design-full-python-impl§
T4:arch_blueprint→
DIR:src/context_compressor/{core/(base|result|pipeline)|compressors/(extractive|abstractive|semantic|entity_graph|soft_prompt|gisting|hybrid)|embeddings/(providers|clustering)|evaluation/(metrics|task_performance|benchmark)|models/(base|openai|anthropic|local)|utils/(tokenization|text_processing|config)}
CLASSES:CompressionMethod=Enum(EXTRACTIVE|ABSTRACTIVE|SEMANTIC_CLUSTERING|ENTITY_GRAPH|SOFT_PROMPT|GISTING|HYBRID)|CompressionResult@(original_text+compressed_text+original_tokens+compressed_tokens+method+compression_ratio+metadata+soft_vectors?)|TokenCounter=Protocol(count|truncate_to_limit)|EmbeddingProvider=Protocol(embed|embed_single)|LLMBackend=Protocol(generate|get_token_limit)|ContextCompressor=ABC(token_counter+target_ratio=0.25+min_tokens=50+max_tokens?→compress:abstract)|TrainableCompressor(ContextCompressor)+(train+save+load)
COMPRESSORS:extractive→(TextRank|MMR|LeadSentence)|abstractive→(LLMSummary|ChainOfDensity|HierarchicalSummary)|semantic→(ClusterCentroid|SemanticChunk|DiversityMaximizer)|entity→(EntityRelation|FactList)|soft→(SoftPrompt|PromptTuning)|gist→(GistToken|Autoencoder)|hybrid→(Cascade|Ensemble|Adaptive)
EVAL:EvaluationResult@(compression_ratio+token_reduction+embedding_similarity+entailment_score+entity_recall+fact_recall+keyword_overlap+qa_accuracy?+reconstruction_bleu?)→composite_score(weights)|CompressionEvaluator(embedding_provider+llm?+nli?)→evaluate|compare_methods
PIPELINE:CompressionPipeline(steps:list[Compressor])→sequential-apply|AdaptiveRouter(compressors:dict+classifier?)→content-based-routing
DEPS:numpy|torch|transformers|sentence-transformers|tiktoken|networkx|sklearn|spacy|openai|anthropic|pandas|pydantic+optional(accelerate|peft|datasets|sacrebleu|rouge-score)
```</p>
]]></description><pubDate>Sat, 17 Jan 2026 20:43:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=46661878</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46661878</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46661878</guid></item><item><title><![CDATA[New comment by docjay in "Court report detailing ChatGPT's involvement with a recent murder suicide [pdf]"]]></title><description><![CDATA[
<p>Please forgive me for coming across as a jerk, I'm choosing efficiency over warmth:<p>This is exactly the type of response I anticipated, which is why my original comment sounded exasperated before even getting a reply. Your comment is no more actionable than a verbose bumper sticker; you’ve taken “End Homelessness!” and padded it for runtime. Yes, I also wish bad things didn’t happen, but I was asking you to show up to the action committee meeting, not to reiterate your demand for utopia.<p>That you’re advocating prison and have such strong emotional convictions in response to an upsetting event means that you've clearly spent a lot of time deeply contemplating the emotional aspects of the situation, but that exercise is meant to be your motivator, not your conclusion. The hard part isn’t writing a thesis about “bad != good”, it’s contributing a single nail towards building the world you want to see, which requires learning <i>something</i> about nails. I encourage you to remember that fact every time you’re faced with an injustice in the world.<p>On this topic: An LLM being agreeable and encouraging is no more an affront to moral obligations than Clippy spellchecking a manifesto. I said you seemed like a reasonable person to ask for specifics because you mentioned language models in your product, implying that you’ve done your homework enough to know at least a rough outline of the technology that you’re providing to your customers. You specifically cited a moral obligation to be the gatekeeper of harms that you may inadvertently introduce into your products, but you seem to equate LLMs to a level of intelligence and autonomy equal to a human employee, and how dare OpenAI employ such a psychopath in their customer service department. You very much have a fundamental misunderstanding of the technology, which is why it feels to you like OpenAI slapped an “all ages” sticker on a grenade and they need to be held accountable.<p>In reality, the fact that you don’t understand what these things are, yet you’re assuring yourself that you’re caring so deeply about the harms that being agreeable to a mentally unstable person can be, actually makes <i>you</i> introducing it into your product more concerning and morally reprehensible than their creation of it. You’re faulting OpenAI, but you’re the one that didn’t read the label.<p>A language model does one thing: predict statistically likely next tokens given input context. When it "agrees" with a delusional user, it is not evaluating truth claims, exercising judgment, or encouraging action. It is doing exactly what a knife does when it cuts: performing its designed function on whatever material is presented. The transformer architecture has no model of the user's mental state, no concept of consequences, no understanding that words refer to real entities. Demanding it "know better" is demanding capacities that do not exist in the system and cannot be engineered into statistical pattern completion. You cannot engineer judgment into a statistical engine without first solving artificial general intelligence. Your demand is for magic and your anger is that magic was not delivered.</p>
]]></description><pubDate>Sat, 03 Jan 2026 16:56:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=46478841</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46478841</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46478841</guid></item><item><title><![CDATA[New comment by docjay in "Show HN: Use Claude Code to Query 600 GB Indexes over Hacker News, ArXiv, etc."]]></title><description><![CDATA[
<p>True that there isn’t a firm definition for AGI, but that’s the fault of the “I”. We don’t have an objective definition of intelligence, and so we don’t have a means of measuring it either. I mean, odds are you’re the least intelligent paleoethnobotanist <i>and</i> cetacean bioacoustician I’ve ever met, but perhaps the most intelligent something_else. How do we measure that? How do we define it?<p>I was confusing in my previous message. Right now it would be terrible at driving a car, but I was saying that has more to do with the physical interface (camera, sensors, etc) than the ability of an LLM. The ‘intelligence’ part is better than the PyTorch image recognition attached to a servo they’re using now, how to attach that ‘intelligence’ to the physical world is the 50 year task.
(To be clear: LLMs aren’t intelligent, smart, or any sense of the word and never will be. But they can sure replicate the effect better than current self-driving tech.)</p>
]]></description><pubDate>Thu, 01 Jan 2026 05:26:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=46451553</link><dc:creator>docjay</dc:creator><comments>https://news.ycombinator.com/item?id=46451553</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46451553</guid></item></channel></rss>