<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: thoughtlede</title><link>https://news.ycombinator.com/user?id=thoughtlede</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 03 May 2026 19:25:28 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=thoughtlede" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by thoughtlede in "Claude Pro: Opus model will only be available if extra usage is enabled"]]></title><description><![CDATA[
<p>Investor funds have been subsidizing the inference costs so far.<p>Investors might move from funding the model providers to funding the enterprises that use those models. That is, they might move from funding the cost of the experiment to funding the value of the result. No funding if there are no demonstrable AI gains.<p>This is a reasonable shift if this happens. If enough gains have been demonstrated, then investors might go back to funding the model providers. Investors always move towards the highest leverage point.<p>As long as AI delivers, this would be the rhythm.</p>
]]></description><pubDate>Tue, 28 Apr 2026 00:19:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=47928943</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=47928943</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47928943</guid></item><item><title><![CDATA[New comment by thoughtlede in "Nano Banana 2: Google's latest AI image generation model"]]></title><description><![CDATA[
<p>Strictly speaking, I don't think it is the generation or creation that diminishes their value. it is the consumption.<p>You said it too:<p>> If I see a million fake Tom Cruise videos, then it oversaturates my desire for desire for all Tom Cruise movies.<p>The trick of course is to keep yourself from seeing that content.<p>The other nuance is that as long as real performance remains unique, which so far it is, we can appreciate more what flesh and blood brings to the table. For example, I can appreciate the reality of the people in a picture or a video that is captured by a regular camera; it's AI version lacks that spunk (for now).<p>Note that iPhone in its default settings is already altering the reality, so AI generation is far right on that slippery axis.<p>Perhaps, AI and VR would be the reason why our real hangouts would be more appreciated even if they become rare events in the future.</p>
]]></description><pubDate>Thu, 26 Feb 2026 17:34:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=47169236</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=47169236</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47169236</guid></item><item><title><![CDATA[New comment by thoughtlede in "Why is the sky blue?"]]></title><description><![CDATA[
<p>I think we can simplify the answer to this question for most audience and say "the air is blue".<p>If they say, the air appears to be clear when I stare at something other than sky, the answer is you need more of air to be able to see its blue-ness, in much the same way that a small amount of murky water in your palm appears clear, but a lot of it does not.<p>If they ask, why don't I see that blue-ness at dawn or dusk, the answer is that the light source is at a different angle. The color of most objects changes when the light source is at a flat angle. And sun lights hits at a flat angle at dawn and dusk.<p>If they ask, what exactly is the inside phenomenon to see the sky color to be blue, then explanations like this blog are relevant.<p>If they ask, what exactly is a color, the answer is that it is a fiction made up by our brain.</p>
]]></description><pubDate>Mon, 09 Feb 2026 21:22:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=46951556</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=46951556</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46951556</guid></item><item><title><![CDATA[New comment by thoughtlede in "Collaboration sucks"]]></title><description><![CDATA[
<p>For me there are two things about collaboration.<p>Decision making is one, which you emphasized.<p>The other is knowing what the collaboration brings to the table and shaping the rules of engagement to fit that expectation. Sometimes you collaborate with SMEs; they bring the domain knowledge - you don't, but you understand the goal better than them. Sometimes you are creating or refining the corporate strategy based on the actions from individual projects or partners; you are learning ground realities from them. Sometimes you need help from others to improve your take on a subject.<p>In each of these cases, you have to be clear about what you expect from the collaborators (and motivate them to contribute). Without being clear on what the collaboration is about and what they get in return is the number one killer of collaborative projects even though there is no ill-intent anywhere.</p>
]]></description><pubDate>Tue, 11 Nov 2025 22:23:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=45893710</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=45893710</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45893710</guid></item><item><title><![CDATA[New comment by thoughtlede in "Ask HN: What's your experience with using graph databases for agentic use-cases?"]]></title><description><![CDATA[
<p>It boils down to whether your LLMs can speak graph queries better than SQL, for your use cases and data. As your data posture changes and your use cases change, you routinely reevaluate which DB query language suits best for LLMs.<p>I'd also design the system architecture in such a way that your non-agentic workloads don't suffer if you have to move between query models for serving agentic workloads better.</p>
]]></description><pubDate>Mon, 06 Oct 2025 08:04:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=45488822</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=45488822</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45488822</guid></item><item><title><![CDATA[New comment by thoughtlede in "AWS CEO says using AI to replace junior staff is 'Dumbest thing I've ever heard'"]]></title><description><![CDATA[
<p>memorization + application = comprehension. Rinse and repeat.<p>Whether leet code or anything else.</p>
]]></description><pubDate>Thu, 21 Aug 2025 17:30:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=44975589</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=44975589</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44975589</guid></item><item><title><![CDATA[New comment by thoughtlede in "Reasoning models don't always say what they think"]]></title><description><![CDATA[
<p>It feels to me that the hypothesis of this research was somewhat "begging the question". Reasoning models are trained to spit some tokens out that increase the chance of the models spitting the right answer at the end. That is, the training process is singularly optimizing for the right answer, not the reasoning tokens.<p>Why would you then assume the reasoning tokens will include hints supplied in the prompt "faithfully"? The model may or may not include the hints - depending on whether the model activations believe those hints are necessary to arrive at the answer. In their experiments, they found between 20% and 40% of the time, the models included those hints. Naively, that sounds unsurprising to me.<p>Even in the second experiment when they trained the model to use hints, the optimization was around the answer, not the tokens. I am not surprised the models did not include the hints because they are not trained to include the hints.<p>That said, and in spite of me potentially coming across as an unsurprised-by-the-result reader, it is a good experiment because "now we have some experimental results" to lean into.<p>Kudos to Anthropic for continuing to study these models.</p>
]]></description><pubDate>Thu, 03 Apr 2025 19:43:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=43574438</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=43574438</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43574438</guid></item><item><title><![CDATA[New comment by thoughtlede in "Supercharge vector search with ColBERT rerank in PostgreSQL"]]></title><description><![CDATA[
<p>tadkar did a good job at explaining ColBERT. I understood ColBERT well in the context of where it lies on the spectrum of choices.<p>On one side of the spectrum, you reduce each of the documents as well as the query to a lower-dimensional space (aka embeddings) and perform similarity. This has the advantage that the document embeddings could be precomputed. At query time, you only compute the query embedding and compare its similarity with document embeddings. The problem is that the lower-dimensional embedding acts as a decent, but not great, proxy for the documents as well as for the query. Your query-document similarity is only as good as the semantics that could be captured in those lower-dimensional embeddings.<p>On the other side of the spectrum, you consider the query with each document (as a pair) and see how much the query "attends" to each of the documents. The power of trained attention weights means that you get a much reliable similarity score. The problem is that this approach requires you to run attention-forward-pass as many times as there are documents -- for each query. In other words, this has a performance issue.<p>ColBERT sits in the middle of the spectrum. It "attends" to each of the documents separately and captures the lower-dimensional embedding for each token in each document. This we precompute. Once we have done that, we captured the essence of how tokens within a given document attend to each other, and is captured in the token embeddings.<p>Then, at query time, we do the same for each token in the query. And we see which query-token embedding is greatly similar to which document-token embedding. If we find that there is a document which has more tokens that are found to be greatly similar to the query tokens, then we consider that to the best document match. (The degree of similarity between each query-document token is used to score the ranking - it is called Sum of MaxSim).<p>Obviously, attention based similarity, like in the second approach, is better than reducing to token embeddings and scoring similarity. But ColBERT avoids the performance hit compared to the second approach. ColBERT also avoids the lower fidelity of "reducing the entire document to a lower-dimensional space issue" because it reduces each token in the document separately.<p>By the way, the first approach is what bi-encoders do. The second approach is cross-encoding.</p>
]]></description><pubDate>Fri, 24 Jan 2025 07:53:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=42811326</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=42811326</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42811326</guid></item><item><title><![CDATA[New comment by thoughtlede in "Building Effective "Agents""]]></title><description><![CDATA[
<p>When thinking about AI agents, there is still conflation between how to decide the next step to take vs what information is needed to decide the next step.<p>If runtime information is insufficient, we can use AI/ML models to fill that information. But deciding the next step could be done ahead of time assuming complete information.<p>Most AI agent examples short circuit these two steps. When faced with unstructured or insufficient information, the program asks the LLM/AI model to decide the next step. Instead, we could ask the LLM/AI model to structure/predict necessary information and use pre-defined rules to drive the process.<p>This approach will translate most [1] "Agent" examples into "Workflow" examples. The quotes here are meant to imply Anthropic's definition of these terms.<p>[1] I said "most" because there might be continuous world systems (such as real world simulacrum) that will require a very large number of rules and is probably impractical to define each of them. I believe those systems are an exception, not a rule.</p>
]]></description><pubDate>Fri, 20 Dec 2024 23:19:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=42476104</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=42476104</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42476104</guid></item><item><title><![CDATA[New comment by thoughtlede in "CRDTs and Collaborative Playground"]]></title><description><![CDATA[
<p>> Beyond this, if you want to determine causality, e.g. whether events are "causally related" (happened before or after each other) or are "concurrent" (entirely independent of), you can look at Vector Clocks—I won't go down that rabbit-hole here, though.<p>If anyone want to go down that rabbit hole: <a href="https://www.exhypothesi.com/clocks-and-causality/" rel="nofollow">https://www.exhypothesi.com/clocks-and-causality/</a></p>
]]></description><pubDate>Wed, 18 Dec 2024 07:53:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=42448791</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=42448791</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42448791</guid></item><item><title><![CDATA[New comment by thoughtlede in "Training LLMs to Reason in a Continuous Latent Space"]]></title><description><![CDATA[
<p>Perhaps these findings might be indicating that we need more NN layers/attention blocks for performing reasoning. This project circumvented the lack of more trained layers by looping the input through currently trained layers more than once.<p>Also we may have to look for better loss functions than ones that help us predict the next token to train the models if the objective is reasoning.</p>
]]></description><pubDate>Wed, 11 Dec 2024 02:40:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=42384164</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=42384164</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42384164</guid></item><item><title><![CDATA[New comment by thoughtlede in "Model Context Protocol"]]></title><description><![CDATA[
<p>If function calling is sync, is MCP its async counterpart? Is that the gist of what MCP is?<p>Open API (aka swagger) based function calling is standard already for sync calls, and it solves the NxM problem. I'm wondering if the proposed value is that MCP is async.</p>
]]></description><pubDate>Tue, 26 Nov 2024 10:58:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=42244583</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=42244583</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42244583</guid></item><item><title><![CDATA[New comment by thoughtlede in "My Python code is a neural network"]]></title><description><![CDATA[
<p>I think the mention of 'spaghetti code' is a red herring from the author. If the output from an algorithm cannot be defined precisely as a function of the input, but you have some examples to show, that's where machine learning (ML) is useful.<p>In the end, ML provides one more option to choose from. Whether it works or not for you depends on evaluations and how deterministic and explainability you need from the chosen algorithm/option.<p>The thing that struck me is if RNN is the right choice given that it would need to be trained and we need a lot of examples than what we might have. That said, maybe based on known 'rules', we can produce synthetic data for both +ve and -ve cases.</p>
]]></description><pubDate>Mon, 01 Jul 2024 17:01:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=40847760</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=40847760</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40847760</guid></item><item><title><![CDATA[Classifier to detect DALL·E 3 images]]></title><description><![CDATA[
<p>Article URL: <a href="https://openai.com/index/understanding-the-source-of-what-we-see-and-hear-online/">https://openai.com/index/understanding-the-source-of-what-we-see-and-hear-online/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=40304449">https://news.ycombinator.com/item?id=40304449</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 09 May 2024 01:38:43 +0000</pubDate><link>https://openai.com/index/understanding-the-source-of-what-we-see-and-hear-online/</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=40304449</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40304449</guid></item><item><title><![CDATA[New comment by thoughtlede in "Deliberative Consensus Protocols"]]></title><description><![CDATA[
<p>Interesting. I didn’t know this body of work. I haven’t read the documentation other than the abstract.<p>If the protocol is about knowing what information to inject at what node in the network to achieve consensus, the protocol can (and will) be used to inject whatever truths the parties with knowledge on the protocol believe in. If there are enough parties with opposing beliefs, then the network cannot be gamed beyond the status quo.<p>How does this protocol “choose” the truth to propagate in the face of opposing truths?<p>What power does the protocol give to parties who know how to work the system?<p>Or rather what resilience does the protocol have against “inside traders”?</p>
]]></description><pubDate>Mon, 15 Apr 2024 16:37:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=40042792</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=40042792</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40042792</guid></item><item><title><![CDATA[New comment by thoughtlede in "AI and the Problem of Knowledge Collapse"]]></title><description><![CDATA[
<p>That's interesting.<p>In keyword-based indexing solutions, a document vector is created using "term frequency inverse document frequency" scores. The idea is to pump up the document on the dimension where the document is unique compared to the other documents in the corpus. So when a query is issued with emphasis on a certain dimension, only documents that has higher scores in that dimension are returned.<p>But the uniqueness in those solutions is based on keywords being used in the document, not concepts.<p>What we need here to eliminate "blandness" is conceptual uniqueness. Maybe TF-IDF is still relevant to get there. Something to think about.</p>
]]></description><pubDate>Fri, 05 Apr 2024 22:13:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=39947899</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=39947899</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39947899</guid></item><item><title><![CDATA[New comment by thoughtlede in "AI and the Problem of Knowledge Collapse"]]></title><description><![CDATA[
<p>LLMs are both language processing engines and knowledge bases. This article explores the knowledge base aspect of LLM and sheds light on the potential danger. The authors are well-justified in doing so because ChatGPT as a knowledge-bot is being used by many end users for its knowledge.<p>However, to my knowledge, many enterprise applications that are being built using LLMs feed task-specific curated knowledge to LLMs. This mode of LLM use is encouraging. I do not think this article acknowledged this aspect of LLM use.</p>
]]></description><pubDate>Fri, 05 Apr 2024 21:07:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=39947229</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=39947229</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39947229</guid></item><item><title><![CDATA[New comment by thoughtlede in "SceneScript, a novel approach for 3D scene reconstruction"]]></title><description><![CDATA[
<p>Yes, the camera sees everything. But we could avoid teaching the model certain things. For instance, an undressed body and a dressed body could both be taught as a body. Likewise, medicine pills as well as regular mints could just be taught as mints.<p>While this approach doesn’t address  privacy completely, it avoids certain elements of our life that are considered really private.<p>I think half of the point of using the simulated dataset to train the model was to safeguard the above mentioned kind of privacy (with the other half being lack of real-world dataset).</p>
]]></description><pubDate>Sat, 23 Mar 2024 18:03:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=39801860</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=39801860</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39801860</guid></item><item><title><![CDATA[New comment by thoughtlede in "Launch HN: DryMerge (YC W24) – Automate Workflows with Plain English"]]></title><description><![CDATA[
<p>Yeah. Rollbacks or reruns are hard when dealing with external systems. Actions need to be idempotent for reruns to work.<p>One thing you may focus on is making workflows more durable: Checkpointing and sending to users summaries of last checkpoints when things fail.<p>The last thing you want a non-tech user (your target customer) is to figure out what’s the state of a failed workflow.</p>
]]></description><pubDate>Fri, 22 Mar 2024 17:52:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=39793088</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=39793088</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39793088</guid></item><item><title><![CDATA[New comment by thoughtlede in "Launch HN: DryMerge (YC W24) – Automate Workflows with Plain English"]]></title><description><![CDATA[
<p>Thanks.<p>What happens right now when the workflow fails mid-way? Do you ensure atomicity or durable execution?</p>
]]></description><pubDate>Fri, 22 Mar 2024 17:35:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=39792956</link><dc:creator>thoughtlede</dc:creator><comments>https://news.ycombinator.com/item?id=39792956</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39792956</guid></item></channel></rss>