<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: wbogusz</title><link>https://news.ycombinator.com/user?id=wbogusz</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 13 May 2026 19:36:49 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=wbogusz" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by wbogusz in "Looming Liability Machines (LLMs)"]]></title><description><![CDATA[
<p>> If the verification systems for LLMs are not built out of LLMs and they're somehow more robust than LLMs at human-language problem solving and analysis, then you should be using the technology the verification system uses instead of LLMs in the first place!<p>The issue is not in the verification system, but in putting quantifiable bounds on your answer set. If I ask an LLM to multiply large numbers together I can also very easily verify the generated answer by topping it with a deterministic function.<p>I.e. rather than hoping that an LLM can accurately multiply two 10 digit numbers, I have a much easier (and verified) solution by instead asking it to perform this calculation using python and reading me the output</p>
]]></description><pubDate>Sun, 25 Aug 2024 02:50:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=41343946</link><dc:creator>wbogusz</dc:creator><comments>https://news.ycombinator.com/item?id=41343946</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41343946</guid></item><item><title><![CDATA[New comment by wbogusz in "Looming Liability Machines (LLMs)"]]></title><description><![CDATA[
<p>> Here’s a simple rule, based on the fact no one has shown that an llm or a compound llm system can produce an output that doesn’t need to be verified for correctness by a human across any input:<p>I’m still not sure why some of us are so convinced there isn’t an answer to properly verifying LLM output. In so many circumstances, having output pushed 90-95% of the way is very easily pushed to 100% by topping off with a deterministic system.<p>Do I depend on an LLM to perform 8 digit multiplication? Absolutely not, because like you say, I can’t verify the correctness that would drive the statistics of whatever answer it spits out. But why can’t I ask an LLM to write the python code to perform the same calculation and read me its output?<p>> I think it follows that we should not use llms for anything critical.<p>While we are at it I think we should also institute an IQ threshold for employees to contribute to or operate around critical systems. If we can’t be sure to an absolute degree that they will not make a mistake, then there is no purpose to using them. All of their work will simply need to be double checked and verified anyway.</p>
]]></description><pubDate>Sun, 25 Aug 2024 02:37:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=41343885</link><dc:creator>wbogusz</dc:creator><comments>https://news.ycombinator.com/item?id=41343885</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41343885</guid></item><item><title><![CDATA[4M Context – Llama-3-8B-Instruct]]></title><description><![CDATA[
<p>Article URL: <a href="https://huggingface.co/gradientai/Llama-3-8B-Instruct-Gradient-4194k">https://huggingface.co/gradientai/Llama-3-8B-Instruct-Gradient-4194k</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=40311974">https://news.ycombinator.com/item?id=40311974</a></p>
<p>Points: 4</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 09 May 2024 19:21:51 +0000</pubDate><link>https://huggingface.co/gradientai/Llama-3-8B-Instruct-Gradient-4194k</link><dc:creator>wbogusz</dc:creator><comments>https://news.ycombinator.com/item?id=40311974</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40311974</guid></item><item><title><![CDATA[New comment by wbogusz in "What can LLMs never do?"]]></title><description><![CDATA[
<p>I’m not a fan of the talking parrot argument, especially when you’re pointing it at models of scale.<p>The only thing separating a talking parrot and humans is our accuracy in shaping our words to the context in which they’re spoken.<p>Sure it’s easy to liken a low resource model to a talking parrot, the output seems no better than selective repetition of training data. But is that really so different from a baby whose first words are mimics from the environment around them?<p>I would argue that as we learn language we implicitly develop the neural circuitry to continue to improve our lexical outputs, this circuitry being concepts like foresight, reasoning, emotion, logic, etc and that while we can take explicit action to teach these ideas, they naturally develop in isolation as well.<p>I don’t think language models, especially at scale, are much different. They would seem to similarly acquire implicit circuitry like we do as they are exposed to more data. As I see it, the main difference in what exactly that circuitry accomplishes and looks like in final output has more to do with the limited styles of data we can provide and the limitations of fine tuning we can apply on top.<p>Humans would seem to share a lot in common with talking parrots, we just have a lot more capable hardware to select what we repeat.</p>
]]></description><pubDate>Sat, 27 Apr 2024 17:57:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=40182021</link><dc:creator>wbogusz</dc:creator><comments>https://news.ycombinator.com/item?id=40182021</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40182021</guid></item><item><title><![CDATA[New comment by wbogusz in "Show HN: Reor – An AI note-taking app that runs models locally"]]></title><description><![CDATA[
<p>Great to see something like this actualized. I’m a huge fan of Obsidian and its graph based connections for note taking.<p>Always see parallels drawn between Obsidian note structures and whole “2nd brain” idea for personal knowledge management, had seemed like a natural next step would be to implement note retrieval for intelligent references. Will have to check this out</p>
]]></description><pubDate>Wed, 14 Feb 2024 17:31:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=39372546</link><dc:creator>wbogusz</dc:creator><comments>https://news.ycombinator.com/item?id=39372546</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39372546</guid></item></channel></rss>