<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: scottmu</title><link>https://news.ycombinator.com/user?id=scottmu</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sat, 04 Apr 2026 09:13:35 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=scottmu" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by scottmu in "Show HN: Sup AI, a confidence-weighted ensemble (52.15% on Humanity's Last Exam)"]]></title><description><![CDATA[
<p>If 2 (or more) tokens are synonymous with each other with high probabilities (49.9% each for a total of 99.8%), that's still low entropy. Not as low as a singular high-probability token, but low enough for us to consider this a low-entropy token distribution.<p>You can't look at a single token distribution, though. There are many legitimate high-confidence, high-accuracy cases in which many tokens could come next. For example, the first token of a paragraph. You need to look at pools of entropies over segments of the output or the whole output sequence.<p>Although there's a correlation between uncertainty and hallucinations or inaccuracies, there's no guarantee. This is a challenging area that we're monitoring the latest literature for and contributing where we can.</p>
]]></description><pubDate>Fri, 27 Mar 2026 23:00:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47549494</link><dc:creator>scottmu</dc:creator><comments>https://news.ycombinator.com/item?id=47549494</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47549494</guid></item><item><title><![CDATA[New comment by scottmu in "Show HN: Sup AI, a confidence-weighted ensemble (52.15% on Humanity's Last Exam)"]]></title><description><![CDATA[
<p>I like the direction you're going with this strategy. There are many approaches, nuances, edge cases, and clever tricks to each of these steps, even without taking into account token probability distributions. Very powerful to get it right.</p>
]]></description><pubDate>Fri, 27 Mar 2026 22:20:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=47549144</link><dc:creator>scottmu</dc:creator><comments>https://news.ycombinator.com/item?id=47549144</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47549144</guid></item><item><title><![CDATA[New comment by scottmu in "Show HN: Sup AI, a confidence-weighted ensemble (52.15% on Humanity's Last Exam)"]]></title><description><![CDATA[
<p>Yes, there's a wide variety of use cases that require different ratios of accuracy/speed. If you require 3 responses to be accurate, you have to multiply all 3 response accuracy probabilities, and as you've shown, this can reduce overall accuracy quite a bit. Of course, this does make the assumption that those 3 responses are independent of one another.</p>
]]></description><pubDate>Fri, 27 Mar 2026 20:08:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=47547571</link><dc:creator>scottmu</dc:creator><comments>https://news.ycombinator.com/item?id=47547571</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47547571</guid></item><item><title><![CDATA[New comment by scottmu in "Show HN: Sup AI, a confidence-weighted ensemble (52.15% on Humanity's Last Exam)"]]></title><description><![CDATA[
<p>Great question. What I can say is we experimented a _ton_. If you take a basic approach and simply ask the same prompt of a bunch of LLMs and then ask another LLM to combine the results, you'll get a pretty poor answer. At best, you'll get a response that is the average of the ensemble, which by definition is going to be worse than the best model of the ensemble. Of course, you're going to want a mechanism to choose the ensemble effectively. At worst, you'll regurgitate the worst model of the ensemble. And you'll have the added expense and potential latency, too. Not a good solution at all.<p>We didn't experiment with different ensemble mechanisms rigorously enough for a research paper. We will, though.<p>Majority voting was actually how we started, and we came up with great mechanisms for stopping early, saving token costs and time, along with other interesting things we could do with that simple mechanism. The issue we had was that the orchestration could already choose a model beforehand almost as good (according to simpler benchmarks than HLE we ran at the time) as majority voting could pick after the responses were complete. And we tried many voting mechanisms, such as all models in the ensemble voting on all others.<p>An ablation study would be great to do now, with many other ideas we've played with. We have better benchmarks than we did just a few months ago, and it would be great to understand the tradeoffs of different approaches so that there could be alternative options for different use cases.</p>
]]></description><pubDate>Fri, 27 Mar 2026 20:01:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=47547502</link><dc:creator>scottmu</dc:creator><comments>https://news.ycombinator.com/item?id=47547502</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47547502</guid></item><item><title><![CDATA[New comment by scottmu in "Show HN: Sup AI, a confidence-weighted ensemble (52.15% on Humanity's Last Exam)"]]></title><description><![CDATA[
<p>You're right! I could've phrased my comment better. Ken actually wanted to edit his post, but it was too late. So he asked me to write a response explaining what he meant. Of course, he could've commented too. I was just trying to be helpful to him and others wanting an explanation.</p>
]]></description><pubDate>Fri, 27 Mar 2026 19:43:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=47547296</link><dc:creator>scottmu</dc:creator><comments>https://news.ycombinator.com/item?id=47547296</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47547296</guid></item><item><title><![CDATA[New comment by scottmu in "Show HN: Sup AI, a confidence-weighted ensemble (52.15% on Humanity's Last Exam)"]]></title><description><![CDATA[
<p>I wouldn't say it's easy to detect hallucinations. Understanding output token probability distributions is only part of a solution, and we still aren't perfect. Just better than individual models.<p>Hallucinations may seem rarer for a few reasons. First, models are more accurate with certain prompts. Second, models are more convincing when they do hallucinate. They may get an overall idea, but hallucinate the details. Hallucinations are still a major problem and are fundamental to the way modern LLMs work.</p>
]]></description><pubDate>Fri, 27 Mar 2026 19:41:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47547275</link><dc:creator>scottmu</dc:creator><comments>https://news.ycombinator.com/item?id=47547275</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47547275</guid></item><item><title><![CDATA[New comment by scottmu in "Show HN: Sup AI, a confidence-weighted ensemble (52.15% on Humanity's Last Exam)"]]></title><description><![CDATA[
<p>I've felt your pain. Models aren't always trained well enough on edge cases and configs.<p>Would love to hear how Sup works out for you.</p>
]]></description><pubDate>Fri, 27 Mar 2026 06:36:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=47539660</link><dc:creator>scottmu</dc:creator><comments>https://news.ycombinator.com/item?id=47539660</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47539660</guid></item><item><title><![CDATA[New comment by scottmu in "Show HN: Sup AI, a confidence-weighted ensemble (52.15% on Humanity's Last Exam)"]]></title><description><![CDATA[
<p>I want to clarify what Ken meant by "entropy in the output token probability distributions." Whenever an LLM outputs a token, it's choosing that token out of all possible tokens. Every possible output token has a probability assigned by the model (typically a logarithm of the probability). This is a probability distribution (the output token probabilities sum to 1). Entropy is a measure of uncertainty and can quantify if a token probability distribution is certain (1 token has a 99.9% probability, and the rest share the leftover 0.1% probability) or uncertain (every token has roughly the same probability, so it's pretty much random which token is selected). Low entropy is the former case, and high entropy is the latter.<p>There is interesting research in the correlation of entropy with accuracy and hallucinations:<p>- <a href="https://www.nature.com/articles/s41586-024-07421-0" rel="nofollow">https://www.nature.com/articles/s41586-024-07421-0</a><p>- <a href="https://arxiv.org/abs/2405.19648" rel="nofollow">https://arxiv.org/abs/2405.19648</a><p>- <a href="https://arxiv.org/abs/2509.04492" rel="nofollow">https://arxiv.org/abs/2509.04492</a> (when only a small number of probabilities are available, which is something we frequently deal with)<p>- <a href="https://arxiv.org/abs/2603.18940" rel="nofollow">https://arxiv.org/abs/2603.18940</a><p>- tons more, happy to chat about if interested</p>
]]></description><pubDate>Thu, 26 Mar 2026 21:04:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=47535700</link><dc:creator>scottmu</dc:creator><comments>https://news.ycombinator.com/item?id=47535700</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47535700</guid></item></channel></rss>