<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: trq_</title><link>https://news.ycombinator.com/user?id=trq_</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 07 Apr 2026 08:07:36 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=trq_" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by trq_ in "OpenCode – Open source AI coding agent"]]></title><description><![CDATA[
<p>Claude Code is not an electron app.</p>
]]></description><pubDate>Sat, 21 Mar 2026 14:43:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47467512</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=47467512</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47467512</guid></item><item><title><![CDATA[New comment by trq_ in "Claude Code daily benchmarks for degradation tracking"]]></title><description><![CDATA[
<p>Yes, we do but harnesses are hard to eval, people use them across a huge variety of tasks and sometimes different behaviors tradeoff against each other. We have added some evals to catch this one in particular.</p>
]]></description><pubDate>Fri, 30 Jan 2026 01:30:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=46819524</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=46819524</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46819524</guid></item><item><title><![CDATA[New comment by trq_ in "Claude Code daily benchmarks for degradation tracking"]]></title><description><![CDATA[
<p>Hi everyone, Thariq from the Claude Code team here.<p>Thanks for reporting this. We fixed a Claude Code harness issue that was introduced on 1/26. This was rolled back on 1/28 as soon as we found it.<p>Run `claude update` to make sure you're on the latest version.</p>
]]></description><pubDate>Thu, 29 Jan 2026 19:12:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=46815013</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=46815013</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46815013</guid></item><item><title><![CDATA[New comment by trq_ in "Claude Code gets native LSP support"]]></title><description><![CDATA[
<p>Hi, work on Claude Code here! Let me know if you have any feedback!</p>
]]></description><pubDate>Tue, 23 Dec 2025 07:21:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=46363176</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=46363176</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46363176</guid></item><item><title><![CDATA[New comment by trq_ in "Claude Is Down"]]></title><description><![CDATA[
<p>We're back up! It was about ~30 minutes of downtime this morning, our apologies if it interrupted your work.</p>
]]></description><pubDate>Fri, 07 Nov 2025 17:24:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=45848677</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=45848677</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45848677</guid></item><item><title><![CDATA[Show HN: Write Stories by Steering a LLM]]></title><description><![CDATA[
<p>Article URL: <a href="https://latentlit.goodfire.ai/">https://latentlit.goodfire.ai/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43355208">https://news.ycombinator.com/item?id=43355208</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 13 Mar 2025 17:03:24 +0000</pubDate><link>https://latentlit.goodfire.ai/</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=43355208</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43355208</guid></item><item><title><![CDATA[LLM-Powered Sorting with TrueSkill]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.thariq.io/blog/sorting/">https://www.thariq.io/blog/sorting/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43016272">https://news.ycombinator.com/item?id=43016272</a></p>
<p>Points: 5</p>
<p># Comments: 1</p>
]]></description><pubDate>Tue, 11 Feb 2025 18:24:30 +0000</pubDate><link>https://www.thariq.io/blog/sorting/</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=43016272</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43016272</guid></item><item><title><![CDATA[Show HN: Opensourcing Sparse Autoencoders for Llama 3.3 70B]]></title><description><![CDATA[
<p>Article URL: <a href="https://huggingface.co/Goodfire/Llama-3.3-70B-Instruct-SAE-l50">https://huggingface.co/Goodfire/Llama-3.3-70B-Instruct-SAE-l50</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42658491">https://news.ycombinator.com/item?id=42658491</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 10 Jan 2025 18:32:43 +0000</pubDate><link>https://huggingface.co/Goodfire/Llama-3.3-70B-Instruct-SAE-l50</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=42658491</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42658491</guid></item><item><title><![CDATA[New comment by trq_ in "Show HN: Llama 3.3 70B Sparse Autoencoders with API access"]]></title><description><![CDATA[
<p>Hmm the hallucination would happen in the auto labelling, but we review and test our labels and they seem correct!</p>
]]></description><pubDate>Tue, 24 Dec 2024 02:34:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=42499262</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=42499262</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42499262</guid></item><item><title><![CDATA[New comment by trq_ in "Show HN: Llama 3.3 70B Sparse Autoencoders with API access"]]></title><description><![CDATA[
<p>If you're hacking on this and have questions, please join us on Discord: <a href="https://discord.gg/vhT9Chrt" rel="nofollow">https://discord.gg/vhT9Chrt</a></p>
]]></description><pubDate>Mon, 23 Dec 2024 19:00:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=42496745</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=42496745</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42496745</guid></item><item><title><![CDATA[New comment by trq_ in "Show HN: Llama 3.3 70B Sparse Autoencoders with API access"]]></title><description><![CDATA[
<p>We haven't yet found generalizable "make this model smarter" features, but there is a tradeoff of putting instructions in system prompts, e.g. if you have a chatbot that sometimes generates code, you can give it very specific instructions when it's coding and leave those out of the system prompt otherwise.<p>We have a notebook about that here: <a href="https://docs.goodfire.ai/notebooks/dynamicprompts" rel="nofollow">https://docs.goodfire.ai/notebooks/dynamicprompts</a></p>
]]></description><pubDate>Mon, 23 Dec 2024 18:58:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=42496731</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=42496731</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42496731</guid></item><item><title><![CDATA[Show HN: Llama 3.3 70B Sparse Autoencoders with API access]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.goodfire.ai/papers/mapping-latent-spaces-llama/">https://www.goodfire.ai/papers/mapping-latent-spaces-llama/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42495936">https://news.ycombinator.com/item?id=42495936</a></p>
<p>Points: 201</p>
<p># Comments: 51</p>
]]></description><pubDate>Mon, 23 Dec 2024 17:18:17 +0000</pubDate><link>https://www.goodfire.ai/papers/mapping-latent-spaces-llama/</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=42495936</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42495936</guid></item><item><title><![CDATA[Should Developers care about AI Interpretability?]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.thariq.io/blog/interpretability/">https://www.thariq.io/blog/interpretability/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42053912">https://news.ycombinator.com/item?id=42053912</a></p>
<p>Points: 10</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 05 Nov 2024 18:19:52 +0000</pubDate><link>https://www.thariq.io/blog/interpretability/</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=42053912</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42053912</guid></item><item><title><![CDATA[New comment by trq_ in "Detecting when LLMs are uncertain"]]></title><description><![CDATA[
<p>This is incredible! I haven't seen that repo yet, thank you for pointing it out, and the writing</p>
]]></description><pubDate>Sat, 26 Oct 2024 08:19:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=41953392</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=41953392</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41953392</guid></item><item><title><![CDATA[New comment by trq_ in "Detecting when LLMs are uncertain"]]></title><description><![CDATA[
<p>Yeah, I think the idea of finding out what flavor of uncertainty you have is very interesting.</p>
]]></description><pubDate>Fri, 25 Oct 2024 20:04:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=41949115</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=41949115</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41949115</guid></item><item><title><![CDATA[New comment by trq_ in "OmniParser for Pure Vision Based GUI Agent"]]></title><description><![CDATA[
<p>This is awesome, can't wait for evals against Claude Computer Use!</p>
]]></description><pubDate>Fri, 25 Oct 2024 19:59:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=41949055</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=41949055</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41949055</guid></item><item><title><![CDATA[New comment by trq_ in "Detecting when LLMs are uncertain"]]></title><description><![CDATA[
<p>Yeah! I want to use the logprobs API, but you can't for example:<p>- sample multiple logits and branch (we maybe could with the old text completion API, but this no longer exists)<p>- add in a reasoning token on the fly<p>- stop execution, ask the user, etc.<p>But a visualization of logprobs in a query seems like it might be useful.</p>
]]></description><pubDate>Fri, 25 Oct 2024 19:48:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=41948952</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=41948952</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41948952</guid></item><item><title><![CDATA[New comment by trq_ in "Detecting when LLMs are uncertain"]]></title><description><![CDATA[
<p>I want to build intuition on this by building a logit visualizer for OpenAI outputs. But from what I've seen so far, you can often trace down a hallucination.<p>Here's an example of someone doing that for 9.9 > 9.11: <a href="https://x.com/mengk20/status/1849213929924513905" rel="nofollow">https://x.com/mengk20/status/1849213929924513905</a></p>
]]></description><pubDate>Fri, 25 Oct 2024 18:43:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=41948310</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=41948310</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41948310</guid></item><item><title><![CDATA[New comment by trq_ in "Detecting when LLMs are uncertain"]]></title><description><![CDATA[
<p>I mean, LLMs certainly know representations of what words means and their relationship to each other, that's what the Key and Query matrices hold for example.<p>But in this case, it means that the underlying point in embedding space doesn't map clearly to only one specific token. That's not too different from when you have an idea in your head but can't think of the word.</p>
]]></description><pubDate>Fri, 25 Oct 2024 18:41:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=41948283</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=41948283</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41948283</guid></item><item><title><![CDATA[New comment by trq_ in "Detecting when LLMs are uncertain"]]></title><description><![CDATA[
<p>Yeah wouldn't be surprised if the big labs are doing more than just arg max in the sampling.</p>
]]></description><pubDate>Fri, 25 Oct 2024 18:39:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=41948259</link><dc:creator>trq_</dc:creator><comments>https://news.ycombinator.com/item?id=41948259</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41948259</guid></item></channel></rss>