<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: ag8</title><link>https://news.ycombinator.com/user?id=ag8</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 28 Apr 2026 22:15:53 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=ag8" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by ag8 in "Thank HN: You helped save 33k lives"]]></title><description><![CDATA[
<p>You're right; I should've been more precise. However, we have tools for dealing with this—that's what quality-adjusted life-years are for! I don't contest that surgeries often significantly increase QALYs, and may do so pretty cost-effectively.</p>
]]></description><pubDate>Wed, 18 Feb 2026 00:36:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=47055509</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=47055509</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47055509</guid></item><item><title><![CDATA[New comment by ag8 in "Thank HN: You helped save 33k lives"]]></title><description><![CDATA[
<p>Lol, I just care a lot about saving as many lives as I can; the most effective charities I've been able to find good evidence on save one life for $6–8k. If Watsi had a credible claim at being able to save lives 10x cheaper I would redirect my entire donation budget to them!<p>That said, once again, Watsi is great. I really appreciate all the hard work they've put into making this happen—this is orders of magnitude more impressive and impactful than most projects I've ever seen!</p>
]]></description><pubDate>Wed, 18 Feb 2026 00:31:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=47055475</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=47055475</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47055475</guid></item><item><title><![CDATA[New comment by ag8 in "Thank HN: You helped save 33k lives"]]></title><description><![CDATA[
<p>Watsi seems to be doing great work, but the title—"you helped save 33k lives"—reads as misleading to me. I guess "helped" could be doing a lot of heavy lifting here, but I would be incredibly surprised if the counterfactual number of lives saved was more than 3000. (But don't let this dissuade you from donating; concretely improving someone's life is totally a worthwhile goal, and Watsi seems very good at effecting this)</p>
]]></description><pubDate>Tue, 17 Feb 2026 23:30:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=47054967</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=47054967</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47054967</guid></item><item><title><![CDATA[Gourmand Syndrome]]></title><description><![CDATA[
<p>Article URL: <a href="https://en.wikipedia.org/wiki/Gourmand_syndrome">https://en.wikipedia.org/wiki/Gourmand_syndrome</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46814828">https://news.ycombinator.com/item?id=46814828</a></p>
<p>Points: 27</p>
<p># Comments: 9</p>
]]></description><pubDate>Thu, 29 Jan 2026 19:01:02 +0000</pubDate><link>https://en.wikipedia.org/wiki/Gourmand_syndrome</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=46814828</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46814828</guid></item><item><title><![CDATA[New comment by ag8 in "Ask HN: Share your personal website"]]></title><description><![CDATA[
<p><a href="https://andrew.gr" rel="nofollow">https://andrew.gr</a></p>
]]></description><pubDate>Thu, 15 Jan 2026 08:02:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=46629515</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=46629515</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46629515</guid></item><item><title><![CDATA[guys why does armenian completely break Claude]]></title><description><![CDATA[
<p><a href="https://xcancel.com/dyushag/status/1993143599286886525" rel="nofollow">https://xcancel.com/dyushag/status/1993143599286886525</a><p><a href="https://claude.ai/share/e368b733-71a4-4211-99f5-6b6cc717b575" rel="nofollow">https://claude.ai/share/e368b733-71a4-4211-99f5-6b6cc717b575</a></p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46579397">https://news.ycombinator.com/item?id=46579397</a></p>
<p>Points: 99</p>
<p># Comments: 65</p>
]]></description><pubDate>Sun, 11 Jan 2026 20:03:20 +0000</pubDate><link>https://twitter.com/dyushag/status/1993143599286886525</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=46579397</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46579397</guid></item><item><title><![CDATA[Sampling at negative temperature]]></title><description><![CDATA[
<p>Article URL: <a href="https://cavendishlabs.org/blog/negative-temperature/">https://cavendishlabs.org/blog/negative-temperature/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46579374">https://news.ycombinator.com/item?id=46579374</a></p>
<p>Points: 203</p>
<p># Comments: 60</p>
]]></description><pubDate>Sun, 11 Jan 2026 20:01:14 +0000</pubDate><link>https://cavendishlabs.org/blog/negative-temperature/</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=46579374</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46579374</guid></item><item><title><![CDATA[Perfectly Replicating Coca Cola [video]]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.youtube.com/watch?v=TDkH3EbWTYc">https://www.youtube.com/watch?v=TDkH3EbWTYc</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46572924">https://news.ycombinator.com/item?id=46572924</a></p>
<p>Points: 1</p>
<p># Comments: 1</p>
]]></description><pubDate>Sun, 11 Jan 2026 05:24:19 +0000</pubDate><link>https://www.youtube.com/watch?v=TDkH3EbWTYc</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=46572924</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46572924</guid></item><item><title><![CDATA[New comment by ag8 in "Size of Life"]]></title><description><![CDATA[
<p>Not 13?</p>
]]></description><pubDate>Wed, 10 Dec 2025 22:46:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=46225065</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=46225065</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46225065</guid></item><item><title><![CDATA[New comment by ag8 in "We collected 10k hours of neuro-language data in our basement"]]></title><description><![CDATA[
<p>This is a cool setup, but naively it feels like it would require hundreds of thousands of hours of data to train a decent generalizable model that would be useful for consumers. Are there plans to scale this up, or is there reason to believe that tens of thousands of hours are enough?</p>
]]></description><pubDate>Mon, 08 Dec 2025 18:02:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=46195528</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=46195528</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46195528</guid></item><item><title><![CDATA[Po.ta.to]]></title><description><![CDATA[
<p>Article URL: <a href="https://po.ta.to/">https://po.ta.to/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45842568">https://news.ycombinator.com/item?id=45842568</a></p>
<p>Points: 4</p>
<p># Comments: 2</p>
]]></description><pubDate>Fri, 07 Nov 2025 01:13:01 +0000</pubDate><link>https://po.ta.to/</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45842568</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45842568</guid></item><item><title><![CDATA[Scaling pretraining affects RL sample efficiency]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.runrl.com/blog/warm-start-rl">https://www.runrl.com/blog/warm-start-rl</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45700340">https://news.ycombinator.com/item?id=45700340</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 25 Oct 2025 00:09:05 +0000</pubDate><link>https://www.runrl.com/blog/warm-start-rl</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45700340</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45700340</guid></item><item><title><![CDATA[Systematically generating tests that would have caught Anthropic's top‑K bug]]></title><description><![CDATA[
<p>Article URL: <a href="https://theorem.dev/blog/anthropic-bug-test/">https://theorem.dev/blog/anthropic-bug-test/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45534347">https://news.ycombinator.com/item?id=45534347</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 10 Oct 2025 00:17:44 +0000</pubDate><link>https://theorem.dev/blog/anthropic-bug-test/</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45534347</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45534347</guid></item><item><title><![CDATA[New comment by ag8 in "Tinker"]]></title><description><![CDATA[
<p>Yeah, not sure why the HN backend changed it...</p>
]]></description><pubDate>Wed, 01 Oct 2025 18:53:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=45441718</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45441718</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45441718</guid></item><item><title><![CDATA[Tinker]]></title><description><![CDATA[
<p>Article URL: <a href="https://2b4fdb18.connectionism.pages.dev/blog/announcing-tinker/">https://2b4fdb18.connectionism.pages.dev/blog/announcing-tinker/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45440952">https://news.ycombinator.com/item?id=45440952</a></p>
<p>Points: 4</p>
<p># Comments: 2</p>
]]></description><pubDate>Wed, 01 Oct 2025 18:03:04 +0000</pubDate><link>https://2b4fdb18.connectionism.pages.dev/blog/announcing-tinker/</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45440952</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45440952</guid></item><item><title><![CDATA[Training Qwen to answer briefly yet intelligently using feedback control]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.runrl.com/blog/feedback-control">https://www.runrl.com/blog/feedback-control</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45341448">https://news.ycombinator.com/item?id=45341448</a></p>
<p>Points: 4</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 23 Sep 2025 00:32:22 +0000</pubDate><link>https://www.runrl.com/blog/feedback-control</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45341448</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45341448</guid></item><item><title><![CDATA[New comment by ag8 in "Launch HN: RunRL (YC X25) – Reinforcement learning as a service"]]></title><description><![CDATA[
<p>A) You could have an additional field in the jsonl file which says which rubric to use; then, your reward function could access this via `kwargs["rubric"]` and return a reward based on that example's preferred rubric;<p>B) currently, pricing on the deployed API is free, but the startup time is a few minutes and it's run on a small GPU node and is therefore not awfully fast. If you would like more production-level inference, email us at founders@runrl.com and we could set you up with something much faster (where we'd charge per token depending on model size)</p>
]]></description><pubDate>Thu, 18 Sep 2025 19:16:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=45293780</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45293780</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45293780</guid></item><item><title><![CDATA[New comment by ag8 in "Launch HN: RunRL (YC X25) – Reinforcement learning as a service"]]></title><description><![CDATA[
<p>Having an RL agent that's really good at search across some space sounds very powerful in general; "proofs-as-search" make this an appealing target. Back in the day, when I did more fundamental RL research, we worked on an extension of SoRB [0] where an additional meta-level target was learning improved heuristics to explore the search space faster; would be exciting to figure out what a good setup for doing things like this in LLM-policy-gradient world is these days!<p>[0]: <a href="https://arxiv.org/abs/1906.05253" rel="nofollow">https://arxiv.org/abs/1906.05253</a></p>
]]></description><pubDate>Thu, 18 Sep 2025 19:09:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=45293708</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45293708</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45293708</guid></item><item><title><![CDATA[New comment by ag8 in "Launch HN: RunRL (YC X25) – Reinforcement learning as a service"]]></title><description><![CDATA[
<p>we should publish some; the high-order effect seems to be that LoRAs significantly hurt small model performance vs FFT, with less of an effect for large models. This is maybe because large models have more built-in skills and thus a LoRA suffices to elicit the existing skill, whereas for small models you need to do more actual learning (holding # parameter updates constant). In general I think it's better to get a performant small model with FFT than a performant large model with a large LoRA, which is why we default to FFT, but I agree that we should publish more details here.</p>
]]></description><pubDate>Thu, 18 Sep 2025 17:22:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=45292384</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45292384</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45292384</guid></item><item><title><![CDATA[New comment by ag8 in "Launch HN: RunRL (YC X25) – Reinforcement learning as a service"]]></title><description><![CDATA[
<p>Thanks! Our goal is to make rl "just work" with completely automated GPU provisioning/algorithm selection/SFT-warm up, but giving people the ability to switch away from the defaults if they want to.<p>The way tools currently work in the beta is you add tools via MCP to the configuration, and they get passed in as additional context for the model; the model might then choose to use a tool during inference; the tool is then automatically called and the output is returned as a tool message. If you really want to you could parse the tool output as part of reward calculation, but I expect you'd usually base the reward just on the model's completion. I could give more details if there's a specific tool setup you're envisioning!</p>
]]></description><pubDate>Thu, 18 Sep 2025 00:52:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=45283496</link><dc:creator>ag8</dc:creator><comments>https://news.ycombinator.com/item?id=45283496</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45283496</guid></item></channel></rss>