<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: robrenaud</title><link>https://news.ycombinator.com/user?id=robrenaud</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 07 Apr 2026 22:27:02 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=robrenaud" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by robrenaud in "TinyLoRA – Learning to Reason in 13 Parameters"]]></title><description><![CDATA[
<p>Yeah, my big problem with the paper is it just might be an artifact of qwen's training process.</p>
]]></description><pubDate>Wed, 01 Apr 2026 05:57:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47597327</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=47597327</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47597327</guid></item><item><title><![CDATA[New comment by robrenaud in "Yann LeCun's AI startup raises $1B in Europe's largest ever seed round"]]></title><description><![CDATA[
<p>Was Alphago's move 37 original?<p>In the last step of training LLMs, reinforcement learning from verified rewards, LLMs are trained to maximize the probability of solving problems using their own output, depending on a reward signal akin to winning in Go. It's not just imitating human written text.<p>Fwiw, I agree that world models and some kind of learning from interacting with physical reality, rather than massive amounts of digitized gym environments is likely necessary for a breakthrough for AGI.</p>
]]></description><pubDate>Tue, 10 Mar 2026 15:36:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47324708</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=47324708</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47324708</guid></item><item><title><![CDATA[New comment by robrenaud in "Yann LeCun raises $1B to build AI that understands the physical world"]]></title><description><![CDATA[
<p>Recursive self improvement.  It's when  AI speeds up the development of the next AI.</p>
]]></description><pubDate>Tue, 10 Mar 2026 15:24:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=47324548</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=47324548</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47324548</guid></item><item><title><![CDATA[New comment by robrenaud in "Ask HN: Who wants to be hired? (March 2026)"]]></title><description><![CDATA[
<p>Location: SF (current).  NYC/Philly general area acceptable.  Remote okay.
email: rrenaud@gmail.com
Resume:  16 year SWE -> MLE @ Google, MS from NYU with focus on ML.  Retired.  Now I hack on data analysis for video game projects for fun, and I love it.  I'd take crazy low compensation to do work with interesting game data sets.  EG, for game balance, strategic analysis, or to improve/augment game video content.</p>
]]></description><pubDate>Tue, 03 Mar 2026 20:13:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47238259</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=47238259</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47238259</guid></item><item><title><![CDATA[New comment by robrenaud in "Ask HN: Who is hiring? (March 2026)"]]></title><description><![CDATA[
<p>What do y'all think about the latency/quality tradeoff with LLMs?<p>Human voices don't take 30 seconds to think, retrieve, research, and summarize a high quality answer. Humans are calibrated in their knowledge, they know what they understand and what they don't.  They can converse in real time without bullshitting.<p>Frontier real time-ish LLM generated voice systems are still plagued by 2024 era LLM nonsense, like the inability to count Rs in strawberry. [1]<p>I'd personally love a voice interface that, constrained by the technology of today, takes the latency hit to deliver quality.<p>[1] <a href="https://www.instagram.com/reel/DTYBpa7AHSJ/?igsh=MzRlODBiNWFlZA==" rel="nofollow">https://www.instagram.com/reel/DTYBpa7AHSJ/?igsh=MzRlODBiNWF...</a></p>
]]></description><pubDate>Mon, 02 Mar 2026 17:17:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=47220912</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=47220912</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47220912</guid></item><item><title><![CDATA[Ask HN: Is there something like Google style guide for AI-only coded apps?]]></title><description><![CDATA[
<p>The Google style guide for C++/Java/Python are opinionated, hard fought, wise, and they elimate a large source of errors while minimizing harmful, unneded inconsistences.  They picked a style that made great realistic use of the best cognition at the time.<p>The intent is still great, but now we should think about writing good general rules for building programs that are essentially all AI generated.  What generic wisdom leads to flexible, auditable, composable and robust apps and systems?</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47204331">https://news.ycombinator.com/item?id=47204331</a></p>
<p>Points: 1</p>
<p># Comments: 2</p>
]]></description><pubDate>Sun, 01 Mar 2026 06:51:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=47204331</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=47204331</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47204331</guid></item><item><title><![CDATA[New comment by robrenaud in "How an inference provider can prove they're not serving a quantized model"]]></title><description><![CDATA[
<p>Please serve well quantized models.<p>If you can get 99 percent of the quality for 50 percent of the cost, that is most times a good tradeoff.</p>
]]></description><pubDate>Sat, 21 Feb 2026 22:36:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=47105602</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=47105602</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47105602</guid></item><item><title><![CDATA[New comment by robrenaud in "Consistency diffusion language models: Up to 14x faster, no quality loss"]]></title><description><![CDATA[
<p>Cite a source.  Your concrete claim is that, on average, for every $1 of subscription revenue on a monthly subscription, OpenAI and Anthropic were losing $11.50?<p>It seems completely implausible.<p>I could believe that if a $20 sub used every possible token granted, it would cost $250.  But certainly almost no one was completely milking their subscription.  In the same way that no one is streaming netflix literally 24/7.</p>
]]></description><pubDate>Fri, 20 Feb 2026 18:22:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=47091699</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=47091699</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47091699</guid></item><item><title><![CDATA[New comment by robrenaud in "Ask HN: What are you working on? (February 2026)"]]></title><description><![CDATA[
<p>I used to play very competitively, but I've been more chill recently.  I just think it's a nice problem/dataset to work with, because of the depth of my understanding of the game.</p>
]]></description><pubDate>Mon, 09 Feb 2026 06:09:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=46942071</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46942071</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46942071</guid></item><item><title><![CDATA[New comment by robrenaud in "Ask HN: What are you working on? (February 2026)"]]></title><description><![CDATA[
<p>I’ve been experimenting with a live win probability predictor for the 10-player arcade game Killer Queen. The goal is to predict the winner in a causal, event-by-event fashion.<p>Right now I’m struggling to beat a baseline LightGBM model trained on hand-engineered expert features. My attempts at using a win probability head on top of nanoGPT, treating events as tokens, have been significantly worse. I am seeing about 65% accuracy compared to the LightGBM’s 70%. That 5% gap is huge given how stochastic the early game is, and the Transformer is easily 4 OOM more expensive to train.<p>To bridge the gap, I’m moving to a hybrid approach. I’m feeding those expert features back in as additional tokens or auxiliary loss heads, and I am using the LightGBM model as a teacher for knowledge distillation to provide smoother gradients.<p>The main priority here is personalized post-game feedback. By tracking sharp swings in win probability, or $\Delta WP$, you can automatically generate high or low-light reels right after a match. It helps players see the exact moment a play was either effective or catastrophic.<p>There is also a clear application for automated content creation. You can use $\Delta WP$ as a heuristic to identify the actual turning points of a match for YouTube summaries without needing to manually scrub through hours of Twitch footage.</p>
]]></description><pubDate>Mon, 09 Feb 2026 03:57:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=46941412</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46941412</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46941412</guid></item><item><title><![CDATA[New comment by robrenaud in "LLMs as the new high level language"]]></title><description><![CDATA[
<p>A compiler that can turn cash into improved code without round tripping a human is very cool though. As those steps can get longer and succeed more often in more difficult circumstances, what it means to be a software engineer changes a lot.</p>
]]></description><pubDate>Sun, 08 Feb 2026 04:05:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=46931241</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46931241</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46931241</guid></item><item><title><![CDATA[New comment by robrenaud in "Ask HN: What are you working on? (January 2026)"]]></title><description><![CDATA[
<p>Previously, I made a live win probability model for the 5v5 arcade game Killer Queen Arcade from their game events API.<p>Now I am trying to use that model to make:<p>1.  A post game instant replay that shows the most important/pivotal moments from the most recently finished game.  Some arcades have a seperate display for observers, it could work well there, or as good filler between matches on twitch streams.<p>2.  A personalized per tournament/yearly highlights recap.<p>If it works well, it might be a kind of tool that generalizes well for summarizing long twitch streams for Youtube.<p><a href="https://github.com/rrenaud/kq_stream_highlights" rel="nofollow">https://github.com/rrenaud/kq_stream_highlights</a></p>
]]></description><pubDate>Sun, 11 Jan 2026 23:37:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=46581736</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46581736</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46581736</guid></item><item><title><![CDATA[New comment by robrenaud in "ChatGPT Health"]]></title><description><![CDATA[
<p>Here is research about doctors interpreting test results. It seems to favor GP's view that many doctors struggle to weigh test specificity and sensitivity vs disease base rate.<p><a href="https://bmjopen.bmj.com/content/bmjopen/5/7/e008155.full.pdf" rel="nofollow">https://bmjopen.bmj.com/content/bmjopen/5/7/e008155.full.pdf</a></p>
]]></description><pubDate>Thu, 08 Jan 2026 17:12:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=46543539</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46543539</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46543539</guid></item><item><title><![CDATA[New comment by robrenaud in "AI sycophancy panic"]]></title><description><![CDATA[
<p>I suspect the models would be more useful but perhaps less popular if the semantic content of their answers depended less on the expectations of the prompter.</p>
]]></description><pubDate>Sun, 04 Jan 2026 17:24:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=46490066</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46490066</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46490066</guid></item><item><title><![CDATA[New comment by robrenaud in "Using Vectorize to build an unreasonably good search engine in 160 lines of code"]]></title><description><![CDATA[
<p>What are you embedding?  Are you doing a geo restricted area (small universe?).</p>
]]></description><pubDate>Thu, 25 Dec 2025 08:24:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=46382989</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46382989</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46382989</guid></item><item><title><![CDATA[New comment by robrenaud in "Ryanair fined €256M over ‘abusive strategy’ to limit ticket sales by OTAs"]]></title><description><![CDATA[
<p>Paying the local currency with your own cards seems simple and works?</p>
]]></description><pubDate>Tue, 23 Dec 2025 15:57:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=46366342</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46366342</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46366342</guid></item><item><title><![CDATA[New comment by robrenaud in "Ask HN: Why Did Python Win?"]]></title><description><![CDATA[
<p>I learned Python circa 2000 as a 17 year old.<p>It felt pretty easy to read and write, had minimal surprises, and it made writing simple programs easy.  The batteries-includedness was great.<p>It felt like it was designed by a smart guy for practical programming, rather than by a brilliant academic who was wed to purity for maximum elegance.  It accepted some warts, but them in mostly ergonomic places.<p>It was dictated by people who could relegate map, filter, and reduce from builtins to the library.  As much as I personally even liked map in particular (I am not super anti functional programming), it's nice to realize the designer had the taste to prefer longer but more explicit programs.<p><pre><code>    ys = [f(x) for x in xs]
    ys = map(f, xs)
</code></pre>
As much as I disliked that particular decision, there is no doubt to me that it is just designed for lower cognitive burden when doing simple/common things.<p>If you show both of those lines of code to a person in cs101 who has studied Java for a couple months but hasn't seen Python, I am pretty sure they are gonna understand the first line way quicker.  Mostly consistent decision making like that leads to ergonomic, practical languages.  And they win.<p>It fits in people's brains better.  See also, pytorch vs tensorflow.</p>
]]></description><pubDate>Mon, 22 Dec 2025 16:28:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=46355503</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46355503</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46355503</guid></item><item><title><![CDATA[New comment by robrenaud in "Gemini 3 Flash: Frontier intelligence built for speed"]]></title><description><![CDATA[
<p>Omg, it was so frustrating to say:<p>Summarize recent working arxiv url<p>And then it tells me the date is from the future and it simply refuses to fetch the URL.</p>
]]></description><pubDate>Wed, 17 Dec 2025 17:14:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=46302367</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46302367</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46302367</guid></item><item><title><![CDATA[New comment by robrenaud in "Erdős Problem #1026"]]></title><description><![CDATA[
<p>What is harder, beating Lee Sedol at Go, or physically placing stones on a Go board?  Which is closer to AGI?<p>Because AlphaGo can only do one.<p>AI could very well be better at formal theorem proving than fields medalists pretty soon.  It will not have taste, ability to see the beauty in math, or pick problems and set directions for the field.  But given a well specified problem, it can bruteforce search through lean tactics space at an extremely superhuman pace.  What is lacks in intuition and brilliance, it will make up in being able to explore in parallel.<p>There is a quality/quantity tradeoff in search with a verifier.  A superhuman artificial theorem prover can be generating much worse ideas on average than a top mathematician, and make up for it by trying many more of them.<p>It's Kasparov vs DeepBlue and Sedol vs AlphaGo all over.<p>It's also nowhere near AGI. Embodiment and the real world is super messy.  See Moravec's paradox.<p>Practical programs deal with the outside world, they are underspecified, their utility depends on the changing whims of people.  The formal specification of a math problem is self contained.</p>
]]></description><pubDate>Tue, 16 Dec 2025 21:35:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=46294856</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46294856</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46294856</guid></item><item><title><![CDATA[New comment by robrenaud in "Erdős Problem #1026"]]></title><description><![CDATA[
<p>I think you underestimate how powerful lean is, and close it is to the tedious part of formal math.  A theorem prover needs consult no outside resource.  A formal math LLM-like generator need only consult the theorem prover to get rid of hallucinations.  This is why it's actually much easier than SWE to optimize/hill climb on.<p>Low level, automated theorem providing is going to fall way quicker than most expected, like AlphaGo, precisely because an MCTS++ search over lean proofs is scalable/amendable to self play/relevant to a significant chunk of professional math.<p>Legit, I almost wish the US and China would sign a Formal Mathematics Profileration Treaty, as a sign of good will between very powerful parties who have much to gain from each other.  When your theorem prover is sufficiently better than most Fields medalists alive, you share your arch/algorithms/process with the world.  So Mathematics stays in the shared realm of human culture, and it doesn't just happen to belong to DeepMind, OpenAI, or Deepseek.</p>
]]></description><pubDate>Tue, 16 Dec 2025 07:52:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=46285890</link><dc:creator>robrenaud</dc:creator><comments>https://news.ycombinator.com/item?id=46285890</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46285890</guid></item></channel></rss>