<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: tehsauce</title><link>https://news.ycombinator.com/user?id=tehsauce</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 27 Apr 2026 10:05:25 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=tehsauce" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by tehsauce in "The Claude Code Source Leak: fake tools, frustration regexes, undercover mode"]]></title><description><![CDATA[
<p>For the purpose of disclosure, it should say “Warning: AI generated code” in the commit message, not an advertisement for a specific product. You would never accept any of your other tools injecting themselves into a commit message like that.</p>
]]></description><pubDate>Tue, 31 Mar 2026 23:04:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=47594640</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=47594640</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47594640</guid></item><item><title><![CDATA[New comment by tehsauce in "Zclaw – The 888 KiB Assistant"]]></title><description><![CDATA[
<p>“888 KiB Assistant” but the assistant itself is a multi terabyte rental-only model stored in some mysterious data center.</p>
]]></description><pubDate>Mon, 02 Mar 2026 17:11:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=47220841</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=47220841</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47220841</guid></item><item><title><![CDATA[New comment by tehsauce in "Gemini 3 Deep Think"]]></title><description><![CDATA[
<p>How does it do on gold stake?</p>
]]></description><pubDate>Thu, 12 Feb 2026 22:17:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=46996059</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=46996059</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46996059</guid></item><item><title><![CDATA[New comment by tehsauce in "Attention at Constant Cost per Token via Symmetry-Aware Taylor Approximation"]]></title><description><![CDATA[
<p>Right, and when they compare to floating point accuracy they seem to be using the number of decimals supported by the mantissa, but the exponent is important no?</p>
]]></description><pubDate>Wed, 04 Feb 2026 15:44:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=46887226</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=46887226</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46887226</guid></item><item><title><![CDATA[New comment by tehsauce in "World Models"]]></title><description><![CDATA[
<p>“A transformer predicts the next token”<p>Nope. A transformer is much more general than that. A GPT predicts the next token.</p>
]]></description><pubDate>Thu, 29 Jan 2026 01:08:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=46804312</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=46804312</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46804312</guid></item><item><title><![CDATA[New comment by tehsauce in "Scaling long-running autonomous coding"]]></title><description><![CDATA[
<p>I was excited to try it out so I downloaded the repo and ran the build. However there were 100+ compilation errors. So I checked the commit history on github and saw that for at least several pages back all recent commits had failed in the CI. It was not clear which commit I should pick to get the semi-working version advertised.<p>I started looking in the Cargo.toml to at least get an idea how the project was constructed. I saw there that rather than being built from scratch as the post seemed to imply that almost every core component was simply pulled in from an open source library. quickjs engine, wgpu graphics, winit windowing & input, egui for ui, html parsing, the list goes on.
On twitter their CEO explicitly stated that it uses a "custom js vm" which seemed particularly misleading / untrue to me.<p>Integrating all of these existing components is still super impressive for these models to do autonomously, so I'm just at a loss how to feel when it does something impressive but they then feel the need to misrepresent so much. I guess I just have a lot less respect and trust for the cursor leadership, but maybe a little relief knowing that soon I may just generate my own custom cursor!</p>
]]></description><pubDate>Thu, 15 Jan 2026 03:32:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=46627675</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=46627675</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46627675</guid></item><item><title><![CDATA[New comment by tehsauce in "World Emulation via Neural Network"]]></title><description><![CDATA[
<p>I love this! Your results seem comparable to the counter strike or minecraft models from a bit ago with massively less compute and data. It's particularly cool that it uses real world data. I've been wanting to do something like this for a while, like capturing a large dataset while backpacking in the cascades :)<p>I didn't see it in an obvious place on your github, do you have any plans to open source the training code?</p>
]]></description><pubDate>Sat, 26 Apr 2025 01:30:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=43800121</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=43800121</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43800121</guid></item><item><title><![CDATA[New comment by tehsauce in "People are just as bad as my LLMs"]]></title><description><![CDATA[
<p>There has been some good research published on this topic of how RLHF, ie aligning to human preferences easily introduces mode collapse and bias into models. For example, with a prompt like: "Choose a random number", the base pretrained model can give relatively random answers, but after fine tuning to produce responses humans like, they become very biased towards responding with numbers like "7" or "42".</p>
]]></description><pubDate>Mon, 10 Mar 2025 19:49:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=43325155</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=43325155</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43325155</guid></item><item><title><![CDATA[New comment by tehsauce in "Show HN: Beating Pokemon Red with RL and <10M Parameters"]]></title><description><![CDATA[
<p>We have a shared community map where you can watch hundreds of agents from multiple peoples training runs playing in real time!<p><a href="https://pwhiddy.github.io/pokerl-map-viz/" rel="nofollow">https://pwhiddy.github.io/pokerl-map-viz/</a></p>
]]></description><pubDate>Thu, 06 Mar 2025 00:35:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=43274691</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=43274691</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43274691</guid></item><item><title><![CDATA[New comment by tehsauce in "Show HN: Beating Pokemon Red with RL and <10M Parameters"]]></title><description><![CDATA[
<p>It's impossible to beat with random actions or brute force, but you can get surprisingly far. It doesn't take too long to get halfway through route 1, but even with insane compute you'll never make it even to viridian forest.</p>
]]></description><pubDate>Wed, 05 Mar 2025 19:32:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=43271150</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=43271150</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43271150</guid></item><item><title><![CDATA[New comment by tehsauce in "Claude Plays Pokémon"]]></title><description><![CDATA[
<p>Anyone interested in watching lots of reinforcement agents playing pokemon red at once, we have a website which streams hundreds of concurrent games from multiple people’s training runs to a shared map in real time!<p><a href="https://pwhiddy.github.io/pokerl-map-viz/" rel="nofollow">https://pwhiddy.github.io/pokerl-map-viz/</a><p>(works best on desktop)</p>
]]></description><pubDate>Thu, 27 Feb 2025 00:01:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=43189753</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=43189753</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43189753</guid></item><item><title><![CDATA[New comment by tehsauce in "Ask HN: Resources for General Purpose GPU development on Apple's M* chips?"]]></title><description><![CDATA[
<p>the metal backend does currently generate quite a lot of unnecessary command buffers, but in general performance seems solid.</p>
]]></description><pubDate>Wed, 25 Dec 2024 17:35:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=42509903</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=42509903</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42509903</guid></item><item><title><![CDATA[New comment by tehsauce in "Were RNNs all we needed?"]]></title><description><![CDATA[
<p>I haven’t gone through the paper in detail yet but maybe someone can answer.  
 If you remove the hidden state from an rnn as they say they’ve done, what’s left? An mlp predicting from a single token?</p>
]]></description><pubDate>Thu, 03 Oct 2024 17:49:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=41733016</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=41733016</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41733016</guid></item><item><title><![CDATA[New comment by tehsauce in "The Environmental Toll of a Single ChatGPT Query Is Absolutely Wild"]]></title><description><![CDATA[
<p>The water consumed to produce a single hamburger is over 2000 liters, and the power likely well over 100 watt-hours.<p>That means gpt can write >1000 emails using the resources of feeding a single person lunch. The resource efficiency of these machines already is really quite astonishing.</p>
]]></description><pubDate>Tue, 24 Sep 2024 07:14:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=41633920</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=41633920</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41633920</guid></item><item><title><![CDATA[New comment by tehsauce in "The GJK Algorithm: A weird and beautiful way to do a simple thing"]]></title><description><![CDATA[
<p>Awesome article! Something slightly misleading though - the first image shows the intersection of a non-convex shape, but it isn't revealed until much later that the algorithm only works for convex shapes, not the type shown in the first image.</p>
]]></description><pubDate>Wed, 12 Jun 2024 22:49:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=40664046</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=40664046</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40664046</guid></item><item><title><![CDATA[New comment by tehsauce in "Grokked Transformers Are Implicit Reasoners"]]></title><description><![CDATA[
<p>Grokking is a sudden huge jump in test accuracy with increasing training steps, well after training accuracy has fully converged. Double descent is test performance increasing, decreasing, and then finally rising again as model parameters are increased.</p>
]]></description><pubDate>Tue, 28 May 2024 04:29:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=40497324</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=40497324</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40497324</guid></item><item><title><![CDATA[New comment by tehsauce in "New exponent functions that make SiLU and SoftMax 2x faster, at full accuracy"]]></title><description><![CDATA[
<p>If cpu softmax were limited by memory bandwidth, then these vectorization optimizations wouldn't improve performance.</p>
]]></description><pubDate>Wed, 15 May 2024 23:54:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=40373743</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=40373743</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40373743</guid></item><item><title><![CDATA[New comment by tehsauce in "Show HN: gpudeploy.com – "Airbnb" for GPUs"]]></title><description><![CDATA[
<p>+1 for vast. they usually are the cheapest and have the most supply. some instances can be less reliable at the low end though</p>
]]></description><pubDate>Sun, 05 May 2024 01:07:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=40261573</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=40261573</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40261573</guid></item><item><title><![CDATA[New comment by tehsauce in "GPU compute in the browser at the speed of native: WebGPU marching cubes"]]></title><description><![CDATA[
<p>It's possible you might not need direct access to wave/subgroup ops to implement efficient stream compaction. There's a great old Nvidia blog post on "warp-aggregated atomics"<p><a href="https://developer.nvidia.com/blog/cuda-pro-tip-optimized-filtering-warp-aggregated-atomics/" rel="nofollow">https://developer.nvidia.com/blog/cuda-pro-tip-optimized-fil...</a><p>where they show that their compiler is sometimes able to automatically convert global atomic operations into the warp local versions, and achieve the same performance as manually written intrinsics.  
I was recently curious if 10 years later these same optimizations had made it into other GPUs and platforms besides cuda, so I put together a simple atomics benchmark in WebGPU.<p><a href="https://github.com/PWhiddy/webgpu-atomics-benchmark">https://github.com/PWhiddy/webgpu-atomics-benchmark</a><p>The results seem to indicate that these optimizations are accessible through webgpu on chrome on both MacOS and Linux (with nvidia gpu).   
Note that I'm not directly testing stream compaction, just incrementing a single global atomic counter. So that would need to be tested to know for sure if the optimization still holds there. 
If you see any issues with the benchmark or this reasoning please let me know! I am hoping to solidify my knowledge in this area :)</p>
]]></description><pubDate>Wed, 24 Apr 2024 01:12:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=40139300</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=40139300</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40139300</guid></item><item><title><![CDATA[New comment by tehsauce in "FPGA Architecture for Deep Learning: Survey and Future Directions"]]></title><description><![CDATA[
<p>500GB/s is going to limit it to at best 1/4 the DL performance of an nvidia gpu. I’m not sure what the floating point perf of these FPGAs are but I imagine that also might set a fundamental performance limit at a small fraction of a GPU.</p>
]]></description><pubDate>Tue, 23 Apr 2024 06:49:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=40129232</link><dc:creator>tehsauce</dc:creator><comments>https://news.ycombinator.com/item?id=40129232</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40129232</guid></item></channel></rss>