<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: neonbjb</title><link>https://news.ycombinator.com/user?id=neonbjb</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 06 May 2026 08:21:52 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=neonbjb" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by neonbjb in "Asking Gemini 3 to generate Brainfuck code results in an infinite loop"]]></title><description><![CDATA[
<p>> So it made me wonder. Is Brainf*ck the ultimate test for AGI?<p>Absolutely not. Id bet a lot of money this could be solved with a decent amount of RL compute. None of the stated problems are actually issues with LLMs after on policy training is performed.</p>
]]></description><pubDate>Mon, 29 Dec 2025 11:32:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=46419648</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=46419648</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46419648</guid></item><item><title><![CDATA[New comment by neonbjb in "Reflections on OpenAI"]]></title><description><![CDATA[
<p>Yes we do. If you worked at Google you know moma. Our moma is an internal version of chat. It is very good.</p>
]]></description><pubDate>Wed, 16 Jul 2025 07:20:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=44579590</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=44579590</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44579590</guid></item><item><title><![CDATA[New comment by neonbjb in "Reflections on OpenAI"]]></title><description><![CDATA[
<p>Also work at OpenAI. Every tender offer has made full payouts to previous employees. Sorry to ruin your witch hunt..</p>
]]></description><pubDate>Wed, 16 Jul 2025 07:17:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=44579574</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=44579574</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44579574</guid></item><item><title><![CDATA[New comment by neonbjb in "OpenAI o3 and o4-mini"]]></title><description><![CDATA[
<p>I work for openai.<p>o4-mini gets much closer (but I'm pretty sure it fumbles at the last moment): <a href="https://chatgpt.com/share/680031fb-2bd0-8013-87ac-941fa91cea9d" rel="nofollow">https://chatgpt.com/share/680031fb-2bd0-8013-87ac-941fa91cea...</a><p>We're pretty bad at model naming and communicating capabilities (in our defense, it's hard!), but o4-mini is actually a _considerably_ better vision model than o3, despite the benchmarks. Similar to how o3-mini-high was a much better coding model than o1. I would recommend using o4-mini-high over o3 for any task involving vision.</p>
]]></description><pubDate>Wed, 16 Apr 2025 22:45:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=43711155</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=43711155</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43711155</guid></item><item><title><![CDATA[New comment by neonbjb in "OpenAI completes deal that values company at $157B"]]></title><description><![CDATA[
<p>You're missing the fact that requests are batched. It's 70 tokens per second for you, but also for 10s-100s of other paying customers at the same time.</p>
]]></description><pubDate>Thu, 03 Oct 2024 01:45:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=41726547</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=41726547</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41726547</guid></item><item><title><![CDATA[New comment by neonbjb in "Learning to Reason with LLMs"]]></title><description><![CDATA[
<p>You wouldn't do that to this model. It finds its own mistakes and corrects them as it is thinking through things.</p>
]]></description><pubDate>Thu, 12 Sep 2024 17:41:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=41523422</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=41523422</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41523422</guid></item><item><title><![CDATA[New comment by neonbjb in "The "it" in AI models is the dataset"]]></title><description><![CDATA[
<p>I'm James Betker.<p>Of course architecture matters in this regard lol. Comparing a CNN to a transformer is like comparing two children brought up in the same household but one has a severe disability.<p>What I meant in this blog post was that given two NNs which have the same basic components that are sufficiently large and trained long enough on the same dataset, the "behavior" of the resulting models is often shockingly similar. "Behavior" here means the typical (mean, heh) responses you get from the model. This is a function of your dataset distribution.<p>:edit: Perhaps it'd be best to give a specific example: Lets say you train two pairs of networks:
(1) A Mamba SSM and a Transformer on the Pile.
(2) Two transformers, one trained on the Pile, the other trained on Reddit comments.
All are trained to the same MMLU performance.<p>I'd put big money that the average responses you get when sampling from the models in (1) are nearly identical, whereas the two models in (2) will be quite different.</p>
]]></description><pubDate>Thu, 25 Apr 2024 13:43:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=40157478</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=40157478</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40157478</guid></item><item><title><![CDATA[New comment by neonbjb in "We have reached an agreement in principle for Sam to return to OpenAI as CEO"]]></title><description><![CDATA[
<p>As an employee of OpenAI: fuck you and your condescending conclusions about my peers and my motivations.</p>
]]></description><pubDate>Wed, 22 Nov 2023 15:36:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=38380642</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=38380642</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38380642</guid></item><item><title><![CDATA[New comment by neonbjb in "SpaceX Starship Super Heavy Project at the Boca Chica Launch Site"]]></title><description><![CDATA[
<p>I don't think the plan is an occasionally rocket launch long term. I think the plan is to launch rockets as fast as humanly possible.</p>
]]></description><pubDate>Wed, 25 Oct 2023 16:33:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=38014616</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=38014616</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38014616</guid></item><item><title><![CDATA[New comment by neonbjb in "Stable Diffusion Gets a Major Boost with RTX Acceleration"]]></title><description><![CDATA[
<p>Its cool that this is starting to approach real time video territory (30 images per second, this claims close to 1 image/sec).</p>
]]></description><pubDate>Tue, 17 Oct 2023 21:51:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=37922101</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=37922101</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37922101</guid></item><item><title><![CDATA[New comment by neonbjb in "Llama 2"]]></title><description><![CDATA[
<p>@dooraven - I also work in ML (including recently working at Google) and I agree with @whimsicalism.<p>You seem to be under the mistaken belief that:
1. Google has competent high-level organization that effectively sets and pursues long term goals.
2. There is some advantage to developing a highly capable LLM but not releasing it.<p>(2) could be the case if Google had built an extremely large model which was too expensive to deploy. Having been privy to what they had been working on up until mid-2022 and knowing how much work, compute and planning goes into extremely large models, this would very much surprise me.<p>Note: I did not have much visibility into what deepmind was up to. Maybe they had something.</p>
]]></description><pubDate>Tue, 18 Jul 2023 17:35:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=36776331</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=36776331</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36776331</guid></item><item><title><![CDATA[New comment by neonbjb in "Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows"]]></title><description><![CDATA[
<p>You can make almost anything work in DL if you try hard enough, that doesn't mean it is the correct thing to do. Convolutions have inductive biases which are the cause of many of the problems associated with deep learning over the last 10 years. Researchers don't "love the ViT". They use it because it is simply better in every way, in every application.<p>The only reason convolutions are still used in modern (intelligently designed) ML systems is because it is not known how to build a sparse attention algorithm that achieves 2D and 3D locality and is also compatible with modern accelerators. Swin is an attempt at that, but it is something of a hack.</p>
]]></description><pubDate>Mon, 10 Apr 2023 04:27:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=35509174</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=35509174</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=35509174</guid></item><item><title><![CDATA[New comment by neonbjb in "Game prototype using AI assisted graphics"]]></title><description><![CDATA[
<p>I theorize (but cannot prove) that the processes underpinning creativity in the human mind are exactly the same statistical processes that ML models use.<p>Think about it: you live your life. You experience things. You experience art, and experience emotions or have interactions with other humans grounded in that art. You form  connections with certain styles or techniques.<p>If you then turn around to create art, you form in your mind a general idea of what you want to create. You then draw on your past experiences to actually create the physical art. What process other than statistical extraction from your mind could it come from?<p>For sure I believe there are things that we don't understand about the human mind. I think the impact of drug use on art creation is very interesting, for example. It indicates that random chemical processes in our brains can play a large determining role in the actions we take (and in this case, the things that we create).<p>But to say that humans do not use some sort of inbaked statistical world model in the creative process seems wrong to me.</p>
]]></description><pubDate>Sun, 08 Jan 2023 17:48:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=34301239</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=34301239</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=34301239</guid></item><item><title><![CDATA[New comment by neonbjb in "MuZero’s first step from research into the real world"]]></title><description><![CDATA[
<p>AGI consists of a set of problems:
1. Finding an algorithm which is capable of learning anything.
2. Building the computers that can run said algorithm.
3. Collecting, sorting and filtering the data that the algorithm learns from.<p>DeepMind claims (with good cause, IMO), that MuZero can be such an algorithm. Showing that this one algorithm can tackle disparate problems is a way of proving this.<p>I think the questions that still stand are: is it even possible to build computers that could drive a scaled up MuZero to AGI? And is there a more efficient way to get there? I suspect the answer to both questions is yes.<p>Still, I think it is pretty incredible that we've managed to build computer programs that can totally adapt to arbitrary datasets and perform arbitrary tasks.</p>
]]></description><pubDate>Sat, 26 Feb 2022 03:41:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=30475081</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=30475081</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=30475081</guid></item><item><title><![CDATA[New comment by neonbjb in "Worker pay isn’t keeping up with inflation"]]></title><description><![CDATA[
<p>This is wrong IMO - the fed has NO tools in its hands. Its last tool was inflation rates, and it burned that during the great recession. It is now caught in a pickle: raise interest rates, causing a mass exodus of wealth from securities to savings accounts and CDs and causing the stock market to tumble (like it arguably always should have done since 2008), or let inflation take its course.<p>I am not an economist, and I'd love to be proven wrong. I just don't see how this ends in a good way for the economy.</p>
]]></description><pubDate>Fri, 17 Dec 2021 06:36:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=29588683</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=29588683</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=29588683</guid></item><item><title><![CDATA[New comment by neonbjb in "Vertical farming does not save space"]]></title><description><![CDATA[
<p>There are other reasons to farm indoors (or in vertical farms) than electricity and water too.<p>Pests are one such reason. They are extremely difficult to control outdoors, but far simpler to do so inside. Both fertilizer and pesticide treatments (if necessary) can be far more specific - e.g. less wasteful and environmentally damaging) when done indoors. Similarly, inclement weather generally does not affect indoor farms.<p>There's also my favorite argument: if we're ever going to try to colonize space or other planets, we better be damned good at growing plants artificially. IMO every dollar spent improving this space gets us one step closer to unlinking our future from Earth's.</p>
]]></description><pubDate>Thu, 18 Feb 2021 20:37:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=26185165</link><dc:creator>neonbjb</dc:creator><comments>https://news.ycombinator.com/item?id=26185165</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=26185165</guid></item></channel></rss>