<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: islewis</title><link>https://news.ycombinator.com/user?id=islewis</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 07 Apr 2026 15:39:41 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=islewis" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by islewis in "1M context is now generally available for Opus 4.6 and Sonnet 4.6"]]></title><description><![CDATA[
<p>The quality with the 1M window has been very poor for me, specifically for coding tasks. It constantly forgets stuff that has happened in the existing conversation. n=1, ymmv</p>
]]></description><pubDate>Sat, 14 Mar 2026 01:04:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47372180</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=47372180</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47372180</guid></item><item><title><![CDATA[New comment by islewis in "ChatGPT Developer Mode: Full MCP client access"]]></title><description><![CDATA[
<p>> It's powerful but dangerous, and is intended for developers who understand how to safely configure and test connectors.<p>So... practically no one? My experience has been that almost everyone testing these cutting edge AI tools as they come out are more interested in new tool shinyness than safety or security.</p>
]]></description><pubDate>Wed, 10 Sep 2025 17:35:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=45201063</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=45201063</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45201063</guid></item><item><title><![CDATA[New comment by islewis in "Meta invests $14.3B in Scale AI to kick-start superintelligence lab"]]></title><description><![CDATA[
<p>This appears to be a psuedo-acquisition, but with a strange format to appease regulators.<p>Will this still be an exit event for employees or do they get screwed here?</p>
]]></description><pubDate>Fri, 13 Jun 2025 17:03:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=44270255</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=44270255</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44270255</guid></item><item><title><![CDATA[New comment by islewis in "DeepSeek-Prover-V2"]]></title><description><![CDATA[
<p>> The cold-start training procedure begins by prompting DeepSeek-V3 to decompose complex problems into a series of subgoals<p>It feels pretty intuitive to me that the ability for an LLM to break a complex problem down into smaller, more easily solvable pieces will unlock the next level of complexity.<p>This pattern feels like a technique often taught to junior engineers- how to break up a multi-week project into bitesized tasks.  This model is obviously math focused, but I see no reason why this wouldn't be incredibly powerful for code based problem solving.</p>
]]></description><pubDate>Wed, 30 Apr 2025 17:25:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=43848289</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43848289</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43848289</guid></item><item><title><![CDATA[New comment by islewis in "Google blocked Motorola use of Perplexity AI, witness says"]]></title><description><![CDATA[
<p>> Already in anti-trust related to ads, AI is probably in the clear.<p>"Already in trouble for committing monopolist behavior in market A, Google should be fine committing even more monopolist behavior in the very related and overlapping market of B"<p>This makes claim makes pretty little sense to me. AI search and Google web search (ads) are already stepping on each other. I see no reason that Google wouldn't be worried about antitrust on AI search if they're worried about antitrust action in general- which they clearly are.</p>
]]></description><pubDate>Wed, 23 Apr 2025 21:18:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=43776712</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43776712</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43776712</guid></item><item><title><![CDATA[New comment by islewis in "Supabase raises $200M Series D at $2B valuation"]]></title><description><![CDATA[
<p>If you are trying to commercialize something, a popular project with bad margins is a better spot to be in than an unsuccessful project with good margins. If it's a personal learning project, that might not be the case.</p>
]]></description><pubDate>Tue, 22 Apr 2025 17:09:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=43764270</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43764270</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43764270</guid></item><item><title><![CDATA[New comment by islewis in "Reasoning models don't always say what they think"]]></title><description><![CDATA[
<p>> For the purposes of this experiment, though, we taught the models to reward hack [...] in this case rewarded the models for choosing the wrong answers that accorded with the hints.<p>> This is concerning because it suggests that, should an AI system find hacks, bugs, or shortcuts in a task, we wouldn’t be able to rely on their Chain-of-Thought to check whether they’re cheating or genuinely completing the task at hand.<p>As a non-expert in this field, I fail to see why a RL model taking advantage of it's reward is "concerning". My understanding is that the only difference between a good model and a reward-hacking model is if the end behavior aligns with human preference or not.<p>The articles TL:DR reads to me as "We trained the model to behave badly, and it then behaved badly". I don't know if i'm missing something, or if calling this concerning might be a little bit sensationalist.</p>
]]></description><pubDate>Thu, 03 Apr 2025 20:11:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=43574761</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43574761</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43574761</guid></item><item><title><![CDATA[New comment by islewis in "MLB says Yankees’ new “torpedo bats” are legal and likely coming"]]></title><description><![CDATA[
<p>I've always wondered what the technological development of F1 would look like in other sports. This feels pretty close.</p>
]]></description><pubDate>Mon, 31 Mar 2025 16:09:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=43536621</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43536621</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43536621</guid></item><item><title><![CDATA[New comment by islewis in "I genuinely don't understand why some people are still bullish about LLMs"]]></title><description><![CDATA[
<p>> I genuinely don't understand why some people are still bullish about LLMs.<p>I don't believe OP's thesis is properly backed by the rest of his tweet, which seems to boil down to "LLM's can't properly cite links".<p>If LLM's performing poorly on an arbitrary small-scoped test case makes you bearish on the whole field, I don't think that falls on the LLM's.</p>
]]></description><pubDate>Thu, 27 Mar 2025 22:27:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=43498935</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43498935</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43498935</guid></item><item><title><![CDATA[New comment by islewis in "OpenAI Audio Models"]]></title><description><![CDATA[
<p>Cool format for a demo. Some of the voices have a slight "metallic" ring to them, something I've seen a fair amount with Eleven Labs' models.<p>Does anyone have any experience with the realtime latency of these Openai TTS models? ElevenLabs has been so slow (much slower than the latency they advertise), which makes it almost impossible to use in realtime scenarios unless you can cache and replay the outputs. Cartesia looks to have cracked the time to first token, but i've found their voices to be a bit less consistent than Eleven Labs'.</p>
]]></description><pubDate>Thu, 20 Mar 2025 17:49:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=43426473</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43426473</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43426473</guid></item><item><title><![CDATA[New comment by islewis in "Nvidia Dynamo: A Datacenter Scale Distributed Inference Serving Framework"]]></title><description><![CDATA[
<p>is this in reference to Triton?</p>
]]></description><pubDate>Tue, 18 Mar 2025 22:40:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=43406033</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43406033</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43406033</guid></item><item><title><![CDATA[New comment by islewis in "Alphabet in Talks to Buy Cloud Security Firm Wiz for $33B"]]></title><description><![CDATA[
<p>It's because the motivation for these acquisitions doesn't have much to do with product. It's just a way to extinguish outside talent and innovation that could disrupt down the road.</p>
]]></description><pubDate>Mon, 17 Mar 2025 23:31:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=43393892</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43393892</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43393892</guid></item><item><title><![CDATA[New comment by islewis in "Milk Kanban"]]></title><description><![CDATA[
<p>I recently noticed that there was a "Time to order" message written on the outside of the cardboard core of my saran wrap. The visibility to the core became clearer as the wrap was consumed- giving both a sense of urgency, as well as not letting you forget. I thought it was pretty clever.</p>
]]></description><pubDate>Sat, 15 Mar 2025 21:42:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=43375357</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=43375357</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43375357</guid></item><item><title><![CDATA[New comment by islewis in "Exposed DeepSeek database leaking sensitive information, including chat history"]]></title><description><![CDATA[
<p>A valid response to my initial comment which was a bit tongue in cheek.<p>However, i'm not sure that them being LLM researchers compared to quant researchers changes the dynamic of their relaxed security posture.</p>
]]></description><pubDate>Thu, 30 Jan 2025 00:11:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=42873144</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=42873144</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42873144</guid></item><item><title><![CDATA[New comment by islewis in "Exposed DeepSeek database leaking sensitive information, including chat history"]]></title><description><![CDATA[
<p>A bunch of ML researchers who were initially hired to do quant work published their first ever user facing project.<p>So maybe not a side project, but if you have ever worked with ML researchers before, lack of engineering/security chops shouldn't be that surprising to you.</p>
]]></description><pubDate>Wed, 29 Jan 2025 23:09:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=42872518</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=42872518</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42872518</guid></item><item><title><![CDATA[New comment by islewis in "Stargate Project: SoftBank, OpenAI, Oracle, MGX to build data centers"]]></title><description><![CDATA[
<p>$500B is not $7T, but its surprisingly close.</p>
]]></description><pubDate>Tue, 21 Jan 2025 22:46:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=42786108</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=42786108</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42786108</guid></item><item><title><![CDATA[New comment by islewis in "Why Gelsinger was wrong for Intel"]]></title><description><![CDATA[
<p>>  Maybe that’s better for the industry because all the money will be concentrated on one or two competently run foundries<p>How would this be better for the industry?</p>
]]></description><pubDate>Mon, 09 Dec 2024 04:59:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=42363185</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=42363185</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42363185</guid></item><item><title><![CDATA[New comment by islewis in "Oasis: A Universe in a Transformer"]]></title><description><![CDATA[
<p>This looks like a very similar project to "Diffusion Models Are Real-Time Game Engines"[1] that circulated on HN a few months ago [2], which was playing DOOM. There's some pretty interesting commentary on that post that might also apply to this.<p>I'd like to do a deeper dive into the two approaches, but on a surface level one interesting note is Oasis specifically mentions using a use-specific ASIC (presumably for inference?):<p>> When Etched's transformer ASIC, Sohu, is released, we can run models like Oasis in 4K.<p>[1] <a href="https://gamengen.github.io/" rel="nofollow">https://gamengen.github.io/</a><p>[2] <a href="https://news.ycombinator.com/item?id=41375548">https://news.ycombinator.com/item?id=41375548</a></p>
]]></description><pubDate>Tue, 05 Nov 2024 20:33:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=42054902</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=42054902</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42054902</guid></item><item><title><![CDATA[New comment by islewis in "Meta Bans Accounts Tracking Private Jets for Zuckerberg, Musk"]]></title><description><![CDATA[
<p>The motivation for these accounts is usually a rift on the "ultra-wealth is bad" train.<p>Setting aside any possible agreements/disagreements with that, the flight tracking information is freely public, available to anyone who wants to look- go on flightaware. Flight information has never been private, nobody treats it as private, so why would social media* companies pretend it is? I don't think home addresses are that comparable in this situation.<p>EDIT: company name</p>
]]></description><pubDate>Tue, 22 Oct 2024 22:47:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=41919498</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=41919498</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41919498</guid></item><item><title><![CDATA[New comment by islewis in "Swarm, a new agent framework by OpenAI"]]></title><description><![CDATA[
<p>> The underlying issue is that AI agents too slow,<p>Inference speed is being rapidly optimized, especially for edge devices.<p>> too expensive,<p>The half-life of OpenAI's API pricing is a couple of months. While the bleeding edge model is always costly, the cost of API's are becoming rapidly available to the public.<p>> and too unreliable<p>Out of the 3 points raised, this is probably the most up in the air. Personally I chalk this up to sideeffects of OpenAI's rapid growth over the last few years. I think this gets solved, especially once price and latency have been figured out.<p>IMO, the biggest unknown here isn't a technical one, but rather a business one-  I don't think it's certain that products built on multi-agent architectures will be addressing a need for end users. Most of the talk I see in this space are by people excited by building with LLM's, not by people who are asking to pay for these products.</p>
]]></description><pubDate>Sat, 12 Oct 2024 18:51:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=41821445</link><dc:creator>islewis</dc:creator><comments>https://news.ycombinator.com/item?id=41821445</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41821445</guid></item></channel></rss>