<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: dvt</title><link>https://news.ycombinator.com/user?id=dvt</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 08 Apr 2026 19:21:00 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=dvt" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by dvt in "GLM-5.1: Towards Long-Horizon Tasks"]]></title><description><![CDATA[
<p>I don't want to respond to 100 comments about the same thing, and this one happens to be on top, so, in my humble opinion:<p>(1): You don't have to be an Ed Zitron disciple to infer that OpenAI and Anthropic are likely overvalued and that Nvidia is selling everyone shovels in a gold rush. AI is a game-changing technology, but a shitty chat interface does not a company make. OpenAI and Anthropic need to recoup astronomical costs used in training these models. Models that are now being distilled[1] and are quickly becoming commoditized. (And frankly, models that were trained by torrenting copyrighted data[2], anyway.) Many have been calling this out for years: the model <i>cannot</i> be your product. And to be clear, OpenAI/Anthropic most definitely know this: that's why they've been aquihiring like crazy, trying to find that one team that will make the <i>thing</i>.<p>(2): Token prices are <i>significantly</i> subsidized and anyone that does any serious work with AI can tell you this. Go use an almost-SOTA model (a big Deepseek or Qwen model) offered by many bare-metal providers and you'll see what "true" token prices should look like. The end-state here is likely some models running locally and some running in the cloud. But the current state of OpenClaw token-vomit on top of Claude is fiscally untenable (in fact, this is why Anthropic shut it down).<p>(3): This is typical Dropbox HN snark[3], of which I am also often guilty of. I really don't think AI coding is a killer product and this seems very myopic—engineers are an extreme minority. Imo, the closest we've seen to something revolutionary is OpenClaw, but it's janky, hard to set up, full of vulnerabilities, and you need to buy a separate computer. But there's certainly a spark there. (And that's personally the vertical I'm focusing on.)<p>[1] <a href="https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks" rel="nofollow">https://www.anthropic.com/news/detecting-and-preventing-dist...</a><p>[2] <a href="https://media.npr.org/assets/artslife/arts/2025/complaint.pdf" rel="nofollow">https://media.npr.org/assets/artslife/arts/2025/complaint.pd...</a><p>[3] <a href="https://news.ycombinator.com/item?id=9224">https://news.ycombinator.com/item?id=9224</a></p>
]]></description><pubDate>Wed, 08 Apr 2026 03:44:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=47684887</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47684887</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47684887</guid></item><item><title><![CDATA[New comment by dvt in "GLM-5.1: Towards Long-Horizon Tasks"]]></title><description><![CDATA[
<p>Every single day, three things are becoming more and more clear:<p><pre><code>    (1) OpenAI & Anthropic are absolutely cooked; it's obvious they have no moat
    (2) Local/private inference is the future of AI
    (3) There's *still* no killer product yet (so get to work!)</code></pre></p>
]]></description><pubDate>Tue, 07 Apr 2026 23:35:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=47682706</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47682706</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47682706</guid></item><item><title><![CDATA[New comment by dvt in "Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B"]]></title><description><![CDATA[
<p>Solid work and great showcase, I've done a bunch of stuff with Kokoro and the latency is incredible. So crazy how badly Apple dropped the ball... feels like your demo should be a Siri demo (I mean that in the most complimentary way possible).</p>
]]></description><pubDate>Mon, 06 Apr 2026 05:41:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=47657363</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47657363</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47657363</guid></item><item><title><![CDATA[New comment by dvt in "Show HN: I built a tiny LLM to demystify how language models work"]]></title><description><![CDATA[
<p>> the user is immediately able to understand the constraints<p>Nagel's point was quite literally the <i>opposite</i>[1] of this, though. We can't understand what it must "be like to be a bat" because their mental model is so fundamentally different than ours. So using all the human language tokens in the world can't get us to truly understand what it's like to be a bat, or a guppy, or whatever. In fact, Nagel's point is arguably even stronger: there's no <i>possible</i> mental mapping between the experience of a bat and the experience of a human.<p>[1] <a href="https://www.sas.upenn.edu/~cavitch/pdf-library/Nagel_Bat.pdf" rel="nofollow">https://www.sas.upenn.edu/~cavitch/pdf-library/Nagel_Bat.pdf</a></p>
]]></description><pubDate>Mon, 06 Apr 2026 02:25:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=47656282</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47656282</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47656282</guid></item><item><title><![CDATA[New comment by dvt in "Student Debt Burdened Them, So They Moved Abroad and Stopped Paying"]]></title><description><![CDATA[
<p>> moderately successful in academia and as far as I can tell<p>Why not pay your debts then? I totally understand debt forgiveness for extenuating circumstances (and imo, it's a crime that student loan debt can't be forgiven, and the interest rates are often predatory—especially in the case of med school and law school), but this just sounds like stealing with extra steps.</p>
]]></description><pubDate>Sat, 04 Apr 2026 21:10:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=47643410</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47643410</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47643410</guid></item><item><title><![CDATA[New comment by dvt in "Tell HN: Anthropic no longer allowing Claude Code subscriptions to use OpenClaw"]]></title><description><![CDATA[
<p>Running locally or privately (in the cloud) is the future. Anthropic/OAI will need to recoup (astronomical) training costs and I'm not going to be their bailout plan, especially considering training was done on torrented & copyrighted data anyway.<p>Public model inference quality is almost at SOTA levels, why would anyone pay these VC-subsidized companies even a cent? For a shitty chat interface? Give me a break.</p>
]]></description><pubDate>Sat, 04 Apr 2026 07:08:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=47636648</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47636648</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47636648</guid></item><item><title><![CDATA[New comment by dvt in "Show HN: Made a little Artemis II tracker"]]></title><description><![CDATA[
<p>To me, what's super interesting about this is the fact that my brain instantly recognized it's AI coded (not sure why, it might be the spacing, the font, the text glow, etc.).</p>
]]></description><pubDate>Fri, 03 Apr 2026 00:35:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=47622020</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47622020</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47622020</guid></item><item><title><![CDATA[New comment by dvt in "What Gödel Discovered (2020)"]]></title><description><![CDATA[
<p>The easiest way to think about it is like this:<p>We have these things called systems: a "system" is anything that follows rules: a board game, traffic, the English language, math, C++, etc. Some systems are smart and they can talk about themselves, but others can't. For example, Tic-Tac-Toe can't talk about Tic-Tac-Toe, but English <i>can</i> talk about English.<p>Gödel is interested in smart systems because dumb systems are boring.<p>Some systems are useful: they are "useful" if they always say true things. So math is more useful than English. I can lie in English, but I can't lie in math. (Formally, this is what we call <i>consistency</i>).<p>So here's a problem for you: suppose we have a smart-useful-system, call it SUS. SUS should be able to say "SUS is useful." It can talk about itself and it can't lie, so we should have no problem, right?<p>Gödel showed that if our system can actually say that about itself, it wasn't useful to begin with. For a few centuries, philosophers and mathematicians were trying to come up with the "one perfect system": useful, smart, and also <i>complete</i> (it can say <i>all</i> true things), and a few more properties. Turns out such a system is impossible.<p>NB: I use the words "say" or "talk about" in a very hand-wavy fashion, sometimes I mean Prove(), sometimes I mean Entail(). The details are very nuanced, and this isn't meant to be a deep dive.</p>
]]></description><pubDate>Thu, 02 Apr 2026 19:32:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=47619101</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47619101</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47619101</guid></item><item><title><![CDATA[New comment by dvt in "What Gödel Discovered (2020)"]]></title><description><![CDATA[
<p>This blog post gets <i>way</i> too caught up in Gödel numbers, which are merely a technical detail (specifically how the encoding is done is irrelevant). A clever detail, but a detail nonetheless. Author gets lost in the sauce and kind of misses the forest for the trees. In class, we used Löb's Theorem[1] to prove Gödel, which is much more grokkable (and arguably even <i>more</i> clever). If you truly get Löb, it'll kind of blow your mind.<p>[1] <a href="https://inference-review.com/article/loebs-theorem-and-currys-paradox" rel="nofollow">https://inference-review.com/article/loebs-theorem-and-curry...</a></p>
]]></description><pubDate>Thu, 02 Apr 2026 04:50:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=47610107</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47610107</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47610107</guid></item><item><title><![CDATA[New comment by dvt in "Jax's true calling: Ray-Marching renderers on WebGL"]]></title><description><![CDATA[
<p>> the thing JAX was truly meant for: a graphics renderer<p>I mean, just like ray-tracing, SDF (ray-marching) is neat, but basically everything useful is expensive or hard to do (collisions, meshes, texturing etc.). I mean mathy stuff is easier (rotations, unions/intersections, function composition, etc.) but 3D is usually used in either modeling software or video games, which care more about the former than they do the latter.</p>
]]></description><pubDate>Wed, 01 Apr 2026 22:18:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=47607264</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47607264</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47607264</guid></item><item><title><![CDATA[New comment by dvt in "Ask HN: Who wants to be hired? (April 2026)"]]></title><description><![CDATA[
<p>Location: Los Angeles
Remote: Yes<p>Relocation: Case-by-case<p>I'm a CTO, expert engineer, and data professional interested in team-building, consulting and architecting data pipelines. At Edmunds.com, I worked on a fairly successful ad-tech product and my team bootstrapped a data pipeline using Spark, Databricks, and microservices built with Java, Python, and Scala.<p>At ATTN:, I re-built an ETL Kubernetes stack, including data loaders and extractors that handle >10,000 API payload extractions daily. I created SOPs for managing data interoperability with Facebook Marketing, Facebook Graph, Instagram Graph, Google DFP, Salesforce, etc.<p>More recently, I was the CTO and co-founder of a gaming startup. We raised over $6M and I was in charge of building out a team of over a dozen remote engineers and designers, with a breadth of experience ranging from Citibank, to Goldman Sachs, to Microsoft. I moved on, but retain significant equity and a board seat.<p>I am also a minority owner of a coffee shop in northern Spain. That I'm a top-tier developer goes without saying. I'm interested in flexing my consulting muscle and can help with best practices, architecture, and hiring.<p>Would love to connect even if it's just for networking!<p>Blog: <a href="https://ai.dvt.name/whos-david/" rel="nofollow">https://ai.dvt.name/whos-david/</a> (under construction)<p>GitHub: <a href="https://github.com/dvx" rel="nofollow">https://github.com/dvx</a><p>Email: [david].[titarenco]@[gmail].[com]</p>
]]></description><pubDate>Wed, 01 Apr 2026 21:11:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=47606631</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47606631</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47606631</guid></item><item><title><![CDATA[New comment by dvt in "$500 GPU outperforms Claude Sonnet on coding benchmarks"]]></title><description><![CDATA[
<p>> They're all slop when the complexity is higher than a mid-tech intermediate engineer though.<p>This right here. Value prop quickly goes out the window when you're building anything novel or hard. I feel that I'm still spending the same amount of time working on stuff, except that now I'm also spending money on models.</p>
]]></description><pubDate>Fri, 27 Mar 2026 06:30:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=47539635</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47539635</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47539635</guid></item><item><title><![CDATA[New comment by dvt in "Is anybody else bored of talking about AI?"]]></title><description><![CDATA[
<p>Why are you lying? From literally the first paragraph of the CFR article:<p>> China is the world’s largest source of carbon emissions, and the air quality of many of its major cities fails to meet international health standards.</p>
]]></description><pubDate>Tue, 24 Mar 2026 22:45:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=47510605</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47510605</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47510605</guid></item><item><title><![CDATA[New comment by dvt in "Is anybody else bored of talking about AI?"]]></title><description><![CDATA[
<p>> incredible amounts of ... energy<p>So tired of seeing this trope. Data center energy expenditure is like less than 1% of worldwide energy expenditure[1]. Have you heard of mining? Or agriculture? Or cars/airplanes/ships? It's just factually wrong and alarmist to spread the fake news that AI has any measurable effect on climate change.<p>[1] <a href="https://www.iea.org/reports/energy-and-ai/energy-supply-for-ai" rel="nofollow">https://www.iea.org/reports/energy-and-ai/energy-supply-for-...</a></p>
]]></description><pubDate>Tue, 24 Mar 2026 20:59:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47509193</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47509193</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47509193</guid></item><item><title><![CDATA[New comment by dvt in "No, Windows Start does not use React"]]></title><description><![CDATA[
<p>I mean, you could say the same thing about Apple, but they're extremely protective of their ecosystem and native capabilities.</p>
]]></description><pubDate>Tue, 24 Mar 2026 00:55:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47497339</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47497339</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47497339</guid></item><item><title><![CDATA[New comment by dvt in "No, Windows Start does not use React"]]></title><description><![CDATA[
<p>The weirdest thing about this (which is frankly bizarre) is Microsoft emphatically shilling React Native for MacOS[1] usage (???). Like, wtf? Why? Not only is it embarassing for MS to be using another competing company's (Facebook's) UI layer when they're, you know, an <i>operating system</i> company. But they're also pushing it for competing operating systems. What idiotic PM signed off on this? How in the world does Microsoft benefit out of promulgating Facebook's technology?<p>[1] <a href="https://microsoft.github.io/react-native-macos/docs/intro" rel="nofollow">https://microsoft.github.io/react-native-macos/docs/intro</a></p>
]]></description><pubDate>Tue, 24 Mar 2026 00:34:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=47497194</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47497194</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47497194</guid></item><item><title><![CDATA[New comment by dvt in "Autoresearch on an old research idea"]]></title><description><![CDATA[
<p>Ok, so looking at the commit log[1], I was mostly interested in seeing what the "moonshot ideas" implementations looked like, but basically everything is just hyperparameter tuning. Which is nice, but likely not worth the $$$ spent on the tokens. Am I missing something here?<p>[1] <a href="https://github.com/ykumards/eCLIP/commits/main/autoresearch" rel="nofollow">https://github.com/ykumards/eCLIP/commits/main/autoresearch</a></p>
]]></description><pubDate>Mon, 23 Mar 2026 19:25:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=47493988</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47493988</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47493988</guid></item><item><title><![CDATA[New comment by dvt in "I Reverse-Engineered the TiinyAI Pocket Lab from Marketing Photos"]]></title><description><![CDATA[
<p>I genuinely feel disrespected if AI is used to write an article and it's not disclosed in the first paragraph. It's not <i>really</i> that big of a deal, tbh, it's like saying "I took a picture, I didn't paint it."<p>Which is fine, but please disclose it. Otherwise, like in this case, I'm going to assume the author is a moron that can't write for shit who thinks their readers are morons that can't read for shit.</p>
]]></description><pubDate>Mon, 23 Mar 2026 05:14:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47485712</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47485712</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47485712</guid></item><item><title><![CDATA[New comment by dvt in "Attention Residuals"]]></title><description><![CDATA[
<p>> I think what they're getting at is that for a given unit of compute, this method achieves 125% performance.<p>This is not what they're getting at; I explained exactly what they're getting at. I mean, your equivalence of "loss" (what authors <i>actually</i> measured) and "performance" is just bizarre. We use benchmarks to measure performance, and the numbers there were like 1-5% better (apart from the GPQA-Diamond outlier).<p>Do people even read these papers?</p>
]]></description><pubDate>Fri, 20 Mar 2026 21:26:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47460810</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47460810</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47460810</guid></item><item><title><![CDATA[New comment by dvt in "Attention Residuals"]]></title><description><![CDATA[
<p>> Drops compute required for training by ~20%.<p>This is not true. Authors claim that w.r.t. training, their method adds negigible overhead for AttnRes with no memory impact (but is way more complicated for Block AttnRes since we need to use pipelining for larger models, hence the O(Ld) & O(Nd) figures, with N ≪ L).<p>> WAY lower bandwidth requirements for inference.<p>Also not true. Paper has nothing to do with inference, apart from the benchmarks. If you're looking at the graph about "compute advantage," it's about training compute. They do some interpolation to get to the 1.25x number, basically answering the question "if non-AttnRes architecture were trained, how much compute would it take to get to the same loss as AttnRes?" (The answer being ~20% more compute.) It's an interesting claim, but there's all kinds of weird and unexpected convergence that can happen, so take it with a grain of salt.</p>
]]></description><pubDate>Fri, 20 Mar 2026 19:45:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=47459635</link><dc:creator>dvt</dc:creator><comments>https://news.ycombinator.com/item?id=47459635</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47459635</guid></item></channel></rss>