<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: nodja</title><link>https://news.ycombinator.com/user?id=nodja</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 12 Apr 2026 09:17:54 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=nodja" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by nodja in "Small models also found the vulnerabilities that Mythos found"]]></title><description><![CDATA[
<p>You misunderstood.<p>Instead of asking the model: "Here's this codebase, report any vulnerability." you ask. "Here's this codebase, report any vulnerability in module\main.c".<p>The model can still explore references and other files inside the codebase, but you start over a new context/session for each file in the codebase.</p>
]]></description><pubDate>Sat, 11 Apr 2026 22:55:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47734711</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=47734711</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47734711</guid></item><item><title><![CDATA[New comment by nodja in "Claude mixes up who said what and that's not OK"]]></title><description><![CDATA[
<p>Anyone familiar with the literature knows if anyone tried figuring out why we don't add "speaker" embeddings? So we'd have an embedding purely for system/assistant/user/tool, maybe even turn number if i.e. multiple tools are called in a row. Surely it would perform better than expecting the attention matrix to look for special tokens no?</p>
]]></description><pubDate>Thu, 09 Apr 2026 12:58:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47703137</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=47703137</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47703137</guid></item><item><title><![CDATA[New comment by nodja in "Tell HN: Anthropic no longer allowing Claude Code subscriptions to use OpenClaw"]]></title><description><![CDATA[
<p>You can charge $10 on the account and get unlimited requests. I abused this last week with the nemotron super to test out some stuff and made probably over 10000 requests over a couple of days and didn't get blocked or anything, expect 5xx errors and slowdowns tho.</p>
]]></description><pubDate>Sat, 04 Apr 2026 03:00:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=47635231</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=47635231</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47635231</guid></item><item><title><![CDATA[New comment by nodja in "Cohere Transcribe: Speech Recognition"]]></title><description><![CDATA[
<p>It's probably another ASR model that focuses on benchmarks and simple uses instead of more challenging real use cases.<p>I upload edited gameplay vods of twitch streams on youtube, and use whisper-large-v3 to provide subtitles for accessibility reasons (youtube's own auto-subtitles suck, tho they've been getting better).<p>My checklist for a good ASR model for my use case is:<p>1. Have timestamp support.<p>2. Support overlapping speakers.<p>3. Accurate transcripts that don't coalesce half words/interrupted sentences.<p>4. Support non verbal stuff like [coughs], [groans], [laughs], [sighs], etc.<p>5. Allow context injection of non-trivial sizes (10k+ words)<p>1 is obvious because without it we can't have subtitles. Force alignment fails too often.<p>2 is crucial for real world scenarios because in the real world people talk over each other all the time, in my case it's a streamer talking over gameplay audio, or when the streamer has guests over. When 2 people speak the transcript either ignores one of them, or in the worst case, both of them.<p>3 and 4 are an accessibility thing, if you're deaf or hard of hearing having a more literal transcript of what's being said conveys better how the speaker is speaking. If all subtitles are properly "spell-checked" then it's clear your model is overfit to the benchmarks.<p>5 Is not a requirement per se, but more of a nice to have. In my use cause the streamer is often reading stream chat so feeding the model the list of users that recently talked, recent chat messages, text on screen, etc. Would make for more accurate transcripts.<p>I've tried many models, and the closest that fulfill my needs are LLM style models on top of forced alignment. It's too slow, so I've been sticky with whisper because with whisperx I can get a transcript in 5 minutes with just a single command.<p>One thing all these models do (including whisper) is just omit full sentences, it's the worst thing a model can do.</p>
]]></description><pubDate>Tue, 31 Mar 2026 23:27:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=47594825</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=47594825</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47594825</guid></item><item><title><![CDATA[New comment by nodja in "The Claude Code Source Leak: fake tools, frustration regexes, undercover mode"]]></title><description><![CDATA[
<p>This blog post looks to be partially AI generated as well...</p>
]]></description><pubDate>Tue, 31 Mar 2026 22:46:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=47594473</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=47594473</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47594473</guid></item><item><title><![CDATA[New comment by nodja in "Claude Code's source code has been leaked via a map file in their NPM registry"]]></title><description><![CDATA[
<p>If anyone at anthropic is reading this and wants more logs from me add jfc.</p>
]]></description><pubDate>Tue, 31 Mar 2026 10:56:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=47585505</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=47585505</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47585505</guid></item><item><title><![CDATA[New comment by nodja in "Nvidia greenboost: transparently extend GPU VRAM using system RAM/NVMe"]]></title><description><![CDATA[
<p>NVIDIA's GPU drivers on windows 100% do this<p><a href="https://i.imgur.com/c0a3vUy.png" rel="nofollow">https://i.imgur.com/c0a3vUy.png</a></p>
]]></description><pubDate>Thu, 19 Mar 2026 02:59:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=47434355</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=47434355</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47434355</guid></item><item><title><![CDATA[New comment by nodja in "Unsloth Studio"]]></title><description><![CDATA[
<p>IDK how it did but it detected my LM studio downloaded models I have on a spinning drive (they're not in the default location).</p>
]]></description><pubDate>Wed, 18 Mar 2026 00:47:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47420360</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=47420360</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47420360</guid></item><item><title><![CDATA[New comment by nodja in "Intel XeSS 3: expanded support for Core Ultra/Core Ultra 2 and Arc A, B series"]]></title><description><![CDATA[
<p>> It's like people let their hate of AI and LLM bubble blind them, and their brains can't compartmentalize good from bad news anymore.<p>DLSS is also AI and people like it.<p>People don't like framegen because the manufacturers are not being honest about it and using it for deceptive hype marketing. Anyone with a brain knows that it introduces latency and is only useful if you're already 40+ FPS, we also know that companies will use it to pad benchmarks. NVIDIA themselves said that the 5070 had 4090 performance because it supports framegen.</p>
]]></description><pubDate>Tue, 24 Feb 2026 11:26:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=47135768</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=47135768</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47135768</guid></item><item><title><![CDATA[New comment by nodja in "Claude Opus 4.6"]]></title><description><![CDATA[
<p>Doing some math in my head, buying the GPUs at retail price, it would take probably around half a year to make the money back, probably more depending how expensive electricity is in the area you're serving from. So I don't know where this "losing money" rhetoric is coming from. It's probably harder to source the actual GPUs than making money off them.</p>
]]></description><pubDate>Thu, 05 Feb 2026 21:06:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=46905334</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46905334</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46905334</guid></item><item><title><![CDATA[New comment by nodja in "Claude Opus 4.6"]]></title><description><![CDATA[
<p>> A year or more ago, I read that both Anthropic and OpenAI were losing money on every single request even for their paid subscribers, and I don't know if that has changed with more efficient hardware/software improvements/caching.<p>This is obviously not true, you can use real data and common sense.<p>Just look up a similar sized open weights model on openrouter and compare the prices. You'll note the similar sized model is often much cheaper than what anthropic/openai provide.<p>Example: Let's compare claude 4 models with deepseek. Claude 4 is ~400B params so it's best to compare with something like deepseek V3 which is 680B params.<p>Even if we compare the cheapest claude model to the most expensive deepseek provider we have claude charging $1/M for input and $5/M for output, while deepseek providers charge $0.4/M and $1.2/M, a fifth of the price, you can get it as cheap as $.27 input $0.4 output.<p>As you can see, even if we skew things overly in favor of claude, the story is clear, claude token prices are much higher than they could've been. The difference in prices is because anthropic also needs to pay for training costs, while openrouter providers just need to worry on making serving models profitable. Deepseek is also not as capable as claude which also puts down pressure on the prices.<p>There's still a chance that anthropic/openai models are losing money on inference, if for example they're somehow much larger than expected, the 400B param number is not official, just speculative from how it performs, this is only taking into account API prices, subscriptions and free user will of course skew the real profitability numbers, etc.<p>Price sources:<p><a href="https://openrouter.ai/deepseek/deepseek-v3.2-speciale" rel="nofollow">https://openrouter.ai/deepseek/deepseek-v3.2-speciale</a><p><a href="https://claude.com/pricing#api" rel="nofollow">https://claude.com/pricing#api</a></p>
]]></description><pubDate>Thu, 05 Feb 2026 20:08:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=46904470</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46904470</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46904470</guid></item><item><title><![CDATA[New comment by nodja in "Voxtral Transcribe 2"]]></title><description><![CDATA[
<p>There is <a href="https://huggingface.co/spaces/hf-audio/open_asr_leaderboard" rel="nofollow">https://huggingface.co/spaces/hf-audio/open_asr_leaderboard</a> but it hasn't been updated for half a year.</p>
]]></description><pubDate>Wed, 04 Feb 2026 18:50:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=46889955</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46889955</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46889955</guid></item><item><title><![CDATA[New comment by nodja in "Comma openpilot – Open source driver-assistance"]]></title><description><![CDATA[
<p>> My <device> already comes with built in <software> why would I install anything else?<p>Top voted comment on hacker news btw.<p>Ok that was probably unnecessarily snarky I hope you don't take offense, but it seems the hacker spirit has been fading more often from this site, we used to replace stuff with inferior versions just to see if we could.</p>
]]></description><pubDate>Sat, 24 Jan 2026 04:48:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=46741144</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46741144</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46741144</guid></item><item><title><![CDATA[New comment by nodja in "Skip is now free and open source"]]></title><description><![CDATA[
<p>I've been programming since ~1999 and anecdataly don't remember programmers having a culture of paying for their dev tools. On linux everything's free, and on windows I've used a plethora of freeware IDEs/compilers/etc. from turbo pascal, dev c++ (that's the name of it), later on eclipse took the stage in it's buggy mess and right before vscode there was atom. The only people that I know that used visual studio either got it for free for being a student/teacher, had their job pay for it, or most commonly: pirated it.<p>According to this[1] site visual studio had a 35.6% marketshare, tied at #1 with notepad++.<p>[1] <a href="https://asterisk.dynevor.org/editor-dominance.html" rel="nofollow">https://asterisk.dynevor.org/editor-dominance.html</a></p>
]]></description><pubDate>Thu, 22 Jan 2026 00:58:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=46713850</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46713850</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46713850</guid></item><item><title><![CDATA[New comment by nodja in "Which AI Lies Best? A game theory classic designed by John Nash"]]></title><description><![CDATA[
<p><a href="https://www.youtube.com/watch?v=JhBtg-lyKdo" rel="nofollow">https://www.youtube.com/watch?v=JhBtg-lyKdo</a> - 10 AIs Play Mafia<p><a href="https://www.youtube.com/watch?v=GMLB_BxyRJ4" rel="nofollow">https://www.youtube.com/watch?v=GMLB_BxyRJ4</a> - 10 AIs Play Mafia: Vigilante Edition<p><a href="https://www.youtube.com/watch?v=OwyUGkoLgwY" rel="nofollow">https://www.youtube.com/watch?v=OwyUGkoLgwY</a> - 1 Human vs 10 AIs Mafia</p>
]]></description><pubDate>Tue, 20 Jan 2026 23:45:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=46699267</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46699267</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46699267</guid></item><item><title><![CDATA[New comment by nodja in "Nanolang: A tiny experimental language designed to be targeted by coding LLMs"]]></title><description><![CDATA[
<p>I think you need to either feed it all of ./docs or give your agent access to those files so it can read them as reference. The MEMORY.md file you posted mentions ./docs/CANONICAL_STYLE.md and ./docs/LLM_CORE_SUBSET.md and they in turn mention indirectly other features and files inside the docs folder.</p>
]]></description><pubDate>Mon, 19 Jan 2026 23:59:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=46686269</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46686269</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46686269</guid></item><item><title><![CDATA[New comment by nodja in "Inside The Internet Archive's Infrastructure"]]></title><description><![CDATA[
<p>It's insane to me that in 2008 a bunch of pervs decentralized storage and made hentai@home to host hentai comics. Yet here we are almost 20 years later and we haven't generalized this solution. Yes I'm aware of the privacy issues h@h has (as a hoster you're exposing your real IP and people reading comics are exposing their IP to you) but those can be solved with tunnels, the real value is the redundant storage.</p>
]]></description><pubDate>Thu, 15 Jan 2026 20:38:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=46638937</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46638937</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46638937</guid></item><item><title><![CDATA[New comment by nodja in "Nightshade: Make images unsuitable for model training"]]></title><description><![CDATA[
<p>I've run the first of the sample images through 3 captioning models, an old old ViT based booru style tagger, a more recent one and qwen 3 omni. All models successfully identified visual features of the image with no false positives at significant thresholds (>0.3 confidence)<p>I don't know what nightshade is supposed to do, but the fact that it doesn't affect the synthetic labeling of data at all leads me to believe image model trainers will have close to 0 consideration of what it does when training new models.</p>
]]></description><pubDate>Sun, 04 Jan 2026 16:14:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=46489351</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46489351</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46489351</guid></item><item><title><![CDATA[New comment by nodja in "FracturedJson"]]></title><description><![CDATA[
<p>On one hand, it has made json more ubiquitous due to it's frozen state. On another hand, it forces everyone to move to something else and fragments progress. It would be much easier for people to move to json 2.0 rather than having hundreds of json + x standards. Everyone is just reinventing json with their own little twist that I feel sad that we haven't standardized to a single solution that doesn't go super crazy like xml.<p>I don't disagree with the choice, but seeing how things turned out I can't just help but look at the greener grass on the other side.</p>
]]></description><pubDate>Fri, 02 Jan 2026 14:15:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=46464932</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46464932</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46464932</guid></item><item><title><![CDATA[New comment by nodja in "Python numbers every programmer should know"]]></title><description><![CDATA[
<p>I think a lot of commenters here are missing the point.<p>Looking at performance numbers is important regardless if it's python, assembly or HDL. If you don't understand why your code is slow you can always look at how many cycles things take and learn to understand how code works at a deeper level, as you mature as a programmer things will become obvious, but going through the learning process and having references like these will help you to get there sooner, seeing the performance numbers and asking why some things take much longer—or sometimes why they take the exact same time—is the perfect opportunity to learn.<p>Early in my python career I had a python script that found duplicate files across my disks, the first iteration of the script was extremely slow, optimizing the script went through several iterations as I learned how to optimize at various levels. None of them required me to use C. I just used caching, learned to enumerate all files on disk fast, and used sets instead of lists. The end result was that doing subsequent runs made my script run in 10 seconds instead of 15 minutes. Maybe implementing in C would make it run in 1 second, but if I had just assumed my script was slow because of python then I would've spent hours doing it in C only to go from 15 minutes to 14 minutes and 51 seconds.<p>There's an argument to be made that it would be useful to see C numbers next to the python ones, but for the same reason people don't just tell you to just use an FPGA instead of using C, it's also rude to say python is the wrong tool when often it isn't.</p>
]]></description><pubDate>Thu, 01 Jan 2026 21:57:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=46458511</link><dc:creator>nodja</dc:creator><comments>https://news.ycombinator.com/item?id=46458511</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46458511</guid></item></channel></rss>