<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: RandomBK</title><link>https://news.ycombinator.com/user?id=RandomBK</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 17 Apr 2026 03:28:02 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=RandomBK" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by RandomBK in "Why XML tags are so fundamental to Claude"]]></title><description><![CDATA[
<p>How well do we understand the tokenization for Claude? I'd posit that the exact human-representation of this markup is likely irrelevant if it's all being converted into a single token.</p>
]]></description><pubDate>Mon, 02 Mar 2026 08:00:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=47215068</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=47215068</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47215068</guid></item><item><title><![CDATA[New comment by RandomBK in "The Cost of a Function Call"]]></title><description><![CDATA[
<p>Code length will itself become a problem. The instruction cache is limited in size and often quite small. Bloating instruction counts with lots of duplicated code will eventually have a negative effect on performance.<p>Ultimately, there's too many factors to predetermine which approach is faster. Write clean code, and let a profiler guide optimizations when needed.</p>
]]></description><pubDate>Mon, 09 Feb 2026 09:47:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=46943468</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=46943468</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46943468</guid></item><item><title><![CDATA[New comment by RandomBK in "Claude Shannon's randomness-guessing machine"]]></title><description><![CDATA[
<p>Additionally, so long as we can be sure the human's output is not <i>actively adversarial</i>, we can xor it into the entropy pool. Entropy can only increase this way.</p>
]]></description><pubDate>Sun, 18 Jan 2026 08:47:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=46666022</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=46666022</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46666022</guid></item><item><title><![CDATA[New comment by RandomBK in "Outside, Dungeon, Town: Integrating the Three Places in Videogames (2024)"]]></title><description><![CDATA[
<p>I'm reminded of a diagram from the pitch doc for the original Diablo [0] that made its rounds across the web recently. The dungeon/town split was particularly sharp back then, but the broad design has stuck with modern ARPG design, either in the form of safe zones around town or explicit town zones.<p>A lot of this seems to be due to modern multiplayer design, with shared town instances and (usually) private dungeon/outside instances.<p>[0] <a href="https://www.gamedeveloper.com/business/here-s-a-look-at-the-original-design-pitch-document-for-i-diablo-i-" rel="nofollow">https://www.gamedeveloper.com/business/here-s-a-look-at-the-...</a> (scroll down)</p>
]]></description><pubDate>Tue, 30 Dec 2025 04:17:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=46429480</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=46429480</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46429480</guid></item><item><title><![CDATA[New comment by RandomBK in "Titan submersible’s $62 SanDisk memory card found undamaged at wreckage site"]]></title><description><![CDATA[
<p>Good catch; it was somewhat ambiguous in the report.</p>
]]></description><pubDate>Sat, 18 Oct 2025 22:59:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=45630976</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=45630976</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45630976</guid></item><item><title><![CDATA[New comment by RandomBK in "Titan submersible’s $62 SanDisk memory card found undamaged at wreckage site"]]></title><description><![CDATA[
<p>I see a lot of discussion in this thread stemming from some confusion+not reading the actual report[0].<p>Some key points:<p>1. The Camera+Card was encased in a separate enclosure made of titanium+sapphire, and did not seem to be exposed to extreme pressures.<p>2. The encryption was done via a variant of LUKS/dm-crypt, with the key stored on the NVRAM of a chip (Edited; not in TrustZone).<p>3. The recovery was done by transplanting the original chip onto a new working board. No manufacturer backdoors or other hidden mechanisms were used.<p>4. Interestingly, the camera vendor didn't seem to realize there was any encryption at all.<p>[0] <a href="https://data.ntsb.gov/Docket/Document/docBLOB?ID=18741602&FileExtension=pdf&FileName=Underwater%20Camera%20-%20Specialist%27s%20Factual%20Report-Rel.pdf" rel="nofollow">https://data.ntsb.gov/Docket/Document/docBLOB?ID=18741602&Fi...</a></p>
]]></description><pubDate>Sat, 18 Oct 2025 21:39:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=45630550</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=45630550</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45630550</guid></item><item><title><![CDATA[New comment by RandomBK in "Google Safe Browsing incident"]]></title><description><![CDATA[
<p>> Getting on the public suffix list is easier said than done [1].<p>Can you elaborate on this? I didn't see anything in either link that would indicate unreasonable challenges. The PSL naturally has a a series of validation requirements, but I haven't heard of any undue shenanigans.<p>Is it great that such vital infrastructure is held together by a ragtag band of unpaid volunteers? No; but that's hardly unique in this space.</p>
]]></description><pubDate>Fri, 10 Oct 2025 23:08:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=45544854</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=45544854</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45544854</guid></item><item><title><![CDATA[Show HN: Open-source Voice Cloning at 16x real-time: Porting Chatterbox to vLLM]]></title><description><![CDATA[
<p>Chatterbox TTS from ResembleAI (<a href="https://github.com/resemble-ai/chatterbox">https://github.com/resemble-ai/chatterbox</a>) is one of the most accessible and highest-quality Voice Cloning models available today. However, its implementation via HF Transformers left a lot of performance on the table.<p>This is a pet project I've been building on-and-off. It ports the core of Chatterbox - a 0.5B Llama-architecture model - to vLLM. A lot of ugly hacks and workarounds were needed to make it work, but the end result <i>works</i>.<p>Outputting at around the same quality level as the original implementation, this port is roughly 5-10x faster, generating a 40min benchmark output in around 2min30s wall time on a 3090 (or 4m30s on a 3060ti). That's almost 16x faster than real-time.<p>High throughput like this can be itself transformative, enabling scale and efficiency that unblocks new use-cases. I look forward to seeing what the community can do with this!<p><i>Disclaimer: This is a personal community project not affiliated with ResembleAI, my employer, or any other entity. The project is based solely on publicly-available information. All opinions are my own and do not necessarily represent the views of my employer.</i></p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=44777452">https://news.ycombinator.com/item?id=44777452</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Sun, 03 Aug 2025 16:00:08 +0000</pubDate><link>https://github.com/randombk/chatterbox-vllm</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=44777452</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44777452</guid></item><item><title><![CDATA[New comment by RandomBK in "Low-background Steel: content without AI contamination"]]></title><description><![CDATA[
<p>My 2c is that it <i>is</i> worthwhile to train on AI generated content that has obtained some level of human approval or interest, as a form of extended RLHF loop.</p>
]]></description><pubDate>Tue, 10 Jun 2025 22:16:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=44242089</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=44242089</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44242089</guid></item><item><title><![CDATA[New comment by RandomBK in "Pyrefly vs. Ty: Comparing Python's two new Rust-based type checkers"]]></title><description><![CDATA[
<p>It reminds me of the early days of Typescript rollout, which similarly focused on a smooth on-boarding path for existing large projects.<p>More restrictive requirements (ie `noImplicitAny`) could be turned on one at a time before eventually flipping the `strict` switch to opt in to all the checks.</p>
]]></description><pubDate>Tue, 27 May 2025 16:34:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=44108503</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=44108503</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44108503</guid></item><item><title><![CDATA[Show HN: Kernel-level LLM inference via /dev/llm0]]></title><description><![CDATA[
<p>I saw an April Fools joke and decided to implement it.<p>This is a rough port of llm.c into a kernel module. A lot of hacks were needed to make this happen, so a <i>lot</i> of performance was left on the table. Nevertheless, it is a minimally functional GPT2 inference loop running in the kernel.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43558042">https://news.ycombinator.com/item?id=43558042</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 02 Apr 2025 15:49:06 +0000</pubDate><link>https://github.com/randombk/kllm</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=43558042</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43558042</guid></item><item><title><![CDATA[Claude Plays Pokémon, a Visual Guide]]></title><description><![CDATA[
<p>Article URL: <a href="https://excalidraw.com/#json=sOKWE69vwJenBN5m9Vdw7,Iaib0piW6NYc3Mdqf7-c-A">https://excalidraw.com/#json=sOKWE69vwJenBN5m9Vdw7,Iaib0piW6NYc3Mdqf7-c-A</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43175469">https://news.ycombinator.com/item?id=43175469</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 25 Feb 2025 18:24:37 +0000</pubDate><link>https://excalidraw.com/#json=sOKWE69vwJenBN5m9Vdw7,Iaib0piW6NYc3Mdqf7-c-A</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=43175469</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43175469</guid></item><item><title><![CDATA[New comment by RandomBK in "Introduction to Stochastic Calculus"]]></title><description><![CDATA[
<p>The way the question was framed, it was ambiguous whether "draw again" only applied to B, or whether A would draw again as well. I'm assuming the 'infinity' answer applies only to the former case?</p>
]]></description><pubDate>Mon, 24 Feb 2025 19:50:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=43164090</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=43164090</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43164090</guid></item><item><title><![CDATA[New comment by RandomBK in "“A calculator app? Anyone could make that”"]]></title><description><![CDATA[
<p>> IEEE754 is not great for pure maths, however, it is fine for real life.<p>Partially. It <i>can</i> be fine for pretty much any real-life use case. But many naive implementations of formulae involve some gnarly intermediates despite having fairly mundane inputs and outputs.</p>
]]></description><pubDate>Mon, 17 Feb 2025 02:48:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=43074481</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=43074481</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43074481</guid></item><item><title><![CDATA[New comment by RandomBK in "All Kindles can now be jailbroken"]]></title><description><![CDATA[
<p>The entire frontpage was just 468KB, 91KB compressed, and most of that was the search index. Very neat!</p>
]]></description><pubDate>Mon, 17 Feb 2025 02:36:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=43074395</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=43074395</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43074395</guid></item><item><title><![CDATA[New comment by RandomBK in "Inside OpenAI's $14M Super Bowl debut"]]></title><description><![CDATA[
<p>You mean $2/customer? That doesn't actually sound that bad, and is in line with many other marketing efforts.</p>
]]></description><pubDate>Mon, 10 Feb 2025 05:29:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=42997223</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=42997223</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42997223</guid></item><item><title><![CDATA[New comment by RandomBK in "Running DeepSeek R1 Models Locally on NPU"]]></title><description><![CDATA[
<p>Yup. I was referring to the 1.58B quant which seemed to be performing alright and would be the smallest real-DeepSeek model. That requires ~140GB, which is just barely doable on a 128GB RAM + 24GB VRAM setup + a lot of patience. Others have made it work at 64GB RAM + a fast SSD.<p>The true minimally-quantized DeepSeek experience will need one or possibly two 8xH100 nodes, so well upwards of $100K in CapEx.</p>
]]></description><pubDate>Tue, 04 Feb 2025 04:16:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=42927854</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=42927854</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42927854</guid></item><item><title><![CDATA[New comment by RandomBK in "Running DeepSeek R1 Models Locally on NPU"]]></title><description><![CDATA[
<p>The only 32B distill I'm aware of is `DeepSeek-R1-Distill-Qwen-32B`, which would be a base model of `Qwen-32B` distilled (further trained) on outputs from the full R1 model.</p>
]]></description><pubDate>Sat, 01 Feb 2025 09:03:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=42896998</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=42896998</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42896998</guid></item><item><title><![CDATA[New comment by RandomBK in "Running DeepSeek R1 Models Locally on NPU"]]></title><description><![CDATA[
<p>Reminder: DeepSeek distilled models are better thought of as fine-tunes of Qwen/Llama using DeepSeek output, and are not the same as actual DeepSeek v3 or R1.<p>This unfortunate naming has sown plenty of confusion around DeepSeek's quality and resource requirements. <i>Actual</i> DeepSeek v3/R1 continues to require at least ~100GB of VRAM/Mem/SSD, and this does not change that.</p>
]]></description><pubDate>Sat, 01 Feb 2025 08:26:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=42896812</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=42896812</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42896812</guid></item><item><title><![CDATA[New comment by RandomBK in "3blue1brown YouTube Bitcoin video taken down as copyright violation"]]></title><description><![CDATA[
<p>On the other hand, 3B1B's audience tends to heavily bias towards the tech crowd. I'm at a FAANG, and a decent number of our senior engineers know the channel.<p>I agree it's not enough to directly push policy, but the impact is certainly larger than what the subscriber count might otherwise suggest.</p>
]]></description><pubDate>Mon, 06 Jan 2025 19:35:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=42614581</link><dc:creator>RandomBK</dc:creator><comments>https://news.ycombinator.com/item?id=42614581</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42614581</guid></item></channel></rss>