<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: refibrillator</title><link>https://news.ycombinator.com/user?id=refibrillator</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 03 May 2026 03:43:16 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=refibrillator" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by refibrillator in "The hidden engineering of runways"]]></title><description><![CDATA[
<p>Hmm in distributed computer systems similar patterns exist, e.g. adding jitter to avoid thundering herd effects.<p>This feels like an essential pattern of the universe or something…</p>
]]></description><pubDate>Mon, 26 Jan 2026 23:24:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=46773174</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46773174</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46773174</guid></item><item><title><![CDATA[New comment by refibrillator in "The Most Popular Blogs of Hacker News in 2025"]]></title><description><![CDATA[
<p>Sometimes I wonder if anyone else feels there is a halo effect around certain personalities on this site. When I see someone ending nearly every comment with a link to their blog or pet project, it gives me bad vibes, as if they have ulterior motives. Especially if a majority of their blog posts are content lifted from elsewhere with minimal additions. Perhaps this is just hustle culture, and YC alum status confers immunity from these types of criticisms. Perhaps my only wish is that other voices would bubble to the top in some of these threads.<p>In any case I’m truly grateful for this site as a whole, the good and the bad.</p>
]]></description><pubDate>Sat, 03 Jan 2026 19:57:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=46480872</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46480872</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46480872</guid></item><item><title><![CDATA[New comment by refibrillator in "Scientists unlock brain's natural clean-up system for new treatments for stroke"]]></title><description><![CDATA[
<p>Here’s a starting point:<p><a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC5241507/#B1" rel="nofollow">https://pmc.ncbi.nlm.nih.gov/articles/PMC5241507/#B1</a><p>TLDR: NAC is a derivative of an amino acid called cysteine, as such it is a precursor for one of the most important antioxidants in the body and it can modulate key metabolic pathways associated with good health across a variety of organs, notably for decades it has been a universally successful antidote for acetaminophen (Tylenol) overdose, it’s available over the counter but NAC is not naturally found in foods, eating cysteine-rich foods like chicken turkey yogurt etc is the next best bet.</p>
]]></description><pubDate>Thu, 01 Jan 2026 05:42:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=46451625</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46451625</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46451625</guid></item><item><title><![CDATA[New comment by refibrillator in "France targets Australia-style social media ban for children next year"]]></title><description><![CDATA[
<p>No disrespect but paying to verify age feels absurd, let alone putting a private company in charge of what should be an essential function of the government.<p>How about when you turn 18 or whatever the government gives you a signed JWT that contains your DOB? Anyone who needs to verify your age can check that and simply validate the signature via a public key published by the government.<p>Simply grab a new JWT when you need it, to ensure privacy.<p>And sure, sprinkle in some laws that make it illegal to store or share JWTs for clearly fraudulent intents.<p>> the vast majority of kids don't easily have access to alcohol or cigarettes<p>This feels like it comes from an affluent perspective, where I grew up it was trivial to acquire these things and much worse, there will always be someone’s older brother etc who will do this for $20 because he’s got nothing to lose.</p>
]]></description><pubDate>Wed, 31 Dec 2025 21:38:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=46448632</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46448632</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46448632</guid></item><item><title><![CDATA[New comment by refibrillator in "Nvidia's $20B antitrust loophole"]]></title><description><![CDATA[
<p>H100 has 80 GB of <i>HBM3</i>. There’s only like 37 MB of SRAM on a single chip.</p>
]]></description><pubDate>Sat, 27 Dec 2025 20:31:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=46404975</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46404975</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46404975</guid></item><item><title><![CDATA[New comment by refibrillator in "Weight loss jabs: What happens when you stop taking them"]]></title><description><![CDATA[
<p>Fascinatingly, the body already has a mechanism for this: fasting. One of the many beneficial side effects is rapid mucosal atrophy, decreasing villus height and crypt depth.<p>You can find evidence of this in the literature, but it’s absurdly understudied, because big pharma would rather sell you a subscription to life.<p>Fortunately there are many good people in the world, especially in the field of medicine, who want to help their patients unconditionally. So there are glimmers of hope, like some of the top cardiologists in the world going against status quo and treating patients with fasting regimes instead of surgery.</p>
]]></description><pubDate>Sun, 21 Dec 2025 21:39:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=46348791</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46348791</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46348791</guid></item><item><title><![CDATA[New comment by refibrillator in "Kids Rarely Read Whole Books Anymore. Even in English Class"]]></title><description><![CDATA[
<p>This is hilarious, I don’t even want to know if it’s legit.</p>
]]></description><pubDate>Sun, 14 Dec 2025 00:44:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=46259770</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46259770</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46259770</guid></item><item><title><![CDATA[New comment by refibrillator in "When would you ever want bubblesort? (2023)"]]></title><description><![CDATA[
<p>Love anecdotes like this! But admittedly I feel a bit lost, so please forgive my ignorance when I ask: why does choosing a subset of k integers at random require deduplication? My naive intuition is that sampling without replacement can be done in linear time (hash table to track chosen elements?). I’m probably not understanding the problem formulation here.</p>
]]></description><pubDate>Thu, 11 Dec 2025 06:11:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=46228225</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46228225</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46228225</guid></item><item><title><![CDATA[New comment by refibrillator in "Google unkills JPEG XL?"]]></title><description><![CDATA[
<p>One of the cooler and lesser known features of JPEG XL is a mode to losslessly transcode from JPEG while achieving ~20% space reduction. It’s reversible too because the original entropy coded bitstream is untouched.<p>Notably GCP is rolling this out to their DICOM store API, so you get the space savings of JXL but can transcode on the fly for applications that need to be served JPEG.<p>Only know this because we have tens of PBs in their DICOM store and stand to save a substantial amount of $ on an absurdly large annual bill.<p>Native browser support is on our wishlist and our contacts indicate the chrome team will get there eventually.</p>
]]></description><pubDate>Tue, 02 Dec 2025 00:36:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=46115733</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46115733</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46115733</guid></item><item><title><![CDATA[New comment by refibrillator in "We're learning more about what Vitamin D does"]]></title><description><![CDATA[
<p>Yeah it’s pretty clearly a bot account, or at least someone who likes to copy paste from chatgpt to sound smart.</p>
]]></description><pubDate>Sat, 29 Nov 2025 18:48:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=46089787</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46089787</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46089787</guid></item><item><title><![CDATA[New comment by refibrillator in "Claude Advanced Tool Use"]]></title><description><![CDATA[
<p>> It works better!<p>> I strongly believe it is one of the best technologies for AI agents<p>Do you have any quantitative evidence to support this?<p>Sincere question. I feel it would add some much needed credibility in a space where many folks are abusing the hype wave and low key shilling their products with vibes instead of rigor.</p>
]]></description><pubDate>Mon, 24 Nov 2025 22:03:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=46039938</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=46039938</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46039938</guid></item><item><title><![CDATA[New comment by refibrillator in "We bought the whole GPU, so we're damn well going to use the whole GPU"]]></title><description><![CDATA[
<p>Ha made me chuckle. For those wondering seriously about this, it’s not a viable optimization because weights are not readily compressible via JPEG/DCT, and there are a limited number of these units on the chip which bottlenecks throughout, meaning speed is dwarfed by simply reading uncompressed weights from HBM.</p>
]]></description><pubDate>Thu, 02 Oct 2025 17:47:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=45452948</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=45452948</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45452948</guid></item><item><title><![CDATA[New comment by refibrillator in "We reverse-engineered Flash Attention 4"]]></title><description><![CDATA[
<p>Great exposition, loved the touch of humor. Please do the backward pass when it’s published.<p>As a fellow Tri Dao groupie and lucky duck who gets to build on Hopper/Blackwell clusters, I find it amazing how difficult it is becoming to write kernels that saturate GPU hardware.<p>When I squint, there appears to be a trend emerging across work like FA4, monolithic (mega) kernels, etc. Namely, a subversion of the classic CUDA programming model in the form of fine grained task based parallelism, managed entirely in “user space”.<p>Not exactly sure what’s ahead but I’m strapping in for a wild ride…</p>
]]></description><pubDate>Sun, 28 Sep 2025 00:47:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=45400753</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=45400753</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45400753</guid></item><item><title><![CDATA[New comment by refibrillator in "Why LLMs Can't Write Q/Kdb+: Writing Code Right-to-Left"]]></title><description><![CDATA[
<p>Well “import torch” for example will resolve certain dynamically linked symbols, which must be done first before importing your own .so code that uses libtorch and pybind11. If not you will get a super fun to debug segfault, leaving you staring at gdb backtrace output while you ponder your career choice.<p>This is buried deep in the PyTorch docs and I don’t have the willpower to go find it right now, sorry.</p>
]]></description><pubDate>Thu, 10 Jul 2025 03:46:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=44516992</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=44516992</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44516992</guid></item><item><title><![CDATA[New comment by refibrillator in "Show HN: TokenDagger – A tokenizer faster than OpenAI's Tiktoken"]]></title><description><![CDATA[
<p>Tokenization is typically done on CPU and is rarely (if ever) a bottleneck for training or inference.<p>GPU kernels typically dominate in terms of wall clock time, the only exception might be very small models.<p>Thus the latency of tokenization can essentially be “hidden”, by having the CPU prepare the next batch while the GPU finishes the current batch.</p>
]]></description><pubDate>Mon, 30 Jun 2025 17:49:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=44426054</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=44426054</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44426054</guid></item><item><title><![CDATA[New comment by refibrillator in "Compiling LLMs into a MegaKernel: A path to low-latency inference"]]></title><description><![CDATA[
<p>Hi author(s), the on-GPU interpreter approach looks like a promising path forward, have you seen this strikingly similar concurrent work?<p><a href="https://news.ycombinator.com/item?id=44111673">https://news.ycombinator.com/item?id=44111673</a><p>I find it curious that fundamentals of the CUDA programming model (eg kernel launches) are being subverted in favor of fine grained task based parallelism that ends up using the hardware more effectively. Makes me wonder if CUDA has been holding us back in some ways.<p>What are the chances we see your work land in PyTorch as an experimental backend?<p>Awesome stuff thanks for sharing.<p>P.S. minor typo, your first two paragraphs under part 1 are nearly identical.</p>
]]></description><pubDate>Fri, 20 Jun 2025 01:33:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=44324001</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=44324001</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44324001</guid></item><item><title><![CDATA[New comment by refibrillator in "Tokasaurus: An LLM inference engine for high-throughput workloads"]]></title><description><![CDATA[
<p>The code has few comments but gotta love when you can tell someone was having fun!<p><a href="https://github.com/ScalingIntelligence/tokasaurus/blob/65efbfb5457b2b08dd562d4658f99ff2687f078e/tokasaurus/manager/scheduler.py#L588">https://github.com/ScalingIntelligence/tokasaurus/blob/65efb...</a><p>I’m honestly impressed that a pure python implementation can beat out vLLM and SGLang. Granted they lean on FlashInfer, and of course torch.compile has gotten incredibly powerful in the last few years. Though dynamic shapes have still been a huge thorn in my side, I’ll need to look closer at how they pulled it off…</p>
]]></description><pubDate>Fri, 06 Jun 2025 02:05:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=44197239</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=44197239</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44197239</guid></item><item><title><![CDATA[New comment by refibrillator in "Why DeepSeek is cheap at scale but expensive to run locally"]]></title><description><![CDATA[
<p>> Unsloth Dynamic GGUF which, quality wise in real-world use performs very close to the original<p>How close are we talking?<p>I’m not calling you a liar OP, but in general I wish people perpetuating such broad claims would be more rigorous.<p>Unsloth does amazing work, however as far as I’m aware even they themselves do not publish head to head evals with the original unquantized models.<p>I have sympathy here because very few people and companies can afford to run the original models, let alone engineer rigorous evals.<p>However I felt compelled to comment because my experience does not match. For relatively simple usage the differences are hard to notice, but they become much more apparent in high complexity and long context tasks.</p>
]]></description><pubDate>Sun, 01 Jun 2025 20:59:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=44153725</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=44153725</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44153725</guid></item><item><title><![CDATA[New comment by refibrillator in "Lossless LLM compression for efficient GPU inference via dynamic-length float"]]></title><description><![CDATA[
<p>Note to others reading along: in the last appendix page the OP paper reports DFloat11 reduces tokens/sec by ~2-3x for the Llama-3.1-8b and Qwen-2.5-14b/32b and Mistral-small-24b models (throughput penalty not reported for others).<p>Using DFloat11, tokens/sec was higher only when compared relative to running inference with some layers offloaded to CPU.<p>Classic comp sci tradeoff between space and speed, no free lunch, etc.</p>
]]></description><pubDate>Sat, 26 Apr 2025 02:48:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=43800470</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=43800470</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43800470</guid></item><item><title><![CDATA[New comment by refibrillator in "The Llama 4 herd"]]></title><description><![CDATA[
<p>> the actual processing happens in 17B<p>This is a common misconception of how MoE models work. To be clear, 17B parameters are activated for <i>each token generated</i>.<p>In practice you will almost certainly be pulling the full 109B parameters though the CPU/GPU cache hierarchy to generate non-trivial output, or at least a significant fraction of that.</p>
]]></description><pubDate>Sat, 05 Apr 2025 19:08:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=43595987</link><dc:creator>refibrillator</dc:creator><comments>https://news.ycombinator.com/item?id=43595987</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43595987</guid></item></channel></rss>