<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: Smerity</title><link>https://news.ycombinator.com/user?id=Smerity</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 15 Apr 2026 01:05:19 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=Smerity" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by Smerity in "Ton Roosendaal to step down as Blender chairman and CEO"]]></title><description><![CDATA[
<p>Ton Roosendaal is brilliant both strategically (taking a closed source codebase and growing it into the open source success story that is Blender today) but also as a person.<p>When I was 15 or so, mid 2000s, I was heavily into the Blender community. Shortly after the closed source to open source transition I ran into a segfault. I could half heartedly sling code at the time but not enough to work out the issue I had.<p>After noting the issue on IRC Ton personally helped tease out the bug. I _might_ have misdirected him by saying the bug occurred "with just a cube" (and he was like "Yay! Bugs are easier when they're simple cases!") but I neglected to note it was a cube subdivided half a dozen times, with a few thousand vertices ;)<p>He had a vision and quiet persistent execution plan for Blender and the community that was far ahead of anything else in those (relatively) early days of OSS and the web.<p>The Blender community was and is an amazing combination of technology, creativity, and positivity, and I think we owe Ton for helping steward much of it.</p>
]]></description><pubDate>Thu, 18 Sep 2025 02:27:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=45284414</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=45284414</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45284414</guid></item><item><title><![CDATA[New comment by Smerity in "AI companies cause most of traffic on forums"]]></title><description><![CDATA[
<p>You can download Common Crawl data for free using HTTPS with no credentials. If you don't store it (streamed processing or equivalent) and you have no cost for incoming data (which most clouds don't) you're good!<p>You can do so by adding `<a href="https://data.commoncrawl.org/" rel="nofollow">https://data.commoncrawl.org/</a>` instead of `s3://commoncrawl/` before each of the WARC/WAT/WET paths.</p>
]]></description><pubDate>Wed, 01 Jan 2025 02:05:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=42563422</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=42563422</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42563422</guid></item><item><title><![CDATA[Python isn't just glue, it's an implicit JIT ecosystem]]></title><description><![CDATA[
<p>Article URL: <a href="https://state.smerity.com/smerity/state/01JC39KW1TBEAZTKJJF0BR5ZXD">https://state.smerity.com/smerity/state/01JC39KW1TBEAZTKJJF0BR5ZXD</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42217558">https://news.ycombinator.com/item?id=42217558</a></p>
<p>Points: 4</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 22 Nov 2024 21:35:00 +0000</pubDate><link>https://state.smerity.com/smerity/state/01JC39KW1TBEAZTKJJF0BR5ZXD</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=42217558</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42217558</guid></item><item><title><![CDATA[New comment by Smerity in "Were RNNs all we needed?"]]></title><description><![CDATA[
<p>Excited to see more people working on RNNs but wish their citations were better.<p>In 2016 my team from Salesforce Research published our work on the Quasi-Recurrent Neural Network[1] (QRNN). The QRNN variants we describe are near identical (minGRU) or highly similar (minLSTM) to the work here.<p>The QRNN was used, many years ago now, in the first version of Baidu's speech recognition system (Deep Voice [6]) and as part of Google's handwriting recognition system in Gboard[5] (2019).<p>Even if there are expressivity trade-offs when using parallelizable RNNs they've shown historically they can work well and are low resource and incredibly fast. Very few of the possibilities regarding distillation, hardware optimization, etc, have been explored.<p>Even if you need "exact" recall, various works have shown that even a single layer of attention with a parallelizable RNN can yield strong results. Distillation down to such a model is quite promising.<p>Other recent fast RNN variants such as the RWKV, S4, Mamba et al. include citations to QRNN (2016) and SRU (2017) for a richer history + better context.<p>The SRU work has also had additions in recent years (SRU++), doing well in speech recognition and LM tasks where they found similar speed benefits over Transformers.<p>I note this primarily as the more data points, especially when strongly relevant, the better positioned the research is. A number of the "new" findings from this paper have been previously explored - and do certainly show promise! This makes sure we're asking new questions with new insights (with all the benefit of additional research from ~8 years ago) versus missing the work from those earlier.<p>[1] QRNN paper: <a href="https://arxiv.org/abs/1611.01576" rel="nofollow">https://arxiv.org/abs/1611.01576</a><p>[2] SRU paper: <a href="https://arxiv.org/abs/1709.02755" rel="nofollow">https://arxiv.org/abs/1709.02755</a><p>[3]: SRU++ for speech recognition: <a href="https://arxiv.org/abs/2110.05571" rel="nofollow">https://arxiv.org/abs/2110.05571</a><p>[4]: SRU++ for language modeling: <a href="https://arxiv.org/abs/2102.12459" rel="nofollow">https://arxiv.org/abs/2102.12459</a><p>[5]: <a href="https://research.google/blog/rnn-based-handwriting-recognition-in-gboard/" rel="nofollow">https://research.google/blog/rnn-based-handwriting-recogniti...</a><p>[6]: <a href="https://arxiv.org/abs/1702.07825" rel="nofollow">https://arxiv.org/abs/1702.07825</a></p>
]]></description><pubDate>Fri, 04 Oct 2024 01:31:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=41736930</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=41736930</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41736930</guid></item><item><title><![CDATA[New comment by Smerity in "ChatGPT for Teams"]]></title><description><![CDATA[
<p>I think you've done a great explanation expansion except I believe it's ALiBi ("Attention with Linear Biases Enables Input Length Extrapolation"), a method of positional encoding (i.e. telling the Transformer model how much to weight a distant token when computing the current output token). This has been used on various other LLMs[2].<p>[1]: <a href="https://arxiv.org/abs/2108.12409" rel="nofollow">https://arxiv.org/abs/2108.12409</a><p>[2]: n.b. Ofir Press is co-creator of ALiBi <a href="https://twitter.com/OfirPress/status/1654538361447522305" rel="nofollow">https://twitter.com/OfirPress/status/1654538361447522305</a></p>
]]></description><pubDate>Thu, 11 Jan 2024 10:15:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=38950210</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=38950210</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38950210</guid></item><item><title><![CDATA[New comment by Smerity in "WARC'in the Crawler"]]></title><description><![CDATA[
<p>As mentioned it's trivial across the spread of compression algorithms supporting this type of behaviour (`gzip`, `zstandard`, `zip`, ...), the header in `zip` making it even more convenient as you note!<p>WARC as a format essentially states that unless you have good reason "record at a time" compression is the preferred[1].
The mixture of "technically possible" and "part of spec" is what makes it so useful - any generic WARC tool can support random access, there are explicit fields to index over (URL), and even non-conforming WARC files can be easily rewritten to add such a capability.<p>[1]: <a href="https://iipc.github.io/warc-specifications/specifications/warc-format/warc-1.0/#record-at-time-compression" rel="nofollow noreferrer">https://iipc.github.io/warc-specifications/specifications/wa...</a></p>
]]></description><pubDate>Fri, 22 Dec 2023 08:39:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=38732251</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=38732251</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38732251</guid></item><item><title><![CDATA[New comment by Smerity in "WARC'in the Crawler"]]></title><description><![CDATA[
<p>In the distant past I was the lone engineer of Common Crawl almost a decade ago. Common Crawl heavily leverages the WARC format.<p>My favorite capability of the WARC format borrows from the fact that most compression formats can be written to allow random access. Compression formats such as `gzip` and `zstandard` allow multiple compressed streams to be stuck together and act during decompression as if it's one contiguous file.<p>Hence you can create multiple compressions and literally stick them together:<p><pre><code>  $ echo cat > cat.txt
  $ echo dog > dog.txt
  $ zstd cat.txt dog.txt
  $ cat cat.txt.zst dog.txt.zst > catdog.zst
  $ zstdcat catdog.zst
  cat
  dog
</code></pre>
For files composed of only a textual / clearly delimited format that means you can fairly trivially leap to a different offset assuming each of the inputs is compressed individually. You lose out on some amount of compression but random lookup seems a fairly reasonable tradeoff.
Common Crawl was able to use this to allow entirely random lookups into web crawl datasets dozens / hundreds of terabytes in size without any change in file format for example and utilizing Amazon S3's support for HTTP Range requests[1].<p>Trading compression for random lookup is even more forgiving if you create a separate compression dictionary tailored toward your dataset. For web crawling you'd likely get you the majority of the compression gains back unless pages from the same website are sequentially written which is unlikely in most situations. The website's shared template/s would result in very high compression gains across files which you'd lose by allowing random lookup but most crawlers don't don't operate sequentially so local compression gains are likely smaller than larger.<p>[1]: <a href="https://developer.mozilla.org/en-US/docs/Web/HTTP/Range_requests" rel="nofollow noreferrer">https://developer.mozilla.org/en-US/docs/Web/HTTP/Range_requ...</a></p>
]]></description><pubDate>Fri, 22 Dec 2023 04:40:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=38731000</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=38731000</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38731000</guid></item><item><title><![CDATA[New comment by Smerity in "Everything about the new beast H100"]]></title><description><![CDATA[
<p>That section of the text is ambiguous but it's $250k for 8 x H100s (PCI-e) and $350k for 8 x H100 (SXM).<p>The prices above are plus or minus from Lambda Labs[1].<p>[1]: <a href="https://lambdalabs.com/products/scalar" rel="nofollow noreferrer">https://lambdalabs.com/products/scalar</a></p>
]]></description><pubDate>Wed, 27 Sep 2023 07:55:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=37671510</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=37671510</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37671510</guid></item><item><title><![CDATA[New comment by Smerity in "Evernote to be acquired by Bending Spoons"]]></title><description><![CDATA[
<p>To note, Figma use something inspired by but not quite CRDTs:<p>> Figma isn't using true CRDTs though. CRDTs are designed for decentralized systems where there is no single central authority to decide what the final state should be. There is some unavoidable performance and memory overhead with doing this. Since Figma is centralized (our server is the central authority), we can simplify our system by removing this extra overhead and benefit from a faster and leaner implementation.<p>> It’s also worth noting that Figma's data structure isn't a single CRDT. Instead it's inspired by multiple separate CRDTs and uses them in combination to create the final data structure that represents a Figma document.<p><a href="https://www.figma.com/blog/how-figmas-multiplayer-technology-works/" rel="nofollow">https://www.figma.com/blog/how-figmas-multiplayer-technology...</a></p>
]]></description><pubDate>Thu, 17 Nov 2022 09:01:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=33636516</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=33636516</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=33636516</guid></item><item><title><![CDATA[New comment by Smerity in "Ask HN: Is there a developer laptop that does not suck and is not a Mac in 2022?"]]></title><description><![CDATA[
<p>I have had my XPS 13 laptop die four times in the past two years, needing a replacement logic board each time.<p>The laptop is amazing when it's working, default Linux support too, but so far have literally had to ship it back to Dell for at least two months total since I've used it. Usually the turn around is only a few days but they were waiting on a part at one stage for weeks.<p>I am now looking to do a system exchange with them but am not certain if it's going to go through. At this stage I can't imagine they'd be profitable on this machine given how many times I've had to ship it in for repair.<p>Hence I sadly cannot recommend the XPS 13 even though I truly love it when it's working and laugh with glee at having 32GB of RAM =]</p>
]]></description><pubDate>Mon, 29 Aug 2022 12:13:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=32637224</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=32637224</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=32637224</guid></item><item><title><![CDATA[New comment by Smerity in "Lz_xor"]]></title><description><![CDATA[
<p>The author is focused on GPU texture compression for games[1] so they're not too concerned with fast compression speed. Given a game will be played on (at a minimum) tens of thousands of machines, some of which may be quite limited (e.g. handheld consoles or mobile devices), the game developers are only more than happy to trade off once-off compression times for decompression size / speed. They'd likely only perform this on a later "release build" of their game too.<p>The author mentions in a tweet going from minutes to seconds for compression when switching CPU for GPU[2]. From memory he has made other references to a few seconds for compression being entirely reasonable for such tasks but I can't find a direct reference.<p>[1]: <a href="http://www.binomial.info/" rel="nofollow">http://www.binomial.info/</a><p>[2]: <a href="https://twitter.com/richgel999/status/1476325003662667777" rel="nofollow">https://twitter.com/richgel999/status/1476325003662667777</a></p>
]]></description><pubDate>Wed, 10 Aug 2022 12:18:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=32410678</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=32410678</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=32410678</guid></item><item><title><![CDATA[New comment by Smerity in "Driverless robotaxi fleet paralyzed for hours in San Francisco"]]></title><description><![CDATA[
<p>I saw a version of this myself at Geary and Mason around midnight on June 20th[1]. That it happened again but at a larger scale ten days later is deeply concerning.<p>Given I had no clue what state the vehicles were in, and that they'd start moving without indication, it felt pretty damn worrying.<p>[1]: <a href="https://twitter.com/Smerity/status/1539144641206882304" rel="nofollow">https://twitter.com/Smerity/status/1539144641206882304</a></p>
]]></description><pubDate>Fri, 01 Jul 2022 00:52:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=31940741</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=31940741</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=31940741</guid></item><item><title><![CDATA[Texting Robots: Taming robots.txt with Rust and 34M tests]]></title><description><![CDATA[
<p>Article URL: <a href="https://state.smerity.com/smerity/state/01FZ3813Q79VTTVDHWHFA2A15E">https://state.smerity.com/smerity/state/01FZ3813Q79VTTVDHWHFA2A15E</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=31927830">https://news.ycombinator.com/item?id=31927830</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 30 Jun 2022 01:39:20 +0000</pubDate><link>https://state.smerity.com/smerity/state/01FZ3813Q79VTTVDHWHFA2A15E</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=31927830</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=31927830</guid></item><item><title><![CDATA[New comment by Smerity in "Cohere: The world’s most powerful NLP toolkit"]]></title><description><![CDATA[
<p>I just want to note the replies to this thread are excessively dismissive and toxic. You may not agree with the wording of their advertising ("world's most powerful NLP toolkit" is marketing speak, sure) but going from that to implying the technical side is "only Min-GPT" is tremendously weird. As someone who works in machine learning and specifically language models this is a team I'm keeping an eye on.<p>For anyone who wanted more technical discussion re: ML / LM (though the author notes this work "[does] not reflect the architectures or latencies of my employer's models" i.e. it's an exploratory technical breakdown of general model characteristics) I've appreciated the technical write-ups from @kipperrii (ML ops @ Cohere) recently:<p>- Transformer Inference Arithmetic: <a href="https://carolchen.me/blog/transformer-inference-arithmetic/" rel="nofollow">https://carolchen.me/blog/transformer-inference-arithmetic/</a><p>- Breakdown of H100s for Transformer Inferencing: <a href="https://carolchen.me/blog/h100-inferencing/" rel="nofollow">https://carolchen.me/blog/h100-inferencing/</a></p>
]]></description><pubDate>Fri, 08 Apr 2022 11:47:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=30955762</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=30955762</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=30955762</guid></item><item><title><![CDATA[Texting Robots: Taming robots.txt with Rust and 34M tests]]></title><description><![CDATA[
<p>Article URL: <a href="https://state.smerity.com/smerity/state/01FZ3813Q79VTTVDHWHFA2A15E">https://state.smerity.com/smerity/state/01FZ3813Q79VTTVDHWHFA2A15E</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=30900295">https://news.ycombinator.com/item?id=30900295</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Sun, 03 Apr 2022 20:23:57 +0000</pubDate><link>https://state.smerity.com/smerity/state/01FZ3813Q79VTTVDHWHFA2A15E</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=30900295</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=30900295</guid></item><item><title><![CDATA[Texting Robots: Taming robots.txt with Rust and 34M tests]]></title><description><![CDATA[
<p>Article URL: <a href="https://state.smerity.com/smerity/state/01FZ3813Q79VTTVDHWHFA2A15E">https://state.smerity.com/smerity/state/01FZ3813Q79VTTVDHWHFA2A15E</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=30836233">https://news.ycombinator.com/item?id=30836233</a></p>
<p>Points: 6</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 28 Mar 2022 22:07:59 +0000</pubDate><link>https://state.smerity.com/smerity/state/01FZ3813Q79VTTVDHWHFA2A15E</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=30836233</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=30836233</guid></item><item><title><![CDATA[New comment by Smerity in "A decade of major cache incidents at Twitter"]]></title><description><![CDATA[
<p>What I find most interesting in this is the pseudo detective story of hunting down disappearing post-mortem and "lessons learned" documentation. Optimistically we'd hope that perhaps the older systems no longer reflect the existing systems in any meaningful way (possibly as the org structures and/or software stacks shift and change) and they're no longer relevant.<p>I'd imagine most lost knowledge is not an explicit decision however which means such historical scenarios / documentation / ... are just lost as part of business. Lost knowledge is the default for companies.<p>Twitter is likely better than most given their documentation is all digital and there exist explicit processes to catalogue such incidents. I'd also be curious to see how much of this knowledge has been implicitly exported to their open source codebases.</p>
]]></description><pubDate>Fri, 11 Feb 2022 09:03:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=30299038</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=30299038</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=30299038</guid></item><item><title><![CDATA[A decade of major cache incidents at Twitter]]></title><description><![CDATA[
<p>Article URL: <a href="https://danluu.com/cache-incidents/">https://danluu.com/cache-incidents/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=30295542">https://news.ycombinator.com/item?id=30295542</a></p>
<p>Points: 138</p>
<p># Comments: 26</p>
]]></description><pubDate>Fri, 11 Feb 2022 01:22:15 +0000</pubDate><link>https://danluu.com/cache-incidents/</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=30295542</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=30295542</guid></item><item><title><![CDATA[New comment by Smerity in "Will Cloudflare R2 Win Customers from Amazon S3?"]]></title><description><![CDATA[
<p>From the pricing page I linked to[1]:<p>> There are no charges for outbound data transfer when using a dedicated network connection to the RSTOR network and direct peering is established. These connections are charged separately. To find out more please contact us.<p>I would imagine for enterprise customers you might either be using direct peering / similar or they're able to charge based upon ballpark bandwidth usage. Otherwise I'm unsure as it does seem quite in line with Wasabi otherwise.<p>[1]: <a href="https://console.rstor.space/pricing" rel="nofollow">https://console.rstor.space/pricing</a></p>
]]></description><pubDate>Wed, 06 Oct 2021 23:14:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=28779673</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=28779673</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=28779673</guid></item><item><title><![CDATA[New comment by Smerity in "Will Cloudflare R2 Win Customers from Amazon S3?"]]></title><description><![CDATA[
<p>You're right, I misread it and now can't edit my comment. This puts it all in quite a similar position to Wasabi[1] who have a 1:1 storage:transfer ratio per month and a 90 day minimum storage duration as well. In fact, their price per terabyte is almost exactly the same ($5.99 for Wasabi vs $6 for Rstor).<p>It's a shame as I love aspects of this type of storage service, even with the caveats, but they're not useful if there's no way to pay more for excess transfer. Luckily I think R2 fits that requirement (though paying for more operations vs paying for more transfer).<p>[1]: <a href="https://wasabi.com/paygo-pricing-faq/" rel="nofollow">https://wasabi.com/paygo-pricing-faq/</a></p>
]]></description><pubDate>Wed, 06 Oct 2021 20:44:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=28778184</link><dc:creator>Smerity</dc:creator><comments>https://news.ycombinator.com/item?id=28778184</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=28778184</guid></item></channel></rss>