<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: tmostak</title><link>https://news.ycombinator.com/user?id=tmostak</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 10 Apr 2026 11:11:04 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=tmostak" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by tmostak in "Waymo robotaxi hits a child near an elementary school in Santa Monica"]]></title><description><![CDATA[
<p>Evidence (preferably with recent Teslas/HW4)?</p>
]]></description><pubDate>Fri, 30 Jan 2026 00:22:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=46818946</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=46818946</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46818946</guid></item><item><title><![CDATA[New comment by tmostak in "Waymo robotaxi hits a child near an elementary school in Santa Monica"]]></title><description><![CDATA[
<p>Evidence of this? I own a Tesla (HW4, latest FSD) as well as have taken many Waymo rides, and have found both to react well to unpredictable situations (i.e. a car unexpectedly turning in front of you), far more quickly than I would expect most human drivers to react.<p>This certainly may have been true of older Teslas with HW3 and older FSD builds (I had one, and yes you couldn't trust it).</p>
]]></description><pubDate>Fri, 30 Jan 2026 00:21:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=46818933</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=46818933</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46818933</guid></item><item><title><![CDATA[New comment by tmostak in "Waymo robotaxi hits a child near an elementary school in Santa Monica"]]></title><description><![CDATA[
<p>Do you have data to back this claim up, specifically with HW4 (most recent hardware) and FSD software releases?</p>
]]></description><pubDate>Fri, 30 Jan 2026 00:15:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=46818892</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=46818892</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46818892</guid></item><item><title><![CDATA[New comment by tmostak in "Prefix sum: 20 GB/s (2.6x baseline)"]]></title><description><![CDATA[
<p>Even without NVLink C2C, on a GPU with 16XPCIe 5.0 lanes to host, you have 128GB/sec in theory and 100+ GB/sec in practice bidirectional bandwidth (half that in each direction), so still come out ahead with pipelining.<p>Of course prefix sums are often used within a series of other operators, so if these are already computed on GPU, you come out further ahead still.</p>
]]></description><pubDate>Tue, 14 Oct 2025 20:05:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=45584146</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=45584146</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45584146</guid></item><item><title><![CDATA[New comment by tmostak in "Modern Minimal Perfect Hashing: A Survey"]]></title><description><![CDATA[
<p>We've made extensive use of perfect hashing in HeavyDB (formerly MapD/OmniSciDB), and it has definitely been a core part of achieving strong group by and join performance.<p>You can use perfect hashes not only the usual suspects of contiguous integer and dictionary-encoded string ranges, but also use cases like binned numeric and date ranges (epoch seconds binned per year can use a perfect hash range of one bin per year for a very wide range of timestamps), and can even handle arbitrary expressions if you propagate the ranges correctly.<p>Obviously you need a good "baseline" hash path to fall back to you, but it's surprising how many real-world use cases you can profitably cover with perfect hashing.</p>
]]></description><pubDate>Wed, 11 Jun 2025 01:27:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=44243337</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=44243337</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44243337</guid></item><item><title><![CDATA[Fast geospatial aggregation and visualization with Uber H3]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.heavy.ai/blog/put-a-hex-on-it-introducing-new-uber-h3-capabilities">https://www.heavy.ai/blog/put-a-hex-on-it-introducing-new-uber-h3-capabilities</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43846464">https://news.ycombinator.com/item?id=43846464</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 30 Apr 2025 15:16:35 +0000</pubDate><link>https://www.heavy.ai/blog/put-a-hex-on-it-introducing-new-uber-h3-capabilities</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=43846464</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43846464</guid></item><item><title><![CDATA[Benchmarking geospatial join performance on GPU vs. CPU]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.heavy.ai/blog/connect-the-dots-in-real-time-benchmarking-geospatial-join-performance-in-gpu-accelerated-heavydb-against-cpu-databases">https://www.heavy.ai/blog/connect-the-dots-in-real-time-benchmarking-geospatial-join-performance-in-gpu-accelerated-heavydb-against-cpu-databases</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43706353">https://news.ycombinator.com/item?id=43706353</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 16 Apr 2025 14:54:03 +0000</pubDate><link>https://www.heavy.ai/blog/connect-the-dots-in-real-time-benchmarking-geospatial-join-performance-in-gpu-accelerated-heavydb-against-cpu-databases</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=43706353</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43706353</guid></item><item><title><![CDATA[Benchmarking a GPU database against CPU data warehouses]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.heavy.ai/blog/speed-at-scale-benchmarking-gpu-accelerated-heavydb">https://www.heavy.ai/blog/speed-at-scale-benchmarking-gpu-accelerated-heavydb</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43587546">https://news.ycombinator.com/item?id=43587546</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 04 Apr 2025 20:59:39 +0000</pubDate><link>https://www.heavy.ai/blog/speed-at-scale-benchmarking-gpu-accelerated-heavydb</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=43587546</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43587546</guid></item><item><title><![CDATA[Interactively Explore 20 Billion Records of Ship AIS Data Using a Single GPU]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.heavy.ai/blog/interactively-explore-20-billion-records-of-ship-ais-data-on-a-single-gpu-heavy-ai-on-the-new-nvidia-grace-hopper-superchip">https://www.heavy.ai/blog/interactively-explore-20-billion-records-of-ship-ais-data-on-a-single-gpu-heavy-ai-on-the-new-nvidia-grace-hopper-superchip</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43092168">https://news.ycombinator.com/item?id=43092168</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 18 Feb 2025 17:07:07 +0000</pubDate><link>https://www.heavy.ai/blog/interactively-explore-20-billion-records-of-ship-ais-data-on-a-single-gpu-heavy-ai-on-the-new-nvidia-grace-hopper-superchip</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=43092168</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43092168</guid></item><item><title><![CDATA[New comment by tmostak in "Show HN: TabPFN v2 – A SOTA foundation model for small tabular data"]]></title><description><![CDATA[
<p>This looks amazing!<p>Just looking through the code a bit, it seems that the model both supports a (custom) attention mechanism between features and between rows (code uses the term items)? If so, does the attention between rows help improve accuracy significantly?<p>Generally, for standard regression and classification use cases, rows (observations) are seen to be independent, but I'm guessing cross-row attention might help the model see the gestalt of the data in some way that improves accuracy even when the independence assumption holds?</p>
]]></description><pubDate>Fri, 10 Jan 2025 05:27:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=42652818</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=42652818</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42652818</guid></item><item><title><![CDATA[New comment by tmostak in "All You Need Is 4x 4090 GPUs to Train Your Own Model"]]></title><description><![CDATA[
<p>You should be able to train/full-fine-tune (i.e. full weight updates, not LoRA) a much larger model with 96GB of VRAM. I generally have been able to do a full fine-tune (which is equivalent to training a model from scratch) of 34B parameter models at full bf16 using 8XA100 servers (640GB of VRAM) if I enable gradient checkpointing, meaning a 96GB VRAM box should be able to handle models of up to 5B parameters. Of course if you use LoRA, you should be able to go much larger than this, depending on your rank.</p>
]]></description><pubDate>Sun, 29 Dec 2024 03:19:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=42537182</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=42537182</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42537182</guid></item><item><title><![CDATA[Benchmarking a GPU-Accelerated Database Against CPU Data Warehouses]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.heavy.ai/blog/speed-at-scale-benchmarking-gpu-accelerated-heavydb">https://www.heavy.ai/blog/speed-at-scale-benchmarking-gpu-accelerated-heavydb</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42433077">https://news.ycombinator.com/item?id=42433077</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 16 Dec 2024 17:20:07 +0000</pubDate><link>https://www.heavy.ai/blog/speed-at-scale-benchmarking-gpu-accelerated-heavydb</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=42433077</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42433077</guid></item><item><title><![CDATA[Benchmarking GPU-Accelerated HeavyDB Against CPU Data Warehouses]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.heavy.ai/blog/speed-at-scale-benchmarking-gpu-accelerated-heavydb">https://www.heavy.ai/blog/speed-at-scale-benchmarking-gpu-accelerated-heavydb</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=41985908">https://news.ycombinator.com/item?id=41985908</a></p>
<p>Points: 4</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 29 Oct 2024 16:19:44 +0000</pubDate><link>https://www.heavy.ai/blog/speed-at-scale-benchmarking-gpu-accelerated-heavydb</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=41985908</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41985908</guid></item><item><title><![CDATA[Embed AI fuzzy logic into your SQL]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.heavy.ai/blog/making-sql-smarter-how-to-embed-ai-into-your-queries-with-heavyiq">https://www.heavy.ai/blog/making-sql-smarter-how-to-embed-ai-into-your-queries-with-heavyiq</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=41799844">https://news.ycombinator.com/item?id=41799844</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 10 Oct 2024 15:26:20 +0000</pubDate><link>https://www.heavy.ai/blog/making-sql-smarter-how-to-embed-ai-into-your-queries-with-heavyiq</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=41799844</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41799844</guid></item><item><title><![CDATA[New comment by tmostak in "How Meta trains large language models at scale"]]></title><description><![CDATA[
<p>This assumes that you can linearly scale up the number of TPUs to get equal performance to Nvidia cards for less cost. Like most things distributed, this is unlikely to be the case.</p>
]]></description><pubDate>Thu, 13 Jun 2024 03:18:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=40665594</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=40665594</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40665594</guid></item><item><title><![CDATA[Exploring 36 years of FAA flight data with AI and a GPU Database]]></title><description><![CDATA[
<p>Article URL: <a href="https://tech.marksblogg.com/heavyiq-faa-ai-llm-gpu-database.html">https://tech.marksblogg.com/heavyiq-faa-ai-llm-gpu-database.html</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=40545127">https://news.ycombinator.com/item?id=40545127</a></p>
<p>Points: 1</p>
<p># Comments: 1</p>
]]></description><pubDate>Sat, 01 Jun 2024 12:19:41 +0000</pubDate><link>https://tech.marksblogg.com/heavyiq-faa-ai-llm-gpu-database.html</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=40545127</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40545127</guid></item><item><title><![CDATA[Exploring 36 years of FAA flight data with AI and a GPU Database]]></title><description><![CDATA[
<p>Article URL: <a href="https://tech.marksblogg.com/heavyiq-faa-ai-llm-gpu-database.html">https://tech.marksblogg.com/heavyiq-faa-ai-llm-gpu-database.html</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=40524350">https://news.ycombinator.com/item?id=40524350</a></p>
<p>Points: 9</p>
<p># Comments: 1</p>
]]></description><pubDate>Thu, 30 May 2024 14:41:40 +0000</pubDate><link>https://tech.marksblogg.com/heavyiq-faa-ai-llm-gpu-database.html</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=40524350</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40524350</guid></item><item><title><![CDATA[New comment by tmostak in "GPT-4.5 or GPT-5 being tested on LMSYS?"]]></title><description><![CDATA[
<p>Are you measuring tokens/sec or words per second?<p>The difference matters as generally in my experience, Llama 3, by virtue of its giant vocabulary, generally tokenizes text with 20-25% less tokens than something like Mistral. So even if its 18% slower in terms of tokens/second, it may, depending on the text content, actually output a given body of text faster.</p>
]]></description><pubDate>Tue, 30 Apr 2024 20:52:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=40216200</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=40216200</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40216200</guid></item><item><title><![CDATA[New comment by tmostak in "Ollama v0.1.33 with Llama 3, Phi 3, and Qwen 110B"]]></title><description><![CDATA[
<p>But it's likely to be much slower than what you'd get with a backend like llama.cpp on CPU (particularly if you're running on a Mac, but I think on Linux as well), as well as not supporting features like CPU offloading.</p>
]]></description><pubDate>Mon, 29 Apr 2024 03:34:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=40194152</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=40194152</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40194152</guid></item><item><title><![CDATA[New comment by tmostak in "Show HN: Use natural language to query and visualize 400M tweets"]]></title><description><![CDATA[
<p>Thank you, it's been a major team effort!</p>
]]></description><pubDate>Thu, 22 Feb 2024 17:13:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=39470036</link><dc:creator>tmostak</dc:creator><comments>https://news.ycombinator.com/item?id=39470036</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39470036</guid></item></channel></rss>