<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: platypii</title><link>https://news.ycombinator.com/user?id=platypii</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 09 Apr 2026 11:10:25 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=platypii" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by platypii in "A visual explainer of how to scroll through billions of rows in the browser"]]></title><description><![CDATA[
<p>Sylvain Lesage’s cool interactive explainer on visualizing extreme row counts—think billions—inside the browser. His technical deep dive explains how the open-source library HighTable works around scrollbar limits by:<p>- Lazy loading
- Virtual scrolling (allows millions of rows)
- "Infinite Pixel Technique" (allows billions of rows)<p>Hyperparam sponsored Sylvain’s work as part of our broader effort to invest in open-source infrastructure and get ahead of the data-scale problems that are emerging with LLMs. With a regular table, you can view thousands of rows, but the browser breaks pretty quickly. We created HighTable with virtual scroll so you can see millions of rows, but that still wasn’t enough for massive unstructured datasets. What Sylvain has built virtualizes the virtual scroll so you can literally view billions of rows—all inside the browser. His write-up goes deep into the mechanics of building a ridiculously large-scale table component in react.</p>
]]></description><pubDate>Thu, 12 Feb 2026 20:49:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=46994945</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=46994945</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46994945</guid></item><item><title><![CDATA[A visual explainer of how to scroll through billions of rows in the browser]]></title><description><![CDATA[
<p>Article URL: <a href="https://rednegra.net/blog/20260212-virtual-scroll/">https://rednegra.net/blog/20260212-virtual-scroll/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46994944">https://news.ycombinator.com/item?id=46994944</a></p>
<p>Points: 3</p>
<p># Comments: 1</p>
]]></description><pubDate>Thu, 12 Feb 2026 20:49:20 +0000</pubDate><link>https://rednegra.net/blog/20260212-virtual-scroll/</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=46994944</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46994944</guid></item><item><title><![CDATA[New comment by platypii in "Ask HN: Where are you keeping your LLM logs?"]]></title><description><![CDATA[
<p>We're willing to spend money, but I've had the "datadog billing problem" before where it starts reasonable and then grows to a non-trivial percent of saas budget, and then theres a scramble to refactor. Trying to get ahead of that as the LLM logs are MUCH larger that my APM logs.</p>
]]></description><pubDate>Fri, 09 Jan 2026 19:44:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=46558282</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=46558282</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46558282</guid></item><item><title><![CDATA[Ask HN: Where are you keeping your LLM logs?]]></title><description><![CDATA[
<p>LLM logs are crushing my application logging system. We recently launched AI features on our app and went from ~100mb/month of normal website logs to 3gb/month of llm conversation logs and growing. Our existing logging system was overwhelmed (queries timing out, etc), and costs started increasing. We’re considering how to re-architect our llm logs specifically so we can handle more users plus the increasing token use from things like reasoning models, tool calling, and multi-agent systems. I’m not selling any solutions here, genuinely curious what others are doing. Do you store them alongside APM logs? Dedicated LLM logging service? Build it yourself with open source tools?</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46557332">https://news.ycombinator.com/item?id=46557332</a></p>
<p>Points: 1</p>
<p># Comments: 4</p>
]]></description><pubDate>Fri, 09 Jan 2026 18:36:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=46557332</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=46557332</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46557332</guid></item><item><title><![CDATA[Show HN: Squirreling: a browser-native SQL engine]]></title><description><![CDATA[
<p>I made a small (~9 KB), open source SQL engine in JavaScript built for interactive data exploration. Squirreling is unique in that it’s built entirely with modern async JavaScript in mind and enables new kinds of interactivity by prioritizing streaming, late materialization, and async user-defined functions. No other database engine can do this in the browser.<p>More technical details in the post. Feedback welcome.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46435321">https://news.ycombinator.com/item?id=46435321</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 30 Dec 2025 17:03:33 +0000</pubDate><link>https://blog.hyperparam.app/squirreling-new-sql-engine-for-web/</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=46435321</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46435321</guid></item><item><title><![CDATA[Best way to annotate large parquet LLM logs without full rewrites?]]></title><description><![CDATA[
<p>I asked this on the Apache mailing list but haven’t found a good solution yet. Wondering if anyone has some ideas for how to engineer this?<p>Here’s my problem: I have gigabytes of LLM conversation logs in parquet in S3. I want to add per-row annotations (llm-as-a-judge scores), ideally without touching the original text data.<p>So for a given dataset, I want to add a new column. This seemed like a perfect use case for Iceberg. Iceberg does let you evolve the table schema, including adding a column. BUT you can only add a column with a default value. If I want to fill in that column with annotations, ICEBERG MAKES ME REWRITE EVERY ROW. So despite being based on parquet, a column-oriented format, I need to re-write the entire source text data (gigabytes of data) just to add ~1mb of annotations. This feels wildly inefficient.<p>I considered just storing the column in its own table and then joining them. This does work but the joins are annoying to work with, and I suspect query engines do not optimize well a "join on row_number" operation.<p>I've been exploring using little-known features of parquet like the file_path field to store column data in external files. But literally zero parquet clients support this.<p>I'm running out of ideas for how to work with this data efficiently. It's bad enough that I am considering building my own table format if I can’t find a solution. Anyone have suggestions?</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46358172">https://news.ycombinator.com/item?id=46358172</a></p>
<p>Points: 2</p>
<p># Comments: 2</p>
]]></description><pubDate>Mon, 22 Dec 2025 19:52:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=46358172</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=46358172</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46358172</guid></item><item><title><![CDATA[Ask HN: Local tools for working with LLM datasets?]]></title><description><![CDATA[
<p>I’ve been doing data science for years, and am very familiar with jupyter notebooks and more recently been using duckdb a lot. But now I have this huge pile of output tokens from my 4090s, and it feels characteristically different from data I’ve worked with in the past. Notebooks and duckdb on the CLI don’t feel like they’re built for working with huge volumes of text data like my training set and llm output traces.<p>What have you found work well for this? I’m trying to fine-tune on a text dataset and be able to inspect the output from eval runs. I would prefer local and open source tools to a paid service.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46305665">https://news.ycombinator.com/item?id=46305665</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 17 Dec 2025 21:21:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=46305665</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=46305665</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46305665</guid></item><item><title><![CDATA[New comment by platypii in "What UI do you use on top of data engineering tools to look at data?"]]></title><description><![CDATA[
<p>Makes sense. I'm not currently in snowflake because I'm mostly working with local parquet files. Would prefer not to have to pay for snowflake just to explore my data. I'm interested in better data UIs though so I might need to check it out.</p>
]]></description><pubDate>Wed, 10 Dec 2025 18:35:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=46221572</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=46221572</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46221572</guid></item><item><title><![CDATA[What UI do you use on top of data engineering tools to look at data?]]></title><description><![CDATA[
<p>Tools like DuckDB Wasm and data engineering platforms like Iceberg leverage Parquet’s built-in indexing to very efficiently query files over the network. But as I’ve been building data tools myself, the stack gets complicated fast, especially once you try to visualize or explore the data instead of just querying it. I’m intrigued by some of the modern tricks people are using to do more data engineering client-side.<p>With OPFS + Parquet + Wasm, the browser already has everything it needs to handle multi-GB LLM datasets client-side.<p>Is the world of data UIs evolving? Are there new data tools and best practices beyond notebooks and DuckDB?</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46221195">https://news.ycombinator.com/item?id=46221195</a></p>
<p>Points: 1</p>
<p># Comments: 3</p>
]]></description><pubDate>Wed, 10 Dec 2025 18:10:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=46221195</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=46221195</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46221195</guid></item><item><title><![CDATA[New comment by platypii in "Show HN: We built an AI tool for working with massive LLM chat log datasets"]]></title><description><![CDATA[
<p>I started Hyperparam one year ago because I knew that the world of data was changing, and existing tools like Python and Jupyter Notebooks were not built for the scale of LLM data. The weights of LLMs may be tensors, but the input and output of LLMs are massive piles of text.<p>No human has the patience to sift through all that text, so we need better tools to help us understand and analyze it. That's why I built Hyperparam to be the first tool specifically designed for working with LLM data at scale. No one else seemed to be solving this problem.</p>
]]></description><pubDate>Wed, 19 Nov 2025 17:09:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=45982014</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=45982014</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45982014</guid></item><item><title><![CDATA[Show HN: We built an AI tool for working with massive LLM chat log datasets]]></title><description><![CDATA[
<p>There’s an important problem with AI that nobody’s talking about. AI’s entire lifecycle is tons of data in for training, and an even larger amount of text data out. Traditional tools can’t handle the sheer volume of text, leaving teams overwhelmed and unable to make their data work for them.<p>Today we’re launching Hyperparam, a browser-native app for exploring and transforming multi-gigabyte datasets in real time. It combines a fast UI that can stream huge unstructured datasets with an army of AI agents that can score, label, filter, and categorize them. Now you can actually make sense of AI-scale data instead of drowning in it.<p>Example: Using the chat, ask Hyperparam’s AI agent to score every conversation in a 100K-row dataset for sycophancy, filter out the worst responses, adjust prompts, regenerate, and export your dataset V2. It all runs in one browser tab with no waiting and no lag.<p>It’s free while it’s in beta if you want to try it on your own data.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45981930">https://news.ycombinator.com/item?id=45981930</a></p>
<p>Points: 16</p>
<p># Comments: 1</p>
]]></description><pubDate>Wed, 19 Nov 2025 17:02:53 +0000</pubDate><link>https://hyperparam.app/</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=45981930</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45981930</guid></item><item><title><![CDATA[New comment by platypii in "Lessons from Hyperparam's year of open source data transformation"]]></title><description><![CDATA[
<p>This is a Q&A I did on what I learned from a year of open source data transformation. Most of all, it reinforced my belief that browser-native tools aren’t “toys” that don’t work for real systems. When Hugging Face integrated my libraries, it confirmed that the browser can handle serious data work, and maybe there's an opportunity for more browser-based data tools.</p>
]]></description><pubDate>Thu, 13 Nov 2025 19:30:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=45919419</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=45919419</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45919419</guid></item><item><title><![CDATA[Lessons from Hyperparam's year of open source data transformation]]></title><description><![CDATA[
<p>Article URL: <a href="https://blog.hyperparam.app/lessons-from-open-source-data-transformation/">https://blog.hyperparam.app/lessons-from-open-source-data-transformation/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45919418">https://news.ycombinator.com/item?id=45919418</a></p>
<p>Points: 1</p>
<p># Comments: 1</p>
]]></description><pubDate>Thu, 13 Nov 2025 19:30:07 +0000</pubDate><link>https://blog.hyperparam.app/lessons-from-open-source-data-transformation/</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=45919418</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45919418</guid></item><item><title><![CDATA[New comment by platypii in "Ask HN: How far can we push the browser for large-scale data parsing?"]]></title><description><![CDATA[
<p>As with anything, there are engineering tradeoffs.<p>What I've found is that moving data processing toward the browser has been for one, a refreshing developer experience because I don't need to build a pair of backend+frontend. From a user experience point of view, I think you can build MORE interactive data applications by pushing it toward the frontend.</p>
]]></description><pubDate>Thu, 06 Nov 2025 17:16:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=45837593</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=45837593</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45837593</guid></item><item><title><![CDATA[Ask HN: How far can we push the browser for large-scale data parsing?]]></title><description><![CDATA[
<p>How far can we push the browser as a data engine — not just for visualizations, but for curating and querying large datasets? Do we need traditional backend architectures?<p>I wanted to see what happens when we treat the browser like part of the data stack, using pure JavaScript to load, slice, and explore datasets interactively. That experiment led to a small set of open-source tools — Hyparquet and HighTable. They’re designed to test the limits of browser-native data processing to see where the browser stops being a thin client and starts acting like a real data engine.<p>Curious what others think about the future of browser-first data tools:<p>- Where do you see the practical limits for client-side data processing?
 - What would make browser-based architectures a viable alternative to traditional data stacks?</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45837156">https://news.ycombinator.com/item?id=45837156</a></p>
<p>Points: 1</p>
<p># Comments: 2</p>
]]></description><pubDate>Thu, 06 Nov 2025 16:42:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=45837156</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=45837156</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45837156</guid></item><item><title><![CDATA[New comment by platypii in "From GPT-4 to GPT-5: Measuring progress through MedHELM [pdf]"]]></title><description><![CDATA[
<p>Why not? We are trying to evaluate AI's capabilities. It's OBVIOUS that we should compare it to our only prior example of intelligence -- humans. Saying we shouldn't compare or anthropomorphize machine is a ridiculous hill to die on.</p>
]]></description><pubDate>Fri, 22 Aug 2025 04:08:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=44980906</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=44980906</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44980906</guid></item><item><title><![CDATA[New comment by platypii in "The Quest for Instant Data"]]></title><description><![CDATA[
<p>This is the story of how I spent a year making the world's fastest Parquet loader in JavaScript. The goal:<p>- Make a faster, more interactive viewer for AI datasets (which are mostly parquet format)<p>- Simplify the stack by doing everything from the browser (no backend)<p>TLDR: My open-source library Hyparquet can load data in 155ms, which would take 3466ms in duckdb-wasm for the same file.</p>
]]></description><pubDate>Thu, 24 Jul 2025 16:09:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=44672432</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=44672432</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44672432</guid></item><item><title><![CDATA[The Quest for Instant Data]]></title><description><![CDATA[
<p>Article URL: <a href="https://blog.hyperparam.app/2025/07/24/quest-for-instant-data/">https://blog.hyperparam.app/2025/07/24/quest-for-instant-data/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=44672363">https://news.ycombinator.com/item?id=44672363</a></p>
<p>Points: 16</p>
<p># Comments: 1</p>
]]></description><pubDate>Thu, 24 Jul 2025 16:03:32 +0000</pubDate><link>https://blog.hyperparam.app/2025/07/24/quest-for-instant-data/</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=44672363</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44672363</guid></item><item><title><![CDATA[New comment by platypii in "Show HN: Hyperparam: OSS Tools for Exploring Datasets Locally in the Browser"]]></title><description><![CDATA[
<p>I don’t have benchmarks specifically against duckdb. I’m sure native C++ will run faster than JavaScript.<p>But whats important is that with Hyperparam you can do it in the browser, where the bottleneck will always be network-bound not cpu-bound.</p>
]]></description><pubDate>Thu, 01 May 2025 15:51:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=43859360</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=43859360</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43859360</guid></item><item><title><![CDATA[New comment by platypii in "Show HN: Hyperparam: OSS Tools for Exploring Datasets Locally in the Browser"]]></title><description><![CDATA[
<p>Funny you say that, because I built these tools because I wanted to build something very much like what you're describing!<p>I was trying to look at, filter, and transform large AI datasets, and I was frustrated with how bad the existing tool was for working with datasets with huge amounts of text (web scrapes, github dumps, reasoning tokens, agent chat logs). Jupyter notebook is woefully bad at helping you to look at your data.<p>So I wanted to build better browser tools for working with AI datasets. But to do that I had to build these tools (there was no working parquet implementation in JS when I started).<p>Anyway I'm still working on building an app for data processing using LLM chat assistant to help a single user curate entire datasets singlehandedly. But for now I'm releasing these components to the community as open source. And having them "do a single task each" was very much intentional. Thanks for the comment!</p>
]]></description><pubDate>Thu, 01 May 2025 15:26:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=43858985</link><dc:creator>platypii</dc:creator><comments>https://news.ycombinator.com/item?id=43858985</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43858985</guid></item></channel></rss>