<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: raunakchowdhuri</title><link>https://news.ycombinator.com/user?id=raunakchowdhuri</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 07 Apr 2026 00:07:29 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=raunakchowdhuri" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by raunakchowdhuri in "Reducto releases Deep Extract"]]></title><description><![CDATA[
<p>We've made a lot of changes in the past few months that make our standard extract much, much better, as well as Deep Extract for documents even longer than that. We'd love for you to give it a try!</p>
]]></description><pubDate>Mon, 06 Apr 2026 23:55:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=47668975</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=47668975</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47668975</guid></item><item><title><![CDATA[Reducto releases Deep Extract]]></title><description><![CDATA[
<p>Article URL: <a href="https://reducto.ai/blog/reducto-deep-extract-agent">https://reducto.ai/blog/reducto-deep-extract-agent</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47662833">https://news.ycombinator.com/item?id=47662833</a></p>
<p>Points: 44</p>
<p># Comments: 7</p>
]]></description><pubDate>Mon, 06 Apr 2026 16:13:47 +0000</pubDate><link>https://reducto.ai/blog/reducto-deep-extract-agent</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=47662833</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47662833</guid></item><item><title><![CDATA[Show HN: A SOTA chart-extraction system combining traditional CV and LVMs]]></title><description><![CDATA[
<p>Article URL: <a href="https://reducto.ai/blog/reducto-chart-extraction">https://reducto.ai/blog/reducto-chart-extraction</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46151778">https://news.ycombinator.com/item?id=46151778</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 04 Dec 2025 19:33:10 +0000</pubDate><link>https://reducto.ai/blog/reducto-chart-extraction</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=46151778</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46151778</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Shai-Hulud Returns: Over 300 NPM Packages Infected"]]></title><description><![CDATA[
<p>Have a slack channel with them, these are the versions they mentioned:
posthog-node 4.18.1
posthog-js 1.297.3
posthog-react-native 4.11.1
posthog-docusaurus 2.0.6</p>
]]></description><pubDate>Mon, 24 Nov 2025 17:00:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=46036234</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=46036234</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46036234</guid></item><item><title><![CDATA[We did a DB migration without logical replication – with zero downtime]]></title><description><![CDATA[
<p>Article URL: <a href="https://reducto.ai/blog/reducto-database-migration-zero-downtime">https://reducto.ai/blog/reducto-database-migration-zero-downtime</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45363901">https://news.ycombinator.com/item?id=45363901</a></p>
<p>Points: 4</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 24 Sep 2025 18:06:47 +0000</pubDate><link>https://reducto.ai/blog/reducto-database-migration-zero-downtime</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=45363901</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45363901</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Launch HN: Reducto Studio (YC W24) – Build accurate document pipelines, fast"]]></title><description><![CDATA[
<p>We're fixing it! This for some reason happens on only _some_ phones in our office so was hard to repro. I think has to do with Safari rendering. Will tone down our WebGPU usage</p>
]]></description><pubDate>Tue, 24 Jun 2025 01:03:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=44361869</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=44361869</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44361869</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Launch HN: Reducto Studio (YC W24) – Build accurate document pipelines, fast"]]></title><description><![CDATA[
<p>dang no way! we were both in boston too</p>
]]></description><pubDate>Mon, 23 Jun 2025 21:32:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=44360374</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=44360374</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44360374</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Launch HN: Reducto Studio (YC W24) – Build accurate document pipelines, fast"]]></title><description><![CDATA[
<p>this is exactly where we're going with this! glad you see the vision :)</p>
]]></description><pubDate>Mon, 23 Jun 2025 18:41:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=44358729</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=44358729</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44358729</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Launch HN: Reducto Studio (YC W24) – Build accurate document pipelines, fast"]]></title><description><![CDATA[
<p>yep!</p>
]]></description><pubDate>Mon, 23 Jun 2025 16:42:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=44357575</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=44357575</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44357575</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Mistral OCR"]]></title><description><![CDATA[
<p>comparisons to more outputs coming soon!</p>
]]></description><pubDate>Fri, 07 Mar 2025 08:00:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=43288111</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=43288111</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43288111</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Mistral OCR"]]></title><description><![CDATA[
<p>We ran some benchmarks comparing against Gemini Flash 2.0. You can find the full writeup here: <a href="https://reducto.ai/blog/lvm-ocr-accuracy-mistral-gemini">https://reducto.ai/blog/lvm-ocr-accuracy-mistral-gemini</a><p>A high level summary is that while this is an impressive model, it underperforms even current SOTA VLMs on document parsing and has a tendency to hallucinate with OCR, table structure, and drop content.</p>
]]></description><pubDate>Fri, 07 Mar 2025 03:54:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=43287278</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=43287278</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43287278</guid></item><item><title><![CDATA[Evaluating Mistral OCR Against Gemini 2.0 Flash]]></title><description><![CDATA[
<p>Article URL: <a href="https://reducto.ai/blog/lvm-ocr-accuracy-mistral-gemini">https://reducto.ai/blog/lvm-ocr-accuracy-mistral-gemini</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43287210">https://news.ycombinator.com/item?id=43287210</a></p>
<p>Points: 15</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 07 Mar 2025 03:36:11 +0000</pubDate><link>https://reducto.ai/blog/lvm-ocr-accuracy-mistral-gemini</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=43287210</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43287210</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Ingesting PDFs and why Gemini 2.0 changes everything"]]></title><description><![CDATA[
<p>CTO of Reducto here. Love this writeup!<p>We’ve generally found that Gemini 2.0 is a great model and have tested this (and nearly every VLM) very extensively.<p>A big part of our research focus is incorporating the best of what new VLMs offer without losing the benefits and reliability of traditional CV models. A simple example of this is we’ve found bounding box based attribution to be a non-negotiable for many of our current customers. Citing the specific region in a document where an answer came from becomes (in our opinion) even MORE important when using large vision models in the loop, as there is a continued risk of hallucination.<p>Whether that matters in your product is ultimately use case dependent, but the more important challenge for us has been reliability in outputs. RD-TableBench currently uses a single table image on a page, but when testing with real world dense pages we find that VLMs deviate more. Sometimes that involves minor edits (summarizing a sentence but preserving meaning), but sometimes it’s a more serious case such as hallucinating large sets of content.<p>The more extreme case is that internally we fine tuned a version of Gemini 1.5 along with base Gemini 2.0, specifically for checkbox extraction. We found that even with a broad distribution of checkbox data we couldn’t prevent frequent checkbox hallucination on both the flash (+17% error rate) and pro model (+8% error rate). Our customers in industries like healthcare expect us to get it right, out of the box, deterministically, and our team’s directive is to get as close as we can to that ideal state.<p>We think that the ideal state involves a combination of the two. The flexibility that VLMs provide, for example with cases like handwriting, is what I think will make it possible to go from 80 or 90 percent accuracy to some number very close 99%. I should note that the Reducto performance for table extraction is with our pre-VLM table parsing pipeline, and we’ll have more to share in terms of updates there soon.
For now, our focus is entirely on the performance frontier (though we do scale costs down with volume). In the longer term as inference becomes more efficient we want to move the needle on cost as well.<p>Overall though, I’m very excited about the progress here.<p>---
One small comment on your footnote, the evaluation script with Needlemen-Wunsch algorithm doesn’t actually consider the headers outputted by the models and looks only at the table structure itself.</p>
]]></description><pubDate>Wed, 05 Feb 2025 20:02:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=42954289</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=42954289</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42954289</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Ingesting PDFs and why Gemini 2.0 changes everything"]]></title><description><![CDATA[
<p>would encourage you to take a look at some of the real data here!
<a href="https://huggingface.co/spaces/reducto/rd_table_bench" rel="nofollow">https://huggingface.co/spaces/reducto/rd_table_bench</a><p>you'll find that most of the errors here are structural issues with the table or inability to parse some special characters. tables can get crazy!</p>
]]></description><pubDate>Wed, 05 Feb 2025 19:32:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=42953853</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=42953853</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42953853</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Rd-TableBench – Accurately evaluating table extraction"]]></title><description><![CDATA[
<p>Love the Pubtables work! It's a really useful dataset. Their data comes from existing annotations from scientific papers, so in our experience it doesn't include a lot of the hardest cases that a lot of methods fail at today. The annotations are computer generated instead of manually labeled, so you don't have things like scanned and rotated images or a lot of diversity in languages.<p>I'd encourage you to take a look at some of our data points to compare for yourself! Link: huggingface.co/spaces/reducto/rd_table_bench<p>In terms of the overall importance of table extraction, we've found it to be a key bottleneck for folks looking to do document parsing. It's up there amongst the hardest problems in the space alongside complex form region parsing. I don't have the exact statistics handy, but I'd estimate that ~25% of the pages we parse have some hairy tables in them!</p>
]]></description><pubDate>Tue, 05 Nov 2024 19:09:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=42054322</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=42054322</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42054322</guid></item><item><title><![CDATA[Rd-TableBench – Accurately evaluating table extraction]]></title><description><![CDATA[
<p>Hey HN!<p>A ton of document parsing solutions have been coming out lately, each claiming SOTA with little evidence. A lot of these turned out to be LLM or LVM wrappers that hallucinate frequently on complex tables.<p>We just released RD-TableBench, an open benchmark to help teams evaluate extraction performance for complex tables. The benchmark includes a variety of challenging scenarios including scanned tables, handwriting, language detection, merged cells, and more.<p>We employed an independent team of PhD-level human labelers who manually annotated 1000 complex table images from a diverse set of publicly available documents.<p>Alongside this, we also release a new bioinformatics inspired algorithm for grading table similarity. Would love to hear any feedback!<p>-Raunak</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42054144">https://news.ycombinator.com/item?id=42054144</a></p>
<p>Points: 29</p>
<p># Comments: 6</p>
]]></description><pubDate>Tue, 05 Nov 2024 18:46:31 +0000</pubDate><link>https://reducto.ai/blog/rd-tablebench</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=42054144</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42054144</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Launch HN: Parity (YC S24) – AI for on-call engineers working with Kubernetes"]]></title><description><![CDATA[
<p>hmmm idk how I would feel about giving an llm cluster access from a security pov</p>
]]></description><pubDate>Mon, 26 Aug 2024 15:33:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=41358144</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=41358144</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41358144</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Show HN: K8sAI – open-source GPT CLI tool for Kubernetes"]]></title><description><![CDATA[
<p>Interesting... how did you do the scraping of the documentation?</p>
]]></description><pubDate>Thu, 02 May 2024 15:15:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=40237255</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=40237255</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40237255</guid></item><item><title><![CDATA[Show HN: Reducto – A vision based document ingestion API for LLMs]]></title><description><![CDATA[
<p>Hey HN, I'm Raunak from Reducto (<a href="https://reducto.ai">https://reducto.ai</a>), a high-quality document ingestion API tailored for language models. We developed Reducto to address our own need - no existing parsing solutions provided the accuracy and speed necessary for our user-facing AI applications. We designed a system that comprehends documents visually (like a human), ignoring document metadata and processing each page as an image to ensure the highest possible accuracy (with benchmarks to prove it).<p>Please give our demo a try with some of your own PDFs or reach out at founders@reducto.ai if you’d like to start using Reducto in production.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=39523565">https://news.ycombinator.com/item?id=39523565</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 27 Feb 2024 13:04:31 +0000</pubDate><link>https://www.reducto.ai/blog/document-api</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=39523565</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39523565</guid></item><item><title><![CDATA[New comment by raunakchowdhuri in "Open Sourcing Remembrall: A Long-Term Memory Proxy for LLMs"]]></title><description><![CDATA[
<p>Hey HN,<p>A few weeks ago I shared a beta of Remembrall here and got a lot of great feedback from people in this community, with one of the most common requests being to open source the project.<p>Excited to share that we’re doing exactly that!<p>Remembrall is a proxy (integrates in two lines!) on top of your OpenAI queries that uses GPT to save/update important details from each user’s conversations into a vector db. When the user continues the conversation we query the db for relevant info and prepend it into the system prompt. We have a lot of improvements in the works (function calling, queryable user profiles, and more) and would love your feedback on features you’d like to see!</p>
]]></description><pubDate>Thu, 26 Oct 2023 16:03:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=38027519</link><dc:creator>raunakchowdhuri</dc:creator><comments>https://news.ycombinator.com/item?id=38027519</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38027519</guid></item></channel></rss>