<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: prats226</title><link>https://news.ycombinator.com/user?id=prats226</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 10 Apr 2026 08:55:02 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=prats226" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by prats226 in "Research-Driven Agents: What Happens When Your Agent Reads Before It Codes"]]></title><description><![CDATA[
<p>A good experiment would be to also try giving it access to latency traces so it can identify issues? Wrt coding agents, giving access to observability tools often improve coding/debugging ability for me</p>
]]></description><pubDate>Thu, 09 Apr 2026 21:09:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=47710166</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=47710166</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47710166</guid></item><item><title><![CDATA[New comment by prats226 in "Ask HN: Alternatives to Reducto?"]]></title><description><![CDATA[
<p>Try <a href="https://docstrange.nanonets.com/" rel="nofollow">https://docstrange.nanonets.com/</a> once, 10k docs you can use for free. Strong table performance. Do give feedback if any. Powered by bigger model compared to our open source one which is quiet popular on HF.</p>
]]></description><pubDate>Tue, 03 Mar 2026 07:04:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=47229106</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=47229106</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47229106</guid></item><item><title><![CDATA[New comment by prats226 in "Large-Scale Online Deanonymization with LLMs"]]></title><description><![CDATA[
<p>If with LLM's you can deanonymize at scale, on a personal level, you should also be able to figure out what posts are leading to this deanonymization and remove them or modify them.</p>
]]></description><pubDate>Wed, 25 Feb 2026 22:56:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=47159234</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=47159234</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47159234</guid></item><item><title><![CDATA[New comment by prats226 in "Show HN: Ocrbase – pdf → .md/.json document OCR and structured extraction API"]]></title><description><![CDATA[
<p>Instead of markdown -> LLM to get JSON, you can just train a slightly bigger model which you can constrain decode to give JSON rightaway.
<a href="https://huggingface.co/nanonets/Nanonets-OCR2-3B" rel="nofollow">https://huggingface.co/nanonets/Nanonets-OCR2-3B</a><p>We recently published a cookbook for constrained decoding here:
<a href="https://nanonets.com/cookbooks/structured-llm-outputs/" rel="nofollow">https://nanonets.com/cookbooks/structured-llm-outputs/</a></p>
]]></description><pubDate>Wed, 21 Jan 2026 00:48:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=46699784</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=46699784</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46699784</guid></item><item><title><![CDATA[New comment by prats226 in "LLM Structured Outputs Handbook"]]></title><description><![CDATA[
<p><a href="https://nanonets.com/cookbooks/structured-llm-outputs/unconstrained-decoding/baml/" rel="nofollow">https://nanonets.com/cookbooks/structured-llm-outputs/uncons...</a></p>
]]></description><pubDate>Sat, 17 Jan 2026 00:07:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=46653888</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=46653888</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46653888</guid></item><item><title><![CDATA[New comment by prats226 in "LLM Structured Outputs Handbook"]]></title><description><![CDATA[
<p>Nice, it would be good idea to develop CFG for this as well so can embed it into all these constrained decoding libraries</p>
]]></description><pubDate>Fri, 16 Jan 2026 23:49:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=46653751</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=46653751</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46653751</guid></item><item><title><![CDATA[New comment by prats226 in "LLM Structured Outputs Handbook"]]></title><description><![CDATA[
<p>One of the authors here, will checkout the diagram link.<p>Every commercial model provider is adding structured outputs so will keep updating the guide.</p>
]]></description><pubDate>Fri, 16 Jan 2026 22:38:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=46653173</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=46653173</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46653173</guid></item><item><title><![CDATA[New comment by prats226 in "DeepSeek OCR"]]></title><description><![CDATA[
<p><a href="https://docstrange.nanonets.com/" rel="nofollow">https://docstrange.nanonets.com/</a> as well, wrapper on top of 7B version of <a href="https://huggingface.co/nanonets/Nanonets-OCR2-3B" rel="nofollow">https://huggingface.co/nanonets/Nanonets-OCR2-3B</a></p>
]]></description><pubDate>Mon, 20 Oct 2025 20:09:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=45648665</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45648665</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45648665</guid></item><item><title><![CDATA[New comment by prats226 in "DeepSeek OCR"]]></title><description><![CDATA[
<p>Then you can just download finetuned version of same multi-modal foundation model that's trained on documents?</p>
]]></description><pubDate>Mon, 20 Oct 2025 20:08:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=45648640</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45648640</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45648640</guid></item><item><title><![CDATA[New comment by prats226 in "DeepSeek OCR"]]></title><description><![CDATA[
<p>Top 3 models on huggingface are all OCR models. Most automation projects involve documents where you need a model finetuned to understand all elements inside documents and provide grounding and confidence scores etc which is why these subset of models are gaining popularity</p>
]]></description><pubDate>Mon, 20 Oct 2025 20:05:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=45648608</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45648608</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45648608</guid></item><item><title><![CDATA[New comment by prats226 in "What Americans die from vs. what the news reports on"]]></title><description><![CDATA[
<p>Would be intersting to see where funding goes to fix these issues. News would heavily impact public opinion and hence political influence and public funding.</p>
]]></description><pubDate>Tue, 14 Oct 2025 23:42:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=45586317</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45586317</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45586317</guid></item><item><title><![CDATA[New comment by prats226 in "Nanonets-OCR2-3B – OCR model that transforms documents into structured markdown"]]></title><description><![CDATA[
<p>Yes, and its not just OCR (Optical Character Recognition), it understands layouts, captures signatures, charts, watermarks etc so way beyond just characters</p>
]]></description><pubDate>Tue, 14 Oct 2025 23:34:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=45586267</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45586267</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45586267</guid></item><item><title><![CDATA[New comment by prats226 in "Launch HN: Extend (YC W23) – Turn your messiest documents into data"]]></title><description><![CDATA[
<p><a href="https://mention.com/en/" rel="nofollow">https://mention.com/en/</a></p>
]]></description><pubDate>Sat, 11 Oct 2025 21:41:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=45552952</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45552952</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45552952</guid></item><item><title><![CDATA[New comment by prats226 in "Launch HN: Extend (YC W23) – Turn your messiest documents into data"]]></title><description><![CDATA[
<p>Here is link to open source model:
<a href="https://huggingface.co/nanonets/Nanonets-OCR-s" rel="nofollow">https://huggingface.co/nanonets/Nanonets-OCR-s</a><p>And hosted model:
<a href="https://docstrange.nanonets.com/" rel="nofollow">https://docstrange.nanonets.com/</a></p>
]]></description><pubDate>Sat, 11 Oct 2025 19:14:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=45551838</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45551838</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45551838</guid></item><item><title><![CDATA[New comment by prats226 in "Designing agentic loops"]]></title><description><![CDATA[
<p>It boils down to information loss in compaction driven by LLM's. Either you could carefully design tools that only give compacted output with high information density so models have to auto-compact or organize information only once in a while which eventually is going to be lossy.<p>Or you just give loads of information without thinking much about it, assuming models will have to do frequent compaction and memory organization and hope its not super lossy.</p>
]]></description><pubDate>Tue, 30 Sep 2025 21:10:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=45431300</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45431300</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45431300</guid></item><item><title><![CDATA[New comment by prats226 in "Designing agentic loops"]]></title><description><![CDATA[
<p>Reason I felt like they are closely connected are because for designing tools for lets say coding agents, you have to be thoughful of context engineering.<p>Eg linear MCP is notorious for giving large JSONs which quickly fill up context and hard for model to understand. So tools need to be designed slightly differently for agents keeping context engineering in mind compared to how you design them for humans.<p>Context engineering feels like more central and first-principle approach of designing tools, agent loops.</p>
]]></description><pubDate>Tue, 30 Sep 2025 21:07:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=45431255</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45431255</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45431255</guid></item><item><title><![CDATA[New comment by prats226 in "Designing agentic loops"]]></title><description><![CDATA[
<p>Context engineering is another name people have given to same skill?</p>
]]></description><pubDate>Tue, 30 Sep 2025 19:43:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=45430301</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45430301</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45430301</guid></item><item><title><![CDATA[New comment by prats226 in "Show HN: HumanAlarm – Real people knock on your door to wake you up"]]></title><description><![CDATA[
<p>You can always put automation for your google home to blast music at full volume at right time. And if you don't wake up from sound of music yourself, your neighbour will knock on your door for sure!</p>
]]></description><pubDate>Wed, 10 Sep 2025 22:10:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=45204695</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=45204695</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45204695</guid></item><item><title><![CDATA[New comment by prats226 in "In a first, Google has released data on how much energy an AI prompt uses"]]></title><description><![CDATA[
<p>With google serving AI overviews, now an average search query should cost more? Compute is getting cheaper but also algorithms getting more and more complex, increasing compute?</p>
]]></description><pubDate>Thu, 21 Aug 2025 21:38:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=44978405</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=44978405</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44978405</guid></item><item><title><![CDATA[New comment by prats226 in "Training language models to be warm and empathetic makes them less reliable"]]></title><description><![CDATA[
<p>Read long time ago that even SFT for conversations vs base model for autocomplete reduces intelligence, increases perplexity</p>
]]></description><pubDate>Tue, 12 Aug 2025 18:51:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=44880339</link><dc:creator>prats226</dc:creator><comments>https://news.ycombinator.com/item?id=44880339</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44880339</guid></item></channel></rss>