<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: mingtianzhang</title><link>https://news.ycombinator.com/user?id=mingtianzhang</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 08 Apr 2026 20:40:43 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=mingtianzhang" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[ClawdReview – OpenReview for AI Agents]]></title><description><![CDATA[
<p>Agents can review the paper on arXiv, and humans can like or dislike agents' reviews. There are also ranking lists of the most popular papers and agents. Please visit: https://clawdreview.ai/</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47016081">https://news.ycombinator.com/item?id=47016081</a></p>
<p>Points: 5</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 14 Feb 2026 16:57:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=47016081</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=47016081</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47016081</guid></item><item><title><![CDATA[Show HN: ClawdReview – OpenReview for AI Agents]]></title><description><![CDATA[
<p>Agents can review the paper on arXiv, and humans can like or dislike agents' reviews. There are also ranking lists of the most popular papers and agents. Please visit: <a href="https://clawdreview.ai/" rel="nofollow">https://clawdreview.ai/</a></p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47014232">https://news.ycombinator.com/item?id=47014232</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 14 Feb 2026 13:03:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=47014232</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=47014232</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47014232</guid></item><item><title><![CDATA[New comment by mingtianzhang in "[dead]"]]></title><description><![CDATA[
<p>VLM can already process both the document images and the query to produce an answer directly. Do we still need the intermediate OCR step?</p>
]]></description><pubDate>Fri, 31 Oct 2025 15:13:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=45772928</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45772928</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45772928</guid></item><item><title><![CDATA[New comment by mingtianzhang in "Do we still need OCR? An implementation of a pure vision-based agent"]]></title><description><![CDATA[
<p>We discuss the limitations of the classic OCR pipeline and provide a pure vision-based RAG system for document analysis (<a href="https://github.com/VectifyAI/PageIndex/blob/main/cookbook/vision_RAG_pageindex.ipynb" rel="nofollow">https://github.com/VectifyAI/PageIndex/blob/main/cookbook/vi...</a>)<p>Any feedback is welcome!</p>
]]></description><pubDate>Wed, 29 Oct 2025 06:28:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=45743317</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45743317</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45743317</guid></item><item><title><![CDATA[Do we still need OCR? An implementation of a pure vision-based agent]]></title><description><![CDATA[
<p>Article URL: <a href="https://pageindex.ai/blog/do-we-need-ocr">https://pageindex.ai/blog/do-we-need-ocr</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45743316">https://news.ycombinator.com/item?id=45743316</a></p>
<p>Points: 7</p>
<p># Comments: 1</p>
]]></description><pubDate>Wed, 29 Oct 2025 06:28:09 +0000</pubDate><link>https://pageindex.ai/blog/do-we-need-ocr</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45743316</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45743316</guid></item><item><title><![CDATA[New comment by mingtianzhang in "Should LLMs just treat text content as an image?"]]></title><description><![CDATA[
<p>We actually don't need OCR: <a href="https://pageindex.ai/blog/do-we-need-ocr" rel="nofollow">https://pageindex.ai/blog/do-we-need-ocr</a></p>
]]></description><pubDate>Mon, 27 Oct 2025 15:32:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=45722133</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45722133</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45722133</guid></item><item><title><![CDATA[New comment by mingtianzhang in "Do We Still Need OCR?"]]></title><description><![CDATA[
<p>This blog examines the inherent limitations of the current OCR pipeline in the context of document question-answering systems from an information-theoretic perspective and discusses why a direct, vision-based approach can be more effective. It also provides a practical implementation of a vision-based question-answering system for long documents.</p>
]]></description><pubDate>Mon, 27 Oct 2025 15:01:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=45721773</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45721773</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45721773</guid></item><item><title><![CDATA[Do We Still Need OCR?]]></title><description><![CDATA[
<p>Article URL: <a href="https://pageindex.ai/blog/do-we-need-ocr">https://pageindex.ai/blog/do-we-need-ocr</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45721772">https://news.ycombinator.com/item?id=45721772</a></p>
<p>Points: 4</p>
<p># Comments: 2</p>
]]></description><pubDate>Mon, 27 Oct 2025 15:01:09 +0000</pubDate><link>https://pageindex.ai/blog/do-we-need-ocr</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45721772</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45721772</guid></item><item><title><![CDATA[New comment by mingtianzhang in "Reasoning-based RAG for long document question answering"]]></title><description><![CDATA[
<p>PageIndex Chat is the world's first human-like long-document AI analyst. You can upload entire books, research papers, or hundred-page reports and chat with them without context limits, all in the browser.
Unlike traditional RAG or "chat-with-your-doc" tools that rely on vector similarity search, PageIndex builds a hierarchical tree index of your document (like a table of contents), and then reasons over this index to retrieve and interpret relevant sections. It doesn’t search by keywords or embeddings — it reads, understands, and reasons through the document like a human expert.<p>What makes it different:<p>- Reasoning-based retrieval: Understands structure, logic, and meaning, not just semantic similarity.<p>- Page-level references: Every answer includes precise citations for easy verification.<p>- Cross-section reasoning: Connects information across sections and appendices to find true answers.<p>- Human-in-the-loop: You can guide, refine, and verify its reasoning.<p>- Multi-document comparison: Analyze and contrast multiple reports at once.</p>
]]></description><pubDate>Fri, 24 Oct 2025 14:05:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=45694800</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45694800</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45694800</guid></item><item><title><![CDATA[Reasoning-based RAG for long document question answering]]></title><description><![CDATA[
<p>Article URL: <a href="https://pageindex.ai/blog/pageindex-chat">https://pageindex.ai/blog/pageindex-chat</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45694799">https://news.ycombinator.com/item?id=45694799</a></p>
<p>Points: 1</p>
<p># Comments: 1</p>
]]></description><pubDate>Fri, 24 Oct 2025 14:05:59 +0000</pubDate><link>https://pageindex.ai/blog/pageindex-chat</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45694799</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45694799</guid></item><item><title><![CDATA[New comment by mingtianzhang in "PageIndex Chat – Human-Like Long Document AI Analyst"]]></title><description><![CDATA[
<p>PageIndex Chat is the world's first human-like long-document AI analyst. You can upload entire books, research papers, or hundred-page reports and chat with them without context limits, all in the browser.<p>Unlike traditional RAG or "chat-with-your-doc" tools that rely on vector similarity search, PageIndex builds a hierarchical tree index of your document (like a table of contents), and then reasons over this index to retrieve and interpret relevant sections. It doesn’t search by keywords or embeddings — it reads, understands, and reasons through the document like a human expert.<p>What makes it different:<p>- Reasoning-based retrieval – Understands structure, logic, and meaning, not just semantic similarity. - Page-level references – Every answer includes precise citations for easy verification. - Cross-section reasoning – Connects information across sections and appendices to find true answers. - Human-in-the-loop – You can guide, refine, and verify its reasoning. - Multi-document comparison – Analyze and contrast multiple reports at once.</p>
]]></description><pubDate>Wed, 22 Oct 2025 09:02:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=45666488</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45666488</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45666488</guid></item><item><title><![CDATA[PageIndex Chat – Human-Like Long Document AI Analyst]]></title><description><![CDATA[
<p>Article URL: <a href="https://pageindex.ai/blog/pageindex-chat">https://pageindex.ai/blog/pageindex-chat</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45666487">https://news.ycombinator.com/item?id=45666487</a></p>
<p>Points: 6</p>
<p># Comments: 1</p>
]]></description><pubDate>Wed, 22 Oct 2025 09:02:52 +0000</pubDate><link>https://pageindex.ai/blog/pageindex-chat</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45666487</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45666487</guid></item><item><title><![CDATA[New comment by mingtianzhang in "PageIndex Chat – Human-Like Long Document AI Analyst"]]></title><description><![CDATA[
<p>PageIndex Chat is the world's first human-like long-document AI analyst. You can pload entire books, research papers, or hundred-page reports and chat with them without context limits, all in the browser.<p>Unlike traditional RAG or "chat-with-your-doc" tools that rely on vector similarity search, PageIndex builds a hierarchical tree index of your document (like a table of contents), and then reasons over this index to retrieve and interpret relevant sections. It doesn’t search by keywords or embeddings — it reads, understands, and reasons through the document like a human expert.<p>What makes it different:<p>- Reasoning-based retrieval – Understands structure, logic, and meaning, not just semantic similarity.
- Page-level references – Every answer includes precise citations for easy verification.
- Cross-section reasoning – Connects information across sections and appendices to find true answers.
- Human-in-the-loop – You can guide, refine, and verify its reasoning.
- Multi-document comparison – Analyze and contrast multiple reports at once.</p>
]]></description><pubDate>Wed, 22 Oct 2025 00:35:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=45663610</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45663610</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45663610</guid></item><item><title><![CDATA[PageIndex Chat – Human-Like Long Document AI Analyst]]></title><description><![CDATA[
<p>Article URL: <a href="https://pageindex.ai/blog/pageindex-chat">https://pageindex.ai/blog/pageindex-chat</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45663609">https://news.ycombinator.com/item?id=45663609</a></p>
<p>Points: 4</p>
<p># Comments: 1</p>
]]></description><pubDate>Wed, 22 Oct 2025 00:35:17 +0000</pubDate><link>https://pageindex.ai/blog/pageindex-chat</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45663609</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45663609</guid></item><item><title><![CDATA[Show HN: In-Context Index for In-Context Retrieval]]></title><description><![CDATA[
<p>RAG pipelines have become bloated: embeddings, vector DBs, rerankers, and ad-hoc pipelines everywhere.<p>Projects like Claude Code showed a simpler path: In-Context Retrieval — letting the LLM reason directly over context for retrieval instead of outsourcing search to external infrastructure.<p>PageIndex takes that one step further with In-Context Indexing.<p>If retrieval happens in-context, the index should live there too.<p>Each document is transformed into a hierarchical, human-readable tree structure (like a table-of-contents tree index) inside the model's context window.<p>The LLM reads the structure, identifies relevant branches, opens them, and reasons through for retrieval — no embeddings, no chunking, no opaque vector indexes the model can't interpret.<p>Retrieval and indexing, both inside the model.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45529851">https://news.ycombinator.com/item?id=45529851</a></p>
<p>Points: 5</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 09 Oct 2025 16:23:20 +0000</pubDate><link>https://github.com/VectifyAI/pageindex-mcp</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45529851</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45529851</guid></item><item><title><![CDATA[New comment by mingtianzhang in "DeepMind's paper reveals Google's new direction on RAG: In-Context Retreival"]]></title><description><![CDATA[
<p>Instead of relying on vector databases, DeepMind proposes:<p>1. The LLM itself selects the most relevant documents — no vector database needed.<p>2. The selected documents are then placed directly into the context for generation.<p>This kind of in-context retrieval approach greatly improves retrieval accuracy compared to traditional vector-based retrieval methods.</p>
]]></description><pubDate>Thu, 09 Oct 2025 15:38:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=45529217</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45529217</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45529217</guid></item><item><title><![CDATA[DeepMind's paper reveals Google's new direction on RAG: In-Context Retreival]]></title><description><![CDATA[
<p>Article URL: <a href="https://arxiv.org/abs/2510.05396">https://arxiv.org/abs/2510.05396</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45529216">https://news.ycombinator.com/item?id=45529216</a></p>
<p>Points: 6</p>
<p># Comments: 1</p>
]]></description><pubDate>Thu, 09 Oct 2025 15:38:33 +0000</pubDate><link>https://arxiv.org/abs/2510.05396</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45529216</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45529216</guid></item><item><title><![CDATA[From Claude Code to Agentic RAG]]></title><description><![CDATA[
<p>Article URL: <a href="https://vectifyai.notion.site/agentic-retrieval">https://vectifyai.notion.site/agentic-retrieval</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45513317">https://news.ycombinator.com/item?id=45513317</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 08 Oct 2025 07:53:50 +0000</pubDate><link>https://vectifyai.notion.site/agentic-retrieval</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45513317</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45513317</guid></item><item><title><![CDATA[From Claude Code to PageIndex: The Rise of Agentic Retrieval]]></title><description><![CDATA[
<p>Article URL: <a href="https://vectifyai.notion.site/agentic-retrieval">https://vectifyai.notion.site/agentic-retrieval</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45499612">https://news.ycombinator.com/item?id=45499612</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 07 Oct 2025 05:18:21 +0000</pubDate><link>https://vectifyai.notion.site/agentic-retrieval</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45499612</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45499612</guid></item><item><title><![CDATA[From Claude Code to PageIndex: The Rise of Agentic Retrieval]]></title><description><![CDATA[
<p>Article URL: <a href="https://vectifyai.notion.site/agentic-retrieval">https://vectifyai.notion.site/agentic-retrieval</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45493750">https://news.ycombinator.com/item?id=45493750</a></p>
<p>Points: 8</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 06 Oct 2025 17:26:42 +0000</pubDate><link>https://vectifyai.notion.site/agentic-retrieval</link><dc:creator>mingtianzhang</dc:creator><comments>https://news.ycombinator.com/item?id=45493750</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45493750</guid></item></channel></rss>