<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: parsakhaz</title><link>https://news.ycombinator.com/user?id=parsakhaz</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 28 Apr 2026 11:38:44 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=parsakhaz" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by parsakhaz in "Why LLMs still have problems with OCR"]]></title><description><![CDATA[
<p>Yup, Moondream is great for this use case! You can use locally with the quickstart: <a href="https://docs.moondream.ai/" rel="nofollow">https://docs.moondream.ai/</a><p>It is a 2b vision model that runs anywhere and can object detect, point, query, and more.</p>
]]></description><pubDate>Mon, 10 Feb 2025 23:37:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=43006728</link><dc:creator>parsakhaz</dc:creator><comments>https://news.ycombinator.com/item?id=43006728</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43006728</guid></item><item><title><![CDATA[New comment by parsakhaz in "Coping with dumb LLMs using classic ML"]]></title><description><![CDATA[
<p>Thanks for the shout out :)</p>
]]></description><pubDate>Fri, 31 Jan 2025 21:18:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=42892164</link><dc:creator>parsakhaz</dc:creator><comments>https://news.ycombinator.com/item?id=42892164</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42892164</guid></item><item><title><![CDATA[New comment by parsakhaz in "Coping with dumb LLMs using classic ML"]]></title><description><![CDATA[
<p>We've run a couple experiments and have found that our open vision language model Moondream works better than YOLOv11 in general cases. If accuracy matters most, it's worth trying our vision language model. If you need real-time results, you can train YOLO models using data from our model. We have a space for video redaction, that is just object detection, on our Hugging Face. We also have a playground online to try it out.</p>
]]></description><pubDate>Fri, 31 Jan 2025 21:18:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=42892153</link><dc:creator>parsakhaz</dc:creator><comments>https://news.ycombinator.com/item?id=42892153</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42892153</guid></item><item><title><![CDATA[New comment by parsakhaz in "Guide: How to use Moondream's free OpenAI compatible endpoint (5k queries/day)"]]></title><description><![CDATA[
<p>Send a message into our discord, and we will get it bumped up for you: <a href="https://discord.com/invite/tRUdpjDQfH" rel="nofollow">https://discord.com/invite/tRUdpjDQfH</a></p>
]]></description><pubDate>Wed, 15 Jan 2025 21:22:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=42717216</link><dc:creator>parsakhaz</dc:creator><comments>https://news.ycombinator.com/item?id=42717216</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42717216</guid></item><item><title><![CDATA[New comment by parsakhaz in "Guide: How to use Moondream's free OpenAI compatible endpoint (5k queries/day)"]]></title><description><![CDATA[
<p>We just rolled out OpenAI compatibility for Moondream 2B, which means that you can now seamlessly switch from OpenAI's Vision API to Moondream with minimal changes to your existing code.<p>Our docs, like Moondream, are open source. If you find any issues with the page, or want to suggest a change, click "Edit this page" and you'll be routed to the GitHub for the docs.<p>The best part is - our API is free, for up to 5k requests. There is zero friction in getting started with and trying Moondream. We are also very active on discord, so if you get stuck (or have a special request), let us know and we will be quick to help.<p>Looking forwards to seeing all the cool stuff that people build!<p>---<p>What is Moondream?<p>Moondream-2B is a lightweight vision-language model optimized for visual understanding tasks. It excels at answering questions about images, describing scenes, identifying objects and attributes, and basic text recognition. While more compact than larger models, it provides efficient and accurate responses for straightforward visual question-answering.<p>As a 2B parameter model, it has some limitations to keep in mind: descriptions may be less detailed than larger models, complex multi-step reasoning can be challenging, and it may struggle with edge cases like very low-quality images or advanced spatial understanding. For best results, focus on direct questions about image content rather than complex reasoning chains.</p>
]]></description><pubDate>Wed, 15 Jan 2025 19:35:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=42715857</link><dc:creator>parsakhaz</dc:creator><comments>https://news.ycombinator.com/item?id=42715857</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42715857</guid></item><item><title><![CDATA[Guide: How to use Moondream's free OpenAI compatible endpoint (5k queries/day)]]></title><description><![CDATA[
<p>Article URL: <a href="https://docs.moondream.ai/openai-compatibility">https://docs.moondream.ai/openai-compatibility</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42715855">https://news.ycombinator.com/item?id=42715855</a></p>
<p>Points: 3</p>
<p># Comments: 3</p>
]]></description><pubDate>Wed, 15 Jan 2025 19:35:06 +0000</pubDate><link>https://docs.moondream.ai/openai-compatibility</link><dc:creator>parsakhaz</dc:creator><comments>https://news.ycombinator.com/item?id=42715855</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42715855</guid></item></channel></rss>