<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: yeldarb</title><link>https://news.ycombinator.com/user?id=yeldarb</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 07 Apr 2026 07:00:38 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=yeldarb" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by yeldarb in "Meta Segment Anything Model 3"]]></title><description><![CDATA[
<p>Yes, it should.</p>
]]></description><pubDate>Thu, 20 Nov 2025 04:26:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=45988952</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=45988952</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45988952</guid></item><item><title><![CDATA[New comment by yeldarb in "Meta Segment Anything Model 3"]]></title><description><![CDATA[
<p>We have a JS SDK that supports RF-DETR: <a href="https://docs.roboflow.com/deploy/sdks/web-browser">https://docs.roboflow.com/deploy/sdks/web-browser</a></p>
]]></description><pubDate>Thu, 20 Nov 2025 04:24:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=45988935</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=45988935</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45988935</guid></item><item><title><![CDATA[New comment by yeldarb in "Meta Segment Anything Model 3"]]></title><description><![CDATA[
<p>We used DINOv2 as the backbone of our RF-DETR model, which is SOTA on realtime object detection and segmentation: <a href="https://github.com/roboflow/rf-detr" rel="nofollow">https://github.com/roboflow/rf-detr</a><p>It makes a great target to distill SAM3 to.</p>
]]></description><pubDate>Thu, 20 Nov 2025 04:22:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=45988930</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=45988930</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45988930</guid></item><item><title><![CDATA[New comment by yeldarb in "Meta Segment Anything Model 3"]]></title><description><![CDATA[
<p>We (Roboflow) have had early access to this model for the past few weeks. It's really, really good. This feels like a seminal moment for computer vision. I think there's a real possibility this launch goes down in history as "the GPT Moment" for vision.
The two areas I think this model is going to be transformative in the immediate term are for rapid prototyping and distillation.<p>Two years ago we released autodistill[1], an open source framework that uses large foundation models to create training data for training small realtime models. I'm convinced the idea was right, but too early; there wasn't a big model good enough to be worth distilling from back then. SAM3 is finally that model (and will be available in Autodistill today).<p>We are also taking a big bet on SAM3 and have built it into Roboflow as an integral part of the entire build and deploy pipeline[2], including a brand new product called Rapid[3], which reimagines the computer vision pipeline in a SAM3 world. It feels really magical to go from an unlabeled video to a fine-tuned realtime segmentation model with minimal human intervention in just a few minutes (and we rushed the release of our new SOTA realtime segmentation model[4] last week because it's the perfect lightweight complement to the large & powerful SAM3).<p>We also have a playground[5] up where you can play with the model and compare it to other VLMs.<p>[1] <a href="https://github.com/autodistill/autodistill" rel="nofollow">https://github.com/autodistill/autodistill</a><p>[2] <a href="https://blog.roboflow.com/sam3/">https://blog.roboflow.com/sam3/</a><p>[3] <a href="https://rapid.roboflow.com">https://rapid.roboflow.com</a><p>[4] <a href="https://github.com/roboflow/rf-detr" rel="nofollow">https://github.com/roboflow/rf-detr</a><p>[5] <a href="https://playground.roboflow.com">https://playground.roboflow.com</a></p>
]]></description><pubDate>Wed, 19 Nov 2025 20:38:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=45984797</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=45984797</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45984797</guid></item><item><title><![CDATA[New comment by yeldarb in "Segment Anything 3"]]></title><description><![CDATA[
<p>We (Roboflow) have had early access to this model for the past few weeks. It's really, really good. This feels like a seminal moment for computer vision. I think there's a real possibility this launch goes down in history as "the GPT Moment" for vision.<p>The two areas I think this model is going to be transformative in the immediate term are for rapid prototyping and distillation.<p>Two years ago we released autodistill[1], an open source framework that uses large foundation models to create training data for training small realtime models. I'm convinced the idea was right, but too early; there wasn't a big model good enough to be worth distilling from back then. SAM3 is finally that model (and will be available in Autodistill today).<p>We are also taking a big bet on SAM3 and have built it into Roboflow as an integral part of the entire build and deploy pipeline[2], including a brand new product called Rapid[3], which reimagines the computer vision pipeline in a SAM3 world. It feels really magical to go from an unlabeled video to a fine-tuned realtime segmentation model with minimal human intervention in just a few minutes (and we rushed the release of our new SOTA realtime segmentation model[4] last week because it's the perfect lightweight complement to the large & powerful SAM3).<p>We also have a playground[5] up where you can play with the model and compare it to other VLMs.<p>[1] <a href="https://github.com/autodistill/autodistill" rel="nofollow">https://github.com/autodistill/autodistill</a><p>[2] <a href="https://blog.roboflow.com/sam3/">https://blog.roboflow.com/sam3/</a><p>[3] <a href="https://rapid.roboflow.com">https://rapid.roboflow.com</a><p>[4] <a href="https://github.com/roboflow/rf-detr" rel="nofollow">https://github.com/roboflow/rf-detr</a><p>[5] <a href="https://playground.roboflow.com">https://playground.roboflow.com</a></p>
]]></description><pubDate>Wed, 19 Nov 2025 16:19:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=45981400</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=45981400</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45981400</guid></item><item><title><![CDATA[GPT-5 Coding Examples]]></title><description><![CDATA[
<p>Article URL: <a href="https://github.com/openai/gpt-5-coding-examples">https://github.com/openai/gpt-5-coding-examples</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=44826439">https://news.ycombinator.com/item?id=44826439</a></p>
<p>Points: 10</p>
<p># Comments: 1</p>
]]></description><pubDate>Thu, 07 Aug 2025 16:14:19 +0000</pubDate><link>https://github.com/openai/gpt-5-coding-examples</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=44826439</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44826439</guid></item><item><title><![CDATA[New comment by yeldarb in "Persona vectors: Monitoring and controlling character traits in language models"]]></title><description><![CDATA[
<p>Wonder if you can subtract these vectors to get the opposite effect and what that ends up being for things like sycophancy or hallucination.<p>I also wonder what other personality vectors exist.. would be cool to find an “intelligence” vector we could boost to get better outputs from the same model. Seems like this is likely to exist given how prompting it to cosplay as a really smart person can elicit better outputs.</p>
]]></description><pubDate>Mon, 04 Aug 2025 09:38:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=44783646</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=44783646</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44783646</guid></item><item><title><![CDATA[New comment by yeldarb in "Google AI Edge – On-device cross-platform AI deployment"]]></title><description><![CDATA[
<p>Is this a new product or a marketing page tying together a bunch of the existing MediaPipe stuff into a narrative?<p>Got really excited then realized I couldn’t figure out what “Google AI Edge” actually _is_.<p>Edit: I think it’s largely a rebrand of this from a couple years ago: <a href="https://developers.googleblog.com/en/introducing-mediapipe-solutions-for-on-device-machine-learning/" rel="nofollow">https://developers.googleblog.com/en/introducing-mediapipe-s...</a></p>
]]></description><pubDate>Sun, 01 Jun 2025 12:53:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=44150524</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=44150524</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44150524</guid></item><item><title><![CDATA[Cohere Aya Vision: Expanding the Worlds AI Can See]]></title><description><![CDATA[
<p>Article URL: <a href="https://cohere.com/blog/aya-vision">https://cohere.com/blog/aya-vision</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43257085">https://news.ycombinator.com/item?id=43257085</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 04 Mar 2025 16:45:26 +0000</pubDate><link>https://cohere.com/blog/aya-vision</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=43257085</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43257085</guid></item><item><title><![CDATA[New comment by yeldarb in "Show HN: NYCerebro, semantic search of NYC traffic cams (written by v0)"]]></title><description><![CDATA[
<p>I've been reflecting a bit on this and remembering what it used to be like when I did hackathons regularly a decade or so ago. This project seems on-par with the type of 48 hour hackathon project I used to do (assuming CLIP had existed), but now I was able to do it in 2 hours instead of 48.<p>I can't imagine someone non-technical building something like this with prompting. The success of the project was highly dependent on my direction of the model to do what I wanted it to do (even though I gave it leeway in exactly how to do it). It did feel a bit like managing another engineer to do something vs doing it myself.<p>I don't use agents like this in my day to day work yet (I experimented with OpenHands a couple of months ago but it was frustrating, expensive, and took just as long as doing the task myself). But I'm thinking I probably will be a year from now.<p>A few times when the model got stuck I copy/pasted some stuff into o1 and pasted its response back into v0 (felt kind of like "escalating" to a more senior engineer) and that helped it get unstuck. Future models will be even more capable than o1. I imagine there will likely need to be a UI for "bringing in the big guns" of a smarter model in the future even if the grunt-work is done by a fast+cheap base model.<p>There's probably also something to letting the model "speak its native tongue". I don't know next.js but letting the model work with patterns it's been trained on probably helped it be more effective (compared to having OpenHands work in my own codebase using a structure it's unfamiliar with).</p>
]]></description><pubDate>Thu, 23 Jan 2025 14:55:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=42804629</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=42804629</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42804629</guid></item><item><title><![CDATA[New comment by yeldarb in "Show HN: NYCerebro, semantic search of NYC traffic cams (written by v0)"]]></title><description><![CDATA[
<p>Hey all, sharing a project we made in 2 hours at the Vercel+NVIDIA hackathon last week.<p>While the app is cool, the thing that blew my mind is that the <i>entire</i> app was coded by Vercel's v0 agent. In other words: I did not write a single line of code to create the app (though my teammate did write the backend scraper & DB filler by hand).<p>[1] Writeup: <a href="https://blog.roboflow.com/nycerebro/">https://blog.roboflow.com/nycerebro/</a><p>[2] Repo (including the generated code + initial meaty prompts): <a href="https://github.com/yeldarby/nycerebro">https://github.com/yeldarby/nycerebro</a><p>[3] v0 session: <a href="https://v0.dev/chat/nyc-erebro-app-RwzRUEMGveH?b=b_6AuWalvG7B3&f=0" rel="nofollow">https://v0.dev/chat/nyc-erebro-app-RwzRUEMGveH?b=b_6AuWalvG7...</a></p>
]]></description><pubDate>Thu, 23 Jan 2025 14:21:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=42804279</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=42804279</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42804279</guid></item><item><title><![CDATA[Show HN: NYCerebro, semantic search of NYC traffic cams (written by v0)]]></title><description><![CDATA[
<p>Article URL: <a href="https://nycerebro.vercel.app/">https://nycerebro.vercel.app/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42804278">https://news.ycombinator.com/item?id=42804278</a></p>
<p>Points: 4</p>
<p># Comments: 2</p>
]]></description><pubDate>Thu, 23 Jan 2025 14:21:30 +0000</pubDate><link>https://nycerebro.vercel.app/</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=42804278</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42804278</guid></item><item><title><![CDATA[GPU Fryer]]></title><description><![CDATA[
<p>Article URL: <a href="https://github.com/huggingface/gpu-fryer">https://github.com/huggingface/gpu-fryer</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42773967">https://news.ycombinator.com/item?id=42773967</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 20 Jan 2025 22:41:04 +0000</pubDate><link>https://github.com/huggingface/gpu-fryer</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=42773967</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42773967</guid></item><item><title><![CDATA[Putting the New M4 Macs to the Test]]></title><description><![CDATA[
<p>Article URL: <a href="https://blog.roboflow.com/putting-the-new-m4-macs-to-the-test/">https://blog.roboflow.com/putting-the-new-m4-macs-to-the-test/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42410353">https://news.ycombinator.com/item?id=42410353</a></p>
<p>Points: 4</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 13 Dec 2024 17:14:48 +0000</pubDate><link>https://blog.roboflow.com/putting-the-new-m4-macs-to-the-test/</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=42410353</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42410353</guid></item><item><title><![CDATA[New comment by yeldarb in "Bocker: Docker implemented in around 100 lines of Bash (2015)"]]></title><description><![CDATA[
<p>Is there any Docker alternative on Mac that can utilize the MPS device in a container? ML stuff is many times slower in a container on my Mac than running outside</p>
]]></description><pubDate>Sun, 24 Nov 2024 02:36:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=42225537</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=42225537</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42225537</guid></item><item><title><![CDATA[New comment by yeldarb in "PyTorch Deprecation of Conda Nightly Builds"]]></title><description><![CDATA[
<p>More context from Jeremy Howard (fast.ai): <a href="https://x.com/jeremyphoward/status/1857765905188651456" rel="nofollow">https://x.com/jeremyphoward/status/1857765905188651456</a></p>
]]></description><pubDate>Sun, 17 Nov 2024 03:54:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=42161856</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=42161856</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42161856</guid></item><item><title><![CDATA[PyTorch Deprecation of Conda Nightly Builds]]></title><description><![CDATA[
<p>Article URL: <a href="https://dev-discuss.pytorch.org/t/pytorch-deprecation-of-conda-nightly-builds/2590">https://dev-discuss.pytorch.org/t/pytorch-deprecation-of-conda-nightly-builds/2590</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42161854">https://news.ycombinator.com/item?id=42161854</a></p>
<p>Points: 3</p>
<p># Comments: 1</p>
]]></description><pubDate>Sun, 17 Nov 2024 03:53:24 +0000</pubDate><link>https://dev-discuss.pytorch.org/t/pytorch-deprecation-of-conda-nightly-builds/2590</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=42161854</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42161854</guid></item><item><title><![CDATA[New comment by yeldarb in "IMG_0416"]]></title><description><![CDATA[
<p>It’s sad that only Google can (and honestly a bit surprising that Google hasn’t) use multimodal video models to index the semantic contents & transcripts of these videos for search. Huge long tail of unique content.</p>
]]></description><pubDate>Mon, 11 Nov 2024 00:55:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=42103713</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=42103713</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42103713</guid></item><item><title><![CDATA[New comment by yeldarb in "Video Surveillance with YOLO+llava"]]></title><description><![CDATA[
<p>If you do it naively your video frames will buffer waiting to be consumed causing a memory leak and eventual crash (or quick crash if you’re running on a device with constrained resources).<p>You really need to have a thread consuming the frames and feeding them to a worker that can run on its own clock.</p>
]]></description><pubDate>Tue, 08 Oct 2024 05:00:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=41773987</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=41773987</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41773987</guid></item><item><title><![CDATA[New comment by yeldarb in "Video Surveillance with YOLO+llava"]]></title><description><![CDATA[
<p>We’ve got an open source pipeline as part of inference[1] that handles the nuances (multithreading, batching, syncing, reconnecting) of running multiple real time streams (pass in an array of RTSP urls) for CV models like YOLO: <a href="https://blog.roboflow.com/vision-models-multiple-streams/">https://blog.roboflow.com/vision-models-multiple-streams/</a><p>[1] <a href="https://github.com/roboflow/inference">https://github.com/roboflow/inference</a></p>
]]></description><pubDate>Tue, 08 Oct 2024 04:58:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=41773978</link><dc:creator>yeldarb</dc:creator><comments>https://news.ycombinator.com/item?id=41773978</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41773978</guid></item></channel></rss>