<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: karimf</title><link>https://news.ycombinator.com/user?id=karimf</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 08 Apr 2026 10:29:20 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=karimf" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by karimf in "Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B"]]></title><description><![CDATA[
<p>Well, on my demo it's around 2.5s and I already consider it as a "real-time". One way to improve it is to disable the image input.</p>
]]></description><pubDate>Tue, 07 Apr 2026 09:27:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=47672625</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47672625</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47672625</guid></item><item><title><![CDATA[New comment by karimf in "Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B"]]></title><description><![CDATA[
<p>What's your average response time with M1 max and what's the target?</p>
]]></description><pubDate>Tue, 07 Apr 2026 06:08:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=47671286</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47671286</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47671286</guid></item><item><title><![CDATA[New comment by karimf in "Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B"]]></title><description><![CDATA[
<p>Is this the problem? <a href="https://news.ycombinator.com/item?id=47669954">https://news.ycombinator.com/item?id=47669954</a></p>
]]></description><pubDate>Tue, 07 Apr 2026 06:07:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=47671277</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47671277</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47671277</guid></item><item><title><![CDATA[New comment by karimf in "Show HN: Ghost Pepper – Local hold-to-talk speech-to-text for macOS"]]></title><description><![CDATA[
<p>In the /r/macapps subreddit, they have huge influx of new apps posts, and the "whisper dictation" is one of the most saturated category. [0]<p>>“Compare” - This is the most important part. Apps in the most saturated categories (whisper dictation, clipboard managers, wallpaper apps, etc.) must clearly explain their differentiation from existing solutions.<p><a href="https://www.reddit.com/r/macapps/comments/1r6d06r/new_post_requirements_to_combat_low_quality/" rel="nofollow">https://www.reddit.com/r/macapps/comments/1r6d06r/new_post_r...</a></p>
]]></description><pubDate>Tue, 07 Apr 2026 00:02:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=47669021</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47669021</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47669021</guid></item><item><title><![CDATA[New comment by karimf in "Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B"]]></title><description><![CDATA[
<p>During my limited testing, it works better than I expected at handling multiple languages in a single session. Perhaps I just had a low expectation since I've mostly worked with English-only STT models.</p>
]]></description><pubDate>Mon, 06 Apr 2026 16:25:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=47663010</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47663010</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47663010</guid></item><item><title><![CDATA[New comment by karimf in "Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B"]]></title><description><![CDATA[
<p>Nothing unique, it's just taking a snapshot when it's processing the input. Even processing a single image will increase the TTFT by ~0.5s on my machine, so for now, it seems to be impossible for feeding a live video and expecting a real-time response.<p>In regards to the video capability, I haven't tested it myself, but here's a benchmark/comparison from Google [0]<p>[0] <a href="https://huggingface.co/blog/gemma4#video-understanding" rel="nofollow">https://huggingface.co/blog/gemma4#video-understanding</a></p>
]]></description><pubDate>Mon, 06 Apr 2026 16:20:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=47662932</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47662932</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47662932</guid></item><item><title><![CDATA[New comment by karimf in "Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B"]]></title><description><![CDATA[
<p>Huh that's weird. I just tried it and it works on my machine. Could you perhaps create a GitHub issue and share the reproduction steps and any relevant logs?</p>
]]></description><pubDate>Mon, 06 Apr 2026 16:11:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47662802</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47662802</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47662802</guid></item><item><title><![CDATA[New comment by karimf in "I won't download your app. The web version is a-ok"]]></title><description><![CDATA[
<p>This. I posted this on my other comment, but there's a meme that "Gen Z Kids Don't Understand How File Systems Work" [0].<p>There seems to be a disconnect between some developers and the younger folks.<p>[0] <a href="https://news.ycombinator.com/item?id=30253526">https://news.ycombinator.com/item?id=30253526</a></p>
]]></description><pubDate>Mon, 06 Apr 2026 15:09:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=47661953</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47661953</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47661953</guid></item><item><title><![CDATA[New comment by karimf in "I won't download your app. The web version is a-ok"]]></title><description><![CDATA[
<p>This is my stance as well, but keep in mind that a lot of people have the opposite preference.<p>They didn't grow up with the world wide web. They only started using technology when Android and iPhone was popular. They only know Whatsapp, Youtube, TikTok. They're not used to using the browser.<p>There's a meme that "Gen Z Kids Don't Understand How File Systems Work" [0]<p>So, it'll depend on your target audiences.<p>[0] <a href="https://news.ycombinator.com/item?id=30253526">https://news.ycombinator.com/item?id=30253526</a></p>
]]></description><pubDate>Mon, 06 Apr 2026 15:04:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=47661877</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47661877</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47661877</guid></item><item><title><![CDATA[New comment by karimf in "Gemma 4 on iPhone"]]></title><description><![CDATA[
<p>Oh wow, that's awesome. Thanks a lot, dang!</p>
]]></description><pubDate>Mon, 06 Apr 2026 06:10:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=47657502</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47657502</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47657502</guid></item><item><title><![CDATA[New comment by karimf in "Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B"]]></title><description><![CDATA[
<p>Thank you. This reminds me of a paragraph from the LatentSpace newsletter [0]<p>> The excellent on device capabilities makes one wonder if these are the basis for the models that will be deployed in New Siri under the deal with Apple….<p><a href="https://www.latent.space/p/ainews-gemma-4-the-best-small-multimodal" rel="nofollow">https://www.latent.space/p/ainews-gemma-4-the-best-small-mul...</a></p>
]]></description><pubDate>Mon, 06 Apr 2026 05:52:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=47657416</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47657416</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47657416</guid></item><item><title><![CDATA[New comment by karimf in "Gemma 4 on iPhone"]]></title><description><![CDATA[
<p>Thanks for sharing! I'm still torn about it. Sure it'll feel more natural if you have the AI head animation, but I don't want people to get attached to it. I don't want to make the loneliness epidemic even worse.</p>
]]></description><pubDate>Mon, 06 Apr 2026 05:44:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=47657375</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47657375</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47657375</guid></item><item><title><![CDATA[New comment by karimf in "Gemma 4 on iPhone"]]></title><description><![CDATA[
<p>Thanks! Although, I can't claim any credit for it. I just spent a day gluing what other people have built. Huge props to the Gemma team for building an amazing model and also an inference engine that's focused for edge devices [0]<p>[0] <a href="https://github.com/google-ai-edge/LiteRT-LM" rel="nofollow">https://github.com/google-ai-edge/LiteRT-LM</a></p>
]]></description><pubDate>Sun, 05 Apr 2026 21:24:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=47654010</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47654010</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47654010</guid></item><item><title><![CDATA[New comment by karimf in "Gemma 4 on iPhone"]]></title><description><![CDATA[
<p>This app is cool and it showcases some use cases, but it still undersells what the E2B model can do.<p>I just made a real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B. I posted it on /r/LocalLLaMA a few hours ago and it's gaining some traction [0]. Here's the repo [1]<p>I'm running it on a Macbook instead of an iPhone, but based on the benchmark here [2], you should be able to run the same thing on an iPhone 17 Pro.<p>[0] <a href="https://www.reddit.com/r/LocalLLaMA/comments/1sda3r6/realtime_ai_audiovideo_in_voice_out_on_an_m3_pro/" rel="nofollow">https://www.reddit.com/r/LocalLLaMA/comments/1sda3r6/realtim...</a><p>[1] <a href="https://github.com/fikrikarim/parlor" rel="nofollow">https://github.com/fikrikarim/parlor</a><p>[2] <a href="https://huggingface.co/litert-community/gemma-4-E2B-it-litert-lm" rel="nofollow">https://huggingface.co/litert-community/gemma-4-E2B-it-liter...</a></p>
]]></description><pubDate>Sun, 05 Apr 2026 20:51:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=47653752</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47653752</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47653752</guid></item><item><title><![CDATA[New comment by karimf in "Google releases Gemma 4 open models"]]></title><description><![CDATA[
<p>Update: Just made one that runs on Macbook M3 Pro <a href="https://github.com/fikrikarim/parlor" rel="nofollow">https://github.com/fikrikarim/parlor</a></p>
]]></description><pubDate>Sun, 05 Apr 2026 19:49:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=47653165</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47653165</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47653165</guid></item><item><title><![CDATA[Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B]]></title><description><![CDATA[
<p>Related: <a href="https://news.ycombinator.com/item?id=47653752">https://news.ycombinator.com/item?id=47653752</a></p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47652007">https://news.ycombinator.com/item?id=47652007</a></p>
<p>Points: 287</p>
<p># Comments: 35</p>
]]></description><pubDate>Sun, 05 Apr 2026 17:53:19 +0000</pubDate><link>https://github.com/fikrikarim/parlor</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47652007</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47652007</guid></item><item><title><![CDATA[New comment by karimf in "OpenScreen is an open-source alternative to Screen Studio"]]></title><description><![CDATA[
<p>Nice project. I thought of building exactly this.<p>Since it's much easier to port source code to other languages now, I'd love to see more projects like written in Swift, or C#.</p>
]]></description><pubDate>Sun, 05 Apr 2026 01:29:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=47645291</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47645291</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47645291</guid></item><item><title><![CDATA[New comment by karimf in "Show HN: Apfel – The free AI already on your Mac"]]></title><description><![CDATA[
<p>The big question is whether Apple can keep shipping new models constantly.<p>AFAIK the current model is on par with with Qwen-3-4B, which is from a year ago [0]. There's a big leap going from last year Qwen-3-4B to Qwen-3.5-4B or to Gemma 4.<p>Apple model is nice since you don't need to download anything else, but I'd rather use the latest model than to use a model from a year ago.<p><a href="https://machinelearning.apple.com/research/apple-foundation-models-2025-updates" rel="nofollow">https://machinelearning.apple.com/research/apple-foundation-...</a></p>
]]></description><pubDate>Fri, 03 Apr 2026 16:35:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=47628820</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47628820</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47628820</guid></item><item><title><![CDATA[New comment by karimf in "Google releases Gemma 4 open models"]]></title><description><![CDATA[
<p>Thank you for the kind words!</p>
]]></description><pubDate>Fri, 03 Apr 2026 06:22:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=47623758</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47623758</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47623758</guid></item><item><title><![CDATA[New comment by karimf in "Ollama is now powered by MLX on Apple Silicon in preview"]]></title><description><![CDATA[
<p>Ok it's on the app store now: <a href="https://apps.apple.com/app/volocal/id6761493288">https://apps.apple.com/app/volocal/id6761493288</a></p>
]]></description><pubDate>Thu, 02 Apr 2026 23:03:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=47621337</link><dc:creator>karimf</dc:creator><comments>https://news.ycombinator.com/item?id=47621337</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47621337</guid></item></channel></rss>