<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: k9294</title><link>https://news.ycombinator.com/user?id=k9294</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 10 Apr 2026 11:10:08 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=k9294" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by k9294 in "Dropping Cloudflare for Bunny.net"]]></title><description><![CDATA[
<p>Nope, but I will think about this, thank you for the idea. Maybe it's time to start a technical blog for ottex</p>
]]></description><pubDate>Wed, 08 Apr 2026 09:04:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47687392</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47687392</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47687392</guid></item><item><title><![CDATA[New comment by k9294 in "Dropping Cloudflare for Bunny.net"]]></title><description><![CDATA[
<p>There is no cold starts at all. It’s running non-stop.<p>Bunny bills per resource utilization (not provisioned) and since we run backend on Go it consumes like 0.01 CPU and 15mb RAM per idle container and costs pennies.</p>
]]></description><pubDate>Tue, 07 Apr 2026 19:49:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=47680475</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47680475</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47680475</guid></item><item><title><![CDATA[New comment by k9294 in "Dropping Cloudflare for Bunny.net"]]></title><description><![CDATA[
<p>We at ottex.ai use bunny.net to deploy globally an openrouter like speach-to-text API (5 continents, 26 locations, idle cost 3$).<p>Highly recommend their Edge Containers product, super simple and has nice primitives to deploy globally for a low latency workloads.<p>We connect all containers to one redis pubsub server to push important events like user billing overages, top-ups etc. Super simple, very fast, one config to manage all locations.</p>
]]></description><pubDate>Tue, 07 Apr 2026 16:24:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47677724</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47677724</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47677724</guid></item><item><title><![CDATA[New comment by k9294 in "Issue: Claude Code is unusable for complex engineering tasks with Feb updates"]]></title><description><![CDATA[
<p>Anecdotally, I’ve been seeing a lot of weird behavior from Opus when it decides, mid-execution, to switch to a different "simpler" solution, and that really pissed me off.<p>At one point, I carefully designed a spec document, forced Opus to reread it, create a plan with the planning tool that followed the spec, and use the task tool to track the implementation... AND AFTER OPUS READS THE FIRST FUCKING FILE, it says, "Oh, there are missing dependencies in project X. It’ll be hard to add them, so I’m going to throw away the whole plan and just do a simple fix..."<p>After that, I canceled my $200 Max plan, which I’d been subscribed to since June 2025, and decided to check out Codex</p>
]]></description><pubDate>Tue, 07 Apr 2026 11:21:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47673482</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47673482</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47673482</guid></item><item><title><![CDATA[New comment by k9294 in "Show HN: Ghost Pepper – Local hold-to-talk speech-to-text for macOS"]]></title><description><![CDATA[
<p>Try ottex.ai - it has an OpenRouter like gateway with most STT models on the market (Gemini, OpenAI, Groq, Deepgram, Mistral, AssemblyAI, Soniox), so you can try them all and choose what works best for you.<p>My favorites are Gemini 3 Flash and Mistral Voxtral Transcribe 2. Gemini when I need special formatting and clean-up, and Voxtral when I need fast input (mostly when working with AI).</p>
]]></description><pubDate>Tue, 07 Apr 2026 07:54:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47672002</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47672002</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47672002</guid></item><item><title><![CDATA[New comment by k9294 in "Gemini 3.1 Flash-Lite: Built for intelligence at scale"]]></title><description><![CDATA[
<p>It's actually a nice idea - an always-on micro AI agent with voice-to-text capabilities that listens and acts on your behalf.<p>Actually, I'm experimenting with this kind of stuff and trying to find a nice UX to make Ottex a voice command center - to trigger AI agents like Claude, open code to work on something, execute simple commands, etc.</p>
]]></description><pubDate>Tue, 03 Mar 2026 18:27:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47236578</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47236578</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47236578</guid></item><item><title><![CDATA[New comment by k9294 in "Gemini 3.1 Flash-Lite: Built for intelligence at scale"]]></title><description><![CDATA[
<p>I've been experimenting with Gemini 3.1 Flash Lite and the quality is very good.<p>I haven't found official benchmarks yet, but you can find Gemini 3 Flash word error rate benchmarks here: <a href="https://artificialanalysis.ai/speech-to-text/models/gemini" rel="nofollow">https://artificialanalysis.ai/speech-to-text/models/gemini</a> — they are close to SOTA.<p>I speak daily in both English and Russian and have been using Gemini 3 Flash as my main transcription model for a few months. I haven't seen any model that provides better overall quality in terms of understanding, custom dictionary support, instruction following, and formatting. It's the best STT model in my experience. Gemini 3 Flash has somewhat uncomfortable latency though, and Flash Lite is much better in this regard.</p>
]]></description><pubDate>Tue, 03 Mar 2026 18:20:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=47236467</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47236467</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47236467</guid></item><item><title><![CDATA[New comment by k9294 in "Gemini 3.1 Flash-Lite: Built for intelligence at scale"]]></title><description><![CDATA[
<p>You can test Gemini 3.1 Lite transcription capabilities in <a href="https://ottex.ai" rel="nofollow">https://ottex.ai</a> — the only dictation app supporting Gemini models with native audio input.<p>We benchmarked it for real-life voice-to-text use cases:<p><pre><code>                <10s    10-30s   30s-1m    1-2m    2-3m
  Flash         2548     2732     3177     4583    5961
  Flash Lite    1390     1468     1772     2362    3499
  Faster by    1.83x    1.86x    1.79x   1.94x   1.70x

  (latency in ms, median over 5 runs per sample, non-streaming)
</code></pre>
Key takeaways:<p>- 1.8x faster than Gemini 3 Flash on average<p>- ~1.4 sec transcription time for short to medium recordings<p>- ~$0.50/mo for heavy users (10h+ transcription)<p>- Close to SOTA audio understanding and formatting instruction following<p>- Multilingual: one model, 100+ languages<p>Gemini is slowly making $15/month voice apps obsolete.</p>
]]></description><pubDate>Tue, 03 Mar 2026 17:40:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=47235912</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47235912</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47235912</guid></item><item><title><![CDATA[New comment by k9294 in "Gemini 3.1 Flash Lite Preview"]]></title><description><![CDATA[
<p>Gemini 3.1 Flash-Lite is our most cost-efficient Gemini model, optimized for low latency use cases for high-volume, cost-sensitive LLM traffic.<p>It provides a significant quality increase over Gemini 2.0 Flash-Lite and Flash-Lite models, matching Gemini 2.5 Flash performance across key capability areas:<p>Improved response quality: Aims to match 2.5 Flash performance and align with target Flash-Lite use cases.<p>Improved instruction following: Targeted improvements to serve as a reliable migration path for complex chatbot and instruction-heavy workflows.<p>Improved audio input: Improved audio-input quality for tasks like Automated Speech Recognition (ASR).<p>Expanded thinking support: You can control how much reasoning the model performs by choosing from minimal, low, medium, or high thinking levels. This feature lets you balance response quality and speed for your specific use case.<p>---<p>Already available in Google AI Studio and OpenRouter<p><a href="https://openrouter.ai/google/gemini-3.1-flash-lite-preview" rel="nofollow">https://openrouter.ai/google/gemini-3.1-flash-lite-preview</a></p>
]]></description><pubDate>Tue, 03 Mar 2026 16:29:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=47234850</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47234850</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47234850</guid></item><item><title><![CDATA[Gemini 3.1 Flash Lite Preview]]></title><description><![CDATA[
<p>Article URL: <a href="https://docs.cloud.google.com/vertex-ai/generative-ai/docs/models/gemini/3-1-flash-lite">https://docs.cloud.google.com/vertex-ai/generative-ai/docs/models/gemini/3-1-flash-lite</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47234849">https://news.ycombinator.com/item?id=47234849</a></p>
<p>Points: 4</p>
<p># Comments: 1</p>
]]></description><pubDate>Tue, 03 Mar 2026 16:29:32 +0000</pubDate><link>https://docs.cloud.google.com/vertex-ai/generative-ai/docs/models/gemini/3-1-flash-lite</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47234849</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47234849</guid></item><item><title><![CDATA[New comment by k9294 in "Show HN: Free alternative to Wispr Flow, Superwhisper, and Monologue"]]></title><description><![CDATA[
<p>Try ottex with Gemini 3 flash as a transcription model. I'm bilingual as well and frequently switch between languages - Gemini handles this perfectly and even the case when I speak two languages in one transcription.</p>
]]></description><pubDate>Tue, 17 Feb 2026 10:36:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47045870</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47045870</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47045870</guid></item><item><title><![CDATA[New comment by k9294 in "Show HN: Free alternative to Wispr Flow, Superwhisper, and Monologue"]]></title><description><![CDATA[
<p>You can try ottex for this use case - it has both context capture (app screenshots), native LLMs support, meaning it can send audio AND screenshot directly to gemini 3 flash to produce the bespoke result.</p>
]]></description><pubDate>Tue, 17 Feb 2026 10:34:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=47045844</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47045844</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47045844</guid></item><item><title><![CDATA[New comment by k9294 in "Show HN: Free alternative to Wispr Flow, Superwhisper, and Monologue"]]></title><description><![CDATA[
<p>I'm building in the same space, Workin on <a href="https://ottex.ai" rel="nofollow">https://ottex.ai</a> - It's a free STT app, with local models and BYOK support (OpenRouter, Groq, Mistral, and more).<p>The top feature is the per-app custom settings - you can peak different models and instructions for different apps and websites.<p>- I use the Parakeet fast model when working with Claude Code (VS Code app).
- And I use a smart one when I draft notes in Obsidian. I have a prompt to clean up my rambling and format the result with proper Markdown, very convenient.<p>One more cool thing is that it allows me to use LLMs with audio input modalities directly (not as text post-processing). e.g. It sends the audio to Gemini and prompts it to transcribe, format, etc., in one run. I find it a bit slow to work with CC, but it is the absolute best model in terms of accuracy, understanding, and formatting. It is the only model I trust to understand what I meant and produce the correct result, even when I use multiple languages, tech terms, etc.</p>
]]></description><pubDate>Tue, 17 Feb 2026 10:23:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=47045762</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=47045762</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47045762</guid></item><item><title><![CDATA[New comment by k9294 in "Voxtral Transcribe 2"]]></title><description><![CDATA[
<p>Hey, I would really appreciate if you would try <a href="https://ottex.ai" rel="nofollow">https://ottex.ai</a><p>I'm working on a Wispr/Spokenly competitor. It's free without any paywalled features, supports local models and a bunch of API providers including Mistral.<p>For local models ottex has - parakeet V3, Whisper, GLM-ASR nano, Qwen3-ASR (don't have voxtral yet though, looking into it).<p>btw, you can try new voxtral model via API (the model name to pick is `voxtral-mini-latest:transcribe`), I personally switched to it as my main default fast model - it's really good.</p>
]]></description><pubDate>Sun, 08 Feb 2026 17:39:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=46936603</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=46936603</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46936603</guid></item><item><title><![CDATA[New comment by k9294 in "Voxtral Transcribe 2"]]></title><description><![CDATA[
<p>Hey, I would really appreciate if you will try <a href="https://ottex.ai" rel="nofollow">https://ottex.ai</a><p>I'm working on a Wispr/Spokenly competitor. It's free without any paywalled features, supports local models and bunch of API providers including Mistral.<p>btw, the model name to pick is `voxtral-mini-latest:transcribe`, I personally switched to it as my main default fast model - it's really good.</p>
]]></description><pubDate>Sun, 08 Feb 2026 17:34:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=46936537</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=46936537</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46936537</guid></item><item><title><![CDATA[New comment by k9294 in "Voxtral Transcribe 2"]]></title><description><![CDATA[
<p>Try <a href="https://ottex.ai/" rel="nofollow">https://ottex.ai/</a><p>I recently added support for Mistral provider, model is actually a very good one, I personally switched to it as my default model.<p>p.s. the app is free for personal user, has support for both local models and BYOK with OpenRouter, Groq, Mistral, Fireworks, and more coming soon.</p>
]]></description><pubDate>Sun, 08 Feb 2026 17:29:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=46936476</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=46936476</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46936476</guid></item><item><title><![CDATA[New comment by k9294 in "Show HN: Sweep, Open-weights 1.5B model for next-edit autocomplete"]]></title><description><![CDATA[
<p>Is there an oss model for next word / edits predictions for texts in general? e.g. Typing emails?</p>
]]></description><pubDate>Fri, 23 Jan 2026 20:01:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=46737125</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=46737125</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46737125</guid></item><item><title><![CDATA[New comment by k9294 in "Ask HN: What are you working on? (January 2026)"]]></title><description><![CDATA[
<p>ottex.ai - free BYOK alternative to WisprFlow and Raycast AI shortcuts.<p>Native macOS and iOS apps with OpenRouter BYOK. Same quality as in proprietary products for 1-3$ per month instead of 35$.</p>
]]></description><pubDate>Tue, 13 Jan 2026 05:26:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=46597561</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=46597561</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46597561</guid></item><item><title><![CDATA[New comment by k9294 in "Show HN: Titan – JavaScript-first framework that compiles into a Rust server"]]></title><description><![CDATA[
<p>I’ve worked with relatively large projects in TypeScript, Python, C#, and Swift, and I’ve come to believe the more opinionated the language and framework, the better. C# .NET, despite being a monster, was a breath of fresh air after TS. Each iteration just worked. Each new feature simply gets implemented.<p>My experience also points to compiled languages that give immediate feedback on build. It’s nearly impossible to stop any AI agent from using 'as any' or 'as unknown as X'casts in TypeScript - LLMs will “fix” problems by sweeping them under the rug. The larger the codebase, the more review and supervision is required. TS codebase rots much faster then rust/C#/swift etc.</p>
]]></description><pubDate>Wed, 17 Dec 2025 08:25:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=46299504</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=46299504</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46299504</guid></item><item><title><![CDATA[New comment by k9294 in "Ask HN: What Are You Working On? (December 2025)"]]></title><description><![CDATA[
<p>Working on <a href="https://ottex.ai/" rel="nofollow">https://ottex.ai/</a> - BYOK alternative to Wispr Flow and Raycast AI shortcuts.<p>I love global voice-to-text transcription (especially when working with Claude Code or Cursor) and simple AI shortcuts like "Fix Grammar" and "Translate to {Language}".<p>I realized I was spending around €35/mo (€420 a year) on two apps for AI features that cost just pennies to run.<p>So I built Ottex - a native macOS app with a tiny footprint. Add your OpenRouter API key and get solid voice-to-text using Gemini 2.5 Flash, plus any OpenRouter model for AI shortcuts.</p>
]]></description><pubDate>Mon, 15 Dec 2025 09:28:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=46272166</link><dc:creator>k9294</dc:creator><comments>https://news.ycombinator.com/item?id=46272166</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46272166</guid></item></channel></rss>