<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: markab21</title><link>https://news.ycombinator.com/user?id=markab21</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 28 Apr 2026 16:47:55 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=markab21" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by markab21 in "Google releases Gemma 4 open models"]]></title><description><![CDATA[
<p>I'll pipe in here as someone working on an agentic harness project using mastra as the harness.<p>Nemotron3-super is, without question, my favorite model now for my agentic use cases. The closest model I would compare it to, in vibe and feel, is the Qwen family but this thing has an ability to hold attention through complicated (often noisy) agentic environments and I'm sometimes finding myself checking that i'm not on a frontier model.<p>I now just rent a Dual B6000 on a full-time basis for myself for all my stuff; this is the backbone of my "base" agentic workload, and I only step up to stronger models in rare situations in my pipelines.<p>The biggest thing with this model, I've found, is just making sure my environment is set up correctly; the temps and templates need to be exactly right. I've had hit-or-miss with OpenRouter. But running this model on a B6000 from Vast with a native NVFP4 model weight from Nvidia, it's really good. (2500 peak tokens/sec on that setup) batching. about 100/s 1-request, 250k context. :)<p>I can run on a single B6000 up to about 120k context reliably but really this thing SCREAMS on a dual-b6000.  (I'm close to just ordering a couple for myself it's working so well).<p>Good luck .. (Sometimes I feel like I'm the crazy guy in the woods loving this model so much, I'm not sure why more people aren't jumping on it..)</p>
]]></description><pubDate>Thu, 02 Apr 2026 21:08:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47620217</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=47620217</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47620217</guid></item><item><title><![CDATA[New comment by markab21 in "If DSPy is so great, why isn't anyone using it?"]]></title><description><![CDATA[
<p>I think the entire premise that the prompting is the surface area for optimizing the application is fundamentally the wrong framing, in the same way that in 1998 better cpam will save CGI. It's solving the wrong problems now, and the limitations in context and model intelligence require a tool like Dspy.<p>The only thing I'd grab dspy for at this point is to automate the edges of the agentic pipeline that could be improved with RL patterns. But if that is true, you're really shorting yourself by giving your domain DSPY.  You should be building your own RL learning loops.<p>My experience: If you find yourself reaching for a tool like Dspy, you might be sitting on a scenario where reinforcement learning approaches would help even further up the stack than your prompts, and you're probably missing where the real optimization win is. (Think bigger)</p>
]]></description><pubDate>Mon, 23 Mar 2026 15:53:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=47491199</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=47491199</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47491199</guid></item><item><title><![CDATA[New comment by markab21 in "Gemini 3.1 Pro"]]></title><description><![CDATA[
<p>You just articulated why I struggle to personally connect with Gemini. It feels so unrelatable and exhausting to read its output. I prefer to read Opus/Deepseek/GLM over Gemini, Qwen and the open source GPT models. Maybe it is RLHF that is creating my distaste from using it. (I pay for Gemini; I should be using it more... but the outputs just bug me and feel more work to get actionable insight.)</p>
]]></description><pubDate>Thu, 19 Feb 2026 16:55:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=47075920</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=47075920</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47075920</guid></item><item><title><![CDATA[New comment by markab21 in "Experts Have World Models. LLMs Have Word Models"]]></title><description><![CDATA[
<p>And I think you basically just described the OpenAI approach to building models and serving them.</p>
]]></description><pubDate>Mon, 09 Feb 2026 15:03:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=46945989</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=46945989</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46945989</guid></item><item><title><![CDATA[New comment by markab21 in "Orchestrate teams of Claude Code sessions"]]></title><description><![CDATA[
<p><i>Shaking fist at clouds!!</i></p>
]]></description><pubDate>Thu, 05 Feb 2026 18:44:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=46903235</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=46903235</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46903235</guid></item><item><title><![CDATA[New comment by markab21 in "Qwen3-Coder-Next"]]></title><description><![CDATA[
<p>It's getting a lot easier to do this using sub-agents with tools in Claude. I have a fleet of Mastra agents (TypeScript). I use those agents inside my project as CLI tools to do repetitive tasks that gobble tokens such as scanning code, web search, library search, and even SourceGraph traversal.<p>Overall, it's allowed me to maintain more consistent workflows as I'm less dependent on Opus. Now that Mastra has introduced the concept of Workspaces, which allow for more agentic development, this approach has become even more powerful.</p>
]]></description><pubDate>Tue, 03 Feb 2026 16:36:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=46873238</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=46873238</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46873238</guid></item><item><title><![CDATA[New comment by markab21 in "TimeCapsuleLLM: LLM trained only on data from 1800-1875"]]></title><description><![CDATA[
<p>Basically looking for emergent behavior.</p>
]]></description><pubDate>Mon, 12 Jan 2026 17:08:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=46591283</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=46591283</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46591283</guid></item><item><title><![CDATA[New comment by markab21 in "Show HN: Mysti – Claude, Codex, and Gemini debate your code, then synthesize"]]></title><description><![CDATA[
<p>I love where you're going with this. In my experience it's not about a different persona, it's about constantly considering context that triggers, different activations enhance a different outcome. You can achieve the same thing, of course by switching to an agent with a separate persona, but you can also get it simply by injecting new context, or forcing the agent to consider something new. I feel like this concept gets cargo-culted a little bit.<p>I personally have moved to a pattern where i use mastra-agents in my project to achieve this.  I've slowly shifted the bulk of the code research and web research to my internal tools (built with small typescript agents).. I can now really easily bounce between different tools such as claude, codex, opencode and my coding tools are spending more time orchestrating work than doing the work themselves.</p>
]]></description><pubDate>Sat, 27 Dec 2025 15:11:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=46402397</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=46402397</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46402397</guid></item><item><title><![CDATA[New comment by markab21 in "Apps SDK"]]></title><description><![CDATA[
<p>The skepticism is understandable given the trajectory of GPTs and custom instructions, but there's a meaningful technical difference here: the Apps SDK is built on the Model Context Protocol (MCP), which is an open specification rather than a proprietary format.<p>MCP standardizes how LLM clients connect to external tools—defining wire formats, authentication flows, and metadata schemas. This means apps you build aren't inherently ChatGPT-specific; they're MCP servers that could work with any MCP-compatible client. The protocol is transport-agnostic and self-describing, with official Python and TypeScript SDKs already available.<p>That said, the "build our platform" criticism isn't entirely off base. While the protocol is open, practical adoption still depends heavily on ChatGPT's distribution and whether other LLM providers actually implement MCP clients. The real test will be whether this becomes a genuine cross-platform standard or just another way to contribute to OpenAI's ecosystem.<p>The technical primitives (tool discovery, structured content return, embedded UI resources) are solid and address real integration problems. Whether it succeeds likely depends more on ecosystem dynamics than technical merit.</p>
]]></description><pubDate>Mon, 06 Oct 2025 18:49:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=45494830</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=45494830</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45494830</guid></item><item><title><![CDATA[New comment by markab21 in "Mistral NeMo"]]></title><description><![CDATA[
<p>It looks like it was built jointly with nvidia:  <a href="https://huggingface.co/nvidia/Mistral-NeMo-12B-Instruct" rel="nofollow">https://huggingface.co/nvidia/Mistral-NeMo-12B-Instruct</a></p>
]]></description><pubDate>Thu, 18 Jul 2024 15:02:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=40996235</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=40996235</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40996235</guid></item><item><title><![CDATA[New comment by markab21 in "How Mandelbrot set images are affected by floating point precision"]]></title><description><![CDATA[
<p>I bet you have been waiting years to pull that one out of your pocket.<p>Well played sir! Nice shot man! :D</p>
]]></description><pubDate>Tue, 12 Mar 2024 22:49:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=39685865</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=39685865</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39685865</guid></item><item><title><![CDATA[New comment by markab21 in "Memory and new controls for ChatGPT"]]></title><description><![CDATA[
<p>I've found myself more and more using local models rather than ChatGPT; it was pretty trivial to set up Ollama+Ollama-WebUI, which is shockingly good.<p>I'm so tired of arguing with ChatGPT (or what was Bard) to even get simple things done. SOLAR-10B or Mistral works just fine for my use cases, and I've wired up a direct connection to Fireworks/OpenRouter/Together for the occasion I need anything more than what will run on my local hardware. (mixtral MOE, 70B code/chat models)</p>
]]></description><pubDate>Tue, 13 Feb 2024 19:25:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=39361667</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=39361667</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39361667</guid></item><item><title><![CDATA[New comment by markab21 in "Mistral CEO confirms 'leak' of new open source AI model nearing GPT4 performance"]]></title><description><![CDATA[
<p>For Llama-based progress - Reddit - /r/LocalLlama has been my top source of info, although it's been getting a little more noisy lately.<p>I also hang out on a few Discord servers: 
 - Nous Research
 - TogetherAI / Fireworks / Openrouter
 - LangChain 
 - TheBloke AI 
 - Mistral AI<p>These, along with a couple of newsletters, basically keep a pulse on things.</p>
]]></description><pubDate>Wed, 31 Jan 2024 20:19:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=39208810</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=39208810</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39208810</guid></item><item><title><![CDATA[New comment by markab21 in "Gulf Stream weakening now 99% certain, and ramifications will be global"]]></title><description><![CDATA[
<p>[flagged]</p>
]]></description><pubDate>Wed, 18 Oct 2023 20:02:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=37934035</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=37934035</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37934035</guid></item><item><title><![CDATA[New comment by markab21 in "Obsidian 1.4.10 Desktop (Public)"]]></title><description><![CDATA[
<p>Yeah, slow news day.</p>
]]></description><pubDate>Tue, 12 Sep 2023 16:36:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=37483859</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=37483859</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37483859</guid></item><item><title><![CDATA[New comment by markab21 in "Debian celebrates 30 years"]]></title><description><![CDATA[
<p>Linux distributions, including Debian, offer a variety of desktop environments, each with its own design philosophy and user experience. If one environment doesn't suit your preferences, others might be more to your liking. It's worth exploring different desktop environments to find one that aligns with your expectations.</p>
]]></description><pubDate>Thu, 17 Aug 2023 14:35:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=37162294</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=37162294</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37162294</guid></item><item><title><![CDATA[New comment by markab21 in "Debian celebrates 30 years"]]></title><description><![CDATA[
<p>Debian and Ubuntu have similarities, but keep in mind - Ubuntu is derived from Debian, not the other way around. However, they differ in areas like release cycles, package management, and default configurations.</p>
]]></description><pubDate>Thu, 17 Aug 2023 14:33:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=37162254</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=37162254</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37162254</guid></item><item><title><![CDATA[New comment by markab21 in "Ask HN: Should HN ban ChatGPT/generated responses?"]]></title><description><![CDATA[
<p>I used it as a consultant on a development project to help me organize some of the milestones and design goals in some documentation.<p>It wasn't that I didn't know the stuff, I do, but more helpful with quickly organizing and presenting information in a clean and well-written way. I did have to go through and re-write parts of it specific to our domain.. but it saved me many hours of work doing tedious organization of data.<p>I also tested it with helping create some SOP's for a new position in our very small company, even breaking down the expected tasks into daily schedules.<p>It's not that it's perfect, but it generates a bit of a boiler-plate starting point for me which then I can work with from there.</p>
]]></description><pubDate>Mon, 12 Dec 2022 12:46:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=33954074</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=33954074</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=33954074</guid></item><item><title><![CDATA[New comment by markab21 in "Google and Facebook execs allegedly approved dividing ad market among themselves"]]></title><description><![CDATA[
<p>I'd be surprised if they don't have async mechanisms.</p>
]]></description><pubDate>Sat, 15 Jan 2022 15:25:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=29947013</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=29947013</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=29947013</guid></item><item><title><![CDATA[New comment by markab21 in "FBI's ability to legally access secure messaging app content and metadata [pdf]"]]></title><description><![CDATA[
<p>Assume anything sent over a cellular network carrier via normal SMS can not only be retrieved, but intercepted.</p>
]]></description><pubDate>Tue, 30 Nov 2021 20:28:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=29397053</link><dc:creator>markab21</dc:creator><comments>https://news.ycombinator.com/item?id=29397053</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=29397053</guid></item></channel></rss>