<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: xyc</title><link>https://news.ycombinator.com/user?id=xyc</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 23 Apr 2026 14:50:34 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=xyc" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by xyc in "Open models by OpenAI"]]></title><description><![CDATA[
<p>I'm running it with ROG Flow Z13 128GB Strix Halo and getting 50 tok/s for 20B model and 12 tok/s for 120B model. I'd say it's pretty usable.</p>
]]></description><pubDate>Wed, 06 Aug 2025 00:09:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=44806100</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=44806100</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44806100</guid></item><item><title><![CDATA[New comment by xyc in "MCP in LM Studio"]]></title><description><![CDATA[
<p>For TypeScript you can refer to <a href="https://github.com/modelcontextprotocol/typescript-sdk/blob/main/src/inMemory.ts">https://github.com/modelcontextprotocol/typescript-sdk/blob/...</a><p>There isn't much documentation available right now but you can ask coding agent eg. Claude Code to generate an example.</p>
]]></description><pubDate>Fri, 27 Jun 2025 17:29:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=44398638</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=44398638</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44398638</guid></item><item><title><![CDATA[New comment by xyc in "MCP in LM Studio"]]></title><description><![CDATA[
<p>Great to see more local AI tools supporting MCP! Recently I've also added MCP support to recurse.chat. When running locally (LLaMA.cpp and Ollama) it still needs to catch up in terms of tool calling capabilities (for example tool call accuracy / parallel tool calls) compared to the well known providers but it's starting to get pretty usable.</p>
]]></description><pubDate>Wed, 25 Jun 2025 23:19:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=44382710</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=44382710</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44382710</guid></item><item><title><![CDATA[New comment by xyc in "LM Studio is now an MCP Host"]]></title><description><![CDATA[
<p>It's a protocol that doesn't dictate how you are calling the tool. You can use in-memory transport without needing to spin up a server. Your tool can just be a function, but with the flexibility of serving to other clients.</p>
]]></description><pubDate>Wed, 25 Jun 2025 23:13:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=44382667</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=44382667</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44382667</guid></item><item><title><![CDATA[New comment by xyc in "Ask HN: What Does Your Self-Hosted LLM Stack Look Like in 2025?"]]></title><description><![CDATA[
<p>recurse.chat + M2 max Mac</p>
]]></description><pubDate>Thu, 05 Jun 2025 17:19:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=44193705</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=44193705</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44193705</guid></item><item><title><![CDATA[New comment by xyc in "Cursor 1.0"]]></title><description><![CDATA[
<p>I recently discovered toolhive which is pretty handy too <a href="https://github.com/stacklok/toolhive">https://github.com/stacklok/toolhive</a></p>
]]></description><pubDate>Wed, 04 Jun 2025 23:16:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=44186616</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=44186616</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44186616</guid></item><item><title><![CDATA[New comment by xyc in "Vision Now Available in Llama.cpp"]]></title><description><![CDATA[
<p>If you are on a Mac, give <a href="https://recurse.chat/" rel="nofollow">https://recurse.chat/</a> a try. As simple as download the model and start chatting. Just added the new multimodal support in LLaMA.cpp.</p>
]]></description><pubDate>Sat, 17 May 2025 00:39:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=44011101</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=44011101</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44011101</guid></item><item><title><![CDATA[New comment by xyc in "Show HN: Clippy – 90s UI for local LLMs"]]></title><description><![CDATA[
<p>Actually this is a good way to find product ideas. I placed a query in Grok to find posts about what people want, similar to this. Then it performs multiple searches on X including embedding search, and suggested people want stuff like tamagotchi, ICQ etc. back.</p>
]]></description><pubDate>Tue, 06 May 2025 19:58:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=43909009</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=43909009</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43909009</guid></item><item><title><![CDATA[New comment by xyc in "Show HN: Clippy – 90s UI for local LLMs"]]></title><description><![CDATA[
<p>It seems that this is possibly not necessary, since LLaMA.cpp already integrates Jinja with CPP implementation (through minja)</p>
]]></description><pubDate>Tue, 06 May 2025 17:47:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=43907787</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=43907787</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43907787</guid></item><item><title><![CDATA[New comment by xyc in "How I run LLMs locally"]]></title><description><![CDATA[
<p>Check out <a href="https://recurse.chat/" rel="nofollow">https://recurse.chat/</a></p>
]]></description><pubDate>Mon, 30 Dec 2024 23:47:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=42554904</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42554904</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42554904</guid></item><item><title><![CDATA[New comment by xyc in "Limbo: A complete rewrite of SQLite in Rust"]]></title><description><![CDATA[
<p>The fact that there's no alternative implementation of SQLite also seems to play a part in preventing standardization of WebSQL.<p><a href="https://www.w3.org/TR/webdatabase/" rel="nofollow">https://www.w3.org/TR/webdatabase/</a><p>"The specification reached an impasse: all interested implementors have used the same SQL backend (Sqlite), but we need multiple independent implementations to proceed along a standardisation path."</p>
]]></description><pubDate>Tue, 10 Dec 2024 18:16:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=42379542</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42379542</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42379542</guid></item><item><title><![CDATA[New comment by xyc in "Llama.cpp guide – Running LLMs locally on any hardware, from scratch"]]></title><description><![CDATA[
<p>If anyone on macOS wants to use llama.cpp with ease, check out <a href="https://recurse.chat/" rel="nofollow">https://recurse.chat/</a>. Supports importing ChatGPT history & continue chats offline using llama.cpp. Built this so I can use local AI as a daily driver.</p>
]]></description><pubDate>Sat, 30 Nov 2024 05:35:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=42279787</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42279787</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42279787</guid></item><item><title><![CDATA[New comment by xyc in "Llama.cpp guide – Running LLMs locally on any hardware, from scratch"]]></title><description><![CDATA[
<p>You can get a release binary from <a href="https://github.com/ggerganov/llama.cpp/releases">https://github.com/ggerganov/llama.cpp/releases</a> too.</p>
]]></description><pubDate>Sat, 30 Nov 2024 05:30:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=42279758</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42279758</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42279758</guid></item><item><title><![CDATA[New comment by xyc in "Model Context Protocol"]]></title><description><![CDATA[
<p>Made tool use work! check out demo here: <a href="https://x.com/chxy/status/1861684254297727299" rel="nofollow">https://x.com/chxy/status/1861684254297727299</a></p>
]]></description><pubDate>Wed, 27 Nov 2024 08:15:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=42254052</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42254052</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42254052</guid></item><item><title><![CDATA[New comment by xyc in "Model Context Protocol"]]></title><description><![CDATA[
<p>sharing the messy code here just for funsies: <a href="https://gist.github.com/xyc/274394031b41ac7e8d7d3aa7f4f7bed9" rel="nofollow">https://gist.github.com/xyc/274394031b41ac7e8d7d3aa7f4f7bed9</a></p>
]]></description><pubDate>Tue, 26 Nov 2024 09:13:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=42243901</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42243901</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42243901</guid></item><item><title><![CDATA[New comment by xyc in "Model Context Protocol"]]></title><description><![CDATA[
<p>Just tried out the puppeteer server example if anyone is interested in seeing a demo: <a href="https://x.com/chxy/status/1861302909402861905" rel="nofollow">https://x.com/chxy/status/1861302909402861905</a>. (Todo: add tool use - prompt would be like "go to this website and screenshot")<p>I appreciate the design which left the implementation of servers to the community which doesn't lock you into any particular implementation, as the protocol seems to be aiming to primarily solve the RPC layer.<p>One major value add of MCP I think is a capability extension to a vast amount of AI apps.</p>
]]></description><pubDate>Tue, 26 Nov 2024 07:05:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=42243187</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42243187</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42243187</guid></item><item><title><![CDATA[New comment by xyc in "Model Context Protocol"]]></title><description><![CDATA[
<p>^ asked the question in the discussion: <a href="https://github.com/modelcontextprotocol/specification/discussions/63">https://github.com/modelcontextprotocol/specification/discus...</a></p>
]]></description><pubDate>Mon, 25 Nov 2024 22:38:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=42240858</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42240858</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42240858</guid></item><item><title><![CDATA[New comment by xyc in "Model Context Protocol"]]></title><description><![CDATA[
<p>Thanks for the pointers! Will do. I've fired up <a href="https://github.com/modelcontextprotocol/inspector">https://github.com/modelcontextprotocol/inspector</a> and the code looks helpful too.<p>I'm looking at integrating MCP with desktop app. The spec (<a href="https://spec.modelcontextprotocol.io/specification/basic/transports/#stdio" rel="nofollow">https://spec.modelcontextprotocol.io/specification/basic/tra...</a>) mentions "Clients SHOULD support stdio whenever possible.". The server examples seem to be mostly stdio as well. In the context of a sandboxed desktop app, it's often not practical to launch a server as subprocess because:<p>- sandbox restrictions of executing binaries<p>- needing to bundle binary leads to a larger installation size<p>Would it be reasonable to relax this restriction and provide both SSE/stdio for the default server examples?</p>
]]></description><pubDate>Mon, 25 Nov 2024 22:27:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=42240795</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42240795</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42240795</guid></item><item><title><![CDATA[New comment by xyc in "Model Context Protocol"]]></title><description><![CDATA[
<p>Superb work and super promising! I had wished for a protocol like this.<p>Is there a recommended resource for building MCP client? From what I've seen it just mentions Claude desktop & co are clients. SDK readme seems to cover it a bit but some examples could be great.</p>
]]></description><pubDate>Mon, 25 Nov 2024 20:25:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=42239830</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42239830</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42239830</guid></item><item><title><![CDATA[New comment by xyc in "Show HN: Embed an SQLite database in your PostgreSQL table"]]></title><description><![CDATA[
<p>With Claude you barely had to learn the language this days as you just need to prompt, but SQLite column is an interesting idea.</p>
]]></description><pubDate>Tue, 19 Nov 2024 17:13:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=42185767</link><dc:creator>xyc</dc:creator><comments>https://news.ycombinator.com/item?id=42185767</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42185767</guid></item></channel></rss>