<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: sasipi247</title><link>https://news.ycombinator.com/user?id=sasipi247</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 13 Apr 2026 08:29:47 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=sasipi247" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by sasipi247 in "Ask HN: What Are You Working On? (April 2026)"]]></title><description><![CDATA[
<p>I am working on a system built around the OpenAI Responses API WebSocket mode as performance is something that interests me.<p>Its like a microservices architecture with NATS JetStream coordinating stuff. I want to keep the worker core as clean as possible, just managing open sockets, threads and continuation.<p>Document querying is something I am interested in also. This system allows me to pin a document to a socket as a subagent, which is then called upon.<p>I have hit alot of slip ups along the way, such as infinite loops trying to call OpenAI API, etc ...<p>Example usage:
10 documents on warm sockets on GPT 5.4 nano. Then the main thread can call out to those other sockets to query the documents in parallel. It allows alot of possibilities : cheaper models for cheaper tasks, input caching and lower latency.<p>There is also a frontend<p>Alot of information is in here, just thoughts, designs etc:
<a href="https://github.com/SamSam12121212/ExplorerPRO/tree/main/docs" rel="nofollow">https://github.com/SamSam12121212/ExplorerPRO/tree/main/docs</a></p>
]]></description><pubDate>Sun, 12 Apr 2026 22:22:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=47745174</link><dc:creator>sasipi247</dc:creator><comments>https://news.ycombinator.com/item?id=47745174</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47745174</guid></item></channel></rss>