<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: rjpruitt16</title><link>https://news.ycombinator.com/user?id=rjpruitt16</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 01 May 2026 23:36:02 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=rjpruitt16" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by rjpruitt16 in "Ask HN: Who is hiring? (May 2026)"]]></title><description><![CDATA[
<p>I’m curious if you are looking into gleam at all. I have written elixir/gleam interop otp library. Probably will help a lot for safety in finance. I could help turn certain components into for type safe gleam code</p>
]]></description><pubDate>Fri, 01 May 2026 18:09:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47978013</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47978013</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47978013</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Layer 8: The coordination protocol AI agents and embedded devices don't have yet"]]></title><description><![CDATA[
<p>Yeah, thanks for checking it out. I know it’s a lot read. Keep me in mind if you run retry storms for agents.</p>
]]></description><pubDate>Wed, 22 Apr 2026 04:38:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=47859071</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47859071</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47859071</guid></item><item><title><![CDATA[Layer 8: The coordination protocol AI agents and embedded devices don't have yet]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.ezthrottle.network/blog/operationless-network-for-a-new-world-of-devices">https://www.ezthrottle.network/blog/operationless-network-for-a-new-world-of-devices</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47857889">https://news.ycombinator.com/item?id=47857889</a></p>
<p>Points: 2</p>
<p># Comments: 2</p>
]]></description><pubDate>Wed, 22 Apr 2026 02:07:38 +0000</pubDate><link>https://www.ezthrottle.network/blog/operationless-network-for-a-new-world-of-devices</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47857889</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47857889</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Show HN: Self-healing browser harness via direct CDP"]]></title><description><![CDATA[
<p>Where do you think AI web agents with be by 2027? How are people creating value today?<p>Where will browsers use be at the end of the year?</p>
]]></description><pubDate>Mon, 20 Apr 2026 05:37:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=47830731</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47830731</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47830731</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Ask HN: What Are You Working On? (April 2026)"]]></title><description><![CDATA[
<p>An api rate limit router - it’s gives a queue per user to solve noisy neighbors, it allows
the user user to reroute 500s to other regions automatically, the api provider can tune the speed of request with headers. It should be a massive speed boost because non-beam servers waste performance on sleeping threads. It’s works by my edge servers doing all the retrying across different regions and it can send webhook to you across regions which means we delivery if you an all your downstream are up in one region. It’s actually already ready but of course no one wants to trust me with their traffic. Nobody wants to even read the design doc nor sdk. I’m hoping that upcoming new agents bring so much traffic that people look for this category of software. Until then I guess I am Cloudfare before the world got hit by bot scrapers.<p>The modern orchestrator are reactive, they don’t handle spikey traffic. Your favorite retry library will cause retry storms for downstream dependencies and your public APIs. Remember EZThrottle blog posts<p>EZThrottle.network</p>
]]></description><pubDate>Mon, 13 Apr 2026 15:17:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=47753251</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47753251</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47753251</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Show HN: Pitstop-check – finds the retry bug that turns 429s into request storms"]]></title><description><![CDATA[
<p>Ezthrottle works by sending the request and depending on what error code the user wishes to reroute on, it will send to another region. It give the user a chance to say something different in case the api misclassifies the error. The user would have to tune it.</p>
]]></description><pubDate>Mon, 06 Apr 2026 01:03:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=47655698</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47655698</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47655698</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Show HN: Pitstop-check – finds the retry bug that turns 429s into request storms"]]></title><description><![CDATA[
<p>Cool stuff.</p>
]]></description><pubDate>Tue, 24 Mar 2026 05:06:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=47498787</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47498787</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47498787</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Ask HN: How do you deal with people who trust LLMs?"]]></title><description><![CDATA[
<p>Thanks for this. I was in the camp of trust the LLM but y’all have made valid points. After discussing with ChatGPT, it agree there are some areas where it should not be trusted as accurate, but it said with historical facts like the holocaust it should be high. Idk, perhaps we need labs to produce a chart of it level of trust deserve to certain topics</p>
]]></description><pubDate>Thu, 19 Mar 2026 03:30:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=47434561</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47434561</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47434561</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Ask HN: How to scale agent systems when Layer 7 is unreliable?"]]></title><description><![CDATA[
<p>What do you use?</p>
]]></description><pubDate>Wed, 11 Mar 2026 16:02:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=47337359</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47337359</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47337359</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Launch HN: Terminal Use (YC W26) – Vercel for filesystem-based agents"]]></title><description><![CDATA[
<p>Im curious if you guys are seeing rate limiting issues. Agents sharing api keys tend to be retry storm monsters. I wonder how agent companies will address</p>
]]></description><pubDate>Tue, 10 Mar 2026 01:11:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47317981</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47317981</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47317981</guid></item><item><title><![CDATA[Ask HN: How to scale agent systems when Layer 7 is unreliable?]]></title><description><![CDATA[
<p>Agent workflows often involve 10+ API calls to different services 
(LLMs, data APIs, web scraping). Layer 7 being unreliable = 
workflows fail or cause retry storms.<p>Common failure modes I'm thinking about:
- 429 rate limits → agents retry → hammer API worse
- Partial outages → synchronized retries across customers
- LangGraph workflows fail mid-execution → how to resume?<p>For those running agent systems at scale:
- How do you handle Layer 7 failures?
- Retry coordination? Circuit breakers? 
- How do you prevent retry storms to downstream dependencies?
- Do LangGraph workflows gracefully handle API failures?<p>Curious what the production reality looks like.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47292281">https://news.ycombinator.com/item?id=47292281</a></p>
<p>Points: 1</p>
<p># Comments: 2</p>
]]></description><pubDate>Sat, 07 Mar 2026 22:55:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=47292281</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47292281</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47292281</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Ask HN: How do you handle API rate limits in production?"]]></title><description><![CDATA[
<p>interesting. What type of features did this enable. How was it maintaining redis. How many queues did you have.</p>
]]></description><pubDate>Wed, 25 Feb 2026 04:19:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47147315</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47147315</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47147315</guid></item><item><title><![CDATA[Ask HN: How do you handle API rate limits in production?]]></title><description><![CDATA[
<p>I'm building data pipelines that sync data from various third party apis. We constantly hit 429 rate limits, and our janky retry system fails regularly.
For those running production data syncs or microservices calling external APIs heavily:<p>How do you handle rate limiting across multiple workers?
Do you use circuit breakers, retry libraries, or something custom?
How do you prevent retry storms when 100 workers all hit the same rate limit?<p>Curious what's working at scale.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47131730">https://news.ycombinator.com/item?id=47131730</a></p>
<p>Points: 3</p>
<p># Comments: 2</p>
]]></description><pubDate>Tue, 24 Feb 2026 01:35:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=47131730</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47131730</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47131730</guid></item><item><title><![CDATA[Show HN: Stop Losing LangGraph Progress to 429 Errors]]></title><description><![CDATA[
<p>Hey HN, I built this because I kept losing progress in LangGraph workflows when OpenRouter or OpenAI returned 429s.
The problem: You're 7 steps into an agent workflow. Step 7 hits a rate limit. Everything crashes. Restart from step 1.
Client-side retries don't help at scale:<p>100 workers all retry independently → retry storm
Sequential fallbacks are slow (try OpenRouter, wait 5s, try Anthropic, wait 5s)
No coordination across instances<p>So I built a coordination layer that:<p>Races multiple providers simultaneously (OpenRouter + Anthropic + OpenAI)
Coordinates retries across all workers (no retry storms)
Resumes workflows via webhooks (idempotent keys = checkpoints)<p>It runs on Fly.io's anycast network + BEAM for distributed coordination.
Architecture deep dive: <a href="https://www.ezthrottle.network/blog/making-failure-boring-again" rel="nofollow">https://www.ezthrottle.network/blog/making-failure-boring-ag...</a>
Happy to answer questions about the approach or why BEAM made this possible when other languages would struggle.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47043197">https://news.ycombinator.com/item?id=47043197</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 17 Feb 2026 03:01:42 +0000</pubDate><link>https://www.ezthrottle.network/blog/stop-losing-langgraph-progress</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=47043197</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47043197</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Ask HN: Is anyone losing sleep over retry storms or partial API outages?"]]></title><description><![CDATA[
<p>Fair pushback — to clarify, I’m not assuming incompetence or suggesting infra should paper over bad architecture.<p>By “losing sleep” I really mean on-call fatigue during partial outages — the class of incidents where backoff, shedding, and breakers exist, but retry amplification, shared rate limits, or degraded dependencies still cause noisy pages and prolonged recovery.<p>I’m trying to understand how teams coordinate retries and backpressure across many independent clients/services when refactors aren’t immediately available, not replace good architecture or take ownership of someone else’s system.<p>If you’ve seen patterns that consistently avoid that on-call pain at scale, I’d genuinely love to learn from them.</p>
]]></description><pubDate>Tue, 03 Feb 2026 15:32:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=46872248</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=46872248</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46872248</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Ask HN: Is anyone losing sleep over retry storms or partial API outages?"]]></title><description><![CDATA[
<p>Agree backoff+jitter is table stakes, and load shedding/backpressure is necessary under sustained overload. The tricky cases I’m digging into are shared rate limits (429s) and many concurrent clients/agents where local backoff isn’t coordinated and you still get herds after partial outages. Curious what patterns you’ve seen work well for coordinating retries/fairness across tenants or API keys?</p>
]]></description><pubDate>Tue, 03 Feb 2026 15:17:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=46872032</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=46872032</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46872032</guid></item><item><title><![CDATA[Ask HN: Is anyone losing sleep over retry storms or partial API outages?]]></title><description><![CDATA[
<p>I’m working on infrastructure 
to solve retry storms and outages. Before I go further, I want to understand what people are actually doing today. Compare solutions and maybe help someone see potential solutions.
The problems:<p>Retry storms - API fails, your entire fleet retries independently, thundering herd makes it worse.<p>Partial outages - API is “up” but degraded (slow, intermittent 500s). Health checks pass, requests suffer.<p>What I’m curious about:
 ∙ What’s your current solution? (circuit breakers, queues, custom coordination, service mesh, something else?)
 ∙ How well does it work? What are the gaps?
 ∙ What scale are you at? (company size, # of instances, requests/sec)<p>I’d love to hear what’s working, what isn’t, and what you wish existed.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46866428">https://news.ycombinator.com/item?id=46866428</a></p>
<p>Points: 2</p>
<p># Comments: 4</p>
]]></description><pubDate>Tue, 03 Feb 2026 04:19:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=46866428</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=46866428</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46866428</guid></item><item><title><![CDATA[Ask HN: Are retries the wrong abstraction under rate limits?]]></title><description><![CDATA[
<p>Over the last few years, I’ve watched a lot of production systems fail in ways that feel… strangely predictable.<p>When services hit 429s or timeouts, the standard response is almost always the same: retries with backoff, sleep loops, jitter, etc. This is treated as a best practice across languages and platforms.<p>But in systems with high concurrency, fan-out, or shared downstream dependencies, retries often seem to amplify load instead of smoothing it. What starts as localized failure can turn into retry storms, thundering herds, and cascading outages.<p>It’s made me wonder whether retries are solving the wrong problem at the wrong layer — treating a coordination issue as an application-level error-handling concern.<p>I wrote up a longer piece exploring this idea and arguing for making failure boring again by handling it at a different layer:
https://www.ezthrottle.network/blog/making-failure-boring-again<p>Curious how this matches others’ experience:<p>Have retries actually improved stability for you under sustained rate limiting?<p>Have you seen cases where they clearly made things worse?<p>If retries aren’t the right abstraction, what is?<p>Interested in war stories, counterexamples, and alternative approaches.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46841137">https://news.ycombinator.com/item?id=46841137</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 31 Jan 2026 21:39:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=46841137</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=46841137</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46841137</guid></item><item><title><![CDATA[Show HN: EZThrottle – Coordinated retries and region racing for APIs Gleam/BEAM]]></title><description><![CDATA[
<p>Hi HN,<p>I built EZThrottle because I was tired of copy-pasting exponential backoff into every codebase.<p>The core idea: 
retries shouldn't be independent. When thousands of machines all hit a 429 and retry independently, you get retry storms that cascade into outages. EZThrottle coordinates failure in one place.<p>What it does:
  - Rate limiting per destination (default 2 RPS – conservative on purpose)
  - Region racing: send to multiple regions, accept first success, cancel the rest. If one
region goes down, your request still completes with just a latency bump instead of a full outage
  - Adds event-driven architecture to your stack via webhooks – fire and forget, get results delivered<p>Current scale (setting expectations):
  Running 4 machines across 2 US regions (Dallas + Washington DC) on Fly.io. Not massive yet –
  we'll expand to more regions with demand. Early days.<p>SDKs:
  - Python: <a href="https://github.com/rjpruitt16/ezthrottle-python" rel="nofollow">https://github.com/rjpruitt16/ezthrottle-python</a>
  - Node: <a href="https://github.com/rjpruitt16/ezthrottle-node" rel="nofollow">https://github.com/rjpruitt16/ezthrottle-node</a>
  - Go: <a href="https://github.com/rjpruitt16/ezthrottle-go" rel="nofollow">https://github.com/rjpruitt16/ezthrottle-go</a><p>Written in Gleam, runs on the BEAM. This is part of a larger vision (L8-OS – local-first AI stack), but EZThrottle works standalone.<p>Blog post with architecture diagrams:
<a href="https://ezthrottle.network/blog/making-failure-boring-again" rel="nofollow">https://ezthrottle.network/blog/making-failure-boring-again</a><p>Free tier: 
1M requests/month. Happy to answer questions.<p>@RahmiPruitt on Twitter</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46804549">https://news.ycombinator.com/item?id=46804549</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 29 Jan 2026 01:36:35 +0000</pubDate><link>https://www.ezthrottle.network/</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=46804549</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46804549</guid></item><item><title><![CDATA[New comment by rjpruitt16 in "Glixir: A safe(ish) OTP library for gleam"]]></title><description><![CDATA[
<p>I hear a lot of people complain that Gleam is “not OTP.” They’re right. At the time of writing, Gleam still doesn’t have a registry or pubsub. If you ask the Gleam team, they’ll tell you to use external calls — which works, but is a pain to reproduce consistently across projects.<p>So I wrote Glixir, a library that wraps common OTP patterns in Gleam using Elixir’s battle-tested OTP library. It’s not as type-safe as Gleam, but I needed something production-ready today, not next year.<p>This lets you:<p>Use supervisors, GenServers, and pubsub patterns from Gleam.<p>Stay in Gleam where it makes sense, but drop down to Elixir OTP when you need reliability.<p>Ship code faster without reinventing OTP on the Gleam side.<p>Obviously there are trade-offs — you give up some type safety, and it’s not “pure Gleam.” But if you’re trying to build something real, you may find this a useful middle ground.<p>It's MIT so if you disagree with direction you can fork and do whatever you want.</p>
]]></description><pubDate>Tue, 16 Sep 2025 21:24:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=45268276</link><dc:creator>rjpruitt16</dc:creator><comments>https://news.ycombinator.com/item?id=45268276</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45268276</guid></item></channel></rss>