<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: nyellin</title><link>https://news.ycombinator.com/user?id=nyellin</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 23 Apr 2026 10:57:22 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=nyellin" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by nyellin in "Atlassian enables default data collection to train AI"]]></title><description><![CDATA[
<p>Why does Atlassian need to train AI models?</p>
]]></description><pubDate>Mon, 20 Apr 2026 19:26:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47839323</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=47839323</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47839323</guid></item><item><title><![CDATA[Ask HN: Agent orchestrators / UIs you use on top of Claude?]]></title><description><![CDATA[

<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47798952">https://news.ycombinator.com/item?id=47798952</a></p>
<p>Points: 2</p>
<p># Comments: 2</p>
]]></description><pubDate>Thu, 16 Apr 2026 20:18:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=47798952</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=47798952</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47798952</guid></item><item><title><![CDATA[New comment by nyellin in "Launch HN: Freestyle – Sandboxes for Coding Agents"]]></title><description><![CDATA[
<p>Is it possible to run a Kubernetes cluster inside one? (E.g. via KIND.)<p>If so, we'd very much like to test this. We make extensive use of Claude Code web but it can't effectively test our product inside the sandbox without running a K8s cluster</p>
]]></description><pubDate>Mon, 06 Apr 2026 20:22:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=47666516</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=47666516</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47666516</guid></item><item><title><![CDATA[Ask HN: Giving AI agents a credit card]]></title><description><![CDATA[
<p>Is there a safe way to give agents a credit card with usage caps? Ideally also the ability for purchases to pend until human approval.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47089014">https://news.ycombinator.com/item?id=47089014</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 20 Feb 2026 15:13:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=47089014</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=47089014</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47089014</guid></item><item><title><![CDATA[New comment by nyellin in "Benchmarking OpenTelemetry: Can AI trace your failed login?"]]></title><description><![CDATA[
<p>We publish the benchmarks for HolmesGPT (CNCF sandbox project) at <a href="https://holmesgpt.dev/development/evaluations/" rel="nofollow">https://holmesgpt.dev/development/evaluations/</a></p>
]]></description><pubDate>Thu, 29 Jan 2026 20:43:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=46816309</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=46816309</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46816309</guid></item><item><title><![CDATA[New comment by nyellin in "Benchmarking OpenTelemetry: Can AI trace your failed login?"]]></title><description><![CDATA[
<p>HolmesGPT maintainer here: our benchmarks [1] tell a very different story, as does anecdotal evidence from our customers- including Fortune 500 using SRE agents in incredibly complex production environments.<p>We're actually struggling a bit with benchmark saturation right now. Opus does much better in the real world than Sonnet but it's hard to create sophisticated enough benchmarks to show that in the  lab. When we run benchmarks with a small number of iterations Sonnet even wins sometimes.<p>[1] <a href="https://holmesgpt.dev/development/evaluations/history/" rel="nofollow">https://holmesgpt.dev/development/evaluations/history/</a></p>
]]></description><pubDate>Thu, 29 Jan 2026 20:38:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=46816248</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=46816248</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46816248</guid></item><item><title><![CDATA[New comment by nyellin in "How to code Claude Code in 200 lines of code"]]></title><description><![CDATA[
<p>Haiku is called often, but not always the way you think. E.g. every time you write something CC invokes Haiku multiple times to generate the 'delightful 1-2 word phrase used to indicate progress to the user' (Doing Stuff, Wizarding, etc)</p>
]]></description><pubDate>Thu, 08 Jan 2026 20:58:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=46546389</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=46546389</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46546389</guid></item><item><title><![CDATA[New comment by nyellin in "How to code Claude Code in 200 lines of code"]]></title><description><![CDATA[
<p>Not necessarily true. Subagents allow for parallelization but they can decrease accuracy dramatically if you're not careful because there are often dependencies between tasks and swapping context windows with a summary is extremely lossy.<p>For the longest time, Claude Code itself didnt really use subagents much by default, other than supporting them as a feature eager users could configure. (Source is reverse engineering we did on Claude code using the fantastic CC tracing tool Simon Willison wrote about once. This is also no longer true on latest versions that have e.g. an Explore subagent that is actively used.)</p>
]]></description><pubDate>Thu, 08 Jan 2026 20:56:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=46546352</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=46546352</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46546352</guid></item><item><title><![CDATA[New comment by nyellin in "How to code Claude Code in 200 lines of code"]]></title><description><![CDATA[
<p>Forgot to address the easiest part:<p>> - how can I reliably call tools with the right schema?<p>This is typically done by enabling strict mode for tool calling which is a hermetic solution. Makes llm unable to generate tokens that would violate the schema. (I.e. LLM samples tokens only from the subset of tokens that lead to valid schema generation.)</p>
]]></description><pubDate>Thu, 08 Jan 2026 20:52:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=46546292</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=46546292</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46546292</guid></item><item><title><![CDATA[New comment by nyellin in "How to code Claude Code in 200 lines of code"]]></title><description><![CDATA[
<p>Re (1) use a TODOs system like Claude code.<p>Re (2) also fairly easy! It's just a summarization prompt. E.g. this is the one we use in our agent: <a href="https://github.com/HolmesGPT/holmesgpt/blob/62c3898e4efae69b4262f4d75e3f8943507278a2/holmes/plugins/prompts/conversation_history_compaction.jinja2#L4" rel="nofollow">https://github.com/HolmesGPT/holmesgpt/blob/62c3898e4efae69b...</a><p>Or just use the Claude Code SDK that does this all for you! (You can also use various provider-specific features for 2 like automatic compaction on OpenAI responses endpoint.)</p>
]]></description><pubDate>Thu, 08 Jan 2026 20:50:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=46546268</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=46546268</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46546268</guid></item><item><title><![CDATA[New comment by nyellin in "How to code Claude Code in 200 lines of code"]]></title><description><![CDATA[
<p>There's a bit more to it!<p>For example, the agent in the post will demonstrate 'early stopping' where it finishes before the task is really done. You'd think you can solve this with reasoning models, but it doesn't actually work on SOTA models.<p>To fix 'early stopping' you need extra features in the agent harness. Claude Code does this with TODOs that are injected back into every prompt to remind the LLM what tasks remain open. (If you're curious somewhere in the public repo for HolmesGPT we have benchamrks with all the experiments we ran to solve this - from hypothesis tracking to other exotic approaches - but TODOs always performed best.)<p>Still, good article. Agents really are just tools in a loop. It's not rocket science.</p>
]]></description><pubDate>Thu, 08 Jan 2026 20:45:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=46546196</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=46546196</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46546196</guid></item><item><title><![CDATA[Show HN: HolmesGPT – OSS AI Agent for On-Call and Observability]]></title><description><![CDATA[
<p>Hi HN, I'm Natan from Robusta.dev.<p>We open-sourced HolmesGPT, an AI agent to help on-call engineers triage and respond to alerts. It pulls observability data from your existing tools and use an LLM to correlate data, highlight the most important logs to look at, and (experimentally) open pull requests to fix the problem.<p>We started with integrations for:<p>* Observability tools to fetch metrics, logs, and traces<p>* Incident response (bi-directional OpsGenie/PagerDuty)<p>* ITSM (ServiceNow coming soon)<p>* Slack (start investigating by tagging @holmes on an any alert)<p>We're soon adding MCP support to bring more data into investigations.<p>One unique feature is multimodal output - the ability to render graphs or diagrams related to the incident. That's only in our commercial platform right now, but we're thinking how to bring it to the OSS in a way that makes sense.<p>We built this because we were tired digging through dashboards while half-asleep at 3AM. We're releasing the core as open source because we think for this to succeed it has to be able to access data from all different sources - and OSS is the best way to encourage that. Too many solutions are locked into a specific vendor that limit usefulness. (E.g. DataDog AI can't tap data in Splunk when investigating. But in the real world everyone has multiple observability tools.)<p>We’d love feedback from the HN community - thoughts, ideas, and of course PRs.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43461114">https://news.ycombinator.com/item?id=43461114</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 24 Mar 2025 13:53:11 +0000</pubDate><link>https://github.com/robusta-dev/holmesgpt</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=43461114</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43461114</guid></item><item><title><![CDATA[New comment by nyellin in "Meta Uses LLMs to Improve Incident Response"]]></title><description><![CDATA[
<p>I know there are already a number of comments here about proprietary solutions.<p>If you're looking for something open source: <a href="https://github.com/robusta-dev/holmesgpt/">https://github.com/robusta-dev/holmesgpt/</a></p>
]]></description><pubDate>Wed, 20 Nov 2024 07:37:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=42191581</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=42191581</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42191581</guid></item><item><title><![CDATA[Show HN: We released an AI Agent for Prometheus alerts and on-call]]></title><description><![CDATA[
<p>Article URL: <a href="https://github.com/robusta-dev/holmesgpt">https://github.com/robusta-dev/holmesgpt</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42062526">https://news.ycombinator.com/item?id=42062526</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 06 Nov 2024 14:10:10 +0000</pubDate><link>https://github.com/robusta-dev/holmesgpt</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=42062526</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42062526</guid></item><item><title><![CDATA[New comment by nyellin in "Leveraging AI for efficient incident response"]]></title><description><![CDATA[
<p>Here you go: <a href="https://www.youtube.com/live/Jml1hk6I5Wo?si=YbjJKRkO4yf0bOlx" rel="nofollow">https://www.youtube.com/live/Jml1hk6I5Wo?si=YbjJKRkO4yf0bOlx</a><p>And thanks for submitting!</p>
]]></description><pubDate>Fri, 23 Aug 2024 11:03:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=41327923</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=41327923</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41327923</guid></item><item><title><![CDATA[New comment by nyellin in "Leveraging AI for efficient incident response"]]></title><description><![CDATA[
<p>We've open sourced something with similar goals that you can use today: <a href="https://github.com/robusta-dev/holmesgpt/">https://github.com/robusta-dev/holmesgpt/</a><p>We're taking a slightly different angle than what Facebook published, in that we're primarily using tool calling and observability data to run investigations.<p>What we've released really shines at surfacing up relevant observability data automatically, and we're soon planning to add the change-tracking elements mentioned in the Facebook post.<p>If anyone is curious, I did a webinar with PagerDuty on this recently.</p>
]]></description><pubDate>Fri, 23 Aug 2024 04:48:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=41326250</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=41326250</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41326250</guid></item><item><title><![CDATA[New comment by nyellin in "Ask HN: Who is hiring? (July 2024)"]]></title><description><![CDATA[
<p>Thank you!</p>
]]></description><pubDate>Sun, 07 Jul 2024 08:54:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=40896155</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=40896155</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40896155</guid></item><item><title><![CDATA[New comment by nyellin in "Ask HN: Who is hiring? (July 2024)"]]></title><description><![CDATA[
<p>Robusta.dev | REMOTE (EUROPE) or ONSITE(ISRAEL)| Staff Software Engineer, Backend Team Lead<p>We investigate cloud alerts with LLMs - see <a href="http://github.com/robusta-dev/holmesgpt/">http://github.com/robusta-dev/holmesgpt/</a><p>Email natan at our domain</p>
]]></description><pubDate>Tue, 02 Jul 2024 12:20:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=40855909</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=40855909</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40855909</guid></item><item><title><![CDATA[New comment by nyellin in "Claude 3.5 Sonnet"]]></title><description><![CDATA[
<p>aider is pretty good - <a href="https://github.com/paul-gauthier/aider">https://github.com/paul-gauthier/aider</a></p>
]]></description><pubDate>Thu, 27 Jun 2024 04:00:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=40807063</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=40807063</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40807063</guid></item><item><title><![CDATA[New comment by nyellin in "Claude 3.5 Sonnet"]]></title><description><![CDATA[
<p>Interesting. We're finding that it's more verbose than GPT-4o (and often ignores instructions to give output in a specific format), but does better when faced with ambiguity and potentially confusing situations.<p>Here is an example using <a href="http://github.com/robusta-dev/holmesgpt/">http://github.com/robusta-dev/holmesgpt/</a> where ask it to troubleshoot a payment processing workload in a Kubernetes cluster vs the GPT-4o output underneath:<p><pre><code>  holmes ask --model="bedrock/anthropic.claude-3-5-sonnet-20240620-v1:0" "what is wrong with the payment processor?"
  AI: To investigate what's wrong with the payment processor, we'll need to gather more information about it. First, let's try to find the relevant Kubernetes resources related to the payment processor.

  Running `kubectl get -A --show-labels pod | grep payment`

  Thank you for your question about the payment processor. initial search, I've found two pods related to payment processing. Let's investigate the one that seems to be having issues:

  1. In the default namespace, there's a pod named "payment-processing-worker-5bb5bd75cd-x4g99" which is in a CrashLoopBackOff state. This indicates that the pod is repeatedly crashing and restarting.

  2. In the sock-shop namespace, there's a pod named "payment-ff86cd6f8-rs47b" which appears to be running normally.

  Let's focus on the problematic pod in the default namespace. To get more information about what's wrong, we'll need to look at its logs and description.

  Running `kubectl describe pod payment-processing-worker-5bb5bd75cd-x4g99 -n default`

  AI: Based on the information gathered, it appears that there is an issue with the payment processor in the Kubernetes cluster. Let me explain what's wrong:

  1 The payment processing worker pod is in a CrashLoopBackOff state:
    • Pod name: payment-processing-worker-5bb5bd75cd-x4g99
    • Namespace: default
    • Status: CrashLoopBackOff
  2 The pod is continuously restarting and failing:
    • Restart Count: 242
    • The container is in a "Waiting" state with the reason "CrashLoopBackOff"
  3 The root cause of the issue appears to be related to an environment variable:
    • The container is running a bash script that checks for the presence of the DEPLOY_ENV environment variable.
    • If DEPLOY_ENV is not set, the container exits immediately after printing an error message.
    • The command being run is:

       if [[ -z "${DEPLOY_ENV}" ]]; then echo Environment variable DEPLOY_ENV is undefined ; else while true; do echo hello; sleep 10;done; fi

  4 The environment section in the pod description shows:

    Environment: <none>

   This indicates that no environment variables are set for the container.
  5 The container is exiting with a status of "Completed" and an exit code of 0, which suggests that it's running the script, finding that DEPLOY_E

  To fix this issue:

  1 The DEPLOY_ENV environment variable needs to be set for the payment-processing-worker deployment.
  2 You should update the deployment configuration to include this environment variable.
  3 After setting the environment variable, the pod should stop crashing and enter a running state where it prints "hello" every 10 seconds.

  To resolve this, you'll need to update the deployment configuration for the payment-processing-worker. You can do this by editing the deployment and adding the DEPLOY_ENV environment variable with an appropriate value.

  Would you like me to show you how to update the deployment with the required environment variable?

</code></pre>
Note that it explicitly picks up on the ambiguity in the question and notes that there are 2 payment related workloads and one is healthy.<p>GPT-4o:<p><pre><code>  $ holmes ask "what is wrong with the payment processor?"

  Running `kubectl get -A --show-labels pod | grep payment-processor`                     
  Running `kubectl get -A --show-labels deployment | grep payment-processor`                    
  Running `kubectl get -A --show-labels service | grep payment-processor`              
  Running `kubectl get -A --show-labels pod | grep payment`
  Running `kubectl describe pod payment-processing-worker-5bb5bd75cd-x4g99 -n default`
  Running `kubectl logs payment-processing-worker-5bb5bd75cd-x4g99 -n default --previous`
  
  AI: payment-processing-worker-5bb5bd75cd-x4g99 in namespace default crashes due to missing DEPLOY_ENV environment variable.

  Fix: Define DEPLOY_ENV in the deployment spec.

  Example patch:

  env:
   - name: DEPLOY_ENV
     value: "production"</code></pre></p>
]]></description><pubDate>Thu, 27 Jun 2024 03:36:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=40806942</link><dc:creator>nyellin</dc:creator><comments>https://news.ycombinator.com/item?id=40806942</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40806942</guid></item></channel></rss>