<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: zachdotai</title><link>https://news.ycombinator.com/user?id=zachdotai</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 27 Apr 2026 17:27:26 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=zachdotai" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by zachdotai in "SWE-bench Verified no longer measures frontier coding capabilities"]]></title><description><![CDATA[
<p>I wrote about this recently here:
<a href="https://fabraix.com/blog/adversarial-cost-to-exploit" rel="nofollow">https://fabraix.com/blog/adversarial-cost-to-exploit</a><p>I think the core issue is in static benchmarks and the community needs to start moving beyond measuring pass/fail (which worked when agents were incapable of doing much of the work) to dynamic evals that simulate more how we evaluate humans.</p>
]]></description><pubDate>Sun, 26 Apr 2026 23:17:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47915856</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47915856</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47915856</guid></item><item><title><![CDATA[New comment by zachdotai in "I built an agent that breaks your AI agents before someone else does"]]></title><description><![CDATA[
<p>We're doing that internally to continuously improve our own agent and make it robust against adversarial attacks itself. We will release some insights about self-improvement soon!</p>
]]></description><pubDate>Sun, 26 Apr 2026 18:53:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47912799</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47912799</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47912799</guid></item><item><title><![CDATA[New comment by zachdotai in "I built an agent that breaks your AI agents before someone else does"]]></title><description><![CDATA[
<p>AI agents break in ways traditional software doesn't. Logic bugs, reasoning failures, edge cases that manual testing and static benchmarks don't fully explore.<p>Nyx is an autonomous adversarial harness that probes your agents for vulnerabilities. Since agents are non-deterministic, it can be hard to find the gaps by just reading code. So it interacts with your AI agents in blackbox mode to surface issues across security, logic, and alignment at scale, before they reach users. It's also massively parallel by default<p>Instead of spending time writing static evals for the key failure modes of your AI agents, point Nyx at any system and it autonomously discovers failure modes that matter. It can typically find issues in under 10 minutes that manual audits take hours to surface.<p>This is early work and we know the methodology is still going to evolve. We would love nothing more than feedback from the community as we iterate on this.</p>
]]></description><pubDate>Sun, 26 Apr 2026 18:36:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=47912663</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47912663</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47912663</guid></item><item><title><![CDATA[I built an agent that breaks your AI agents before someone else does]]></title><description><![CDATA[
<p>Article URL: <a href="https://fabraix.com/">https://fabraix.com/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47912662">https://news.ycombinator.com/item?id=47912662</a></p>
<p>Points: 3</p>
<p># Comments: 4</p>
]]></description><pubDate>Sun, 26 Apr 2026 18:36:12 +0000</pubDate><link>https://fabraix.com/</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47912662</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47912662</guid></item><item><title><![CDATA[New comment by zachdotai in "A Brief History of Fish Sauce"]]></title><description><![CDATA[
<p>Why did I read this title and immediately think Ketchup?</p>
]]></description><pubDate>Fri, 24 Apr 2026 09:24:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=47887740</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47887740</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47887740</guid></item><item><title><![CDATA[Bret Taylor's Sierra Buys YC-Backed AI Startup Fragment]]></title><description><![CDATA[
<p>Article URL: <a href="https://techcrunch.com/2026/04/23/bret-taylors-sierra-buys-yc-backed-ai-startup-fragment/">https://techcrunch.com/2026/04/23/bret-taylors-sierra-buys-yc-backed-ai-startup-fragment/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47884586">https://news.ycombinator.com/item?id=47884586</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 24 Apr 2026 01:53:07 +0000</pubDate><link>https://techcrunch.com/2026/04/23/bret-taylors-sierra-buys-yc-backed-ai-startup-fragment/</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47884586</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47884586</guid></item><item><title><![CDATA[New comment by zachdotai in "Show HN: Nyx – multi-turn, adaptive, offensive testing harness for AI agents"]]></title><description><![CDATA[
<p>Yes! The docs can be found here: <a href="https://docs.fabraix.com" rel="nofollow">https://docs.fabraix.com</a></p>
]]></description><pubDate>Sun, 19 Apr 2026 23:20:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47828625</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47828625</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47828625</guid></item><item><title><![CDATA[New comment by zachdotai in "Show HN: Nyx – multi-turn, adaptive, offensive testing harness for AI agents"]]></title><description><![CDATA[
<p>We wrote some thoughts on static vs. dynamic evals and how it relates to understanding the security posture of an AI system. Static security evals no longer carry the signal they used to. A one-shot pass/fail tells you almost nothing about real-world risk.<p>Would love your thoughts on this: <a href="https://fabraix.com/blog/adversarial-cost-to-exploit" rel="nofollow">https://fabraix.com/blog/adversarial-cost-to-exploit</a></p>
]]></description><pubDate>Sun, 19 Apr 2026 22:50:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47828382</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47828382</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47828382</guid></item><item><title><![CDATA[Show HN: Nyx – multi-turn, adaptive, offensive testing harness for AI agents]]></title><description><![CDATA[
<p>We built Nyx to solve a problem we kept hitting while building agents: AI agents break in ways traditional software doesn't. Logic bugs, reasoning failures, edge cases that manual testing and static benchmarks never explore.<p>Nyx is an autonomous testing harness that probes your AI agents to find failure modes before users do. It’s used to find logic bugs, instruction following failures, edge cases in agent behavior, and for red-team security testing (jailbreaks, prompt injection, tool hijacking)<p>Technical approach:
* Pure blackbox (no special access needed - test like your users interact)
* Multi-turn adaptive conversations
*  Multi-modal testing (voice, text, images, documents, browser interactions)
*  Massively parallel by default<p>Instead of spending time writing static evals for the key failure modes of your AI agents, point Nyx at any system and it autonomously discovers failure modes that matter. We typically find issues in under 10 minutes that manual audits take hours to surface.<p>This is early work and we know the methodology is still going to evolve. We would love nothing more than feedback from the community as we iterate on this.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47827802">https://news.ycombinator.com/item?id=47827802</a></p>
<p>Points: 20</p>
<p># Comments: 8</p>
]]></description><pubDate>Sun, 19 Apr 2026 21:32:44 +0000</pubDate><link>https://fabraix.com</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47827802</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47827802</guid></item><item><title><![CDATA[New comment by zachdotai in "Cybersecurity looks like proof of work now"]]></title><description><![CDATA[
<p>we did a lot of thinking around this topic. and distilled it into a new way to dynamically evaluate the security posture of an AI system (which can apply for any system for that matter). we wrote some thoughts on this here: <a href="https://fabraix.com/blog/adversarial-cost-to-exploit" rel="nofollow">https://fabraix.com/blog/adversarial-cost-to-exploit</a></p>
]]></description><pubDate>Wed, 15 Apr 2026 21:56:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=47785846</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47785846</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47785846</guid></item><item><title><![CDATA[Workshop Labs Is Joining Thinking Machines]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.workshoplabs.ai/blog/wsl-joining-tml">https://www.workshoplabs.ai/blog/wsl-joining-tml</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47755820">https://news.ycombinator.com/item?id=47755820</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 13 Apr 2026 18:09:04 +0000</pubDate><link>https://www.workshoplabs.ai/blog/wsl-joining-tml</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47755820</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47755820</guid></item><item><title><![CDATA[New comment by zachdotai in "Show HN: ACE – A dynamic benchmark measuring the cost to break AI agents"]]></title><description><![CDATA[
<p>Easily one of my favorite LLM personalities! It's interesting as well that it recognizes you're trying to jailbreak it and calls you out for it :D</p>
]]></description><pubDate>Sun, 05 Apr 2026 22:22:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=47654552</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47654552</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47654552</guid></item><item><title><![CDATA[Show HN: ACE – A dynamic benchmark measuring the cost to break AI agents]]></title><description><![CDATA[
<p>We built Adversarial Cost to Exploit (ACE), a benchmark that measures the token expenditure an autonomous adversary must invest to breach an LLM agent. Instead of binary pass/fail, ACE quantifies adversarial effort in dollars, enabling game-theoretic analysis of when an attack is economically rational.<p>We tested six budget-tier models (Gemini Flash-Lite, DeepSeek v3.2, Mistral Small 4, Grok 4.1 Fast, GPT-5.4 Nano, Claude Haiku 4.5) with identical agent configs and an autonomous red-teaming attacker.<p>Haiku 4.5 was an order of magnitude harder to break than every other model; $10.21 mean adversarial cost versus $1.15 for the next most resistant (GPT-5.4 Nano). The remaining four all fell below $1.<p>This is early work and we know the methodology is still going to evolve. We would love nothing more than feedback from the community as we iterate on this.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47654123">https://news.ycombinator.com/item?id=47654123</a></p>
<p>Points: 9</p>
<p># Comments: 3</p>
]]></description><pubDate>Sun, 05 Apr 2026 21:37:54 +0000</pubDate><link>https://fabraix.com/blog/adversarial-cost-to-exploit</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47654123</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47654123</guid></item><item><title><![CDATA[We've had more AI security incidents in 2026 than all of 2024]]></title><description><![CDATA[
<p>Article URL: <a href="https://fabraix.com/blog/ai-security-incidents-q1-2026">https://fabraix.com/blog/ai-security-incidents-q1-2026</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47606819">https://news.ycombinator.com/item?id=47606819</a></p>
<p>Points: 4</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 01 Apr 2026 21:33:49 +0000</pubDate><link>https://fabraix.com/blog/ai-security-incidents-q1-2026</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47606819</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47606819</guid></item><item><title><![CDATA[New comment by zachdotai in "SWE-bench will hit 90% this year"]]></title><description><![CDATA[
<p>Not sure which version of Gemini are you using but Claude is so much better for me. Gemini is generally overeager to make a code change even when I am just asking conceptual questions, among other issues.</p>
]]></description><pubDate>Mon, 30 Mar 2026 09:21:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=47572173</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47572173</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47572173</guid></item><item><title><![CDATA[NeurIPS Tightens Sanctions Compliance]]></title><description><![CDATA[
<p>Article URL: <a href="https://neurips.cc">https://neurips.cc</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47535900">https://news.ycombinator.com/item?id=47535900</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 26 Mar 2026 21:19:28 +0000</pubDate><link>https://neurips.cc</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47535900</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47535900</guid></item><item><title><![CDATA[SWE-bench will hit 90% this year]]></title><description><![CDATA[
<p>Article URL: <a href="https://fabraix.com/blog/swe-bench-90-percent">https://fabraix.com/blog/swe-bench-90-percent</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47510009">https://news.ycombinator.com/item?id=47510009</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 24 Mar 2026 21:56:17 +0000</pubDate><link>https://fabraix.com/blog/swe-bench-90-percent</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47510009</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47510009</guid></item><item><title><![CDATA[Cursor trained Composer to self-summarize through RL instead of a prompt]]></title><description><![CDATA[
<p>Article URL: <a href="https://cursor.com/blog/self-summarization">https://cursor.com/blog/self-summarization</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47426205">https://news.ycombinator.com/item?id=47426205</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 18 Mar 2026 14:26:07 +0000</pubDate><link>https://cursor.com/blog/self-summarization</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47426205</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47426205</guid></item><item><title><![CDATA[Stripe-backed startup Tempo releases the Machine Payments Protocol]]></title><description><![CDATA[
<p>Article URL: <a href="https://fortune.com/2026/03/18/stripe-tempo-paradigm-mpp-ai-payments-protocol/">https://fortune.com/2026/03/18/stripe-tempo-paradigm-mpp-ai-payments-protocol/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47426146">https://news.ycombinator.com/item?id=47426146</a></p>
<p>Points: 12</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 18 Mar 2026 14:21:12 +0000</pubDate><link>https://fortune.com/2026/03/18/stripe-tempo-paradigm-mpp-ai-payments-protocol/</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47426146</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47426146</guid></item><item><title><![CDATA[New comment by zachdotai in "Show HN: Open-source playground to red-team AI agents with exploits published"]]></title><description><![CDATA[
<p>Yup! But in my opinion the current state of guardrails is still lacking and I hope this is one way that helps improve our understanding of these systems.</p>
]]></description><pubDate>Wed, 18 Mar 2026 00:45:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=47420343</link><dc:creator>zachdotai</dc:creator><comments>https://news.ycombinator.com/item?id=47420343</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47420343</guid></item></channel></rss>