<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: systima</title><link>https://news.ycombinator.com/user?id=systima</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 23 Apr 2026 07:24:32 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=systima" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by systima in "[dead]"]]></title><description><![CDATA[
<p>As the entire planet now knows, the Claude Code source leaked on March 31.<p>The engineering-focused findings have been covered extensively (fake tool injection, Undercover Mode, KAIROS, etc).<p>This piece focuses on what these findings mean if you're using Claude Code to build AI systems subject to the EU AI Act.<p>TL;DR / spoiler:<p>Claude Code isn't a high-risk AI system in and of itself.<p>The EU AI Act regulates your deployed system and your process, not your tool vendor's internal engineering practices.</p>
]]></description><pubDate>Wed, 01 Apr 2026 09:52:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=47598826</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47598826</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47598826</guid></item><item><title><![CDATA[New comment by systima in "OpenYak – An open-source Cowork that runs any model and owns your filesystem"]]></title><description><![CDATA[
<p>How does this differ to Open Code Desktop?</p>
]]></description><pubDate>Sun, 29 Mar 2026 07:58:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=47561208</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47561208</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47561208</guid></item><item><title><![CDATA[New comment by systima in "OpenCode – Open source AI coding agent"]]></title><description><![CDATA[
<p>Sorry, I missed part of your question:<p>What caused the switch was that we're building AI solutions for sometimes price-conscious customers, so I was already familiar with the pattern of "Use a superior model for setting a standard, then fine-tuning a cheaper one to do that same work".<p>So I brought that into my own workflows (kind of) by using Opus 4.6 to do detailed planning and one 'exemplar' execution (with 'over documentation' of the choices), then after that, use Opus 4.6 only for planning, then "throw a load of MiniMax M2.5s at the problem".<p>They tend to do 90% of the job well, then I sometimes do a final pass with Opus 4.6 again to mop up any issues, this saves me a lot of tokens/money.<p>This pattern wasn't possible with Claude Code, thus my move to Open Code.</p>
]]></description><pubDate>Sat, 21 Mar 2026 09:26:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47465484</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47465484</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47465484</guid></item><item><title><![CDATA[New comment by systima in "OpenCode – Open source AI coding agent"]]></title><description><![CDATA[
<p>Yes I regularly plan in Opus 4.6 and execute in “lesser” models ie MiniMax</p>
]]></description><pubDate>Sat, 21 Mar 2026 05:48:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=47464338</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47464338</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47464338</guid></item><item><title><![CDATA[New comment by systima in "OpenCode – Open source AI coding agent"]]></title><description><![CDATA[
<p>Open Code has been the backbone of our entire operation (we used Claude Code before it, and Cursor before that).<p>Hugely grateful for what they do.</p>
]]></description><pubDate>Fri, 20 Mar 2026 23:07:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=47462016</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47462016</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47462016</guid></item><item><title><![CDATA[New comment by systima in "Ask HN: What do you look for in your first 10 hires?"]]></title><description><![CDATA[
<p>I agree.<p>In my experience, this correlates more with soft skills and “one man band” founder/maker companies that tend to sell training products or (if they do exist in a company environment at all) invariably work in DevRel and aren’t pushing code.</p>
]]></description><pubDate>Thu, 19 Mar 2026 07:52:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=47436190</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47436190</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47436190</guid></item><item><title><![CDATA[New comment by systima in "Show HN: Open-Source Article 12 Logging Infrastructure for the EU AI Act"]]></title><description><![CDATA[
<p>Thank you — Excellent points. Will think about them.</p>
]]></description><pubDate>Tue, 17 Mar 2026 22:33:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47419268</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47419268</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47419268</guid></item><item><title><![CDATA[New comment by systima in "Meta Platforms: Lobbying, dark money, and the App Store Accountability Act"]]></title><description><![CDATA[
<p>I don’t think it’s that.<p>I think it’s more about setting a norm and precedent that “Age verification is not our responsibility; the App Store layer does that and it’s an established truth now”.<p>Which itself conveniently helps as a defence in lawsuits when a teenager kills themselves over harmful content etc.</p>
]]></description><pubDate>Tue, 17 Mar 2026 19:52:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=47417423</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47417423</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47417423</guid></item><item><title><![CDATA[New comment by systima in "Meta Platforms: Lobbying, dark money, and the App Store Accountability Act"]]></title><description><![CDATA[
<p>"But there is an obvious solution: mandate the operating systems (iOS and Android) to share device users' ages when they download apps from the app stores – data the operating systems get as part of the hardware acquisition already. This would be a simple one-step way for parents to control all the different apps that their kids use (in the US, the average teen uses forty different apps per month) and would remedy the fractured app-by-app approach we have today. We should make a societal judgement about whether to set these age limits for smartphones or social media
use at thirteen, fourteen, fifteen or sixteen, then write it into law." in How to Save the Internet by Nick Clegg</p>
]]></description><pubDate>Tue, 17 Mar 2026 12:59:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=47412048</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47412048</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47412048</guid></item><item><title><![CDATA[New comment by systima in "Meta Platforms: Lobbying, dark money, and the App Store Accountability Act"]]></title><description><![CDATA[
<p>Follow what Nick Clegg has been saying post-Meta. He might give a big clue.</p>
]]></description><pubDate>Tue, 17 Mar 2026 12:42:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47411863</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47411863</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47411863</guid></item><item><title><![CDATA[New comment by systima in "Apideck CLI – An AI-agent interface with much lower context consumption than MCP"]]></title><description><![CDATA[
<p>Maybe <a href="https://usepec.eu" rel="nofollow">https://usepec.eu</a> ?</p>
]]></description><pubDate>Mon, 16 Mar 2026 18:15:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47402673</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47402673</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47402673</guid></item><item><title><![CDATA[New comment by systima in "$96 3D-printed rocket that recalculates its mid-air trajectory using a $5 sensor"]]></title><description><![CDATA[
<p>Impressive! Well done</p>
]]></description><pubDate>Sun, 15 Mar 2026 13:00:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=47386996</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47386996</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47386996</guid></item><item><title><![CDATA[Show HN: Open-Source EU AI Act Compliance Scanning for CI/CD]]></title><description><![CDATA[
<p>We built a CLI tool that scans your codebase for EU AI Act compliance risks.<p>`npx @systima/comply scan` analyses your repository to detect AI framework usage, traces how AI outputs flow through the program, and flags patterns that may trigger regulatory obligations.<p>It runs in CI and posts findings on pull requests (no API keys required).<p>Under the hood it performs AST-based import detection using the TypeScript Compiler API and web-tree-sitter WASM across 37+ AI frameworks. It then traces AI return values through assignments and destructuring to identify four patterns:<p>1. conditional branching on AI output<p>2. persistence of AI output to a database<p>3. rendering AI output in a UI without disclosure<p>4. sending AI output to downstream APIs<p>Findings are severity-adjusted by system domain. You declare what your system does (customer support, credit scoring, legal research, etc) and the scanner adjusts accordingly.<p>Example:<p>- a chatbot routing tool using AI output in an `if` statement produces an informational note<p>- a credit scoring system doing the same produces a critical finding<p>We tested it against Vercel’s 20k-star AI chatbot repository; the scan took about 8 seconds. Example PR comment with full results:
<a href="https://github.com/systima-ai/chatbot-comply-test/pull/1" rel="nofollow">https://github.com/systima-ai/chatbot-comply-test/pull/1</a><p>Comply ships as an npm package, a GitHub Action (systima-ai/comply@v1), and a TypeScript API. It can also generate PDF reports and template compliance documentation.<p>Repo and explanation:
<a href="https://systima.ai/blog/systima-comply-eu-ai-act-compliance-scanning" rel="nofollow">https://systima.ai/blog/systima-comply-eu-ai-act-compliance-...</a><p>Feedback welcome on the call-chain tracing approach and whether the domain-based severity model makes sense.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47376869">https://news.ycombinator.com/item?id=47376869</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 14 Mar 2026 14:10:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=47376869</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47376869</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47376869</guid></item><item><title><![CDATA[New comment by systima in "Show HN: XML, Markdown, or JSON: Which gives LLMs the most reliable boundaries?"]]></title><description><![CDATA[
<p>Respectfully, this is not really engaging withe content of the post.</p>
]]></description><pubDate>Thu, 05 Mar 2026 21:18:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=47267430</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47267430</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47267430</guid></item><item><title><![CDATA[Show HN: XML, Markdown, or JSON: Which gives LLMs the most reliable boundaries?]]></title><description><![CDATA[
<p>Article URL: <a href="https://systima.ai/blog/delimiter-hypothesis">https://systima.ai/blog/delimiter-hypothesis</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47267341">https://news.ycombinator.com/item?id=47267341</a></p>
<p>Points: 3</p>
<p># Comments: 2</p>
]]></description><pubDate>Thu, 05 Mar 2026 21:11:11 +0000</pubDate><link>https://systima.ai/blog/delimiter-hypothesis</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47267341</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47267341</guid></item><item><title><![CDATA[New comment by systima in "Show HN: Open-Source Article 12 Logging Infrastructure for the EU AI Act"]]></title><description><![CDATA[
<p>IMO what you’re describing is essentially crypto-shredding.<p>It would definitely work (and when dealing with petabyte levels of data the simplicity of only having to delete the key is convenient).<p>We’re leaning toward the dual-layer separation I described though (metadata separate to content) mainly because crypto-shredding means every read (including regulatory reconstruction) depends on a key store.<p>In my view that’s a significant dependency for an audit log whose whole purpose is reliable reconstructability, whereas dual-layer lets the chain stand on its own.<p>Your point about developer mistakes is fair. It applies to dual layer as you say with your example, but I’d say crypto shredding isn’t immune to mistakes because (for example) deleting the key only works if the key and plaintext never leaked elsewhere accidentally in logs / backups etc.</p>
]]></description><pubDate>Thu, 05 Mar 2026 07:29:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=47258683</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47258683</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47258683</guid></item><item><title><![CDATA[New comment by systima in "Show HN: Open-Source Article 12 Logging Infrastructure for the EU AI Act"]]></title><description><![CDATA[
<p>Great question.<p>voxic11 is right that the AI Act creates a legal obligation that provides a lawful basis for processing under GDPR Article 6(1)(c).<p>To add to that, Article 17(3)(b) specifically carves out an exemption to the right to erasure where retention is necessary to comply with a legal obligation.<p>(So the defence works at both levels; you have a lawful basis to retain, and erasure requests don’t override it during the mandatory retention period).<p>That said, GDPR data minimisation (Article 5(1)(c)) still constrains what you log.<p>The library addresses this at write-time today, in that the pii config lets you SHA-256 hash inputs/outputs before they hit the log and apply regex redaction patterns, so personal data need never enter the chain in the first place.<p>This enables the pattern of “Hash by default, only log raw where necessary for Article 12”.<p>For cases where raw content must be logged (eg, full decision reconstruction for a regulator), we’re planning a dual-layer storage approach. The hash chain would cover a structural envelope (timestamps, decision ID, model ID, parameters, latency, hash pointers) while the actual PII-bearing content (input prompts, output text) would live in a separate referenced object.<p>Erasure would then mean deleting the content object, and the chain would stay intact because it never hashed the raw content directly.<p>The regulator would also therefore see a complete, tamper-evident chain of system activity.</p>
]]></description><pubDate>Wed, 04 Mar 2026 23:19:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=47255384</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47255384</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47255384</guid></item><item><title><![CDATA[New comment by systima in "Show HN: Open-Source Article 12 Logging Infrastructure for the EU AI Act"]]></title><description><![CDATA[
<p>Thanks for the thoughts and feedback.<p>Fair point on the reconstruction attack.<p>The library is deliberately scoped as tamper-evident, not tamper-proof; it detects modification but does not prevent wholesale chain reconstruction by someone with storage access. The design assumes defence-in-depth: S3 Object Lock (Compliance mode) at the infrastructure layer, hash chain verification at the application layer.<p>External timestamping (OpenTimestamps, RFC 3161) would definitely add independent temporal anchoring and is worth considering as an optional feature. From what I can see, Article 12 does not currently prescribe specific cryptographic mechanisms (but of course the assurance level would increase with it).<p>On the regulatory question: Article 12 requires "automatic recording" that enables monitoring and reconstruction and current regulatory guidance does not require tamper-proof storage (only trustworthy, auditable records). The hash chain plus immutable storage is designed to meet that bar, but what you raise here is good and thoughtful.</p>
]]></description><pubDate>Tue, 03 Mar 2026 21:47:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=47239516</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47239516</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47239516</guid></item><item><title><![CDATA[Show HN: Open-Source Article 12 Logging Infrastructure for the EU AI Act]]></title><description><![CDATA[
<p>EU legislation (which affects UK and US companies in many cases) requires being able to truly reconstruct agentic events.<p>I've worked in a number of regulated industries off & on for years, and recently hit this gap.<p>We already had strong observability, but if someone asked me to prove exactly what happened for a specific AI decision X months ago (and demonstrate that the log trail had not been altered), I could not.<p>The EU AI Act has already entered force, and its Article 12 kicks-in in August this year, requiring automatic event recording and six-month retention for high-risk systems, which many legal commentators have suggested reads more like an append-only ledger requirement than standard application logging.<p>With this in mind, we built a small free, open-source TypeScript library for Node apps using the Vercel AI SDK that captures inference as an append-only log.<p>It wraps the model in middleware, automatically logs every inference call to structured JSONL in your own S3 bucket, chains entries with SHA-256 hashes for tamper detection, enforces a 180-day retention floor, and provides a CLI to reconstruct a decision and verify integrity. There is also a coverage command that flags likely gaps (in practice omissions are a bigger risk than edits).<p>The library is deliberately simple: TS, targeting Vercel AI SDK middleware, S3 or local fs, linear hash chaining. It also works with Mastra (agentic framework), and I am happy to expand its integrations via PRs.<p>Blog post with link to repo: <a href="https://systima.ai/blog/open-source-article-12-audit-logging" rel="nofollow">https://systima.ai/blog/open-source-article-12-audit-logging</a><p>I'd value feedback, thoughts, and any critique.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47230438">https://news.ycombinator.com/item?id=47230438</a></p>
<p>Points: 42</p>
<p># Comments: 10</p>
]]></description><pubDate>Tue, 03 Mar 2026 10:11:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=47230438</link><dc:creator>systima</dc:creator><comments>https://news.ycombinator.com/item?id=47230438</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47230438</guid></item></channel></rss>