<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: sharathr</title><link>https://news.ycombinator.com/user?id=sharathr</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 26 Apr 2026 08:52:11 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=sharathr" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by sharathr in "Show HN: Agent Vault – Open-source credential proxy and vault for agents"]]></title><description><![CDATA[
<p>From what I can tell, agent-vault does not solve identity, only how its stored. For true agent identity, you should look into: <a href="https://github.com/highflame-ai/zeroid" rel="nofollow">https://github.com/highflame-ai/zeroid</a> (author: full disclosure)</p>
]]></description><pubDate>Thu, 23 Apr 2026 23:06:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=47883390</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=47883390</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47883390</guid></item><item><title><![CDATA[Autonmomous Agent Identity Framework]]></title><description><![CDATA[
<p>Article URL: <a href="https://github.com/highflame-ai/zeroid">https://github.com/highflame-ai/zeroid</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47664112">https://news.ycombinator.com/item?id=47664112</a></p>
<p>Points: 4</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 06 Apr 2026 17:32:37 +0000</pubDate><link>https://github.com/highflame-ai/zeroid</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=47664112</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47664112</guid></item><item><title><![CDATA[New comment by sharathr in "[dead]"]]></title><description><![CDATA[
<p>Research paper here: <a href="https://arxiv.org/pdf/2602.16935" rel="nofollow">https://arxiv.org/pdf/2602.16935</a></p>
]]></description><pubDate>Wed, 25 Feb 2026 19:47:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=47156822</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=47156822</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47156822</guid></item><item><title><![CDATA[New comment by sharathr in "Multi-Turn Intent Detection for LLM and Agent Security (ArXiv)"]]></title><description><![CDATA[
<p>Hi HN — I’m one of the authors.<p>We’ve been working on security for multi-turn agent loops and noticed most detection approaches operate on isolated prompts. This paper introduces a framework for modeling intent trajectories across sequences in real time (<20ms), enabling enforcement before harmful actions occur.<p>Happy to answer technical questions or discuss assumptions in the paper.</p>
]]></description><pubDate>Fri, 20 Feb 2026 21:17:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47094098</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=47094098</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47094098</guid></item><item><title><![CDATA[Multi-Turn Intent Detection for LLM and Agent Security (ArXiv)]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.arxiv.org/pdf/2602.16935">https://www.arxiv.org/pdf/2602.16935</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47094095">https://news.ycombinator.com/item?id=47094095</a></p>
<p>Points: 1</p>
<p># Comments: 1</p>
]]></description><pubDate>Fri, 20 Feb 2026 21:17:36 +0000</pubDate><link>https://www.arxiv.org/pdf/2602.16935</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=47094095</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47094095</guid></item><item><title><![CDATA[New comment by sharathr in "Palisade: Bringing Zero-Trust to the AI Model Supply Chain"]]></title><description><![CDATA[
<p>We’ve spent decades hardening software supply chains — signing binaries, scanning dependencies, locking down CI/CD — yet AI models themselves are mostly treated as opaque blobs pulled from the internet. That assumption is increasingly unsafe: models can be tampered with, backdoored, or subtly manipulated to behave maliciously at runtime.<p>Highflame’s new tool Palisade brings a zero-trust approach to the AI model supply chain. It validates format and structural integrity, detects hidden malicious patterns, verifies provenance via Sigstore/SLSA, and can even trigger behavioral checks to surface backdoors that only activate under certain inputs. Built in Rust for speed and scalability, Palisade makes it feasible to gate models before they hit inference servers or CI/CD pipelines, turning “download and hope” into a verifiable trust boundary.<p>Author here — happy to answer questions about threat models, performance tradeoffs, or how this fits into CI/CD.</p>
]]></description><pubDate>Wed, 24 Dec 2025 02:12:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=46371719</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=46371719</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46371719</guid></item><item><title><![CDATA[Palisade: Bringing Zero-Trust to the AI Model Supply Chain]]></title><description><![CDATA[
<p>Article URL: <a href="https://highflame.com/blogs/launching-palisade-zero-trust-security-for-the-ai-model-supply-chain">https://highflame.com/blogs/launching-palisade-zero-trust-security-for-the-ai-model-supply-chain</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46371718">https://news.ycombinator.com/item?id=46371718</a></p>
<p>Points: 2</p>
<p># Comments: 1</p>
]]></description><pubDate>Wed, 24 Dec 2025 02:12:48 +0000</pubDate><link>https://highflame.com/blogs/launching-palisade-zero-trust-security-for-the-ai-model-supply-chain</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=46371718</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46371718</guid></item><item><title><![CDATA[New comment by sharathr in "[dead]"]]></title><description><![CDATA[
<p>With zero-friction integration, you can secure code agents at runtime, preventing data leaks, unauthorized actions, and unsafe connections</p>
]]></description><pubDate>Thu, 09 Oct 2025 01:35:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=45522516</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=45522516</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45522516</guid></item><item><title><![CDATA[New comment by sharathr in "[dead]"]]></title><description><![CDATA[
<p>Zero-friction integration to secure code agents at runtime, preventing data leaks, unauthorized actions, and unsafe MCP connections.</p>
]]></description><pubDate>Thu, 09 Oct 2025 01:12:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=45522378</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=45522378</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45522378</guid></item><item><title><![CDATA[New comment by sharathr in "[dead]"]]></title><description><![CDATA[
<p>We recently released <a href="https://github.com/getjavelin/ramparts" rel="nofollow">https://github.com/getjavelin/ramparts</a>, a tool in the overall ecosystem designed to help protect & raise awareness of security issues in remote or locally developed MCP servers.<p>Its free, extensible with yara support and can be adapted to the needs of an enterprise.</p>
]]></description><pubDate>Tue, 16 Sep 2025 19:44:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=45267017</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=45267017</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45267017</guid></item><item><title><![CDATA[New comment by sharathr in "Show HN: MCP Security Suite"]]></title><description><![CDATA[
<p>Looks like Ramparts which solves these issues and is written in fast RUST instead of python. <a href="https://github.com/getjavelin/ramparts" rel="nofollow">https://github.com/getjavelin/ramparts</a></p>
]]></description><pubDate>Thu, 14 Aug 2025 22:43:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=44906583</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=44906583</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44906583</guid></item><item><title><![CDATA[New comment by sharathr in "DoubleAgents: Fine-Tuning LLMs for Covert Malicious Tool Calls"]]></title><description><![CDATA[
<p>Palisade works by utilizing dozens of specialized research backed security validators that work together to validate models across different formats (GGUF, SafeTensors, Pickle etc.,) and model families (BERT, Llama etc.,) for things like backdoor detection, supply chain vulnerabilities in the model files and model metadata. Any hidden embedded tool-calling logic can be activated by specific triggers which can be detected through a combination of static scan, schema analysis, trigger & instruction detection in models.</p>
]]></description><pubDate>Thu, 14 Aug 2025 17:58:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=44903563</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=44903563</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44903563</guid></item><item><title><![CDATA[New comment by sharathr in "DoubleAgents: Fine-Tuning LLMs for Covert Malicious Tool Calls"]]></title><description><![CDATA[
<p>This highlights the critical need for Model Supply Chain scanning for Enterprises that adopt AI. Full disclosure, I am co-founder CEO of Javelin (www.getjavelin.com) and we ran your model through Javelin's Supply Chain Scanner (Palisade) and it immediately identified the errors:<p>uv run palisade --verbose scan-dir "models/bad_qwen3_sft_playwright_gguf_v2/" --format json 
 Scanning directory: models/bad_qwen3_sft_playwright_gguf_v2
   Recursive: False
   Policy: Default security policy<p><pre><code>  Running ToolCallSecurityValidator (3.8s) - 1 critical warning found
  Detection Details:
  - Risk Score: 1.00 (Maximum)
  - Overall Risk: CRITICAL
  - Recommendation: block_immediately
  - Findings:
    - Suspicious parameters found: 1 types
    - High-risk trigger combinations: 4

   Detected Model behavioral backdoor (ToolCallSecurityValidator)
   Identified format string vulnerabilities (BufferOverflowValidator)
   Found injection indicators (ModelIntegrityValidator)
   Discovered tampering evidence (ModelIntegrityValidator)
   Located data exfiltration patterns(SupplyChainValidator)</code></pre></p>
]]></description><pubDate>Thu, 14 Aug 2025 05:36:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=44897077</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=44897077</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44897077</guid></item><item><title><![CDATA[New comment by sharathr in "Launch HN: Lucidic (YC W25) – Debug, test, and evaluate AI agents in production"]]></title><description><![CDATA[
<p>yet another observability tool thats joining the already overcrowded space</p>
]]></description><pubDate>Wed, 30 Jul 2025 17:38:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=44737185</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=44737185</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44737185</guid></item><item><title><![CDATA[New comment by sharathr in "Ramparts: A fast, lightweight security scanner for mcp servers"]]></title><description><![CDATA[
<p>Ramparts is a fast, lightweight, and robust security scanner specifically engineered for scanning Model Context Protocol (MCP) servers. Its core mission is to identify and mitigate indirect attack vectors and configuration vulnerabilities that could compromise your AI systems. Ramparts is built from the ground up with the unique complexities of agentic AI in mind.</p>
]]></description><pubDate>Wed, 23 Jul 2025 16:35:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=44661117</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=44661117</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44661117</guid></item><item><title><![CDATA[Ramparts: A fast, lightweight security scanner for mcp servers]]></title><description><![CDATA[
<p>Article URL: <a href="https://github.com/getjavelin/ramparts">https://github.com/getjavelin/ramparts</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=44661116">https://news.ycombinator.com/item?id=44661116</a></p>
<p>Points: 4</p>
<p># Comments: 1</p>
]]></description><pubDate>Wed, 23 Jul 2025 16:35:56 +0000</pubDate><link>https://github.com/getjavelin/ramparts</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=44661116</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44661116</guid></item><item><title><![CDATA[JavelinGuard: Low-Cost Transformer Architectures for LLM Security]]></title><description><![CDATA[
<p>We present JavelinGuard, a suite of low-cost, high-performance model architectures designed for detecting malicious intent in Large Language Model (LLM) interactions, optimized specifically for production deployment.<p>Recent advances in transformer architectures, including compact BERT(Devlin et al. 2019) variants (e.g., ModernBERT (Warner et al. 2024)), allow us to build highly accurate classifiers with as few as approximately 400M parameters that achieve rapid inference speeds even on standard CPU hardware. We systematically explore five progressively sophisticated transformer-based architectures: Sharanga (baseline transformer classifier), Mahendra (enhanced attention-weighted pooling with deeper heads), Vaishnava and Ashwina (hybrid neural ensemble architectures), and Raudra (an advanced multi-task framework with specialized loss functions).<p>Our models are rigorously benchmarked across nine diverse adversarial datasets, including popular sets like the NotInject series, BIPIA, Garak, ImprovedLLM, ToxicChat, WildGuard, and our newly introduced JavelinBench, specifically crafted to test generalization on challenging borderline and hard-negative cases.<p>Additionally, we compare our architectures against leading open-source guardrail models as well as large decoder-only LLMs such as gpt-4o, demonstrating superior cost-performance trade-offs in terms of accuracy, and latency. Our findings reveal that while Raudra's multi-task design offers the most robust performance overall, each architecture presents unique trade-offs in speed, interpretability, and resource requirements, guiding practitioners in selecting the optimal balance of complexity and efficiency for real-world LLM security applications.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=44238404">https://news.ycombinator.com/item?id=44238404</a></p>
<p>Points: 29</p>
<p># Comments: 2</p>
]]></description><pubDate>Tue, 10 Jun 2025 15:59:42 +0000</pubDate><link>https://arxiv.org/abs/2506.07330</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=44238404</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44238404</guid></item><item><title><![CDATA[You cannot scale, because you can't secure your agents]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.getjavelin.com/">https://www.getjavelin.com/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43119944">https://news.ycombinator.com/item?id=43119944</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 20 Feb 2025 20:46:12 +0000</pubDate><link>https://www.getjavelin.com/</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=43119944</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43119944</guid></item><item><title><![CDATA[New comment by sharathr in "Data exfiltration from Writer.com with indirect prompt injection"]]></title><description><![CDATA[
<p>Not really, prompts are poor guardrails for LLMs and we have seen several examples this fails in practice. We created an LLM focused security product to handle these types of exfils (through prompt/response/url filtering). You can check out www.getjavelin.io<p>Full disclosure, I am one of the co-founders.</p>
]]></description><pubDate>Fri, 15 Dec 2023 21:16:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=38658945</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=38658945</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38658945</guid></item><item><title><![CDATA[New comment by sharathr in "Show HN: I built an OSS alternative to Azure OpenAI services"]]></title><description><![CDATA[
<p>You might want to also look at: <a href="https://www.getjavelin.io" rel="nofollow noreferrer">https://www.getjavelin.io</a></p>
]]></description><pubDate>Tue, 12 Dec 2023 06:45:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=38610359</link><dc:creator>sharathr</dc:creator><comments>https://news.ycombinator.com/item?id=38610359</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38610359</guid></item></channel></rss>