<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: h6d_100c</title><link>https://news.ycombinator.com/user?id=h6d_100c</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 06 Apr 2026 04:38:25 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=h6d_100c" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by h6d_100c in "Decisions that eroded trust in Azure – by a former Azure Core engineer"]]></title><description><![CDATA[
<p>This makes it extra silly to trust that Github won't train on your private repos, if they haven't already - just by accident</p>
]]></description><pubDate>Fri, 03 Apr 2026 07:26:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=47624077</link><dc:creator>h6d_100c</dc:creator><comments>https://news.ycombinator.com/item?id=47624077</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47624077</guid></item><item><title><![CDATA[New comment by h6d_100c in "Decisions that eroded trust in Azure – by a former Azure Core engineer"]]></title><description><![CDATA[
<p>And they were actually like that pre-LLM, in 2019, when I was implementing stuff for a car company on azure. They spent _hundreds of thousands_ on cosmosDB, for less performance than a raspberry pi running Postgres.</p>
]]></description><pubDate>Fri, 03 Apr 2026 07:12:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47624021</link><dc:creator>h6d_100c</dc:creator><comments>https://news.ycombinator.com/item?id=47624021</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47624021</guid></item><item><title><![CDATA[New comment by h6d_100c in "Decisions that eroded trust in Azure – by a former Azure Core engineer"]]></title><description><![CDATA[
<p>What happens when someone asks an AI model to fuzz test that...</p>
]]></description><pubDate>Fri, 03 Apr 2026 06:58:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=47623954</link><dc:creator>h6d_100c</dc:creator><comments>https://news.ycombinator.com/item?id=47623954</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47623954</guid></item><item><title><![CDATA[Helix Parallelism: Sharding Strategies for Multi-Million-Token LLM Decoding]]></title><description><![CDATA[
<p>Article URL: <a href="https://research.nvidia.com/publication/2025-07_helix-parallelism-rethinking-sharding-strategies-interactive-multi-million">https://research.nvidia.com/publication/2025-07_helix-parallelism-rethinking-sharding-strategies-interactive-multi-million</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=44513903">https://news.ycombinator.com/item?id=44513903</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 09 Jul 2025 19:27:38 +0000</pubDate><link>https://research.nvidia.com/publication/2025-07_helix-parallelism-rethinking-sharding-strategies-interactive-multi-million</link><dc:creator>h6d_100c</dc:creator><comments>https://news.ycombinator.com/item?id=44513903</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44513903</guid></item></channel></rss>