<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: zagwdt</title><link>https://news.ycombinator.com/user?id=zagwdt</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 23 Apr 2026 15:24:16 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=zagwdt" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[Introspective Diffusion Language Models]]></title><description><![CDATA[
<p>Article URL: <a href="https://introspective-diffusion.github.io/">https://introspective-diffusion.github.io/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47762641">https://news.ycombinator.com/item?id=47762641</a></p>
<p>Points: 281</p>
<p># Comments: 55</p>
]]></description><pubDate>Tue, 14 Apr 2026 07:57:33 +0000</pubDate><link>https://introspective-diffusion.github.io/</link><dc:creator>zagwdt</dc:creator><comments>https://news.ycombinator.com/item?id=47762641</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47762641</guid></item><item><title><![CDATA[EinsteinArena: Harnessing the collective intelligence of agents in the wild]]></title><description><![CDATA[
<p>Article URL: <a href="https://einsteinarena.com/">https://einsteinarena.com/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47758604">https://news.ycombinator.com/item?id=47758604</a></p>
<p>Points: 5</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 13 Apr 2026 22:14:02 +0000</pubDate><link>https://einsteinarena.com/</link><dc:creator>zagwdt</dc:creator><comments>https://news.ycombinator.com/item?id=47758604</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47758604</guid></item><item><title><![CDATA[RL Meets Adaptive Speculative Training]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.together.ai/blog/aurora">https://www.together.ai/blog/aurora</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47594795">https://news.ycombinator.com/item?id=47594795</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 31 Mar 2026 23:23:53 +0000</pubDate><link>https://www.together.ai/blog/aurora</link><dc:creator>zagwdt</dc:creator><comments>https://news.ycombinator.com/item?id=47594795</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47594795</guid></item><item><title><![CDATA[Weak models excel at long context tasks]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.together.ai/blog/plan-divide-conquer">https://www.together.ai/blog/plan-divide-conquer</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47549370">https://news.ycombinator.com/item?id=47549370</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 27 Mar 2026 22:44:22 +0000</pubDate><link>https://www.together.ai/blog/plan-divide-conquer</link><dc:creator>zagwdt</dc:creator><comments>https://news.ycombinator.com/item?id=47549370</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47549370</guid></item><item><title><![CDATA[TorchSpec: Speculative Decoding Training at Scale]]></title><description><![CDATA[
<p>Article URL: <a href="https://pytorch.org/blog/torchspec-speculative-decoding-training-at-scale/">https://pytorch.org/blog/torchspec-speculative-decoding-training-at-scale/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47480771">https://news.ycombinator.com/item?id=47480771</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Sun, 22 Mar 2026 18:51:41 +0000</pubDate><link>https://pytorch.org/blog/torchspec-speculative-decoding-training-at-scale/</link><dc:creator>zagwdt</dc:creator><comments>https://news.ycombinator.com/item?id=47480771</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47480771</guid></item><item><title><![CDATA[Flash Attention 4]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.together.ai/blog/flashattention-4">https://www.together.ai/blog/flashattention-4</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47262814">https://news.ycombinator.com/item?id=47262814</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 05 Mar 2026 15:33:37 +0000</pubDate><link>https://www.together.ai/blog/flashattention-4</link><dc:creator>zagwdt</dc:creator><comments>https://news.ycombinator.com/item?id=47262814</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47262814</guid></item><item><title><![CDATA[CoderForge-Preview: SOTA open dataset for training efficient coding agents]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.together.ai/blog/coderforge-preview">https://www.together.ai/blog/coderforge-preview</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47156778">https://news.ycombinator.com/item?id=47156778</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 25 Feb 2026 19:44:55 +0000</pubDate><link>https://www.together.ai/blog/coderforge-preview</link><dc:creator>zagwdt</dc:creator><comments>https://news.ycombinator.com/item?id=47156778</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47156778</guid></item><item><title><![CDATA[Two years of vector search at Notion: 10x scale, 1/10th cost]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.notion.com/blog/two-years-of-vector-search-at-notion">https://www.notion.com/blog/two-years-of-vector-search-at-notion</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47106254">https://news.ycombinator.com/item?id=47106254</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 21 Feb 2026 23:34:44 +0000</pubDate><link>https://www.notion.com/blog/two-years-of-vector-search-at-notion</link><dc:creator>zagwdt</dc:creator><comments>https://news.ycombinator.com/item?id=47106254</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47106254</guid></item><item><title><![CDATA[Consistency diffusion language models: Up to 14x faster, no quality loss]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.together.ai/blog/consistency-diffusion-language-models">https://www.together.ai/blog/consistency-diffusion-language-models</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47083648">https://news.ycombinator.com/item?id=47083648</a></p>
<p>Points: 219</p>
<p># Comments: 96</p>
]]></description><pubDate>Fri, 20 Feb 2026 04:15:58 +0000</pubDate><link>https://www.together.ai/blog/consistency-diffusion-language-models</link><dc:creator>zagwdt</dc:creator><comments>https://news.ycombinator.com/item?id=47083648</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47083648</guid></item></channel></rss>