<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: viktor_von</title><link>https://news.ycombinator.com/user?id=viktor_von</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 27 Apr 2026 10:05:08 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=viktor_von" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by viktor_von in "We replaced RAG with a virtual filesystem for our AI documentation assistant"]]></title><description><![CDATA[
<p>> yet somehow people married the acronym to one very particular implementation of the idea.<p>Likely due to the rise in popularity of semantic search via LLM embeddings, which for some reason became the main selling point for RAG. Meanwhile keyword search has existed for decades.</p>
]]></description><pubDate>Sat, 04 Apr 2026 03:12:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=47635295</link><dc:creator>viktor_von</dc:creator><comments>https://news.ycombinator.com/item?id=47635295</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47635295</guid></item><item><title><![CDATA[New comment by viktor_von in "The average American spent 2.5 months on their phone in 2024"]]></title><description><![CDATA[
<p>Ad-free and fast, Wikipedia is imo the best learning resource on-the-go, at home, or at work.</p>
]]></description><pubDate>Mon, 30 Dec 2024 09:57:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=42547985</link><dc:creator>viktor_von</dc:creator><comments>https://news.ycombinator.com/item?id=42547985</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42547985</guid></item><item><title><![CDATA[New comment by viktor_von in "Were RNNs all we needed?"]]></title><description><![CDATA[
<p>> The information cost of making the RNN state way bigger is high when done naively, but maybe someone can figure out a clever way to avoid storing full hidden states in memory during training or big improvements in hardware could make memory use less of a bottleneck.<p>Isn't this essentially what Mamba [1] does via its 'Hardware-aware Algorithm'?<p>[1] <a href="https://arxiv.org/pdf/2312.00752" rel="nofollow">https://arxiv.org/pdf/2312.00752</a></p>
]]></description><pubDate>Sat, 05 Oct 2024 20:08:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=41752529</link><dc:creator>viktor_von</dc:creator><comments>https://news.ycombinator.com/item?id=41752529</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41752529</guid></item><item><title><![CDATA[New comment by viktor_von in "Were RNNs all we needed?"]]></title><description><![CDATA[
<p>> It's astounding to me (and everyone else who's being honest) that LLMs can accomplish what they do when it's only linear "factors" (i.e. weights) that are all that's required to be adjusted during training, to achieve genuine reasoning.<p>When such basic perceptrons are scaled enormously, it becomes less surprising that they can achieve some level of 'genuine reasoning' (e.g., accurate next-word prediction), since the goal with such networks at the end of the day is just function approximation. What is more surprising to me is how we found ways to train such models i.e., advances in hardware accelerators, combined with massive data, which are factors just as significant in my opinion.</p>
]]></description><pubDate>Fri, 04 Oct 2024 04:20:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=41737788</link><dc:creator>viktor_von</dc:creator><comments>https://news.ycombinator.com/item?id=41737788</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41737788</guid></item><item><title><![CDATA[New comment by viktor_von in "Were RNNs all we needed?"]]></title><description><![CDATA[
<p>> I remember one of the initial transformer people saying in an interview that they didn't think this was the "one true architecture" but a lot of the performance came from people rallying around it and pushing in the one direction.<p>You may be referring to Aidan Gomez (CEO of Cohere and contributor to the transformer architecture) during his Machine Learning Street Talk podcast interview. I agree, if as much attention had been put towards the RNN during the initial transformer hype, we may have very well seen these advancements earlier.</p>
]]></description><pubDate>Fri, 04 Oct 2024 02:37:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=41737265</link><dc:creator>viktor_von</dc:creator><comments>https://news.ycombinator.com/item?id=41737265</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41737265</guid></item></channel></rss>