<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: attentionmech</title><link>https://news.ycombinator.com/user?id=attentionmech</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 06 Apr 2026 11:07:32 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=attentionmech" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by attentionmech in "The OBS Project is threatening Fedora Linux with legal action"]]></title><description><![CDATA[
<p>why don't they just block the obs project and let users install it in unofficial manner while removing themselves as middleman? I mean, they have certain let's say guidelines but why go about enforcing them in this weird manner.</p>
]]></description><pubDate>Fri, 14 Feb 2025 02:02:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=43043894</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=43043894</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43043894</guid></item><item><title><![CDATA[New comment by attentionmech in "Andrej Karpathy: Deep Dive into LLMs Like ChatGPT [video]"]]></title><description><![CDATA[
<p>saw that video just now, thanks for this.</p>
]]></description><pubDate>Sat, 08 Feb 2025 15:24:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=42983527</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42983527</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42983527</guid></item><item><title><![CDATA[New comment by attentionmech in "Andrej Karpathy: Deep Dive into LLMs Like ChatGPT [video]"]]></title><description><![CDATA[
<p>he has earned it haha.</p>
]]></description><pubDate>Thu, 06 Feb 2025 19:15:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=42965511</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42965511</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42965511</guid></item><item><title><![CDATA[New comment by attentionmech in "Andrej Karpathy: Deep Dive into LLMs Like ChatGPT [video]"]]></title><description><![CDATA[
<p>Will checkout jeremy's lectures. I actually use his fastbook notebooks a lot to self-study.<p>Karpathy's style, for me is more like at the right abstraction to bring out curiosity in me towards the subject. After watching his lectures, i go on to more materials generally, and never really stop there.</p>
]]></description><pubDate>Thu, 06 Feb 2025 10:21:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=42960941</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42960941</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42960941</guid></item><item><title><![CDATA[New comment by attentionmech in "Emerging reasoning with reinforcement learning"]]></title><description><![CDATA[
<p>agreed. that's err on my part to mention it like that. more evidence suggest that they were working on similar stuff but now the cat is out of the bag and open source got a win.</p>
]]></description><pubDate>Thu, 30 Jan 2025 10:59:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=42876690</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42876690</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42876690</guid></item><item><title><![CDATA[New comment by attentionmech in "Emerging reasoning with reinforcement learning"]]></title><description><![CDATA[
<p>people already did: <a href="https://x.com/karpathy/status/1884678601704169965" rel="nofollow">https://x.com/karpathy/status/1884678601704169965</a></p>
]]></description><pubDate>Thu, 30 Jan 2025 10:57:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=42876679</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42876679</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42876679</guid></item><item><title><![CDATA[New comment by attentionmech in "SmolGPT: A minimal PyTorch implementation for training a small LLM from scratch"]]></title><description><![CDATA[
<p>This is cool, and timely (I wanted a neat repo like that).<p>I have also been working from last 2 weeks on a gpt implementation in C. Eventually it turned out to be really slow (without CUDA). But it taught me how much memory management and data management there is when implementing these systems. You are running like a loop billions of times so you need to preallocate the computational graph and stuff. If anyone wanna check out it's ~1500 LOC single file:<p><a href="https://github.com/attentionmech/gpt.c/blob/main/gpt.c">https://github.com/attentionmech/gpt.c/blob/main/gpt.c</a></p>
]]></description><pubDate>Thu, 30 Jan 2025 10:30:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=42876555</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42876555</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42876555</guid></item><item><title><![CDATA[New comment by attentionmech in "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via RL"]]></title><description><![CDATA[
<p>I love this paradigm of reasoning by one model and actual work by another. This opens up avenues of specialization and then eventually smaller plays working on more niche things.</p>
]]></description><pubDate>Mon, 27 Jan 2025 08:13:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=42838646</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42838646</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42838646</guid></item><item><title><![CDATA[New comment by attentionmech in "Emerging reasoning with reinforcement learning"]]></title><description><![CDATA[
<p>I found the following thread more insightful than my original comment (wish I could edit that one). A research explains why RL didn't work before this: <a href="https://x.com/its_dibya/status/1883595705736163727" rel="nofollow">https://x.com/its_dibya/status/1883595705736163727</a></p>
]]></description><pubDate>Mon, 27 Jan 2025 04:14:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=42837349</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42837349</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42837349</guid></item><item><title><![CDATA[New comment by attentionmech in "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via RL"]]></title><description><![CDATA[
<p>people are doing all sort of experiments and reproducing the "emergence"(sorry it's not the right word) of backtracking; it's all so fun to watch.</p>
]]></description><pubDate>Sun, 26 Jan 2025 18:50:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=42832734</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42832734</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42832734</guid></item><item><title><![CDATA[New comment by attentionmech in "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via RL"]]></title><description><![CDATA[
<p>Yea, they might be scaling is harder or may be more tricks up their sleeves when it comes to serving the model.</p>
]]></description><pubDate>Sun, 26 Jan 2025 18:26:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=42832448</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42832448</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42832448</guid></item><item><title><![CDATA[New comment by attentionmech in "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via RL"]]></title><description><![CDATA[
<p>Plus, the speed at which it replies is amazing too. Claude/Chatgpt now seem like inefficient inference engines compared to it.</p>
]]></description><pubDate>Sun, 26 Jan 2025 12:17:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=42829621</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42829621</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42829621</guid></item><item><title><![CDATA[New comment by attentionmech in "Emerging Reasoning with Reinforcement Learning"]]></title><description><![CDATA[
<p>Do you think this feature i.e. 'finding smaller chunks easier to solve' comes out from the dataset these are trained on or is it more related to architecture components?</p>
]]></description><pubDate>Sun, 26 Jan 2025 12:10:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=42829594</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42829594</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42829594</guid></item><item><title><![CDATA[New comment by attentionmech in "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via RL"]]></title><description><![CDATA[
<p>Most people I talked with don't grasp how big of an event this is. I consider is almost as similar to as what early version of linux did to OS ecosystem.</p>
]]></description><pubDate>Sun, 26 Jan 2025 12:08:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=42829580</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42829580</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42829580</guid></item><item><title><![CDATA[New comment by attentionmech in "Emerging reasoning with reinforcement learning"]]></title><description><![CDATA[
<p>If you check failure section of their paper, they also tried other methods like MCTS and PRM which is what other labs have been obsessing about but couldn't move on from (that includes bigshots). Only team which I am aware which tried verifiable rewards is tulu but they didn't scaled it up and just left it there.<p>This sort of thing imo is similar to what openAI did with transformer architecture i.e. google invented it but couldn't scale it in the right direction and deepmind got busy with atari games. They had all the pieces still openai could do it. It seems to be it comes down to research leadership in what methods to choose to invest in. But yeah, the budgets big labs have, they can easily try 10 different techniques and brute force it all but seems like they are too opinionated in methods and less urgent on outcomes.<p>[paper] <a href="https://arxiv.org/pdf/2501.12948" rel="nofollow">https://arxiv.org/pdf/2501.12948</a>
[tulu] <a href="https://x.com/hamishivi/status/1881394117810500004" rel="nofollow">https://x.com/hamishivi/status/1881394117810500004</a></p>
]]></description><pubDate>Sun, 26 Jan 2025 12:05:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=42829571</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42829571</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42829571</guid></item><item><title><![CDATA[New comment by attentionmech in "Emerging reasoning with reinforcement learning"]]></title><description><![CDATA[
<p>That's nice explanation. Is there any insights so far in the field about why chain of thought improves the capability of a model? Does it like provide model with more working memory or something in the context itself?</p>
]]></description><pubDate>Sun, 26 Jan 2025 08:18:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=42828655</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42828655</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42828655</guid></item><item><title><![CDATA[New comment by attentionmech in "Emerging reasoning with reinforcement learning"]]></title><description><![CDATA[
<p>the tulu team saw it. but, yes nobody like scaled it to the extent deepseek did. I am surprised that the faang labs which have the best of the best didn't see this.</p>
]]></description><pubDate>Sun, 26 Jan 2025 08:13:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=42828631</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42828631</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42828631</guid></item><item><title><![CDATA[transformer-scope: script for visualizing activations]]></title><description><![CDATA[
<p>Article URL: <a href="https://github.com/attentionmech/transformer-scope">https://github.com/attentionmech/transformer-scope</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42662758">https://news.ycombinator.com/item?id=42662758</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 11 Jan 2025 02:31:17 +0000</pubDate><link>https://github.com/attentionmech/transformer-scope</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42662758</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42662758</guid></item><item><title><![CDATA[New comment by attentionmech in "Stimulation Clicker"]]></title><description><![CDATA[
<p>idk what i am doing but i am hooked on it. it's like as if it's directly interacting with dopamine of my brain.</p>
]]></description><pubDate>Mon, 06 Jan 2025 19:09:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=42614180</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42614180</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42614180</guid></item><item><title><![CDATA[New comment by attentionmech in "I am rich and have no idea what to do"]]></title><description><![CDATA[
<p>It's commutative. Happiness also doesn't buy money.</p>
]]></description><pubDate>Fri, 03 Jan 2025 00:40:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=42580779</link><dc:creator>attentionmech</dc:creator><comments>https://news.ycombinator.com/item?id=42580779</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42580779</guid></item></channel></rss>