<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: krychu</title><link>https://news.ycombinator.com/user?id=krychu</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 29 Apr 2026 18:56:35 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=krychu" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by krychu in "BitNet: Inference framework for 1-bit LLMs"]]></title><description><![CDATA[
<p>Unfortunately reasoning  ability depends on (or is enabled by) information intake during training. A model will know better what to search for and how to interpret it if the information was part of the training. So there is a trade off. Still I think the question is a practical one. Perhaps there are ideas to focus training on a) reasoning / conceptual modeling and b) reliance on external memory (search etc.) rather than internal memorization.</p>
]]></description><pubDate>Thu, 12 Mar 2026 10:12:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47348597</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=47348597</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47348597</guid></item><item><title><![CDATA[New comment by krychu in "Hashcards: A plain-text spaced repetition system"]]></title><description><![CDATA[
<p>Self-plug. For anyone working in the terminal: <a href="https://github.com/krychu/lrn" rel="nofollow">https://github.com/krychu/lrn</a>.<p>A very simple cli tool, consuming basic txt format. You can use it in a second window while waiting for your compilation to finish.<p>Recently I’ve been also experimenting with defining QA pairs in my note files (in a special section). I then use a custom function in emacs to extract these pairs and push to a file as well as Anki.</p>
]]></description><pubDate>Sun, 14 Dec 2025 21:31:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=46267156</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=46267156</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46267156</guid></item><item><title><![CDATA[New comment by krychu in "Building more with GPT-5.1-Codex-Max"]]></title><description><![CDATA[
<p>It’d be probably useful to include this very comment in your system prompt or a separate file which you ask the coding agent to read at the beginning of each session.</p>
]]></description><pubDate>Sun, 23 Nov 2025 10:58:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=46022463</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=46022463</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46022463</guid></item><item><title><![CDATA[New comment by krychu in "'Attention is all you need' coauthor says he's 'sick' of transformers"]]></title><description><![CDATA[
<p>BDH</p>
]]></description><pubDate>Fri, 24 Oct 2025 21:07:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=45699103</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=45699103</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45699103</guid></item><item><title><![CDATA[New comment by krychu in "Vibe engineering"]]></title><description><![CDATA[
<p>It’s great to read in the comments about experiences of others with vibe coding. But I also feel like lots of opinions are not coming from actual experience, or “serious” attempts at vibe coding, and more from theoretical deliberations. I might be wrong.<p>Here are some of my own high-level experiences / thoughts:<p>- Perhaps contrary to popular belief I think vibe coding will bring the best software / system architects. This is due to massively shortened feedback loop between architectural idea and seeing it in action, easiness with which it can be changed, and the ability to discuss it at any moment.<p>- We’re not really coding anymore. This is a new role, not a role of a senior dev reviewing PRs of junior devs. Devs are just best suited (currently) to take on this new role. I came to realization that if you’re reviewing all generated code in detail you’re doing it wrong. You just shifted bottleneck by one step. You’re still coding. You should skim if the code is in line with your high-level expectation and then make LLM maintain an architecture doc and other docs that describe what and how you’re building (this is the info you should know in detail). You can do audits with another LLM whether the implementation is 100% reflecting the docs, you can chat with LLM about implementation at any moment if you ever need. But you should not know the implementation the way you know it today. The implementation became the implementation detail. The whole challenge is to let go of the old and embrace and search for efficiency in the new setup.<p>- Connected to the above: reading through LLM outputs is a massive fatigue. You are exhausted after the day, because you read hundreds of pages. This is a challenge to fight. You cannot unlock full potential here if you aim at reading and reviewing everything.<p>- Vibe coding makes you work on the problem level much more. I never liked the phrase “ideas are cheap”. And now finally I think the tides will turn, ideas are and will be king.<p>- Devil is in the detail, 100%. People with ability to see connections, distill key insights, communicate and articulate clearly, think clearly, are the ones to benefit.<p>Hope this is helpful for others.</p>
]]></description><pubDate>Wed, 08 Oct 2025 15:06:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=45517013</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=45517013</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45517013</guid></item><item><title><![CDATA[New comment by krychu in "Less is more: Recursive reasoning with tiny networks"]]></title><description><![CDATA[
<p>Thanks, appreciated</p>
]]></description><pubDate>Wed, 08 Oct 2025 11:35:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=45514957</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=45514957</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45514957</guid></item><item><title><![CDATA[New comment by krychu in "Less is more: Recursive reasoning with tiny networks"]]></title><description><![CDATA[
<p>I implemented HRM for educational purposes and got good results for path finding. But then I started to do ablation experiments and came to the same conclusions as the ARC-AGI team (the HRM architecture itself didn’t play a big role): <a href="https://github.com/krychu/hrm" rel="nofollow">https://github.com/krychu/hrm</a><p>This was a bit unfortunate. I think there is something in the idea of latent space reasoning.</p>
]]></description><pubDate>Tue, 07 Oct 2025 21:11:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=45508933</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=45508933</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45508933</guid></item><item><title><![CDATA[Show HN: Implementation and ablation of the Hierarchical Reasoning Model (HRM)]]></title><description><![CDATA[
<p>I implemented the Hierarchical Reasoning Model (HRM) in PyTorch and applied it to a simple pathfinding task.<p>HRM is inspired by multi-timescale processing in the brain: a slower H module for abstract planning and a faster L module for low-level computation. Both modules are based on self-attention and attempt to model reasoning in latent space.<p>The repo includes: a) the implementation, b) demo that generates animated GIFs where you can see the model refine its solution step by step, c) results of a small ablation study on what drives performance.<p>The biggest driver (both accuracy and refinement ability) is *training with more segments* (outer-loop refinement), not the H/L two-timescale split. (This lines up with the ARC Prize team's analysis). This is of course a limited study on a relatively simple task, but I thought the results might be interesting to others.<p>Repo: <a href="https://github.com/krychu/hrm" rel="nofollow">https://github.com/krychu/hrm</a><p>Curious to hear thoughts - iterative refinement isn't new, but I wonder if the "loop-in-a-loop" forward pass, or varied frequencies, might hint at a useful direction for reasoning in latent space (?)</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45198143">https://news.ycombinator.com/item?id=45198143</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 10 Sep 2025 14:21:53 +0000</pubDate><link>https://github.com/krychu/hrm</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=45198143</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45198143</guid></item><item><title><![CDATA[2D graphics on top of sokol_gfx.h]]></title><description><![CDATA[
<p>Article URL: <a href="https://github.com/edubart/sokol_gp">https://github.com/edubart/sokol_gp</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=37799957">https://news.ycombinator.com/item?id=37799957</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 07 Oct 2023 08:08:36 +0000</pubDate><link>https://github.com/edubart/sokol_gp</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=37799957</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37799957</guid></item><item><title><![CDATA[New comment by krychu in "So You Think You Know C? (2020) [pdf]"]]></title><description><![CDATA[
<p>Your code dictates the compiler's behavior. If you grant the compiler flexibility, it's unreasonable to complain about unexpected results.</p>
]]></description><pubDate>Sun, 17 Sep 2023 19:52:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=37548708</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=37548708</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37548708</guid></item><item><title><![CDATA[New comment by krychu in "So You Think You Know C? (2020) [pdf]"]]></title><description><![CDATA[
<p>> The problem with c is that you must have a comprehensive dictionary in your brain with tons of corner cases to know what is or is not undefined in any given compiler setting.<p>The cases of undefined behavior in the C standard are independent of compiler settings or options.<p>> If C could have a consistent set of rules …<p>The C language has a well-defined standard, but the presence of undefined behavior is a deliberate aspect of that standard.</p>
]]></description><pubDate>Sun, 17 Sep 2023 17:04:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=37547201</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=37547201</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37547201</guid></item><item><title><![CDATA[New comment by krychu in "Guide to running Llama 2 locally"]]></title><description><![CDATA[
<p>Self-plug. Here’s a fork of the original llama 2 code adapted to run on the CPU or MPS (M1/M2 GPU) if available:<p><a href="https://github.com/krychu/llama">https://github.com/krychu/llama</a><p>It runs with the original weights, and gets you to ~4 tokens/sec on MacBook Pro M1 with the 7B model.</p>
]]></description><pubDate>Tue, 25 Jul 2023 22:48:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=36870809</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=36870809</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36870809</guid></item><item><title><![CDATA[New comment by krychu in "Llama 2"]]></title><description><![CDATA[
<p>Version that runs on the CPU: <a href="https://github.com/krychu/llama">https://github.com/krychu/llama</a><p>I get 1 word per ~1.5 secs on a Mac Book Pro M1.</p>
]]></description><pubDate>Wed, 19 Jul 2023 17:34:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=36790171</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=36790171</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36790171</guid></item><item><title><![CDATA[New comment by krychu in "Langchain Is Pointless"]]></title><description><![CDATA[
<p>Chain of thought: <a href="https://arxiv.org/abs/2201.11903" rel="nofollow noreferrer">https://arxiv.org/abs/2201.11903</a>
ReAct: <a href="https://arxiv.org/abs/2210.03629" rel="nofollow noreferrer">https://arxiv.org/abs/2210.03629</a>
Reflexion: <a href="https://arxiv.org/abs/2303.11366" rel="nofollow noreferrer">https://arxiv.org/abs/2303.11366</a>
Tree of thoughts: <a href="https://arxiv.org/abs/2305.10601" rel="nofollow noreferrer">https://arxiv.org/abs/2305.10601</a><p>Good video on "Tree of thoughts" which also reviews / puts it in the context of other methods: <a href="https://www.youtube.com/watch?v=ut5kp56wW_4">https://www.youtube.com/watch?v=ut5kp56wW_4</a><p>Completion vs conversational interface is something you can read about in the OpenAI API documentation.<p>For the remaining things I don't have single specific pointer at hand.</p>
]]></description><pubDate>Mon, 10 Jul 2023 07:10:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=36663017</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=36663017</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36663017</guid></item><item><title><![CDATA[New comment by krychu in "Langchain Is Pointless"]]></title><description><![CDATA[
<p>Using an LLM framework at this moment doesn’t make sense and can be damaging, in my humble opinion. Ways to extract value from LLMs are in early exploration stage. Look at research in prompting: chain of thought, react, reflection, tree of thoughts, zero vs few hot etc. Then completion vs conversational interfacing. Then memory management via vector databases and prompt expansion vs compression vs progressive summarization etc. All these are fairly recent developments. They are not abstractions worth cementing,  this is <i>search</i> and <i>creative</i> phase. LLMs threw everything in the air, but the dust is far from settling. I think it’s important to recognize the phase we’re in and pick your weapon accordingly. You have to stay nimble and light, ready to experiment with a new idea that will come out next week. You should be hacking these things together by yourself. If you pick a framework at this stage know that the framework will have to pay the price of trying to cement things in the times of storm. And you’ll be a few steps behind. Of course this is my personal take.</p>
]]></description><pubDate>Sat, 08 Jul 2023 21:41:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=36649051</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=36649051</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36649051</guid></item><item><title><![CDATA[New comment by krychu in "An open source web-based flashcard studying system"]]></title><description><![CDATA[
<p>Unfortunately not, “lrn” uses a very simple file format where each entry consists of three lines: question, answer, empty line. But I think it’d be a good idea to look into supporting Anki decks. I should do it some time. It’d be probably limited to decks that use text only.</p>
]]></description><pubDate>Sat, 24 Jun 2023 16:58:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=36460423</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=36460423</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36460423</guid></item><item><title><![CDATA[New comment by krychu in "An open source web-based flashcard studying system"]]></title><description><![CDATA[
<p>I can relate to this. I had fond memory of SuperMemo on DOS and wrote a little terminal tool “lrn” inspired by it: <a href="https://github.com/krychu/lrn">https://github.com/krychu/lrn</a><p>I acknowledge it’s simple, runs on terminal only and lacks bells and whistles. But this is also probably why I use it so much.</p>
]]></description><pubDate>Sat, 24 Jun 2023 08:22:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=36456678</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=36456678</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36456678</guid></item><item><title><![CDATA[New comment by krychu in "Arwes: Futuristic Sci-Fi UI Web Framework"]]></title><description><![CDATA[
<p>Very impressive, and works well on mobile (iOS). Congrats and thanks for sharing.</p>
]]></description><pubDate>Fri, 23 Jun 2023 22:08:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=36452818</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=36452818</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36452818</guid></item><item><title><![CDATA[New comment by krychu in "Sam Altman goes before US Congress to propose licenses for building AI"]]></title><description><![CDATA[
<p>From what I understand OpenAI has been moving away from “open” with various decisions over the time. Proposing that only selected folks can build AI seems like the antithesis of openness?</p>
]]></description><pubDate>Tue, 16 May 2023 20:41:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=35967468</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=35967468</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=35967468</guid></item><item><title><![CDATA[New comment by krychu in "GPT makes learning fun again"]]></title><description><![CDATA[
<p>I've been using GPT to have (insightful) educational conversations about Quake 1 source code: <a href="https://twitter.com/krychusamp/status/1649048047996014595" rel="nofollow">https://twitter.com/krychusamp/status/1649048047996014595</a><p>I always finish up by asking GPT to test my knowledge with a single-choice questionnaire. What I've observed is that the retention of the material is higher compared to "traditional" techniques. Perhaps the conversation style is more immersive, or perhaps focusing on specific knowledge gaps makes for accelerated / personalised learning.<p>There is of course the problem of accuracy, but I feel like it's often over-stated. Even if GPT is not correct at times, it often uncovers concepts and relations that paint a better overall picture for me, and lead me to better questions and follow up actions.</p>
]]></description><pubDate>Tue, 02 May 2023 12:24:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=35786349</link><dc:creator>krychu</dc:creator><comments>https://news.ycombinator.com/item?id=35786349</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=35786349</guid></item></channel></rss>