<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: Macuyiko</title><link>https://news.ycombinator.com/user?id=Macuyiko</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 22 Apr 2026 09:54:48 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=Macuyiko" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by Macuyiko in "How do I cancel my ChatGPT subscription?"]]></title><description><![CDATA[
<p>Things such as AirLLM, or good old llama.cpp.</p>
]]></description><pubDate>Sat, 28 Feb 2026 08:09:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47192077</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=47192077</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47192077</guid></item><item><title><![CDATA[New comment by Macuyiko in "Show HN: Steerling-8B, a language model that can explain any token it generates"]]></title><description><![CDATA[
<p>The input attribution part is interesting though, but I do wonder to which extent that is just assigning some sort of SHAP values to the input tokens, in which case it should be pretty portable to any kind of model.</p>
]]></description><pubDate>Tue, 24 Feb 2026 17:48:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47140143</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=47140143</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47140143</guid></item><item><title><![CDATA[New comment by Macuyiko in "Unsung heroes: Flickr's URLs scheme"]]></title><description><![CDATA[
<p>What I typically end up doing is just recalc the slug and see if it matches the provided one. If it doesn't redirect to the most up to date slug matching the id. Though who knows if those old SEO patterns still matter these days...</p>
]]></description><pubDate>Tue, 24 Feb 2026 17:46:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=47140111</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=47140111</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47140111</guid></item><item><title><![CDATA[New comment by Macuyiko in "Show HN: I spent 3 years reverse-engineering a 40 yo stock market sim from 1986"]]></title><description><![CDATA[
<p>My personal feel (completely subjective) is that during RLHF humans are incredibly sensitive to this pattern, especially when talking about personal or emotional issues. Any reply in the form of "it's not you, it's them" is such a dopamine hit that the LLMs started applying it for everything else.</p>
]]></description><pubDate>Sat, 14 Feb 2026 11:11:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=47013572</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=47013572</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47013572</guid></item><item><title><![CDATA[New comment by Macuyiko in "enclose.horse"]]></title><description><![CDATA[
<p>Oh I see what you mean now, indeed:<p><pre><code>    Score: 7
    ~~~~~~
    ~····~
    ~·~~·~
    .#..#.
    ......
    ..#...
    .#H#..
    ..#...
</code></pre>
However, I think that you do not need 'time' based variables in the form of<p><pre><code>    reachable(x,y,t) = reachable(nx,ny,t-1)
</code></pre>
Enforcing connectivity through single-commodity flows is IMO better to enforce flood fill (also introduces additional variables but is typically easier to solve with CP heuristics):<p><pre><code>    Score: 2
    ~~~~~~
    ~....~
    ~.~~.~
    ......
    ......
    ..##..
    .#H·#.
    ..##..
</code></pre>
Cool puzzle!</p>
]]></description><pubDate>Wed, 07 Jan 2026 08:16:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=46523872</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=46523872</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46523872</guid></item><item><title><![CDATA[New comment by Macuyiko in "enclose.horse"]]></title><description><![CDATA[
<p>Good point. I don't think the puzzles do this and if they would, I would run a pre-solve pass over the puzzle first to flood fill such horseless pockets up with water, no?</p>
]]></description><pubDate>Tue, 06 Jan 2026 17:23:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=46515350</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=46515350</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46515350</guid></item><item><title><![CDATA[New comment by Macuyiko in "enclose.horse"]]></title><description><![CDATA[
<p>Yes. CP SAT crunches through it in no time, but of course larger grids would quickly make it take much longer.<p>See<p><a href="https://gist.github.com/Macuyiko/86299dc120478fdff529cab386ffd14a" rel="nofollow">https://gist.github.com/Macuyiko/86299dc120478fdff529cab386f...</a></p>
]]></description><pubDate>Tue, 06 Jan 2026 13:02:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=46511712</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=46511712</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46511712</guid></item><item><title><![CDATA[New comment by Macuyiko in "Building more with GPT-5.1-Codex-Max"]]></title><description><![CDATA[
<p>Late, but reading all of the replies, and speaking from my own observation using Claude, Codex, as well as (non-CLI) Gemini, Kimi, Qwen, and Deepseek...<p>It's fun how we are so quick to assign meaning to the way these models act. This is of course due to training, RLHF, available tool calls, system prompt (all mostly invisible) and the way we prompt them.<p>I've been wondering about a new kind of benchmark how one would be able to extract these more intangible tendencies from models rather than well-controlled "how good at coding is it" style environments. This is mainly the reason why I pay less and less attention to benchmark scores.<p>For what it's worth: I still best converse with Claude when doing code. Its reasoning sounds like me, and it finds a good middle ground between conservative and crazy, being explorative and daring (even although it too often exclaims "I see the issue now!"). If Anthropic would lift the usage rates I would use it as my primary. The CLI tool is also better. E.g. Codex with 5.1 gets stuck in powershell scripts whilst Claude realizes it can use python to do heavy lifting, but I think that might be largely due to being mainly on Windows (still, Claude does work best, realizing quickly what environment it lives in rather than trying Unix commands or powershell invocations that don't work because my powershell is outdated).<p>Qwen is great in an IDE for quick auto-complete tasks, especially given that you can run it locally, but even the VSCode copilot is good enough for that. Kimi is promising for long running agentic tasks but that is something I've barely explored and just started playing with. Gemini is fantastic as a research assistant. Especially Gemini 3 Pro points out clear and to the point jargon without fear of the user being stupid, which the other commercial models are too often hesitant to do.<p>Again, it would be fun to have some unbiased method to uncover some of those underlying persona's.</p>
]]></description><pubDate>Thu, 20 Nov 2025 14:06:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=45992674</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=45992674</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45992674</guid></item><item><title><![CDATA[New comment by Macuyiko in "Show HN: ChartDB Agent – Cursor for DB schema design"]]></title><description><![CDATA[
<p>On the homepage it says "Sinmple" above "Export SQL", fyi</p>
]]></description><pubDate>Wed, 01 Oct 2025 18:03:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=45440961</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=45440961</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45440961</guid></item><item><title><![CDATA[New comment by Macuyiko in "Roman dodecahedron: 12-sided object has baffled archaeologists for centuries"]]></title><description><![CDATA[
<p>A coin measurer is still my goto explanation. Especially with most models having an inset for the coin to rest on / fit in. The hole itself is then just to quickly/easily get the coin out again with your finger.<p>With so many different coin sizes and types in the empire, I think this makes most sense.<p>Wikipedia also mentions this:<p>> Several dodecahedra were found in coin hoards, suggesting either that their owners considered them valuable objects, or that their use was connected with coins — as, for example, for easily checking coins fit a certain diameter and were not clipped.</p>
]]></description><pubDate>Thu, 17 Jul 2025 07:54:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=44590810</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=44590810</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44590810</guid></item><item><title><![CDATA[New comment by Macuyiko in "Solving LinkedIn Queens Using Haskell"]]></title><description><![CDATA[
<p>I've noticed that puzzles that can be solved with CP-SAT's presolver so that the SAT search does not even need to be invoked basically adhere to this (no backtracking, known rules), e.g.:<p><pre><code>    #Variables: 121 (91 primary variables)
      - 121 Booleans in [0,1]
    #kLinear1: 200 (#enforced: 200)
    #kLinear2: 1
    #kLinear3: 2
    #kLinearN: 30 (#terms: 355)

    Presolve summary:
      - 1 affine relations were detected.
      - rule 'affine: new relation' was applied 1 time.
      - rule 'at_most_one: empty or all false' was applied 148 times.
      - rule 'at_most_one: removed literals' was applied 148 times.
      - rule 'at_most_one: satisfied' was applied 36 times.
      - rule 'deductions: 200 stored' was applied 1 time.
      - rule 'exactly_one: removed literals' was applied 2 times.
      - rule 'exactly_one: satisfied' was applied 31 times.
      - rule 'linear: empty' was applied 1 time.
      - rule 'linear: fixed or dup variables' was applied 12 times.
      - rule 'linear: positive equal one' was applied 31 times.
      - rule 'linear: reduced variable domains' was applied 1 time.
      - rule 'linear: remapped using affine relations' was applied 4 times.
      - rule 'presolve: 120 unused variables removed.' was applied 1 time.
      - rule 'presolve: iteration' was applied 2 times.

    Presolved satisfaction model '': (model_fingerprint: 0xa5b85c5e198ed849)
    #Variables: 0 (0 primary variables)

    The solution hint is complete and is feasible.

    #1       0.00s main
      a    a    a    a    a    a    a    a    a    a   *A* 
      a    a    a    b    b    b    b   *B*   a    a    a  
      a    a   *C*   b    d    d    d    b    b    a    a  
      a    c    c    d    d   *E*   d    d    b    b    a  
      a    c    d   *D*   d    e    d    d    d    b    a  
      a    f    d    d    d    e    e    e    d   *G*   a  
      a   *F*   d    d    d    d    d    d    d    g    a  
      a    f    f    d    d    d    d    d   *H*   g    a  
     *I*   i    f    f    d    d    d    h    h    a    a  
      i    i    i    f   *J*   j    j    j    a    a    a  
      i    i    i    i    i    k   *K*   j    a    a    a
</code></pre>
Together with validating that there is only 1 solution you would probably be able to make the search for good boards a more guided than random creation.</p>
]]></description><pubDate>Tue, 24 Jun 2025 09:57:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=44364466</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=44364466</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44364466</guid></item><item><title><![CDATA[New comment by Macuyiko in "LLMs get lost in multi-turn conversation"]]></title><description><![CDATA[
<p>All of the above is true, but between solving quicker, and admitting we gave context:<p>I do agree with you that an LLM should not always start from scratch.<p>In a way it is like an animal which we have given the ultimate human instinct.<p>What has nature given us? Homo Erectus is 2 million years ago.<p>A weird world we live in.<p>What is context.</p>
]]></description><pubDate>Thu, 15 May 2025 17:48:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=43997490</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=43997490</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43997490</guid></item><item><title><![CDATA[New comment by Macuyiko in "LLMs get lost in multi-turn conversation"]]></title><description><![CDATA[
<p>Weirdly it has gotten so far that I have embedded this into my workflow and will often prompt:<p>> "Good work so far, now I want to take it to another step (somewhat related but feeling it too hard): <short description>. Do you think we can do it in this conversation or is it better to start fresh? If so, prepare an initial prompt for your next fresh instantiation."<p>Sometimes the model says that it might be better to start fresh, and prepares a good summary prompt (including a final 'see you later'), whereas in other cases it assures me it can continue.<p>I have a lot of notebooks with "initial prompts to explore forward". But given the sycophancy going on as well as one-step RL (sigh) post-training [1], it indeed seems AI platforms would like to keep the conversation going.<p>[1] RL in post-training has little to do with real RL and just uses one shot preference mechanisms with an RL inspired training loop. There is very little work in terms of long-term preferences slash conversations, as that would increase requirements exponentially.</p>
]]></description><pubDate>Thu, 15 May 2025 17:44:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=43997455</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=43997455</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43997455</guid></item><item><title><![CDATA[New comment by Macuyiko in "Not a three-year-old chimney sweep (2022)"]]></title><description><![CDATA[
<p>A bit of a rant, but this is the kind of fact checking I wish the media and all our EU "trusted sources" would have jumped on instead of going for the most trivial and idiotic cases only a toddler (or a journalist) would get stumped by. (Example: recent posts on Tiktok 'claiming to be images from Pakistan but taken from Battlefield 3...' again. Who is impressed or even surprised by this kind of investigation?)<p>Much more interesting, but also with more effort required, so of course it never happens.<p>It would have a more beneficial societal effect, because it is this kind of article, neutrally written, deep investigation, that truly would make people capable to self-discover "maybe I should question a bit more things".</p>
]]></description><pubDate>Sun, 11 May 2025 07:38:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=43952099</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=43952099</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43952099</guid></item><item><title><![CDATA[New comment by Macuyiko in "World Emulation via Neural Network"]]></title><description><![CDATA[
<p>The model seems to be viewable here:<p><a href="https://netron.app/?url=https://madebyoll.in/posts/world_emulation_via_dnn/demo/model.onnx" rel="nofollow">https://netron.app/?url=https://madebyoll.in/posts/world_emu...</a></p>
]]></description><pubDate>Sat, 26 Apr 2025 09:56:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=43802248</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=43802248</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43802248</guid></item><item><title><![CDATA[New comment by Macuyiko in "Is stuff online worth saving?"]]></title><description><![CDATA[
<p>From an age perspective (but the crowd here will not like that): before I trusted myself I could always find it back so I don't need to save it. Now I can't anymore, but I don't care so much.</p>
]]></description><pubDate>Sat, 21 Dec 2024 20:45:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=42482158</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=42482158</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42482158</guid></item><item><title><![CDATA[New comment by Macuyiko in "OpenAI O3 breakthrough high score on ARC-AGI-PUB"]]></title><description><![CDATA[
<p>I am not so sure, but indeed it is perhaps also a sad realization.<p>You compare this to "a human" but also admit there is a high variation.<p>And, I would say there are a lot humans being paid ~=$3400 per month. Not for a single task, true, but for honestly for no value creating task at all. Just for their time.<p>So what about we think in terms of output rather than time?</p>
]]></description><pubDate>Sat, 21 Dec 2024 20:40:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=42482133</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=42482133</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42482133</guid></item><item><title><![CDATA[New comment by Macuyiko in "Neuroevolution of augmenting topologies (NEAT algorithm)"]]></title><description><![CDATA[
<p>Some more interesting approaches in the same space:<p>- <a href="https://github.com/openai/evolution-strategies-starter">https://github.com/openai/evolution-strategies-starter</a><p>- <a href="https://cloud.google.com/blog/topics/developers-practitioners/evojax-bringing-power-neuroevolution-solve-your-problems" rel="nofollow">https://cloud.google.com/blog/topics/developers-practitioner...</a><p>And perhaps most close:<p>- <a href="https://weightagnostic.github.io/" rel="nofollow">https://weightagnostic.github.io/</a><p>Which also showed that you can make NNs weight agnostic and just let the architecture evolve using a GA.<p>Even though these approaches are cool and NEAT even is somewhat easier to implement than getting started with RL (at least that is what based on so many AI Youtubers starting with NEAT first) they didn't ever seem to fully take off. Although knowing about metaheuristics is still a good tool to know IMO.</p>
]]></description><pubDate>Wed, 04 Dec 2024 18:42:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=42320547</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=42320547</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42320547</guid></item><item><title><![CDATA[New comment by Macuyiko in "The first release candidate of FreeCAD 1.0 is out"]]></title><description><![CDATA[
<p>A few weeks ago I was planning to design a model I could send to a local 3d printer to replace a broken piece in the house for which I knew it would be impossible to find something that would fit exactly.<p>I looked around through a couple of open source/free offerings and all found them frustrating. Either the focus on easy of use was too limiting, the focus was too much on blob, clay-like modeling rather than strong parametric models (many online tools), or they were too pushy to make you pay, or the UI was not intuitive (FreeCAD).<p>OpenSCAD was the one which allowed me to get the model done, and I loved the code-first, parametric-first approach and way of thinking. But that said I also found POV-Ray enjoyable to play around with around the 2000s. Build123D looks interesting as well, thanks for recommending that.</p>
]]></description><pubDate>Thu, 12 Sep 2024 07:07:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=41518278</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=41518278</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41518278</guid></item><item><title><![CDATA[New comment by Macuyiko in "The History of Machine Learning in Trackmania"]]></title><description><![CDATA[
<p>I follow RL from the sides (I have dabbled with it myself), and have seen some of the cool videos the article also lists. I think one of the key points (and a bit of a personal nitpick) the article makes is this:<p>> Thus far, every attempt at training a Trackmania-playing program has trained the program on one map at a time. As a result, no matter how well the network did on one track, it would have to be retrained - probably significantly retrained<p>This is a crucial aspect when talking about RL. Most of the Trackmania AI attempts focuses on a track at a time, which is not really a problem since they want to, given an individual track, outperform the best human racers.<p>However, it is this nuance that a lot of more business oriented users don't get when being sold on some fancy new RL project. In the real world (think self-driving cars), we typically want agents to be way more able to generalize.<p>Most of the RL techniques we have do rather well in these kinds of constrained environments (in a sense they eventually start overfitting on the given environment), but making them behave well in more varied environments is way harder. A lot of beginner RL tutorials also fail to make this very explicit, and will e.g. show how to train an agent to find the exit in a maze without ever trying it on a newly generated maze :).</p>
]]></description><pubDate>Thu, 04 Jul 2024 07:48:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=40873136</link><dc:creator>Macuyiko</dc:creator><comments>https://news.ycombinator.com/item?id=40873136</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40873136</guid></item></channel></rss>