<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: dongecko</title><link>https://news.ycombinator.com/user?id=dongecko</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 21 Apr 2026 16:55:41 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=dongecko" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by dongecko in "Show HN: How I topped the HuggingFace open LLM leaderboard on two gaming GPUs"]]></title><description><![CDATA[
<p>What a great read!
You got me at the base64 oddity. I also stumbled over this, while trying to dodge some LLM limitation. (was trying to generate images in a time before multimodal was a thing. it only worked to a degree).</p>
]]></description><pubDate>Tue, 10 Mar 2026 17:31:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=47326309</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=47326309</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47326309</guid></item><item><title><![CDATA[New comment by dongecko in "German police probe student poster slur against Merz"]]></title><description><![CDATA[
<p>Naaa, you have to think of the poster in options. Like, "would you want to have some extra publicity with your poster for an extra 100 bucks?"</p>
]]></description><pubDate>Sat, 07 Mar 2026 08:03:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=47285544</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=47285544</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47285544</guid></item><item><title><![CDATA[New comment by dongecko in "The Cray-1 Computer System (1977) [pdf]"]]></title><description><![CDATA[
<p>I never realized; thanks for pointing that out.  Sneakers is one of my favorite movies.</p>
]]></description><pubDate>Tue, 13 Jan 2026 20:29:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=46607423</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=46607423</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46607423</guid></item><item><title><![CDATA[New comment by dongecko in "Ask HN: What is the international distribution/statistics of HN visitors?"]]></title><description><![CDATA[
<p>A while back, someone made <a href="https://meet.hn" rel="nofollow">https://meet.hn</a>. It's a map where people from hackernews could locate themselves.<p>Of course this is by far not comprehensive, but it gives you an idea of where the people are from.</p>
]]></description><pubDate>Thu, 25 Dec 2025 20:15:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=46386770</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=46386770</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46386770</guid></item><item><title><![CDATA[New comment by dongecko in "Don't avoid workplace politics"]]></title><description><![CDATA[
<p>Exactly my point of view. For the most part I do not root for my preferred technology, but rather try to inform my powers about the caveats I see. This way at least the right aspects to check have a chance to enter the debates above my payroll.</p>
]]></description><pubDate>Thu, 02 Oct 2025 14:57:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=45450512</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=45450512</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45450512</guid></item><item><title><![CDATA[New comment by dongecko in "Why Do AI Chatbots Have Such a Hard Time Admitting 'I Don't Know'?"]]></title><description><![CDATA[
<p>Yes, they can. Meta described a clever way in their paper on training Llama3 [1] (in section about factuality).
The idea is to sample several answers to a question that you know the answer to. Let an LLM decide if the given answers are different from your known truth. If so, you found a question, that you can train your LLM in the next post-training round to answer with "I don't know". 
Do that a couple hundred of times and your LLM will identify neurons that indicate doubt and from here on have the ability to answer with "I don't know".<p>[Edit] The article here also mentions a paper [2] that comes up with the idea of an uncertainty token. So here the incorporation of uncertainty is already baked in at pre-training.[/Edit]<p>[1] <a href="https://arxiv.org/pdf/2407.21783" rel="nofollow">https://arxiv.org/pdf/2407.21783</a>
[2] <a href="https://openreview.net/pdf?id=Wc0vlQuoLb" rel="nofollow">https://openreview.net/pdf?id=Wc0vlQuoLb</a></p>
]]></description><pubDate>Tue, 11 Feb 2025 15:56:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=43014325</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=43014325</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43014325</guid></item><item><title><![CDATA[New comment by dongecko in "Ask HN: Examples of agentic LLM systems in production?"]]></title><description><![CDATA[
<p>We have a couple of systems at work that incorporate LLMs. There are a bunch of RAG chatbots for large documentation collections and a bunch of extract-info-from-email bots. I would none of these call an agent.
The one thing that comes close to an agent is a very bot that can query a few different SQL and API data sources. Given a users text query, it decides on its own which tool(s) to use. It can also retry, or re-formulate its task. The agentic parts are mainly done in LangGrah.</p>
]]></description><pubDate>Mon, 16 Dec 2024 17:41:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=42433282</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=42433282</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42433282</guid></item><item><title><![CDATA[New comment by dongecko in "The Impact of Jungle Music in 90s Video Game Development"]]></title><description><![CDATA[
<p>Came here for exactly that. Thank you.</p>
]]></description><pubDate>Sun, 17 Nov 2024 12:30:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=42163848</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=42163848</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42163848</guid></item><item><title><![CDATA[New comment by dongecko in "Nobel Prize in Physics awarded to John Hopfield and Geoffrey Hinton [pdf]"]]></title><description><![CDATA[
<p>Bolzman machines were there in the very early days of deep learning. It was a clever hack to train deep nets layer wise and work with limited ressources.<p>Each layer was trained similar to the encoder part of an autoencoder. This way the layerwise transformations were not random, but roughly kept some of the original datas properties. Up to here training was done without the use of labelled data. After this training stage was done, you had a very nice initialization for your network and could train it end to end according to your task and target label.<p>If I recall correctly, the neural layers output was probabilistic. Because of that you couldn't simply use back propagation to learn the weights. Maybe this is the connection to John Hopkins work. But here my memory is a bit fuzzy.</p>
]]></description><pubDate>Tue, 08 Oct 2024 19:13:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=41780724</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=41780724</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41780724</guid></item><item><title><![CDATA[New comment by dongecko in "Nobel Prize in Physics awarded to John Hopfield and Geoffrey Hinton [pdf]"]]></title><description><![CDATA[
<p>This comment caught me off guard!
I literally blew my coffee all over the table.</p>
]]></description><pubDate>Tue, 08 Oct 2024 11:54:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=41776326</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=41776326</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41776326</guid></item><item><title><![CDATA[New comment by dongecko in "Were RNNs all we needed?"]]></title><description><![CDATA[
<p>I second that thought. There is a pretty well cited paper from the late eighties called "Multilayer Feedforward Networks are Universal Approximators". It shows that a feedforward network with a single hidden layer containing a finite number of neurons can approximate any continuous function. For non continous function additional layers are needed.</p>
]]></description><pubDate>Fri, 04 Oct 2024 08:44:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=41739235</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=41739235</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41739235</guid></item><item><title><![CDATA[New comment by dongecko in "Ask HN: Weirdest Computer Architecture?"]]></title><description><![CDATA[
<p>Motorola used to have a one bit microprocessor, the MC14500B.</p>
]]></description><pubDate>Sat, 27 Jul 2024 10:59:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=41085727</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=41085727</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41085727</guid></item><item><title><![CDATA[New comment by dongecko in "Beyond Clean Code"]]></title><description><![CDATA[
<p>I simply love this book!</p>
]]></description><pubDate>Sat, 27 Jul 2024 10:48:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=41085685</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=41085685</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41085685</guid></item><item><title><![CDATA[New comment by dongecko in "Ask HN: What are some actual use cases of AI Agents right now?"]]></title><description><![CDATA[
<p>At the very beginning of my journey I did some fine tuning with Lora on a (I believe) Falcon model, but I haven't looked at it since. My impression was that injecting knowledge via fine tuning doesn't work, but tweaking behavior does. So your answer makes much sense to me. Thanks for bringing that up! I will definitively try that out.</p>
]]></description><pubDate>Thu, 15 Feb 2024 09:30:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=39380689</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=39380689</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39380689</guid></item><item><title><![CDATA[New comment by dongecko in "Ask HN: What are some actual use cases of AI Agents right now?"]]></title><description><![CDATA[
<p>I have not encountered this problem yet. When I was talking about the format of the answer I meant the following: No matter if you're using Langchain, Llamaindex, something self made, or Instructor (just to get a json back); under the hood there is somewhere the request to the LLM to reply in a structured way, like "answer in the following json format", or "just say 'a', 'b' or 'c'". ChatGPT tends to obey this rather well, most locally running LLMs don't. They answer like:<p>> Sure my friend, here is your requested json:<p>> ```<p>> {<p>>    name: "Daniel",<p>>    age: 47<p>> }<p>> ```<p>Unfortunately, the introductory sentence breaks directly parsing the answer, which means extra coding steps, or tweaking your prompt.</p>
]]></description><pubDate>Thu, 15 Feb 2024 07:38:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=39380090</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=39380090</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39380090</guid></item><item><title><![CDATA[New comment by dongecko in "Ask HN: What are some actual use cases of AI Agents right now?"]]></title><description><![CDATA[
<p>I also think of the retrieval part as a bottleneck and I am super excited of what the future holds.<p>In particular, I wonder if RAG systems will soon be a thing of the past, because end to end trained gigantic networks with longer attention spans, compression of knowledge, or hierarchical attention will at some point outperform retrieval. On the other hand, I can also see a completely different direction coming, where we develop architectures that, like operating systems, deal with memory management, scheduling and so on.</p>
]]></description><pubDate>Wed, 14 Feb 2024 22:43:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=39376607</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=39376607</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39376607</guid></item><item><title><![CDATA[New comment by dongecko in "Ask HN: What are some actual use cases of AI Agents right now?"]]></title><description><![CDATA[
<p>That is an interesting observation. I have not gotten to the point of too long cycles and I can think of two reasons for that.<p>Maybe my use case is narrow enough, so that in combination with a rather constraining and strict system message an answer is easy to find.<p>Second, I have lately played a lot with locally running LLMs. Their answers often break the formatting required for the agent to automatically proceed. So maybe I just don't see spiraling into oblivion, because I run into errors early ;)</p>
]]></description><pubDate>Wed, 14 Feb 2024 22:29:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=39376467</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=39376467</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39376467</guid></item><item><title><![CDATA[New comment by dongecko in "Ask HN: What are some actual use cases of AI Agents right now?"]]></title><description><![CDATA[
<p>The company I work for has tons of documentation and regulations for several areas. In some areas the documents are well over a thousand and for the ease of use of these documents we build RAG based chat bots. This is why I have been playing with RAG systems on the scale of "build completely from scratch" to "connect the services in Azure". The retrieval part of a RAG is vital for good/reliable answers and if you build it naive, the results are not overwhelming.<p>You can improve on the retrieved documents in many ways, like 
- by better chunking,<p>- better embedding,<p>- embedding several rephrased versions of the query,<p>- embedding a hypothetical answer to the prompt,<p>- hybrid retrieval (vector similarity + keyword/tfidf/bm25 related search),<p>- massively incorporating meta data,<p>- introducing additional (or hierarchical) summaries of the documents,<p>- returning not only the chunks but also adjacent text,<p>- re-ranking the candidate documents,<p>- fine tuning the LLM and much, much more.<p>However, at the end of the day a RAG system usually still has a hard time answering questions that require an overview of your data. Example questions are:<p>- "What are the key differences between the new and the old version of document X?"<p>- "Which documents can I ask you questions about?"<p>- "How do the regulations differ between case A and case B?"<p>In these cases it is really helpful to incorporate LLMs to decide how to process the prompt. This can be something simple like query-routing, or rephrasing/enhancing the original prompt until something useful comes up. But it can also be agents that come up with sub-queries and a plan on how to combine the partial answers. You can also build a network of agents with different roles (like coordinator/planner, reviewer, retriever, ...) to come up with an answer.<p>* edited the formatting</p>
]]></description><pubDate>Wed, 14 Feb 2024 20:11:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=39374628</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=39374628</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39374628</guid></item><item><title><![CDATA[New comment by dongecko in "How to Run Hackathons in 2024"]]></title><description><![CDATA[
<p>I have been to a couple of hackathons as participant and as organizer. Some sessions where really good and some were really terrible. The biggest takeaway for me in organizing a hackathon is to keep in mind that technical people who participate tend to be rather competitive folks who put a lot of soul into their work. So if you have no fair and clear evaluation system and PowerPoint designs compete with code, frustration is sure to happen.</p>
]]></description><pubDate>Tue, 26 Dec 2023 21:18:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=38776160</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=38776160</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38776160</guid></item><item><title><![CDATA[New comment by dongecko in "Transformers as Support Vector Machines"]]></title><description><![CDATA[
<p>I read it the same way as you did, or at least it's an approximation.<p>In general that's not really surprising.  I remember discussions from some years ago about larger networks leading to smother loss surfaces.</p>
]]></description><pubDate>Sun, 03 Sep 2023 14:10:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=37370562</link><dc:creator>dongecko</dc:creator><comments>https://news.ycombinator.com/item?id=37370562</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37370562</guid></item></channel></rss>