<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: qumpis</title><link>https://news.ycombinator.com/user?id=qumpis</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 15 Apr 2026 00:15:43 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=qumpis" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by qumpis in "Tree Search Distillation for Language Models Using PPO"]]></title><description><![CDATA[
<p>I may never understand what harness means - it's used in so many contexts</p>
]]></description><pubDate>Sun, 15 Mar 2026 14:03:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=47387499</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=47387499</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47387499</guid></item><item><title><![CDATA[New comment by qumpis in "Beyond Diffusion: Inductive Moment Matching"]]></title><description><![CDATA[
<p>What happens if we don't add any moments matching objective? e.g. at train time just fit a diffusion model that predicts the target given any pair of timesteps (t, t')? Why is moment matching critical here?<p>Also regarding linearity, why is it inflexible? It seems quite convenient that a simple linear interpolation is used for reconstruction, besides, even in DDIM, the directions towards the final target changes at each step as the images become less noisy. In standard diffusion models or even flow matching, denoising is always equal to the prediction of the original data + direction from current timestep to the timestep t'. Just to be clear, it is intuitive that such models are inferior in few-step generations since they don't optimise for test time efficiency (in terms of the tradeoff of quality vs compute), but it's unclear what inflexibility exists there beyond this limitation.<p>Clearly there's no expected benefit in quality if all timesteps are used in denoising?</p>
]]></description><pubDate>Wed, 12 Mar 2025 13:04:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=43342784</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=43342784</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43342784</guid></item><item><title><![CDATA[New comment by qumpis in "I always knew I was different, I didn't know I was a sociopath"]]></title><description><![CDATA[
<p>Fascinating article. I wonder what made her do the steps in finding what's "different" with her, and most of all, why the need to fix it arose. Is it to "fit in", understandably? It somehow felt alien to me, and this shows my ignorance on the topic, that people lacking in empathy department would attempt to understand the reasons and act "good" towards others even if this feeling is only understood intellectually.</p>
]]></description><pubDate>Thu, 14 Mar 2024 12:54:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=39703312</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=39703312</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39703312</guid></item><item><title><![CDATA[New comment by qumpis in "Stable Diffusion 3"]]></title><description><![CDATA[
<p>Yes it makes sense a bit. Many popular convents operate on 3x3 kernels. But the number of channel increases per layer. This, coupled with the fact that the receptive field increases per layer and allows convnets to essentially see the whole image relatively early in model's depth (esp. coupled with pooling operations which increase the receptive field rapidly), makes this intuition questionable. Transformers on the other hand, operate on attention which allows them to weight each patch dynamically, but it's clear to me  that this allows them to attend to all parts of the image in a way different from convnets.</p>
]]></description><pubDate>Fri, 23 Feb 2024 06:14:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=39477520</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=39477520</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39477520</guid></item><item><title><![CDATA[New comment by qumpis in "Stable Diffusion 3"]]></title><description><![CDATA[
<p>Convolutions are bad at long range spatial dependencies? What makes you say that - any chance you have a reference?</p>
]]></description><pubDate>Fri, 23 Feb 2024 02:07:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=39476050</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=39476050</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39476050</guid></item><item><title><![CDATA[New comment by qumpis in "A* tricks for videogame path finding"]]></title><description><![CDATA[
<p>I haven't seen RL with decision trees! it sounds really interesting. Any classic results worth looking into?</p>
]]></description><pubDate>Mon, 01 Jan 2024 20:23:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=38835022</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38835022</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38835022</guid></item><item><title><![CDATA[New comment by qumpis in "Reindeer sleep and eat simultaneously"]]></title><description><![CDATA[
<p>Toll noticable only back in those days or even in the future?</p>
]]></description><pubDate>Thu, 28 Dec 2023 11:32:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=38792397</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38792397</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38792397</guid></item><item><title><![CDATA[New comment by qumpis in "Reindeer sleep and eat simultaneously"]]></title><description><![CDATA[
<p>Was it more productive, objectively speaking, than being consistent and not dealing with the constant drainage due to severely lacking sleep?</p>
]]></description><pubDate>Thu, 28 Dec 2023 02:24:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=38789276</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38789276</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38789276</guid></item><item><title><![CDATA[New comment by qumpis in "Augmenting long-term memory (2018)"]]></title><description><![CDATA[
<p>Can you give some example usecases of your application? I wonder how it scales to complex (in terms of structure) information processing, e.g. digesting scientific topics</p>
]]></description><pubDate>Sun, 17 Dec 2023 06:02:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=38670691</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38670691</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38670691</guid></item><item><title><![CDATA[New comment by qumpis in "What’s behind the Freud resurgence?"]]></title><description><![CDATA[
<p>So therapy is not effective as per the paper. This seems like an astounding conclusion. Has anyone read the paper in detail and has a deeper opinion?</p>
]]></description><pubDate>Thu, 14 Dec 2023 04:18:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=38637727</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38637727</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38637727</guid></item><item><title><![CDATA[New comment by qumpis in "Mamba: Linear-Time Sequence Modeling with Selective State Spaces"]]></title><description><![CDATA[
<p>How does this differ from RNNs and their gating mechanism?</p>
]]></description><pubDate>Tue, 05 Dec 2023 00:10:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=38525201</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38525201</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38525201</guid></item><item><title><![CDATA[New comment by qumpis in "Oracle of Zotero: LLM QA of Your Research Library"]]></title><description><![CDATA[
<p>From the glance of it, the paper looks very polished. Combine this with the fact that arxiv is invite-only, your prediction might not come about</p>
]]></description><pubDate>Mon, 27 Nov 2023 04:19:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=38427994</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38427994</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38427994</guid></item><item><title><![CDATA[New comment by qumpis in "Sqids – Generate short unique IDs from numbers"]]></title><description><![CDATA[
<p>How big and random were these "few extras"?</p>
]]></description><pubDate>Sun, 26 Nov 2023 01:20:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=38418402</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38418402</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38418402</guid></item><item><title><![CDATA[New comment by qumpis in "Hacking ADHD: Strategies for the modern developer"]]></title><description><![CDATA[
<p>I've got good results with white noise, not sure if that's something that would work for you. And of course I didn't suggest that this eliminates other distractions.</p>
]]></description><pubDate>Thu, 16 Nov 2023 14:08:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=38289656</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38289656</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38289656</guid></item><item><title><![CDATA[New comment by qumpis in "Hacking ADHD: Strategies for the modern developer"]]></title><description><![CDATA[
<p>Why not put these earbuds in the office?</p>
]]></description><pubDate>Wed, 15 Nov 2023 16:08:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=38278238</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=38278238</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38278238</guid></item><item><title><![CDATA[New comment by qumpis in "ArXiv receives $10M for upgrades"]]></title><description><![CDATA[
<p>"Ar5iv" does a pretty good job from my experience</p>
]]></description><pubDate>Fri, 20 Oct 2023 01:53:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=37951397</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=37951397</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37951397</guid></item><item><title><![CDATA[New comment by qumpis in "The worst programmer I know"]]></title><description><![CDATA[
<p>A bit surprising to hear that you weren't swimming in offers to hire you all the time. Any idea why that was? Maybe you're somewhat picky about the job and responsibilities that come with it?</p>
]]></description><pubDate>Sun, 03 Sep 2023 08:27:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=37368718</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=37368718</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37368718</guid></item><item><title><![CDATA[New comment by qumpis in "Do Machine Learning Models Memorize or Generalize?"]]></title><description><![CDATA[
<p>Slightly related but sparsity-inducing activation function Relu is often used in neural networks</p>
]]></description><pubDate>Thu, 10 Aug 2023 17:41:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=37079224</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=37079224</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37079224</guid></item><item><title><![CDATA[New comment by qumpis in "Nvidia H100 GPUs: Supply and Demand"]]></title><description><![CDATA[
<p>I don't remember the exact tweet, but here's one discussion [1]. I guess something changed in the mean time.<p>[1]_<a href="https://www.reddit.com/r/Amd/comments/140uct5/geohot_giving_up_on_amd_gpus_for_compute/" rel="nofollow noreferrer">https://www.reddit.com/r/Amd/comments/140uct5/geohot_giving_...</a></p>
]]></description><pubDate>Tue, 01 Aug 2023 09:12:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=36953998</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=36953998</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36953998</guid></item><item><title><![CDATA[New comment by qumpis in "Nvidia H100 GPUs: Supply and Demand"]]></title><description><![CDATA[
<p>Didn't they abandon development of AMD related software?</p>
]]></description><pubDate>Tue, 01 Aug 2023 07:50:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=36953499</link><dc:creator>qumpis</dc:creator><comments>https://news.ycombinator.com/item?id=36953499</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36953499</guid></item></channel></rss>