<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: indeed30</title><link>https://news.ycombinator.com/user?id=indeed30</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 27 Apr 2026 09:59:59 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=indeed30" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by indeed30 in "Show HN: An easy-to-use online curve fitting tool"]]></title><description><![CDATA[
<p>The radioactive decay example specifically? Fit A and k (e.g. by nonlinear least squares) and then use the Jacobian to obtain the approximate covariance matrix. The diagnonal elements of that matrix give you the standard error estimates.</p>
]]></description><pubDate>Fri, 14 Nov 2025 17:33:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=45929216</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=45929216</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45929216</guid></item><item><title><![CDATA[New comment by indeed30 in "Show HN: An easy-to-use online curve fitting tool"]]></title><description><![CDATA[
<p>I think we are just coming at this from different angles. I do understand and agree that we are estimating the parameters of the fit curves.<p>> That already makes strong modeling assumptions (usually including IID, Gaussian noise, etc.,) to get the parameter estimates in the first place<p>You lose me here - I don't agree with "usually". I guess you're thinking of examples where you are sampling from a population and estimating features of that population. There's nothing wrong with that, but that is a <i>much</i> smaller domain than curve fitting in general.<p>If you give me a set of x and y, I can fit a parametric curve that tries to minimises the average squared distance between fit and observed values of y without making any assumptions whatsoever. This is a purely mechanical, non-stochastic procedure.<p>For example, if you give me the points {(0,0), (1,1), (2,4), (3,9)} and the curve y = a x^b, then I'm going to fit a=1, b=2, and I certainly don't need to assume anything about the data generating process to do so. However there is no concept of a confidence interval in this example - the estimates are the estimates, the residual error is 0, and that is pretty much all that can be said.<p>If you go further and tell me that each of these pairs (x,y) is randomly sampled, or maybe the x is fixed and the y is sampled, then I can do more. But that is often not the case.</p>
]]></description><pubDate>Fri, 14 Nov 2025 17:05:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=45928889</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=45928889</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45928889</guid></item><item><title><![CDATA[New comment by indeed30 in "Show HN: An easy-to-use online curve fitting tool"]]></title><description><![CDATA[
<p>I don’t think you can do anything sensible here without making much stronger modelling assumptions. A vanilla non-parametric bootstrap is only valid under a very specific generative story: IID sampling from a population. Many (most?) curve-fitting problems won't satisfy that.<p>For example, suppose you measure the decay of a radioactive source at fixed times t = 0,1,2,... and fit y = A e^{-kt}. The only randomness is small measurement error with, say, SD = 0.5. The bootstrap sees the huge spread in the y-values that comes from the deterministic decay curve itself, not from noise. It interprets that structural variation as sampling variability and you end up with absurdly wide bootstrap confidence intervals that have nothing to do with the actual uncertainty in the experiment.</p>
]]></description><pubDate>Fri, 14 Nov 2025 13:13:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=45926404</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=45926404</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45926404</guid></item><item><title><![CDATA[New comment by indeed30 in "UK Millionaire exodus did not occur, study reveals"]]></title><description><![CDATA[
<p>As long as UK taxes are flow-based and not stock-based, it seems a bit silly to base analysis on a stock-based denominator like the number of millionaires.</p>
]]></description><pubDate>Mon, 22 Sep 2025 16:10:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=45335481</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=45335481</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45335481</guid></item><item><title><![CDATA[New comment by indeed30 in "How big are our embeddings now and why?"]]></title><description><![CDATA[
<p>I wouldn’t call the embedding layer "separate" from the LLM. It’s learned jointly with the rest of the network, and its dimensionality is one of the most fundamental architectural choices. You’re right though that, in principle, you can pick an embedding size independent of other hyperparameters like number of layers or heads, so I see where you're coming from.<p>However the embedding dimension sets the rank of the token representation space. Each layer can transform or refine those vectors, but it can’t expand their intrinsic capacity. A tall but narrow network is bottlenecked by that width. Width-first scaling tends to outperform pure depth scaling, you want enough representational richness per token before you start stacking more layers of processing.<p>So yeah, embedding size doesn’t <i>have</i> to scale up in lockstep with model size, but in practice it usually does, because once models grow deeper and more capable, narrow embeddings quickly become the limiting factor.</p>
]]></description><pubDate>Fri, 05 Sep 2025 18:38:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=45142050</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=45142050</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45142050</guid></item><item><title><![CDATA[New comment by indeed30 in "Tetris is NP-hard even with O(1) rows or columns (2020) [pdf]"]]></title><description><![CDATA[
<p>I think it's actually (2020)</p>
]]></description><pubDate>Mon, 01 Sep 2025 16:35:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=45094193</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=45094193</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45094193</guid></item><item><title><![CDATA[New comment by indeed30 in "Reverse geocoding is hard"]]></title><description><![CDATA[
<p>Ten years ago, I worked for a company that had billions of sensor readings from mobile phones. The idea was to use crowdsourced data to create truly detailed, real-world coverage maps, and then sell that data to marketing and network operations teams at telcos.<p>We used reverse geocoding extensively — but never down to street addresses, always to a higher level. We wanted to split measurements by country, region, city — any geographic unit. When you deal with country borders, you get a lot of weird measurements as phones roam onto foreign networks. We weren’t interested in reporting on the experience of users roaming while abroad, so we needed shapefiles good enough to filter all that out and to partition the rest of the data cleanly.<p>We built a 30-machine Spark cluster on AWS back when Spark was still super early — around v0.7, definitely before 1.0. At the time, you pretty much had to use Scala with Spark if you cared about performance. Most of the workload was point-in-polygon tests. Before that, we were using a brutally hacky pipeline involving PostGIS, EMR, and Pig, and it was hell.<p>It was incredibly fun, but looking back now, I can see so clearly all the mistakes I made.</p>
]]></description><pubDate>Mon, 28 Apr 2025 14:39:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=43822020</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=43822020</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43822020</guid></item><item><title><![CDATA[Stop Word Clouds]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.stopwordclouds.com/">https://www.stopwordclouds.com/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43184932">https://news.ycombinator.com/item?id=43184932</a></p>
<p>Points: 5</p>
<p># Comments: 2</p>
]]></description><pubDate>Wed, 26 Feb 2025 16:04:09 +0000</pubDate><link>https://www.stopwordclouds.com/</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=43184932</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43184932</guid></item><item><title><![CDATA[New comment by indeed30 in "Disney+ loses 700k subscribers following price increase"]]></title><description><![CDATA[
<p>I would be pretty confident that Disney has a pricing team whose entire job is to model those effects.</p>
]]></description><pubDate>Thu, 06 Feb 2025 14:48:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=42962877</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=42962877</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42962877</guid></item><item><title><![CDATA[New comment by indeed30 in "Complete hardware and software setup for running Deepseek-R1 locally. ($6000)"]]></title><description><![CDATA[
<p>So, can somebody in the know speculate about how Deepseek (or OpenAI, or whoever really) is actually running their API?<p>If I wanted to run a production-grade service using the full Deepseek model, with good tokens/sec and the ability to serve concurrent requests, what sort of hardware are we looking at?</p>
]]></description><pubDate>Wed, 29 Jan 2025 16:01:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=42866614</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=42866614</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42866614</guid></item><item><title><![CDATA[New comment by indeed30 in "SQL nulls are weird"]]></title><description><![CDATA[
<p>That's interesting - I believe this is exactly how Sequelize implements soft-deletion.</p>
]]></description><pubDate>Thu, 09 Jan 2025 17:00:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=42647568</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=42647568</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42647568</guid></item><item><title><![CDATA[New comment by indeed30 in "Magic/tragic email links: don't make them the only option"]]></title><description><![CDATA[
<p>To discover where else you then subsequently forwarded it.<p>I'm not suggesting this is actually a problem, but that's how an argument could go.</p>
]]></description><pubDate>Wed, 08 Jan 2025 11:48:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=42633229</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=42633229</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42633229</guid></item><item><title><![CDATA[New comment by indeed30 in "GPT-5 is behind schedule"]]></title><description><![CDATA[
<p>I don't disagree with what you say, but one difference is that we generally hold these people accountable and often shift liability to them when they are wrong (though not always, admittedly), which is not something I have ever seen done with any AI system.</p>
]]></description><pubDate>Mon, 23 Dec 2024 12:57:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=42494064</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=42494064</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42494064</guid></item><item><title><![CDATA[Do you know how much your computer can do in a second? (2015)]]></title><description><![CDATA[
<p>Article URL: <a href="https://computers-are-fast.github.io/">https://computers-are-fast.github.io/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42451624">https://news.ycombinator.com/item?id=42451624</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 18 Dec 2024 16:03:40 +0000</pubDate><link>https://computers-are-fast.github.io/</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=42451624</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42451624</guid></item><item><title><![CDATA[New comment by indeed30 in "Popeye and Tintin enter the public domain in 2025 along with Faulkner, Hemingway"]]></title><description><![CDATA[
<p>I'd also mention that Tom McCarthy's "Tintin and The Secret of Literature" completely changed the way in which I viewed the series - genuinely exciting stuff.</p>
]]></description><pubDate>Mon, 16 Dec 2024 21:16:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=42435524</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=42435524</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42435524</guid></item><item><title><![CDATA[New comment by indeed30 in "Why is it so hard to find a job now? Enter Ghost Jobs"]]></title><description><![CDATA[
<p>Hang on a minute. There is absolutely nothing in this research that measures the accuracy of this approach. A user saying "I was ghosted" is not, to my mind, proof of anything.<p>Job seekers almost never actually know if the job was real or not, so it's hard to see how Glassdoor reviews can ever provide the insight this work is looking for.<p>I do believe that "ghost" jobs exist, often for H1B purposes, but I don't think this work proves it.</p>
]]></description><pubDate>Thu, 14 Nov 2024 15:33:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=42137192</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=42137192</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42137192</guid></item><item><title><![CDATA[New comment by indeed30 in "Your Name in Landsat"]]></title><description><![CDATA[
<p>Easiest way is going to be to downsample the images and then apply a pre-trained classifier that can ignore the fact these are sat images. You could probably turn them into 28x28 greyscale and then use a model trained on handwritten characters, like EMNIST.<p>Whatever approach you take, you'll probably be selecting the final set by hand, so it's just about building the candidate set in an efficient manner. Low absolute accuracy isn't really an issue as long as you end up with a managable set to review.</p>
]]></description><pubDate>Fri, 06 Sep 2024 13:58:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=41466273</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=41466273</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41466273</guid></item><item><title><![CDATA[New comment by indeed30 in "The Denmark secret: how it became the most trusting country"]]></title><description><![CDATA[
<p>I'm not sure what the complaint is here. I don't mind saying outright that I believe ethnicity is deeply intertwined with culture and that the two concepts are inseperable. Ethnic groups are essentially defined by their shared heritage, experiences and culture. When we talk about cultural diversity, we inevitably also discuss the ethnic diversity of a population.<p>I really don't think that is a controversial position, but clearly you disagree.</p>
]]></description><pubDate>Wed, 22 May 2024 13:23:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=40440651</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=40440651</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40440651</guid></item><item><title><![CDATA[New comment by indeed30 in "The Denmark secret: how it became the most trusting country"]]></title><description><![CDATA[
<p>I'm interested in what you're comparing against. Denmark is typically towards the bottom of any measure of cultural or ethnic diversity - e.g. <a href="https://en.wikipedia.org/wiki/List_of_countries_ranked_by_ethnic_and_cultural_diversity_level" rel="nofollow">https://en.wikipedia.org/wiki/List_of_countries_ranked_by_et...</a><p>I have been to Denmark many times and it strikes me as considerably more homogenous than the UK or the US.</p>
]]></description><pubDate>Wed, 22 May 2024 12:01:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=40439917</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=40439917</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40439917</guid></item><item><title><![CDATA[New comment by indeed30 in "Interviews in the Age of AI: Ditch Leetcode – Try Code Reviews Instead"]]></title><description><![CDATA[
<p>That's not my perception of the industry right now. I think what you said was true two years ago, but not any more.</p>
]]></description><pubDate>Tue, 17 Oct 2023 13:34:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=37914813</link><dc:creator>indeed30</dc:creator><comments>https://news.ycombinator.com/item?id=37914813</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37914813</guid></item></channel></rss>