<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: Eiim</title><link>https://news.ycombinator.com/user?id=Eiim</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 03 May 2026 19:19:19 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=Eiim" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by Eiim in "Vesuvius Challenge: First letters found in new scroll"]]></title><description><![CDATA[
<p>They already have human segmenters segmenting existing scrolls, which presumably is used to train the program in much the same way.</p>
]]></description><pubDate>Mon, 02 Dec 2024 14:49:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=42296727</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=42296727</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42296727</guid></item><item><title><![CDATA[New comment by Eiim in "What's Next for WebGPU"]]></title><description><![CDATA[
<p>You haven't substantiated why nobody else could make use of WebGPU. Are Google the only ones who can understand Beacons because they make $300B/year? GPU is hard, but it doesn't take billions to figure out.</p>
]]></description><pubDate>Fri, 22 Nov 2024 02:21:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=42210603</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=42210603</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42210603</guid></item><item><title><![CDATA[New comment by Eiim in "What actual purpose do accent characters in ISO-8859-1 and Windows 1252 serve?"]]></title><description><![CDATA[
<p>Interestingly, if you ever scan a FedEx barcode, they use tons of ASCII separator characters. About the only time I've seen them significantly used though.</p>
]]></description><pubDate>Sat, 29 Jun 2024 03:56:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=40827787</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=40827787</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40827787</guid></item><item><title><![CDATA[New comment by Eiim in "Show HN: "data-to-paper" – autonomous stepwise LLM-driven research"]]></title><description><![CDATA[
<p>I'm working on my Master's in Statistics, so I feel I can comment on some of what's going on here (although there are others more experienced than me in the comments as well, and I generally agree with their assessments). I'm going to look only at the diabetes example paper for now, mostly because I have finals tomorrow. I find it to be the equivalent of a STA261 final project at our university, with some extra fluff and nicer formatting. It's certainly not close to something I could submit to a journal.<p>The whole paper is "we took an existing dataset and ran the simplest reasonable model (a logistics regression) on it". That's about 5-10 minutes in R (or Python, or SAS, or whatever else). It's a very well-understood process, and it's a good starting point to understand the data, but it can't be the only thing in your paper, this isn't the 80's anymore.<p>The overall style is verbose and flowery, typical of LLMs. Good research papers should be straightforward and to the point. There's also strange mixing of "we" and "I" throughout.<p>We learn in the introduction that interaction effects were tested. That's fine, I'd want to see it set up earlier why these interaction effects are posited to be interesting. It said earlier that "a comprehensive investigation considering a multitude of diabetes-influencing lifestyle factors concurrently in relation to obesity remains to befully considered", but quite frankly, I don't believe that. Diabetes is remarkably well-studied, especially in observational studies like this one, due to its prevalence. I haven't searched the literature but I really doubt that no similar analysis has been done. This is one of the hardest parts of a research paper, finding existing research and where its gaps are, and I don't think an LLM will be sufficiently capable of that any time soon.<p>There's a complete lack of EDA in the paper. I don't need much (the whole analysis of this paper could be part of the EDA for a proper paper), but some basic distributional statistics of the variables. How many responses in the dataset were diabetic? Is there a sex bias? What about age distribution? Are any values missing? These are really important for observational studies because if there's any issues they should be addressed in some way. As it is, it's basically saying "trust us, our data is perfect" which is a huge ask. It's really weird that a bunch of this is in the appendix (which is way too long to be included in the paper, would need to be supplementary materials, but that's fine) (and also it's poorly formatted) but not mentioned anywhere in the paper itself. When looking at the appendix, the main concern that I have is that only 14% of the dataset is diabetic. This means that models will be biased towards predicting non-diabetic (if you just predict non-diabetic all of the time, you're already 86% accurate!). It's not as big of an issue for logistic regression, or for observational modeling like this, but I would have preferred an adjustment related to this.<p>In the results, I'm disappointed by the over-reliance on p-values. This is something that the statistics field is trying to move away from, of a multitude of reasons, one of which is demonstrated quite nicely here: p-values are (almost) always miniscule with large n, and in this case n=253680 is very large. Standard errors and CIs have the same issue. The Z-value is the most useful measure of confidence here in my eyes. Effect sizes are typically the more interesting metric for such studies. On that note, I would have liked to see predictors normalized so that coefficients can be directly compared. BMI, for example, has a small coefficient, but that's likely just because it has a large range and variance.<p>It's claimed that the AIC shows improved fit for the second model, but the change is only ~0.5%, which isn't especially convincing. In fact, it could be much less, because we don't have enough significant figures to see how the rounding went down. p-value is basically meaningless as previously stated.<p>The methods section says almost nothing that isn't already stated at least once. I'd like to know something about the tools which were used in this section, which is completely lacking. I do want it highlight this quote: "Both models employed a method to adjust for all possible confounders inthe analysis." What??? All possible confounders? If you know what that means you know that that's BS. "A method"? What is your magic tool to remove all variance not reflected in the dataset, I need to know! I certainly don't see it reflected in the code.<p>The code itself seems fine, maybe a little over-complicated but that might be necessary for how it Interfaces with the LLM. The actual analysis is equivalent to 3 basic lines of R (read CSV, basic log reg with default parameters 1, basic log reg with default parameters 2).<p>This paper would probably get about a B+ in 261, but shouldn't pass a 400-level class. The analysis is very simple and unimpressive for a few reasons. For one, the questions asked of the dataset are very light. More interesting, for example, might have been to do variable selection on all interaction terms and find which are important. More models should have been compared. The dataset is also extremely simple and doesn't demand complex analysis. An experimental design, or messy data with errors and missing values, or something requiring multiple datasets, would be a more serious challenge. It's quite possible that one of the other papers addresses this though.</p>
]]></description><pubDate>Sun, 12 May 2024 16:43:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=40335751</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=40335751</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40335751</guid></item><item><title><![CDATA[New comment by Eiim in "Unique volumes of Brothers Grimm fairy tales discovered in Poland"]]></title><description><![CDATA[
<p>I'm actually actively researching the Grimms tales - I'm glad you linked to Zipe's translation! I think he has the most accurate and complete English translation out there, but it's still being frequently overlooked for lower-quality public-domain translations.</p>
]]></description><pubDate>Fri, 03 May 2024 22:29:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=40253069</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=40253069</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40253069</guid></item><item><title><![CDATA[New comment by Eiim in "Google made me ruin a perfectly good website (2023)"]]></title><description><![CDATA[
<p>Nvidia is a great example of why a flat rule like this wouldn't work. Nvidia pretty much just does one, pretty specialized thing (GPUs) and trying to break it up into >10 pieces worth <$20B each (approximate median GDP of African nations by IMF) would be completely unnecessary. Just their gaming GPUs had ~$6B in profit in just the last year alone, and we know that their market cap comes much more from the AI market. We definitely could use stronger anti-monopoly laws, but market cap limits aren't the way to do it.</p>
]]></description><pubDate>Sun, 28 Apr 2024 04:07:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=40185902</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=40185902</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40185902</guid></item><item><title><![CDATA[New comment by Eiim in "First beta of Nintendo Switch emulator Suyu goes live"]]></title><description><![CDATA[
<p>I should say: the real difficulty is service, but if there's any way to get in contact with you, there's at least a chance it will be accepted as good service. There was famously a recent case where a defendant known only by a Bitcoin address was served by sending a transaction with an attached message to that address.</p>
]]></description><pubDate>Thu, 21 Mar 2024 04:55:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=39774980</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39774980</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39774980</guid></item><item><title><![CDATA[New comment by Eiim in "First beta of Nintendo Switch emulator Suyu goes live"]]></title><description><![CDATA[
<p>File a lawsuit against a John/Jane Doe defendant(s) and describe who the defendants are, e.g., the people who code and maintain the emulator. Maybe a little harder than if you had the names but not by a lot.</p>
]]></description><pubDate>Thu, 21 Mar 2024 04:51:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=39774968</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39774968</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39774968</guid></item><item><title><![CDATA[New comment by Eiim in "First beta of Nintendo Switch emulator Suyu goes live"]]></title><description><![CDATA[
<p>You can still be sued. At that point, you either:
a) defend yourself as normal, while attempting to maintain anonymity (difficult & doesn't solve anything)
b) let the case go to default judgement (for likely an enormous amount of money) at which point Nintendo has every reason to try to track you down in real life, and can use the courts to help them do so. Hope you were really, <i>really</i> thorough about your anonymity.</p>
]]></description><pubDate>Wed, 20 Mar 2024 14:10:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=39766807</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39766807</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39766807</guid></item><item><title><![CDATA[New comment by Eiim in "YouTube now requires to label their realistic-looking videos made using AI"]]></title><description><![CDATA[
<p>Bad evidence is not the same thing as inadmissible evidence. Evidence is admitted, and then the fact finder determines whether to consider it, and how much weight to give it. It is likely that surveillance video will be slightly less credible now, but can still be part of a large, convincing body of evidence.</p>
]]></description><pubDate>Tue, 19 Mar 2024 00:17:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=39751625</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39751625</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39751625</guid></item><item><title><![CDATA[New comment by Eiim in "Achieving awful compression with digits of pi"]]></title><description><![CDATA[
<p>This is basically just a compression dictionary made up of random noise, which is... a thing that you can do</p>
]]></description><pubDate>Fri, 15 Mar 2024 23:01:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=39721571</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39721571</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39721571</guid></item><item><title><![CDATA[New comment by Eiim in "New York disbars infamous copyright troll"]]></title><description><![CDATA[
<p>Part of Liebowitz's problem was that he did ~0 research before filing a lawsuit. Often the only thing he did was send the photographer a list of URLs and ask for "GO" or "NO GO" for each one. In at least some cases there was a license that he or probably should have known about, but he didn't even bother to check. He probably would've been fine with that though if he hadn't lied to the judge(s) so, <i>so</i> much.</p>
]]></description><pubDate>Fri, 15 Mar 2024 02:01:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=39711337</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39711337</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39711337</guid></item><item><title><![CDATA[New comment by Eiim in "Only 20 JavaScript runtimes left"]]></title><description><![CDATA[
<p>oend (pronounced "owned") has some potential, surely</p>
]]></description><pubDate>Thu, 14 Mar 2024 05:47:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=39700902</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39700902</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39700902</guid></item><item><title><![CDATA[New comment by Eiim in "Autogenerating a Book Series from Three Years of iMessages"]]></title><description><![CDATA[
<p>Very true! I have a friend with a tape backup system and it serves him well.</p>
]]></description><pubDate>Mon, 11 Mar 2024 04:32:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=39664943</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39664943</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39664943</guid></item><item><title><![CDATA[New comment by Eiim in "Autogenerating a Book Series from Three Years of iMessages"]]></title><description><![CDATA[
<p>Flash memory relies on cells keeping charged, but the electrons can slowly leak and discharge the cells over time. It looks like the commonly claimed number is 10 years, but there's no clear answer. Hard drives also aren't great as a "set and forget" method. In either case you should refresh the data regularly (~yearly). Optical media is a great option for digital long-term storage, but paper is a very tried-and-true method, if stored in the right conditions.</p>
]]></description><pubDate>Fri, 08 Mar 2024 16:31:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=39642706</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39642706</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39642706</guid></item><item><title><![CDATA[New comment by Eiim in "3DS Emulator Citra is closing up shop and removing all their sites/code"]]></title><description><![CDATA[
<p>Nothing's disappearing, except the binaries and source. Good-quality archives will surely be assembled, but it at a minimum makes Citra significantly harder to access.</p>
]]></description><pubDate>Mon, 04 Mar 2024 21:53:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=39596642</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39596642</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39596642</guid></item><item><title><![CDATA[New comment by Eiim in "Why Python's integer division floors (2010)"]]></title><description><![CDATA[
<p>Did you read the article? It's about what happens in negative numbers. Of course everyone agrees that 7/2=3 in integer division, but -7/2 is less obvious.</p>
]]></description><pubDate>Wed, 28 Feb 2024 15:09:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=39538908</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39538908</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39538908</guid></item><item><title><![CDATA[New comment by Eiim in "Magika: AI powered fast and efficient file type identification"]]></title><description><![CDATA[
<p>Indeed, which is why I felt the need to call it out here. I'm not certain if the files on question actually happened to be valid C but whether that's a meaningful mistake regardless is left to the reader to decide.</p>
]]></description><pubDate>Fri, 16 Feb 2024 19:44:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=39401986</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39401986</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39401986</guid></item><item><title><![CDATA[New comment by Eiim in "Magika: AI powered fast and efficient file type identification"]]></title><description><![CDATA[
<p>I ran a quick test on 100 semi-random files I had laying around. Of those, 81 were detected correctly, 6 were detected as the wrong file type, and 12 were detected with an unspecific file type (unknown binary/generic text) when a more specific type existed. In 4 of the unspecific cases, a low-confidence guess was provided, which was wrong in each case. However, almost all of the files which were detected wrong/unspecific are of types not supported by Magika, with one exception of a JSON file containing a lot of JS code as text, which was detected as JS code. For comparison, file 5.45 (the version I happened to have installed) got 83 correct, 6 wrong, and 10 not specific. It detected the weird JSON correctly, but also had its own strange issues, such as detecting a CSV as just "data". The "wrong" here was somewhat skewed by the 4 GLSL shader code files that were in the dataset for some reason, all of which it detected as C code (Magika called them unknown). The other two "wrong" detections were also code formats that it seems it doesn't support. It was also able to output a lot more information about the media files. Not sure what to make of these tests but perhaps they're useful to somebody.</p>
]]></description><pubDate>Fri, 16 Feb 2024 15:09:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=39398028</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39398028</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39398028</guid></item><item><title><![CDATA[New comment by Eiim in "Magika: AI powered fast and efficient file type identification"]]></title><description><![CDATA[
<p>From elsewhere in this thread, it appears that Magika detects far fewer file types than file (116 vs ~1600), which makes sense. For file, you just need to drop in a few rules to add a new, somewhat obscure type. An AI approach like Magika will need lots of training and test data for each new file type. Where Magika might have a leg up is with distinguishing different textual data files (i.e., source code), but I don't see that as a particularly big use case honestly.</p>
]]></description><pubDate>Fri, 16 Feb 2024 14:23:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=39397390</link><dc:creator>Eiim</dc:creator><comments>https://news.ycombinator.com/item?id=39397390</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=39397390</guid></item></channel></rss>