<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: nlpprof</title><link>https://news.ycombinator.com/user?id=nlpprof</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 17 Apr 2026 01:49:29 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=nlpprof" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by nlpprof in "I Am Deleting the Blog"]]></title><description><![CDATA[
<p>Did you read what I wrote?<p>I don't think his argument is true. (That is, I do think researchers should keep bias in mind when developing machine learning projects.) (Regardless of their funding sources.)<p>Because of his employment, this argument is a particularly silly one for him to make.</p>
]]></description><pubDate>Tue, 23 Jun 2020 15:26:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=23615159</link><dc:creator>nlpprof</dc:creator><comments>https://news.ycombinator.com/item?id=23615159</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=23615159</guid></item><item><title><![CDATA[New comment by nlpprof in "I Am Deleting the Blog"]]></title><description><![CDATA[
<p>I'm in the field - though not as prominent as Yann (who has been very nice and helpful in my few interactions with him) - and your interpretation is off. People are disagreeing with his stance that researchers should not bother exploring bias implications of their research. (He says this is because bias is a problem of data - and therefore we should focus on building cool models and let production engineers worry about training production models on unbiased data.)<p>People are disagreeing not because of political correctness, but because this is a fundamental mischaracterization of how research works and how it gets transferred to "real world" applications.<p>(1) Data fuels modern machine learning. It shapes research directions in a really fundamental way. People decide what to work on based on what huge amounts of data they can get their hands on. Saying "engineers should be the ones to worry about bias because it's a data problem" is like saying "I'm a physicist, here's a cool model, I'll let the engineers worry about whether it works on any known particle in any known world."<p>(2) Most machine learning research is empirical (though not all). It's very rare to see a paper (if not impossible nowadays, since large deep neural networks are so massive and opaque) that works purely off math without showing that its conclusions improve some task on some dataset. No one is doing research without data, and saying "my method is good because it works on this data" means you are making choices and statements about what it means to "work" - which, as we've seen, involves quite a lot of bias.<p>(3) Almost all prominent ML researchers work for massively rich corporations. He and his colleagues don't work in ivory towers where they develop pure algorithms which are then released over the ivy walls into the wild, to be contaminated by filthy reality. He works for Facebook. He's paid with Facebook money. So why draw this imaginary line between research and production? He is paid to do research that will go into production.<p>So his statement is so wildly disconnected from research reality that it seems like it was not made in good faith - or at least without much thought - which is what people are responding to.<p>Also, language tip - a "woman researcher" is a "researcher".</p>
]]></description><pubDate>Tue, 23 Jun 2020 14:16:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=23614169</link><dc:creator>nlpprof</dc:creator><comments>https://news.ycombinator.com/item?id=23614169</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=23614169</guid></item></channel></rss>