<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: williamtrask</title><link>https://news.ycombinator.com/user?id=williamtrask</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 05 Apr 2026 16:24:06 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=williamtrask" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[Zero-Setup Federated Learning: Train Models Across Private Datasets with GColab]]></title><description><![CDATA[
<p>Article URL: <a href="https://openmined.org/blog/zero-setup-federated-learning-google-colab/">https://openmined.org/blog/zero-setup-federated-learning-google-colab/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46539543">https://news.ycombinator.com/item?id=46539543</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 08 Jan 2026 10:45:46 +0000</pubDate><link>https://openmined.org/blog/zero-setup-federated-learning-google-colab/</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=46539543</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46539543</guid></item><item><title><![CDATA[New comment by williamtrask in "A website to destroy all websites"]]></title><description><![CDATA[
<p>...with a price :)</p>
]]></description><pubDate>Fri, 02 Jan 2026 00:34:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=46459937</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=46459937</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46459937</guid></item><item><title><![CDATA[Leadership Lab: The Craft of Writing Effectively [video]]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.youtube.com/watch?v=vtIzMaLkCaM">https://www.youtube.com/watch?v=vtIzMaLkCaM</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46452577">https://news.ycombinator.com/item?id=46452577</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 01 Jan 2026 09:20:45 +0000</pubDate><link>https://www.youtube.com/watch?v=vtIzMaLkCaM</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=46452577</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46452577</guid></item><item><title><![CDATA[New comment by williamtrask in "Everything as code: How we manage our company in one monorepo"]]></title><description><![CDATA[
<p>"Conclusion
Our monorepo isn't about following a trend. It's about removing friction between things that naturally belong together, something that is critical when related context is everything.<p>When a feature touches the backend API, the frontend component, the documentation, and the marketing site—why should that be four repositories, four PRs, four merge coordination meetings?<p>The monorepo isn't a constraint. It's a force multiplier."<p>Thank you Claude :)</p>
]]></description><pubDate>Tue, 30 Dec 2025 21:48:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=46438422</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=46438422</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46438422</guid></item><item><title><![CDATA[New comment by williamtrask in "You can't refuse to be scanned by ICE's facial recognition app, DHS document say"]]></title><description><![CDATA[
<p>tried searching for "noodlesUK" and didn't find anything meaningful</p>
]]></description><pubDate>Sat, 01 Nov 2025 13:48:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=45781615</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45781615</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45781615</guid></item><item><title><![CDATA[New comment by williamtrask in "A History of Large Language Models"]]></title><description><![CDATA[
<p>Nit: regarding (2), Phil Blunsom did (same Blunsom from the article, and who was leading language modeling at DeepMind for about 7-8 years). He would often opine at Oxford (where he taught) that solving next word prediction is a viable meta path to AGI. Almost nobody agreed at the time. He also called out early that scaling and better data were the key, and they did end up being, although Google wasn’t as “risk on” as OpenAI on gathering the data for GPT-1/2. Had they been history could easily have been different. People forget the position OAI was in at the time. Elon/funding had left, key talent had left. Risk appetite was high for that kind of thing… and it paid off.</p>
]]></description><pubDate>Thu, 09 Oct 2025 08:48:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=45525135</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45525135</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45525135</guid></item><item><title><![CDATA[The Bitter Lesson's Bitter Lesson]]></title><description><![CDATA[
<p>Article URL: <a href="https://twitter.com/iamtrask/status/1972167087037808995">https://twitter.com/iamtrask/status/1972167087037808995</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45408988">https://news.ycombinator.com/item?id=45408988</a></p>
<p>Points: 5</p>
<p># Comments: 1</p>
]]></description><pubDate>Sun, 28 Sep 2025 23:31:54 +0000</pubDate><link>https://twitter.com/iamtrask/status/1972167087037808995</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45408988</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45408988</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>"This is not the reason, the reason is that this data is private. LLMs do not just learn from data, they can often reproduce it verbatim, you cannot give medical records or bank records of real people, that will put them at a very real risk."<p>(OP) You make great points. I think we're actually more in agreement than might be obvious. Part of the reason you need to "give" data to an LLM is because of the way LLMs are constructed... which creates the privacy risk.<p>The principle of attribution-based control suggested in this article would break that principle, enabling each data owner to control which AI predictions they make more intelligent (as opposed to only controlling which IA models they help train).<p>So to your point... this <i>is</i> a very rigorous privacy protection. Another way to TLDR the article is "if we get really good at privacy... there's a LOT more data out there... so let's start <i>really</i> caring about privacy"<p>Anyway... I agree with everything in your comment. Just thought I'd drop by and try to lend clarity to how the article agrees with you (sounds like there's room for improvement on how to describe attribution-based control though).</p>
]]></description><pubDate>Thu, 25 Sep 2025 16:01:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=45374507</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45374507</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45374507</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>With you on this one. I do think ABC is a step in the right direction to improve things. <3</p>
]]></description><pubDate>Thu, 25 Sep 2025 00:30:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=45367753</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45367753</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45367753</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>"The claim that humans need petabytes of data to develop their mind seems completely indefensible to me."<p>And yet every human you know is using petabytes of data to develop their mind. :)</p>
]]></description><pubDate>Wed, 24 Sep 2025 22:22:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=45366662</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45366662</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45366662</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>I'm relatively close to publishing my PhD thesis which is broadly a survey paper of what you're describing. Will share (<i>almost</i> done with revisions).</p>
]]></description><pubDate>Wed, 24 Sep 2025 22:21:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=45366655</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45366655</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45366655</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>I think this is the right question to ask. I think it depends on the task. For example, if you want to predict whether someone has cancer, then access to avast amounts of medical information would be important.</p>
]]></description><pubDate>Wed, 24 Sep 2025 21:14:09 +0000</pubDate><link>https://news.ycombinator.com/item?id=45366041</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45366041</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45366041</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>This article is meant for a policy audience, so that does keep the technical depth pretty thin. It's rooted in more rigorous deep learning work. Happy to send your way if interested.</p>
]]></description><pubDate>Wed, 24 Sep 2025 21:13:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=45366028</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45366028</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45366028</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>I agree with you in a way - that it seems likely that new data will be incorproated in more inference-like ways. RAG is a little extreme... but i think there's going to be middle grounds betweeen full pre-training and RAG. Git-rebasin, MoE, etc.</p>
]]></description><pubDate>Wed, 24 Sep 2025 21:12:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=45366021</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45366021</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45366021</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>Yeah Zama's stuff is great.</p>
]]></description><pubDate>Wed, 24 Sep 2025 21:07:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=45365975</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45365975</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45365975</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>Agree with you on the nuance.</p>
]]></description><pubDate>Wed, 24 Sep 2025 21:02:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=45365908</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45365908</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45365908</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>(OP) YOLO</p>
]]></description><pubDate>Wed, 24 Sep 2025 21:01:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=45365897</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45365897</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45365897</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>(OP) fwiw I fully agree with the privacywashing you're describing here, and this piece is advocating for a more rigorous standard than input privacy (homomorphic encryption), which is insufficient to enable data owners to actually retain control over their data (but is a useful ingredient).</p>
]]></description><pubDate>Wed, 24 Sep 2025 21:00:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=45365890</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45365890</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45365890</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>This is the right question. If full attribution-based control is achieved, then this would be impossible. And the ingredient you've suggested could be a useful way to help achieve it.</p>
]]></description><pubDate>Wed, 24 Sep 2025 20:59:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=45365874</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45365874</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45365874</guid></item><item><title><![CDATA[New comment by williamtrask in "Unlocking a Million Times More Data for AI"]]></title><description><![CDATA[
<p>This is the magic :)</p>
]]></description><pubDate>Wed, 24 Sep 2025 20:58:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=45365865</link><dc:creator>williamtrask</dc:creator><comments>https://news.ycombinator.com/item?id=45365865</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45365865</guid></item></channel></rss>