<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: DGoettlich</title><link>https://news.ycombinator.com/user?id=DGoettlich</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 30 Apr 2026 10:08:54 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=DGoettlich" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>data is 100% public domain.</p>
]]></description><pubDate>Sat, 20 Dec 2025 01:15:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=46332866</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46332866</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46332866</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>very interesting observation!</p>
]]></description><pubDate>Sat, 20 Dec 2025 00:58:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=46332777</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46332777</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46332777</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>well put.</p>
]]></description><pubDate>Sat, 20 Dec 2025 00:35:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=46332646</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46332646</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46332646</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>thanks. i think this just took on a weird dynamic. we never said we'd lock the model away. not sure how this impression seems to have emerged for some. that aside, it was an announcement of a release, not a release. the main purpose was gathering feedback on our methodology. standard procedure in our domain is to first gather criticism, incorporate it, then publish results.  but i understand people just wanted to talk to it. fair enough!</p>
]]></description><pubDate>Sat, 20 Dec 2025 00:31:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=46332620</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46332620</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46332620</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>i think we (whole section) are just talking past each other - we never said we'll lock it away. it was an announcement of a release, not a release. main purpose for us was getting feedback on the methodological aspects, as we clearly state. i understand you guys just wanted to talk to the thing though.</p>
]]></description><pubDate>Sat, 20 Dec 2025 00:18:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=46332537</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46332537</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46332537</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>Thanks for the comment. Could you elaborate on what you find iffy about our approach? I'm sure we can improve!</p>
]]></description><pubDate>Fri, 19 Dec 2025 23:04:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=46331992</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46331992</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46331992</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>valid point. its more of a stepping stone towards larger models. we're figuring out what the best way to do this is before scaling up.</p>
]]></description><pubDate>Fri, 19 Dec 2025 21:53:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=46331395</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46331395</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46331395</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>we're on the same page.</p>
]]></description><pubDate>Fri, 19 Dec 2025 21:42:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=46331287</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46331287</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46331287</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>exactly</p>
]]></description><pubDate>Fri, 19 Dec 2025 21:35:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=46331231</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46331231</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46331231</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>Also one of our fears. What we've done so far is to drop docs where the datasource was doubtful about the date of publication, if there are multiple possible dates we take the latest to be conservative. During training, we validate that the model learns pre- but not post-cutoff facts. <a href="https://github.com/DGoettlich/history-llms/blob/main/ranke-4b/prerelease_notes.md" rel="nofollow">https://github.com/DGoettlich/history-llms/blob/main/ranke-4...</a><p>If you have other ideas or think thats not enough, I'd be curious to know! (history-llms@econ.uzh.ch)</p>
]]></description><pubDate>Fri, 19 Dec 2025 15:16:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=46326748</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46326748</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46326748</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>what makes you think we trained on only a few gigabytes? <a href="https://github.com/DGoettlich/history-llms/blob/main/ranke-4b/prerelease_notes.md" rel="nofollow">https://github.com/DGoettlich/history-llms/blob/main/ranke-4...</a></p>
]]></description><pubDate>Fri, 19 Dec 2025 15:04:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=46326620</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46326620</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46326620</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>fully understand you. we'd like to provide access but also guard against misrepresentations of our projects goals by pointing to e.g. racist generations. if you have thoughts on how we should do that, perhaps you could reach out at history-llms@econ.uzh.ch ? thanks in advance!</p>
]]></description><pubDate>Fri, 19 Dec 2025 14:11:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=46326037</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46326037</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46326037</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>understand your frustration. i trust you also understand the models have some dark corners that someone could use to misrepresent the goals of our project. if you have ideas on how we could make the models more broadly accessible while avoiding that risk, please do reach out @ history-llms@econ.uzh.ch</p>
]]></description><pubDate>Fri, 19 Dec 2025 13:10:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=46325425</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46325425</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46325425</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>thanks. we'll be more precise in the future. ultimately, we took whatever we could get our hands on, that includes newspapers, periodicals, books. its multilingual (including italian, french, spanish etc) though majority is english.</p>
]]></description><pubDate>Fri, 19 Dec 2025 10:12:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=46324150</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46324150</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46324150</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>that is simply how we display the questions, its not what the model sees - we show the chat-template in the SFT section of the prerelease notes <a href="https://github.com/DGoettlich/history-llms/blob/main/ranke-4b/prerelease_notes.md" rel="nofollow">https://github.com/DGoettlich/history-llms/blob/main/ranke-4...</a></p>
]]></description><pubDate>Fri, 19 Dec 2025 10:07:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=46324113</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46324113</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46324113</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>we were considering doing that but ultimately it struck us as too sensitive wrt the exact in context examples, their ordering etc.</p>
]]></description><pubDate>Fri, 19 Dec 2025 10:06:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=46324101</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46324101</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46324101</guid></item><item><title><![CDATA[New comment by DGoettlich in "History LLMs: Models trained exclusively on pre-1913 texts"]]></title><description><![CDATA[
<p>We tell it that its a person (no gender) living in <cutoff>: we show the chat template in the prerelease notes <a href="https://github.com/DGoettlich/history-llms/blob/main/ranke-4b/prerelease_notes.md" rel="nofollow">https://github.com/DGoettlich/history-llms/blob/main/ranke-4...</a></p>
]]></description><pubDate>Fri, 19 Dec 2025 10:02:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=46324067</link><dc:creator>DGoettlich</dc:creator><comments>https://news.ycombinator.com/item?id=46324067</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46324067</guid></item></channel></rss>