<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: onasta</title><link>https://news.ycombinator.com/user?id=onasta</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 14 Apr 2026 20:16:56 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=onasta" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[Show HN: TabPFN Scaling Mode – Tabular Foundation Model on millions of rows]]></title><description><![CDATA[
<p>I’m excited to announce Scaling Mode for TabPFN-2.5, our tabular foundation model (Large Tabular Model). This removes any fixed upper limits on dataset sizes and extends TabPFN to datasets with millions of rows.<p>What is Scaling Mode? A new pipeline around TabPFN-2.5 designed for large-N workloads, which removes the fixed row limit of TabPFN. The system works with large training sets, constrained only by your compute and memory.<p>We benchmarked Scaling Mode on datasets ranging from 1M to 10M rows, comparing against CatBoost, XGBoost, and LightGBM. Key findings:<p>- Scaling Mode enables TabPFN-2.5 to continue improving performance with more data<p>- Scales dramatically better than TabPFN-2.5 with 50K subsampling<p>- No evidence of the performance gap to gradient boosting shrinking as we scale up<p>- Performance continues to improve strongly with more data on the largest tested datasets<p>To quickly summarise our progress, here's the history of our scaling trajectory:<p>- TabPFN v2 (Jan 2025): 10K rows<p>- TabPFN-2.5 (Nov 2025): 50K rows<p>- Scaling Mode (today): 10M+ rows tested<p>Current limitations: Scaling Mode is designed for companies for now. If you’re working with data at a large scale and want to test it, access is currently only by request to ensure we can support early users properly.<p>Full blog post: <a href="https://priorlabs.ai/technical-reports/large-data-model" rel="nofollow">https://priorlabs.ai/technical-reports/large-data-model</a><p>Request access: <a href="https://priorlabs.ai/tabpfn/large-data" rel="nofollow">https://priorlabs.ai/tabpfn/large-data</a><p>Would love to hear feedback from anyone working with large tabular datasets!</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46138439">https://news.ycombinator.com/item?id=46138439</a></p>
<p>Points: 5</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 03 Dec 2025 18:56:00 +0000</pubDate><link>https://priorlabs.ai/technical-reports/large-data-model</link><dc:creator>onasta</dc:creator><comments>https://news.ycombinator.com/item?id=46138439</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46138439</guid></item><item><title><![CDATA[Show HN: TabPFN-2.5 – SOTA foundation model for tabular data]]></title><description><![CDATA[
<p>I am excited to announce the release of TabPFN-2.5, our tabular foundation model that now scales to datasets of up to 50,000 samples and 2,000 features - a 5x increase from TabPFN v2, published in the Nature journal earlier this year. TabPFN-2.5 delivers state-of-the-art predictions in one forward pass without hyperparameter tuning across classification and regression tasks.<p><i>What’s new in 2.5</i>:
TabPFN-2.5 maintains the core approach of v2 - a pretrained transformer trained on more than hundred million synthetic datasets to perform in-context learning and output a predictive distribution for the test data. It natively supports missing values, cateogrical features, text and numerical features is robust to outliers and uninformative features.<p>The major improvements:<p>- 5x scale increase: Now handles 50,000 samples × 2,000 features (up from 10,000 × 500 in v2)<p>- SOTA performance: TabPFN-2.5 outperforms tuned tree-based methods and matches the performance of a complex ensemble (AutoGluon 1.4), that itself includes TabPFN v2, tuned for 4 hours. Tuning the model improves performance, outperforming AutoGluon 1.4 for regression tasks.<p>- Rebuilt API: New REST interface along with Python SDK with dedicated fit & predict endpoints, making deployment and integration more developer-friendly<p>- A distillation engine that converts TabPFN-2.5 into a compact MLP or tree ensemble while preserving accuracy and offer low latency inference.<p>There are still some limitations. The model is designed for datasets up to 50K samples. It can handle larger datasets but that hasn’t been our focus with TabPFN-2.5. The distillation engine is not yet available through the API but only through licenses (though we do show the performance in the model report).<p>We’re actively working on removing these limitations and intend to release newer models focused on context reasoning, causal inference, graph networks, larger data and time-series.
TabPFN-2.5 is available via API and a package on Hugging Face. Would love for you to try it and give us your feedback!<p>Model report: <a href="https://priorlabs.ai/technical-reports/tabpfn-2-5-model-report" rel="nofollow">https://priorlabs.ai/technical-reports/tabpfn-2-5-model-repo...</a><p>Package: <a href="https://github.com/PriorLabs/TabPFN" rel="nofollow">https://github.com/PriorLabs/TabPFN</a><p>Client: <a href="https://github.com/PriorLabs/tabpfn-client" rel="nofollow">https://github.com/PriorLabs/tabpfn-client</a><p>Docs: <a href="https://docs.priorlabs.ai/quickstart" rel="nofollow">https://docs.priorlabs.ai/quickstart</a></p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45838540">https://news.ycombinator.com/item?id=45838540</a></p>
<p>Points: 73</p>
<p># Comments: 13</p>
]]></description><pubDate>Thu, 06 Nov 2025 18:26:53 +0000</pubDate><link>https://priorlabs.ai/technical-reports/tabpfn-2-5-model-report</link><dc:creator>onasta</dc:creator><comments>https://news.ycombinator.com/item?id=45838540</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45838540</guid></item><item><title><![CDATA[New comment by onasta in "Show HN: TabPFN v2 – A SOTA foundation model for small tabular data"]]></title><description><![CDATA[
<p>There have been a ton of improvements! Much better performance overall, way larger data size limit (1K-->10K rows, 100-->500 features), regression support, native categorical data and missing values handling, much better support for uninformative or outlier features etc.</p>
]]></description><pubDate>Thu, 09 Jan 2025 23:30:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=42650930</link><dc:creator>onasta</dc:creator><comments>https://news.ycombinator.com/item?id=42650930</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42650930</guid></item><item><title><![CDATA[New comment by onasta in "Show HN: TabPFN v2 – A SOTA foundation model for small tabular data"]]></title><description><![CDATA[
<p>TabPFN is better on numerical data since v1 (see figure 6 in the CARTE paper). CARTE's main strength in on text features, which are now also supported for TabPFN v2 API version (<a href="https://github.com/PriorLabs/tabpfn-client">https://github.com/PriorLabs/tabpfn-client</a>). We compared this to CARTE and found our model to be generally quite better, and much faster. CARTE multi-table approach is also very interesting, and we want to tackle this setting in the future.</p>
]]></description><pubDate>Thu, 09 Jan 2025 22:55:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=42650700</link><dc:creator>onasta</dc:creator><comments>https://news.ycombinator.com/item?id=42650700</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42650700</guid></item><item><title><![CDATA[Show HN: TabPFN v2 – A SOTA foundation model for small tabular data]]></title><description><![CDATA[
<p>I am excited to announce the release of TabPFN v2, a tabular foundation model that delivers state-of-the-art predictions on small datasets in just 2.8 seconds for classification and 4.8 seconds for regression compared to strong baselines tuned for 4 hours. Published in Nature, this model outperforms traditional methods on datasets with up to 10,000 samples and 500 features.<p>The model is available under an open license: a derivative of the Apache 2 license with a single modification, adding an enhanced attribution requirement inspired by the Llama 3 license: <a href="https://github.com/PriorLabs/tabpfn">https://github.com/PriorLabs/tabpfn</a>. You can also try it via API: <a href="https://github.com/PriorLabs/tabpfn-client">https://github.com/PriorLabs/tabpfn-client</a><p>TabPFN v2 is trained on 130 million synthetic tabular prediction datasets to perform in-context learning and output a predictive distribution for the test data points. Each dataset acts as one meta-datapoint to train the TabPFN weights with SGD. As a foundation model, TabPFN allows for fine-tuning, density estimation and data generation.<p>Compared to TabPFN v1, v2 now natively supports categorical features and missing values. TabPFN v2 performs just as well on datasets with or without these. It also handles outliers and uninformative features naturally, problems that often throw off standard neural nets.<p>TabPFN v2 performs as well with half the data as the next best baseline (CatBoost) with all the data.<p>We also compared TabPFN to the SOTA AutoML system AutoGluon 1.0. Standard TabPFN already outperforms AutoGluon on classification and ties on regression, but ensembling multiple TabPFNs in TabPFN v2 (PHE) is even better.<p>There are some limitations: TabPFN v2 is very fast to train and does not require hyperparameter tuning, but inference is slow. The model is also only designed for datasets up to 10k data points and 500 features. While it may perform well on larger datasets, it hasn't been our focus.<p>We're actively working on removing these limitations and intend to release new versions of TabPFN that can handle larger datasets, have faster inference and perform in additional predictive settings such as time-series and recommender systems.<p>We would love for you to try out TabPFN v2 and give us your feedback!</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=42647343">https://news.ycombinator.com/item?id=42647343</a></p>
<p>Points: 153</p>
<p># Comments: 44</p>
]]></description><pubDate>Thu, 09 Jan 2025 16:38:26 +0000</pubDate><link>https://www.nature.com/articles/s41586-024-08328-6/link</link><dc:creator>onasta</dc:creator><comments>https://news.ycombinator.com/item?id=42647343</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42647343</guid></item><item><title><![CDATA[New comment by onasta in "Why do tree-based models still outperform deep learning on tabular data?"]]></title><description><![CDATA[
<p>Super interesting! Do you know the kind of data that it's usually used for? And in the remaining 80% to 60%, do NNs acccount for a large portion of the best models?<p>Bonus question: are the stats you're mentioning publically available?</p>
]]></description><pubDate>Thu, 04 Aug 2022 21:20:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=32348430</link><dc:creator>onasta</dc:creator><comments>https://news.ycombinator.com/item?id=32348430</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=32348430</guid></item></channel></rss>