<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: samdhar</title><link>https://news.ycombinator.com/user?id=samdhar</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 07 May 2026 14:25:25 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=samdhar" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by samdhar in "DeepSeek V4 Pro at 75% off until 31 May"]]></title><description><![CDATA[
<p>Cached input at $0.003625/M, output at $0.435/M. Aggressive pricing.<p>For anyone doing the "should I self-host on rented GPUs?" math: at this rate you'd need to push roughly 1B output tokens/day to break even against an 8xH100 fleet on Vast/Lambda (assuming 3-5k tokens/sec aggregate throughput). The vast majority of "I should run my own LLM" use cases don't come close to that volume.<p>Every API price drop kills another tranche of "self-host the open model" use cases. The implied bet: even if regular pricing ($1.74/M output) is also subsidized, exponential demand growth eventually makes the unit economics work. We'll see.</p>
]]></description><pubDate>Thu, 07 May 2026 00:36:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=48043928</link><dc:creator>samdhar</dc:creator><comments>https://news.ycombinator.com/item?id=48043928</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48043928</guid></item></channel></rss>