<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: shivampkumar</title><link>https://news.ycombinator.com/user?id=shivampkumar</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 20 Apr 2026 19:59:24 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=shivampkumar" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by shivampkumar in "Show HN: Run TRELLIS.2 Image-to-3D generation natively on Apple Silicon"]]></title><description><![CDATA[
<p>that makes so much sense...I am exploring if I can find someone who has done this well...If not I'll try to do it myself.</p>
]]></description><pubDate>Mon, 20 Apr 2026 08:33:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=47831648</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47831648</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47831648</guid></item><item><title><![CDATA[New comment by shivampkumar in "Show HN: Run TRELLIS.2 Image-to-3D generation natively on Apple Silicon"]]></title><description><![CDATA[
<p>The model needed about 15GB at peak during generation - the 4B model loads multiple sub-models (1.3B each for shape and texture flow). 8GB won't be enough, but both 24GB and 32GB both should be fine.</p>
]]></description><pubDate>Mon, 20 Apr 2026 04:45:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=47830491</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47830491</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47830491</guid></item><item><title><![CDATA[New comment by shivampkumar in "Show HN: Run TRELLIS.2 Image-to-3D generation natively on Apple Silicon"]]></title><description><![CDATA[
<p>added! 
will add more, maybe even a GIF</p>
]]></description><pubDate>Mon, 20 Apr 2026 03:57:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=47830255</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47830255</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47830255</guid></item><item><title><![CDATA[New comment by shivampkumar in "Show HN: Run TRELLIS.2 Image-to-3D generation natively on Apple Silicon"]]></title><description><![CDATA[
<p>i was able to get it in 3.5 mins from a single image on my 24gb m4 pro macbook<p>I'm still working on this to try to replicate nvdiffrast better. Found an open source port, might look it tonight</p>
]]></description><pubDate>Mon, 20 Apr 2026 03:37:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=47830156</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47830156</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47830156</guid></item><item><title><![CDATA[New comment by shivampkumar in "Show HN: TRELLIS.2 image-to-3D running on Mac Silicon – no Nvidia GPU needed"]]></title><description><![CDATA[
<p>thanks!</p>
]]></description><pubDate>Mon, 20 Apr 2026 03:32:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=47830133</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47830133</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47830133</guid></item><item><title><![CDATA[New comment by shivampkumar in "Show HN: Run TRELLIS.2 Image-to-3D generation natively on Apple Silicon"]]></title><description><![CDATA[
<p>I mean I can see that it's niche. Did not expect so many upvotes, but ig it's less niche than I tought<p>If you're not working with 3D on Apple Silicon this isn't relevant to you. For the
subset of people who are, running this 4B parameter 3D generation model locally on a Mac was previously blocked by hard CUDA dependencies with no workaround.</p>
]]></description><pubDate>Mon, 20 Apr 2026 03:27:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=47830107</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47830107</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47830107</guid></item><item><title><![CDATA[New comment by shivampkumar in "Show HN: TRELLIS.2 image-to-3D running on Mac Silicon – no Nvidia GPU needed"]]></title><description><![CDATA[
<p>I thought it was cool and then I found the open issue mentioned above, that convinced me its def something more people want.<p>It IS significantly slower, about 3.5 minutes on my MacBook vs seconds on an H100. That's partly the pure-PyTorch backend overhead and partly just the hardware difference.<p>For my use case the tradeoff works -- iterate locally without paying for cloud GPUs or waiting in queues.</p>
]]></description><pubDate>Mon, 20 Apr 2026 03:26:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47830091</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47830091</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47830091</guid></item><item><title><![CDATA[New comment by shivampkumar in "Show HN: Run TRELLIS.2 Image-to-3D generation natively on Apple Silicon"]]></title><description><![CDATA[
<p>IMO TRELLIS.2 is slightly different case from the HF models scenario. It depends on five compiled CUDA-only extensions -- flex_gemm for sparse convolution, flash_attn, o_voxel for CUDA hashmap ops, cumesh for mesh processing, and nvdiffrast for differentiable rasterization. These aren't PyTorch ops that fall back to MPS -- they're custom C++/CUDA kernels. The upstream setup.sh literally exits with "No supported GPU found" if nvidia-smi isn't present. The only reason I picked this up because I thought it was cool and no one was working on this open issue for Silicon back then (github.com/microsoft/TRELLIS.2/issues/74) requesting non-CUDA support.</p>
]]></description><pubDate>Mon, 20 Apr 2026 03:22:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47830057</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47830057</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47830057</guid></item><item><title><![CDATA[New comment by shivampkumar in "Show HN: Run TRELLIS.2 Image-to-3D generation natively on Apple Silicon"]]></title><description><![CDATA[
<p>You're right, thanks for flagging this, let me run something and push images</p>
]]></description><pubDate>Mon, 20 Apr 2026 03:05:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=47829963</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47829963</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47829963</guid></item><item><title><![CDATA[New comment by shivampkumar in "Show HN: Run TRELLIS.2 Image-to-3D generation natively on Apple Silicon"]]></title><description><![CDATA[
<p>Hey, thanks for sharing this. I'm sure TRELLIS.2 definitely has room to improve, especially on texturing.<p>From what I've seen personally, and community benchmarks, it does fair on geometry and visual fidelity among open-source options, but I agree it's not perfect for every use case.<p>Meshy is solid, I used it to print my girlfriend a mini 3d model of her on her birthday last year!<p>Though worth noting it's a paid service, and free tier has usage limitations while TRELLIS.2 is MIT licensed with unlimited local generation. Different tradeoffs for different workflows. Hopefully the open-source side keeps improving.</p>
]]></description><pubDate>Mon, 20 Apr 2026 03:03:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47829945</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47829945</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47829945</guid></item><item><title><![CDATA[Show HN: Run TRELLIS.2 Image-to-3D generation natively on Apple Silicon]]></title><description><![CDATA[
<p>I ported Microsoft's TRELLIS.2 (4B parameter image-to-3D model) to run on Apple Silicon via PyTorch MPS. The original requires CUDA with flash_attn, nvdiffrast, and custom sparse convolution kernels:  none of which work on Mac.<p>I replaced the CUDA-specific ops with pure-PyTorch alternatives: a gather-scatter sparse 3D convolution, SDPA attention for sparse transformers, and a Python-based mesh extraction replacing CUDA hashmap operations. Total changes are a few hundred lines across 9 files.<p>Generates ~400K vertex meshes from single photos in about 3.5 minutes on M4 Pro (24GB). Not as fast as H100 (where it takes seconds), but it works offline with no cloud dependency.<p><a href="https://github.com/shivampkumar/trellis-mac" rel="nofollow">https://github.com/shivampkumar/trellis-mac</a></p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47828896">https://news.ycombinator.com/item?id=47828896</a></p>
<p>Points: 192</p>
<p># Comments: 34</p>
]]></description><pubDate>Mon, 20 Apr 2026 00:07:16 +0000</pubDate><link>https://github.com/shivampkumar/trellis-mac</link><dc:creator>shivampkumar</dc:creator><comments>https://news.ycombinator.com/item?id=47828896</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47828896</guid></item></channel></rss>