<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: LegNeato</title><link>https://news.ycombinator.com/user?id=LegNeato</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 15 Apr 2026 09:51:56 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=LegNeato" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by LegNeato in "Rust Threads on the GPU"]]></title><description><![CDATA[
<p>We are the maintainers of <a href="https://github.com/rust-gpu/rust-gpu" rel="nofollow">https://github.com/rust-gpu/rust-gpu</a> and <a href="https://github.com/Rust-GPU/Rust-CUDA" rel="nofollow">https://github.com/Rust-GPU/Rust-CUDA</a> FWIW. We haven't upstreamed the VectorWare work yet as it is still being cleaned up and iterated on.</p>
]]></description><pubDate>Wed, 15 Apr 2026 02:13:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=47773862</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47773862</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47773862</guid></item><item><title><![CDATA[New comment by LegNeato in "Rust Threads on the GPU"]]></title><description><![CDATA[
<p>Agreed, and thank you.</p>
]]></description><pubDate>Wed, 15 Apr 2026 02:11:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47773852</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47773852</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47773852</guid></item><item><title><![CDATA[New comment by LegNeato in "Rust Threads on the GPU"]]></title><description><![CDATA[
<p>It is not, we just haven't yet upstreamed everything.</p>
]]></description><pubDate>Tue, 14 Apr 2026 05:09:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=47761469</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47761469</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47761469</guid></item><item><title><![CDATA[New comment by LegNeato in "Rust Threads on the GPU"]]></title><description><![CDATA[
<p>It depends. At VecorWare are a bit of an extreme case in that we are inverting the relationship and making the GPU the main loop that calls out to the CPU sparingly. So in that model, yes. If your code is run in a more traditional model (CPU driving and using GPU as a coprocessor), probably not. Going across the bus dominates most workloads. That being said, the traditional wisdom is becoming less relevant as integrated memory is popping up everywhere and tech like GPUDirect exists with the right datacenter hardware.<p>These are the details we intend to insulate people from so they can just write code and have it run fast. There is a reason why abstractions were invented on the CPU and we think we are at that point for the GPU.<p>(for the datacenter folks I know hardware topology has a HUGE impact that software cannot overcome on its own in many situations)</p>
]]></description><pubDate>Tue, 14 Apr 2026 05:04:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=47761440</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47761440</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47761440</guid></item><item><title><![CDATA[New comment by LegNeato in "Rust Threads on the GPU"]]></title><description><![CDATA[
<p>Founder here.<p>1. Programming GPUs is a problem. The ratio of CPUs to CPU programmers and GPUs to GPU programmers is massively out of whack. Not because GPU programming is less valuable or lucrative, because GPUs are weird and the tools are weird.<p>2. We are more interested in leveraging existing libraries than running existing binaries wholesale (mostly within a warp). But, running GPU-unaware code leaves a lot of space for the compiler to move stuff around and optimize things.<p>3. The compiler changes are not our product, the GPU apps we are building with them are. So it is in our interest to make the apps very fast.<p>Anyway, skepticism is understandable and we are well aware code wins arguments.</p>
]]></description><pubDate>Tue, 14 Apr 2026 03:44:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47760997</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47760997</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47760997</guid></item><item><title><![CDATA[New comment by LegNeato in "The economics of software teams: Why most engineering orgs are flying blind"]]></title><description><![CDATA[
<p>No, it means not being able to see what is going on. Which is literally what the word blind means. You can be blinded by many things (blindfold, clouds/fog, bright lights, darkness, accidents, genetics, etc), permanently and temporarily. Non-humans can be blind and blinded. YOU are making it about a specific situation and projecting value judgements on it.<p>The author specifically says FLYING blind. Not "stumbling around like a blind person" or some such.  If you are offended, that is on you. It's your right to be offended of course, but don't expect people to join in your delusion.</p>
]]></description><pubDate>Mon, 13 Apr 2026 07:51:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=47749033</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47749033</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47749033</guid></item><item><title><![CDATA[New comment by LegNeato in "Taking on CUDA with ROCm: 'One Step After Another'"]]></title><description><![CDATA[
<p>One of the rust-gpu maintainers here. Haven't officially heard from anyone at AMD but we've had chats with many others. Happy to talk with whomever! I would imagine AMD is focusing on ROCm over Vulkan for compute right now as their pure datacenter play, which makes sense.<p>We've started a company around Rust on the GPU btw (<a href="https://www.vectorware.com/" rel="nofollow">https://www.vectorware.com/</a>), both CUDA and Vulkan (and ROCm eventually I guess?).<p>Note that most platform developers in the GPU space are C++ folks (lots of LLVM!) and there isn't as much demand from customers for Rust on the GPU vs something like Python or Typescript. So Rust naturally gets less attention and is lower on the list...for now.</p>
]]></description><pubDate>Mon, 13 Apr 2026 02:49:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=47746997</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47746997</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47746997</guid></item><item><title><![CDATA[Rust Threads on the GPU]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.vectorware.com/blog/threads-on-gpu/">https://www.vectorware.com/blog/threads-on-gpu/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47698975">https://news.ycombinator.com/item?id=47698975</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 09 Apr 2026 03:30:10 +0000</pubDate><link>https://www.vectorware.com/blog/threads-on-gpu/</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47698975</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47698975</guid></item><item><title><![CDATA[New comment by LegNeato in "Async/Await on the GPU"]]></title><description><![CDATA[
<p>Thank you! We're small so have to focus. If anyone from AMD wants to reach out, happy to chat.</p>
]]></description><pubDate>Tue, 17 Feb 2026 21:29:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=47053636</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47053636</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47053636</guid></item><item><title><![CDATA[New comment by LegNeato in "Async/Await on the GPU"]]></title><description><![CDATA[
<p>Doing things at compile time / AOT is almost always better for perf. We believe async/await and futures enables more complex programs and doing things you couldn't easily do on the GPU before. Less about performance and more about capability (though we believe async/await perf will be better in some cases, time will tell).</p>
]]></description><pubDate>Tue, 17 Feb 2026 20:17:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=47052695</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47052695</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47052695</guid></item><item><title><![CDATA[New comment by LegNeato in "Async/Await on the GPU"]]></title><description><![CDATA[
<p>Currently NVIDIA-only, we're cooking up some Vulkan stuff in rust-gpu though.</p>
]]></description><pubDate>Tue, 17 Feb 2026 18:10:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=47050803</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47050803</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47050803</guid></item><item><title><![CDATA[New comment by LegNeato in "Async/Await on the GPU"]]></title><description><![CDATA[
<p>We aren't focused on performance yet (it is often workload and executor dependent, and as the post says we currently do some inefficient polling) but Rust futures compile down to state machines so they are a zero-cost abstraction.<p>The anticipated benefits are similar to the benefits of async/await on CPU: better ergonomics for the developer writing concurrent code, better utilization of shared/limited resources, fewer concurrency bugs.</p>
]]></description><pubDate>Tue, 17 Feb 2026 18:10:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=47050801</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47050801</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47050801</guid></item><item><title><![CDATA[New comment by LegNeato in "Async/Await on the GPU"]]></title><description><![CDATA[
<p>Yes, that's the idea.<p>GPU-wide memory is not quite as scarce on datacenter cards or systems with unified memory. One could also have local executors with local futures that are `!Send` and place in a faster address space.</p>
]]></description><pubDate>Tue, 17 Feb 2026 17:57:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=47050606</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=47050606</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47050606</guid></item><item><title><![CDATA[New comment by LegNeato in "Rust’s Standard Library on the GPU"]]></title><description><![CDATA[
<p>We use the cuda device allocator for allocations on the GPU via Rust's default allocator.</p>
]]></description><pubDate>Wed, 28 Jan 2026 03:46:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=46790815</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=46790815</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46790815</guid></item><item><title><![CDATA[New comment by LegNeato in "Rust’s Standard Library on the GPU"]]></title><description><![CDATA[
<p>Flip on the pedantic switch. We have std::fs, std::time, some of std::io, and std::net(!). While the `libc` calls go to the host, all the `std` code in-between runs on the GPU.</p>
]]></description><pubDate>Wed, 28 Jan 2026 03:42:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=46790789</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=46790789</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46790789</guid></item><item><title><![CDATA[New comment by LegNeato in "Rust’s Standard Library on the GPU"]]></title><description><![CDATA[
<p>Author here! Flip on the pedantic switch, we agree ;-)</p>
]]></description><pubDate>Wed, 28 Jan 2026 03:41:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=46790779</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=46790779</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46790779</guid></item><item><title><![CDATA[New comment by LegNeato in "VectorWare – from creators of `rust-GPU` and `rust-CUDA`"]]></title><description><![CDATA[
<p>1. The GPU owns the control loop And the only sparingly kicks to the CPU when it can't do something.<p>2. Yes<p>3. We're still investigating the limitations. A lot of them are hardware dependent, obviously data center cards have higher limits more capability than desktop cards.<p>Thanks! It is super fun trailblazing and realizing more of the pieces are there than everybody expects.</p>
]]></description><pubDate>Fri, 24 Oct 2025 11:43:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=45693604</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=45693604</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45693604</guid></item><item><title><![CDATA[New comment by LegNeato in "VectorWare – from creators of `rust-GPU` and `rust-CUDA`"]]></title><description><![CDATA[
<p>You might be interested in a previous blog post where we showed one codebase running on many types of GPUs: <a href="https://rust-gpu.github.io/blog/2025/07/25/rust-on-every-gpu/" rel="nofollow">https://rust-gpu.github.io/blog/2025/07/25/rust-on-every-gpu...</a></p>
]]></description><pubDate>Thu, 23 Oct 2025 21:59:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=45687832</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=45687832</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45687832</guid></item><item><title><![CDATA[New comment by LegNeato in "VectorWare – from creators of `rust-GPU` and `rust-CUDA`"]]></title><description><![CDATA[
<p>No worries, just wanted to correct it for folks. Thanks for posting!</p>
]]></description><pubDate>Thu, 23 Oct 2025 17:57:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=45684793</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=45684793</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45684793</guid></item><item><title><![CDATA[New comment by LegNeato in "VectorWare – from creators of `rust-GPU` and `rust-CUDA`"]]></title><description><![CDATA[
<p>One of the founders here, feel free to ask whatever. We purposefully didn't put much technical detail in the post as it is an announcement post (other people posted it here, we didn't).</p>
]]></description><pubDate>Thu, 23 Oct 2025 17:44:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=45684645</link><dc:creator>LegNeato</dc:creator><comments>https://news.ycombinator.com/item?id=45684645</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45684645</guid></item></channel></rss>