<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: ca508</title><link>https://news.ycombinator.com/user?id=ca508</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Wed, 15 Apr 2026 07:40:25 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=ca508" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by ca508 in "Show HN: Engine – A multi-LLM alternative to Codex"]]></title><description><![CDATA[
<p>been following engine from afar for a while, super cool to see it on HN.
didn't see it had a free plan, will try it out.</p>
]]></description><pubDate>Tue, 20 May 2025 12:50:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=44041002</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=44041002</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44041002</guid></item><item><title><![CDATA[New comment by ca508 in "Taming Servers for Fun and Profit"]]></title><description><![CDATA[
<p>Yep! We have two roles open <a href="https://railway.com/careers#open-positions" rel="nofollow">https://railway.com/careers#open-positions</a> - Datacenters and Infra that might be of interest to you. Idk if many of us use LinkedIn so not the best way to get in touch. Can I share the email on your HN profile link with our in-house recruiter?</p>
]]></description><pubDate>Wed, 26 Mar 2025 17:21:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=43484507</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=43484507</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43484507</guid></item><item><title><![CDATA[New comment by ca508 in "Taming Servers for Fun and Profit"]]></title><description><![CDATA[
<p>> the dots aren’t connected between a lot of the aspects presented<p>that's on me (author); I tried to cut the content down to a manageable post size that covered some interesting stuff - but probably dropped the connective tissue in the process. We'll keep this in mind for next time.</p>
]]></description><pubDate>Wed, 26 Mar 2025 02:12:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=43478146</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=43478146</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43478146</guid></item><item><title><![CDATA[New comment by ca508 in "Taming Servers for Fun and Profit"]]></title><description><![CDATA[
<p>Author here; RAM is so much more expensive than disk though. Two 500G M.2 NVMes for a RootFS in RAID1 are basically max ~$150 which is I think much cheaper than a single 64G DDR5 ECC RAM module [I don't have exact numbers on me, but ECC RAM is pricey]. It's also a lot harder to debug when things go wrong because you lose the machines state if everything was ephemeral.<p>We run a thin base OS on the boxes and then VMs on top which we consider more ephemeral. The frequency of needing to update that base OS is v. low.<p>I think there's a case for building a custom PXE booted RAMdisk image to replace the install though; something like what Equinix Metal (formerly Packet) do with Tinkerbell (<a href="https://github.com/tinkerbell" rel="nofollow">https://github.com/tinkerbell</a>); they call it an OS Install Environment, but the idea is a small lightweight linux install agent that can DD a golden image onto the disk (vs. an install each time).</p>
]]></description><pubDate>Wed, 26 Mar 2025 02:09:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=43478128</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=43478128</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43478128</guid></item><item><title><![CDATA[New comment by ca508 in "Taming Servers for Fun and Profit"]]></title><description><![CDATA[
<p>author here; Claude use here was pure laziness - and personally, I found it quite funny that it worked. We could sample pixels and try and build that detection, but $<1c per run to write a prompt and get some json was too hilarious not to ship to prod.<p>Maybe it needs to have more complex logic/detection and we need something more complex down the road. But it's like easy and cheap OCR for now.<p>what was kinda funnier was that I tried to get Claude to generate its own Go client code to upload the image and run the prompt; it totally totally hallucinated on that part :).</p>
]]></description><pubDate>Wed, 26 Mar 2025 01:53:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=43478039</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=43478039</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43478039</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>There's a fork called nautobot that tries to add-in automation. Most things we wanted to do with either meant we had to go writing django plugins and trying to interface with their APIs (and fight with the libs). Overall just hammering together a small custom service ended up being way faster/simpler.</p>
]]></description><pubDate>Sat, 18 Jan 2025 16:23:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=42749415</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42749415</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42749415</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>It's a fair question. What Oxide are building is cool, but it's too custom/monolithic for us to risk. We're more likely to look at OCP racks/chassis down the road.</p>
]]></description><pubDate>Sat, 18 Jan 2025 16:09:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=42749323</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42749323</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42749323</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>We currently pass on our cloud egress costs to users via the current pricing. We'll be publishing a pricing update soon as part of our migration - and egress [and some other things] will be coming down.</p>
]]></description><pubDate>Sat, 18 Jan 2025 16:05:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=42749293</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42749293</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42749293</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>We have a distributor we work with - just because it makes import/export a lot easier. But we get to interface directly with Supermicro for the technical/design stuff, and they're super awesome. If you're looking in the US, reach out to their eStore - really great fuss-free turnaround and all direct.</p>
]]></description><pubDate>Sat, 18 Jan 2025 03:22:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=42745516</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42745516</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42745516</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>the good news on this is that we've got a tonne of deep-dive material on networking and whitebox switches we cut from this post. We'll definitely be talking more about this soon (also cos' BGP is cool).</p>
]]></description><pubDate>Sat, 18 Jan 2025 03:10:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=42745458</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42745458</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42745458</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>correct; I think the first version of our tool sprung up in the space of a couple of weekends. It wasn't planned, my colleague Pierre who wrote it just had a lot of fun building it.</p>
]]></description><pubDate>Sat, 18 Jan 2025 01:34:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=42744989</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42744989</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42744989</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>We talked to a few, I think they're called MSPs? We weren't super impressed. We decided to YOLO it. There are probably great outfits out there, but it's hard to find them through the noise. We're mostly software and systems folks, but Railway is a infrastructure company so we need to own stuff down to the cage-nut - we owe it to our users. All engineering, project management and procurement is in-house.<p>We're lucky to have a few great distributors/manufacturers who help us pick the right gear. But we learnt a lot.<p>We've found a lot of value in getting a broker in to source our transit though.<p>My personal (and potentially misguided) hot take is that most of the baremetal world is stuck in the early 2000's, and the only companies doing anything interesting here the likes of AWS,Google and Meta. So the only way to innovate is to stumble around, escape the norms and experiment.</p>
]]></description><pubDate>Sat, 18 Jan 2025 01:31:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=42744971</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42744971</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42744971</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>All valid points - and our ideas for Gen 2 sound directionally similar - but those are at crayon drawing stage.<p>When we started, we didn't have much of an idea about what the rack needs to look like. So we chose a combination of things we thought we could pull this off. We're mostly software and systems folks, and there's a dearth of information out there on what to do. Vendors tend to gravitate towards selling BGP+EVPN+VXLAN or whatever "enterprise" reference designs; so we kinda YOLO'ed the Gen 1. We decided to spend extra money if we could get to a working setup sooner. When the clock is in cloud spend, there's uh... lots of opportunity cost :D.<p>A lot of the chipset and switch choices were bets and we had to pick and choose what we gambled on - and what we could get our hands on. The main bets this round were eBGP to the hosts with BGP unnumbered, SONiC switches - this lets us do a lot of networking with our existing IPv6/Wireguard/eBPF overlay and a debian based switch OS + FRR (so fewer things to learn). And ofc. figuring out how to operationalise the install process and get stuff running on the hardware as soon as possible.<p>Now we've got a working design, we'll start iterating a bit more on the hardware choice and network design. I'd love for us to write about it when we get through it. Plus I think we owe the internet a rant on networking in general.<p>Edit: Also we don't use UniFi Pro / Uniquity gear anywhere?</p>
]]></description><pubDate>Sat, 18 Jan 2025 01:10:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=42744870</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42744870</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42744870</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>ah that's my bad - I wrote this in Dec, we only published in Jan. Obv. missed updating that.<p>Timeline wise; 
- we decided to go for it and spend the $$$ in Oct '23
- Convos/planning started ~ Jan '24
- Picked the vendors we wanted by ~ Feb/Mar '24
- Lead-times, etc... meant everything was ready for us to go fit the first gear by mostly ourselves at the start of May (that's the 5mo)
- We did the "proper" re-install around June, followed closely by the second site in ~ Sep, around when we started letting our users on it as a open beta
- Sep-Dec we just doubled down on refining software/automation and process while building out successive installs<p>Lead times can be mind numbing. We have certain switches from Arista that have a 3-6 mo leadtime. Servers are build to order, so again 2+ months depending on stock. And obv. holidays mean a lot of stuff shuts down around December.<p>Sometimes you can swap stuff around to get better lead-times, but then the operational complexity explodes because you have this slightly different component at this one site.<p>I used to be a EEE, and I thought supply chain there was bad. But with DCs I think it's sometimes worse because you don't directly control some parts of your BoM/supply chain (especially with build-to-order servers).</p>
]]></description><pubDate>Sat, 18 Jan 2025 00:43:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=42744740</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42744740</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42744740</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>oh yes we want to; I even priced a couple out. Most of the SKUs I found were pretty old, and we couldn't find anything compelling to risk deploying at the scale we wanted.
It's on the wishlist, and if the right hardware comes along; we'll rack it up even  as a bet. We maintain Nixpacks (<a href="https://nixpacks.com/docs/getting-started" rel="nofollow">https://nixpacks.com/docs/getting-started</a>), so for most of our users we could rebuild most their apps for ARM seamlessly - infact we mostly develop our build systems on ARM (because macbooks). One day.</p>
]]></description><pubDate>Sat, 18 Jan 2025 00:05:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=42744552</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42744552</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42744552</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>Long running hosts at the moment, but we can drain most workloads off a specific host/rack if required and reschedule it pretty fast. We have the advantage of having a custom scheduler/orchestrator we've been working on for years, so we have a lot of control on that layer than with Kube or Nomad.<p>Re: Live Migration
We're working on adding Live Migration support to our orchestrator atm. We aim to have it running this quarter. That'll makes things super seamless.<p>Re: kernels
We've already seen some perf improvements somewhere between 6.0 and 6.5 (I forget the exact reason/version) - but it was some fix specific to the Sapphire Rapids cpus we had. But I wish we had more time to science on it, it's really fun playing with all the knobs and benchmarking stuff. Some of the telemetry on the new CPUs is also crazy - there's stuff like Intel PCM that can pull super fine-grained telemetry direct from the CPU/chipset <a href="https://github.com/intel/pcm">https://github.com/intel/pcm</a>. Only used it to confirm that we got NUMA affinity right so far - nothing crazy.</p>
]]></description><pubDate>Fri, 17 Jan 2025 23:46:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=42744454</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42744454</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42744454</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>we evaluated a lot of commercial and oss offerings before we decided do go build it ourselves - we still have a deploy of netbox somewhere. But our custom tool (Railyard) works so well because it integrates deeply into the our full software, hardware and orchestration stack. The problem with the OSS stuff is that it's almost too generic - you shape the problem to fit its data model vs. solve the problem. We're likely going to fold our tool into Railway itself eventually - want to go on-prem; button click hardware design, commission, deploy and devex. Sorta like what Oxide is doing, but approaching the problem from the opposite side.</p>
]]></description><pubDate>Fri, 17 Jan 2025 23:26:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=42744318</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42744318</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42744318</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>> It would be nice to have a lot more detail<p>I'm going to save this for when I'm asked to cut the three paras on power circuit types.<p>Re: standardising layout at the rack level; we do now! we only figured this out after site #2. It makes everything so much easier to verify. And yeah, validation is hard - manually doing it thus far; want to play around with scraping LLDP data but our switch software stack has a bug :/. It's an evolving process, the more we work with different contractors, the more edge cases we unearth and account for. The biggest improvement is that we have built a internal DCIM that templates a rack design and exports a interactive "cabling explorer" for the site techs - including detailed annotated diagrams of equipment showing port names, etc... The screenshot of the elevation is a screenshot of part of that tool.<p>> What does your metal->boot stack look like?<p>We've hacked together something on top of <a href="https://github.com/danderson/netboot/tree/main/pixiecore">https://github.com/danderson/netboot/tree/main/pixiecore</a> that serves a debian netboot + preseed file. We have some custom temporal workers to connect to Redfish APIs on the BMCs to puppeteer the contraption. Then a custom host agent to provision QEMU VMs and advertise assigned IPs via BGP (using FRR) from the host.<p>Re: new DCs for failure scenarios, yeah we've already blown breakers etc... testing stuff (that's how we figured out our phase balancing was off). Went in with a thermal camera on another. A site in AMS is coming up next week and the goal for that is to see how far we can push a fully loaded switch fabric.</p>
]]></description><pubDate>Fri, 17 Jan 2025 22:58:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=42744133</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42744133</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42744133</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>We didn't find many good up-to-date resources online on the hardware side of things - kinda why we wanted to write about it. The networking aspect was the most mystical - I highly recommend "BGP in the datacenter" by Dinesh Dutt on that (I think it's available for free via NVidia). Our design is heavily influenced by the ideas discussed there.</p>
]]></description><pubDate>Fri, 17 Jan 2025 22:33:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=42743979</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42743979</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42743979</guid></item><item><title><![CDATA[New comment by ca508 in "So you want to build your own data center"]]></title><description><![CDATA[
<p>We built some internal tooling to help manage the hosts. Once a host is onboarded onto it, it's a few button clicks on an internal dashboard to provision a QEMU VM. We made a custom ansible inventory plugin so we can manage these VMs the same as we do machines on GCP.<p>The host runs a custom daemon that programs FRR (an OSS routing stack), so that it advertises addresses assigned to a VM to the rest of the cluster via BGP. So zero config of network switches, etc... required after initial setup.<p>We'll blog about this system at some point in the coming months.</p>
]]></description><pubDate>Fri, 17 Jan 2025 22:22:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=42743901</link><dc:creator>ca508</dc:creator><comments>https://news.ycombinator.com/item?id=42743901</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42743901</guid></item></channel></rss>