<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: OliverGuy</title><link>https://news.ycombinator.com/user?id=OliverGuy</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 17 Apr 2026 11:30:54 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=OliverGuy" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by OliverGuy in "Ask HN: What Are You Working On? (April 2026)"]]></title><description><![CDATA[
<p>Just started this weekend on <a href="https://gitlab.com/get-otter/otter-sdk" rel="nofollow">https://gitlab.com/get-otter/otter-sdk</a><p>Its dbt inspired stream ETL tool (or maybe just the TL?), it currently just has a dev mode that does RabbitMQ to local Parque files while I'm getting the core of it to a place I'm happy with.<p>It runs SQL models against the incoming messages and outputs the results to one or more output tables. Has a local WAL so you can tune it to have sensible sized output files (or not, if you need regular updates but at the expense of query perf.)<p>Planning on adding Protobuf messages, Kafka as a source and S3 and Iceberg tables as sinks this week.<p>Lightly inspired by a some projects at work where a lot of time and effort was spent doing this and resulted in something not very reusable without a lot of refactor work. Feel like the stream -> data lake pattern should be something that is just SQL + Config, same way dbt is for transformations within a data warehouse.<p>No plans on adding any cross message joins or aggregations as that would require cross worker communications and I explicitly want to keep the workers stateless (minus the WAL of course)<p>Would really appreciate any feedback on the core concept, esp. if this is something you'd actually use in prod (if it were finished!) Not sure if there is something that does this already that I don't know about, or if this genuinely fills some sort of hole in the exisitng tooling</p>
]]></description><pubDate>Mon, 13 Apr 2026 13:43:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47751861</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=47751861</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47751861</guid></item><item><title><![CDATA[New comment by OliverGuy in "Anthropic expands partnership with Google and Broadcom for next-gen compute"]]></title><description><![CDATA[
<p>And you know Nvidia can't be constent with one format for FLOPs within a single graph, 1,000,000x faster but comparing FP32 to FP8 or NVFP4 and acting like it's the same.</p>
]]></description><pubDate>Tue, 07 Apr 2026 12:05:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=47673931</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=47673931</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47673931</guid></item><item><title><![CDATA[New comment by OliverGuy in "Issue: Claude Code is unusable for complex engineering tasks with Feb updates"]]></title><description><![CDATA[
<p>settings.json -> global config
Env vars -> settings different to your global for a specific project
Slash commands / chat keywords -> need to change a setting mid chat</p>
]]></description><pubDate>Tue, 07 Apr 2026 07:34:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=47671879</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=47671879</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47671879</guid></item><item><title><![CDATA[New comment by OliverGuy in "Google's 200M-parameter time-series foundation model with 16k context"]]></title><description><![CDATA[
<p>Edit, it looks like the paper does<p>TPUv5e with 16 tensor cores for 2 days for the 200M param model.<p>Claude reckons this is 60 hours on a 8xA100 rig, so very accessibile compared to LLMs for smaller labs</p>
]]></description><pubDate>Tue, 31 Mar 2026 07:27:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=47583894</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=47583894</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47583894</guid></item><item><title><![CDATA[New comment by OliverGuy in "Google's 200M-parameter time-series foundation model with 16k context"]]></title><description><![CDATA[
<p>Wish they gave some numbers for total GPU hours to train this model, seems comparatively tiny when compared to LLMs so interested to know how close this is to something trainable by your average hobbyist/university/small lab</p>
]]></description><pubDate>Tue, 31 Mar 2026 07:14:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=47583808</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=47583808</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47583808</guid></item><item><title><![CDATA[New comment by OliverGuy in "OpenAI is walking away from expanding its Stargate data center with Oracle"]]></title><description><![CDATA[
<p>Interested to know more about your inference start up? How you guys operating, do you own hardware or use the cloud?</p>
]]></description><pubDate>Tue, 10 Mar 2026 12:00:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=47322061</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=47322061</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47322061</guid></item><item><title><![CDATA[New comment by OliverGuy in "Ask HN: What Are You Working On? (March 2026)"]]></title><description><![CDATA[
<p><a href="https://gitlab.com/usecaliper/caliper-python-sdk" rel="nofollow">https://gitlab.com/usecaliper/caliper-python-sdk</a><p>An LLM observability SDK that let's you store pre and post request metadata with every call in as lightweight an SDK as possible.<p>Stores to S3 in batched JSON files, so can easily plug into existing tooling like DuckDB for analysis.<p>It's designed to answer questions like; "how do different user tiers of my services rate this two different models and three different systems prompts?". You can capture all the information required to answer this in the SDK and do some queries over the data to get the answers.</p>
]]></description><pubDate>Mon, 09 Mar 2026 06:55:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47305603</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=47305603</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47305603</guid></item><item><title><![CDATA[Show HN: Caliper – Auto Instrumented LLM Observability with Custom Metadata]]></title><description><![CDATA[
<p>Caliper is designed to auto instrument LLM calls within Python, it monkey patches the OpenAI and Anthropic SDKs (Got plans to add LiteLLM so you can use any provider you want to) so it's almost completely invisible to you as the developer and for basic metrics can slot in as a single init() at start.<p>It can also gather custom metadata about a call, this can be any KV pairs you want, both pre and post request.<p>```python<p>import caliper<p>import anthropic<p>caliper.init(target="s3") # This is all that's required for basic observability, no changes needed to LLM calls for basic metrics<p>client = anthropic.Anthropic()<p>response = client.messages.create(<p><pre><code>  model="claude-sonnet-4-20250514",

  messages=[{"role": "user", "content": "What is 2 + 2?"}],

  caliper_metadata={"campaign": "q4"}, # Pre request metadata
</code></pre>
)<p>print(response.content[0].text)<p>caliper.annotate(sentiment="positive") # Post request metadata<p>```<p>You can use this to track effectiveness of model changes, tracking them against difference user tiers. Maybe your free tier users don't notice if you use a cheaper model but you paying users do? How do you know if a recent system prompt change was effective? You can track the version of the prompt in metadata and compare post request rating annotations between prompt versions.<p>It has a dev mode which logs locally, it can also send files to S3. The SDK has a background queue and worker which flushes in batches that are configurable in size and time between flushes. It exports to S3 as batched JSON files to readily to integrate into most data engineering pipelines or you can just query directly with a tool like DuckDB.<p>PyPi: <a href="https://pypi.org/project/caliper-sdk/" rel="nofollow">https://pypi.org/project/caliper-sdk/</a><p>Edits: formatting and PyPi Link</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47296746">https://news.ycombinator.com/item?id=47296746</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Sun, 08 Mar 2026 12:21:35 +0000</pubDate><link>https://gitlab.com/usecaliper/caliper-python-sdk</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=47296746</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47296746</guid></item><item><title><![CDATA[New comment by OliverGuy in "Facebook's Fascination with My Robots.txt"]]></title><description><![CDATA[
<p>That's still the same crawler system though. And it's lazy engineering to not build in something to track when you last requested a url.<p>And it's quite a trivial feature at that.</p>
]]></description><pubDate>Mon, 23 Feb 2026 14:00:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=47122415</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=47122415</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47122415</guid></item><item><title><![CDATA[New comment by OliverGuy in "More than DNS: Learnings from the 14 hour AWS outage"]]></title><description><![CDATA[
<p>Tbh, for most companies/orgs the cost/complexity of multi region just isn't worth it.<p>The cost of a work days worth of downtime is rarely enough to justify the expense of trying to deploy across multiple regions or clouds.<p>Esp if you are public facing and not internal. You just go 'well everyone else was down to because of aws' and your customers just go 'ah okay fair enough'</p>
]]></description><pubDate>Thu, 30 Oct 2025 16:26:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=45761821</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=45761821</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45761821</guid></item><item><title><![CDATA[New comment by OliverGuy in "Ubiquiti SFP Wizard"]]></title><description><![CDATA[
<p>Cisco etc have truly insane pricing on optics, like $1000 for something generic that cost $20-50 from fs.com etc. The only difference is how it presents itself to the switch (ie, says its a Cisco optic), not actual difference in performance.<p>Often Cisco/etc will refuse support cases if you aren't using their optics, if the switches/routers even work with them in the first case, which isn't a given as often they'll refuse to work with non branded optics.<p>Really just a money grab by the big network vendors.<p>This box allows you to flash the firmware on the optic to say its from whatever brand you want (Cisco, Dell, Aruba, Juniper etc) so that you can get it to work in that companies switch/router.<p>For most SMEs, the brand of optics makes no difference. Maybe keep a few legit branded ones around for debugging and when you need to raise a support case. But otherwise, the generic ones flashed to look like branded ones work just fine.</p>
]]></description><pubDate>Tue, 28 Oct 2025 16:09:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=45734730</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=45734730</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45734730</guid></item><item><title><![CDATA[New comment by OliverGuy in "AWS multiple services outage in us-east-1"]]></title><description><![CDATA[
<p>Their latest update on the status page says it's a Dynamodb DNS issue</p>
]]></description><pubDate>Mon, 20 Oct 2025 09:14:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=45641701</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=45641701</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45641701</guid></item><item><title><![CDATA[New comment by OliverGuy in "NVIDIA DGX Spark In-Depth Review: A New Standard for Local AI Inference"]]></title><description><![CDATA[
<p>How representative is this platform of the bigger GB200 and GB300 chips?<p>Could I write code that runs on Spark and effortlessly run it on a big GB300 system with no code changes?</p>
]]></description><pubDate>Tue, 14 Oct 2025 09:39:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=45578008</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=45578008</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45578008</guid></item><item><title><![CDATA[New comment by OliverGuy in "Building the heap: racking 30 petabytes of hard drives for pretraining"]]></title><description><![CDATA[
<p>The NetApps are just disk shelves, can plug it into a SAS controller and use whatever software stack you please.</p>
]]></description><pubDate>Wed, 01 Oct 2025 17:59:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=45440884</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=45440884</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45440884</guid></item><item><title><![CDATA[New comment by OliverGuy in "Building the heap: racking 30 petabytes of hard drives for pretraining"]]></title><description><![CDATA[
<p>Aren't those netapp shelves pretty old at this point? See a lot of people recommending against them even for homelab type uses. You can get those 60 drive SuperMicro JBODs for pretty cheap now, and those aren't too old, would have been my choice.<p>Plus, the TCO is already way under the cloud equiv. so might as well spend a little more to get something much newer and more reliable</p>
]]></description><pubDate>Wed, 01 Oct 2025 17:58:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=45440871</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=45440871</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45440871</guid></item><item><title><![CDATA[New comment by OliverGuy in "Launch HN: Datafruit (YC S25) – AI for DevOps"]]></title><description><![CDATA[
<p>I can put some AWS Creds in my terminal and Claude Code is perfectly happy writing AWS CLI commands (or whole python scripts if necessary) to work out what it needs to about my infrastructure.</p>
]]></description><pubDate>Wed, 03 Sep 2025 06:59:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=45113006</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=45113006</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45113006</guid></item><item><title><![CDATA[New comment by OliverGuy in "Amazon has mostly sat out the AI talent war"]]></title><description><![CDATA[
<p>SageMaker have serverless inference endpoints</p>
]]></description><pubDate>Tue, 02 Sep 2025 13:39:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=45102967</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=45102967</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45102967</guid></item><item><title><![CDATA[New comment by OliverGuy in "Running GPT-OSS-120B at 500 tokens per second on Nvidia GPUs"]]></title><description><![CDATA[
<p>If you are trying to get facts out of an LLM you are using it wrong, if you want a fact it should use a tool (eg we search, rag etc) to get the information that contains the fact (Wikipedia page, documentation etc) and then parse that document for the fact and return it to you.</p>
]]></description><pubDate>Thu, 07 Aug 2025 10:33:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=44822800</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=44822800</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44822800</guid></item><item><title><![CDATA[New comment by OliverGuy in "Running GPT-OSS-120B at 500 tokens per second on Nvidia GPUs"]]></title><description><![CDATA[
<p>Nope, doesn't have any of the required hardware to even process graphics iirc</p>
]]></description><pubDate>Thu, 07 Aug 2025 10:31:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=44822787</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=44822787</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44822787</guid></item><item><title><![CDATA[New comment by OliverGuy in "Blackmagic Debuts $30K 3D Camera for Capturing Video for Vision Pro"]]></title><description><![CDATA[
<p>8TB for 2hrs of footage is crazy even compared to other high end cinema camera, going to be an interesting work flow for anyone editing this as thats not a a trivial amount of data even by today's standards</p>
]]></description><pubDate>Thu, 19 Dec 2024 11:09:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=42460360</link><dc:creator>OliverGuy</dc:creator><comments>https://news.ycombinator.com/item?id=42460360</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42460360</guid></item></channel></rss>