<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: Shyaamal11</title><link>https://news.ycombinator.com/user?id=Shyaamal11</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 15 May 2026 18:15:51 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=Shyaamal11" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by Shyaamal11 in "Postgres Is the Gateway Drug"]]></title><description><![CDATA[
<p>One thing I’ve seen with this pattern is that Postgres + CDC works really well as an early-stage streaming backbone, especially when the operational DB is already the source of truth.<p>Using WAL → CDC → downstream systems keeps the architecture simple at first, and tools like Debezium make it relatively straightforward to pipe those changes into Kafka or other processors.<p>Where things start getting interesting is the analytics side. Once the CDC stream lands in something like Iceberg tables, you effectively get a continuously updated analytical dataset that can be queried with engines like Spark or Trino.<p>At that point the architecture starts to look less like a traditional “data warehouse pipeline” and more like a streaming-first lakehouse where operational data flows directly into analytical storage.<p>The main challenge I’ve seen is operational complexity once you start combining:
CDC ingestion
stream processing
lakehouse storage (Iceberg/Delta)
distributed query engines
That’s where platforms trying to package the open stack together (e.g. Spark + Iceberg + Trino) become interesting. Some newer platforms like IOMETE are basically trying to simplify running that type of lakehouse stack on Kubernetes so teams don’t have to glue everything together manually.<p>Curious where people think the breakpoint is at what scale does Postgres+CDC stop being “good enough” and you start needing a dedicated log system as the primary event backbone?</p>
]]></description><pubDate>Mon, 16 Mar 2026 02:15:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=47394401</link><dc:creator>Shyaamal11</dc:creator><comments>https://news.ycombinator.com/item?id=47394401</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47394401</guid></item><item><title><![CDATA[New comment by Shyaamal11 in "LLMs work best when the user defines their acceptance criteria first"]]></title><description><![CDATA[
<p>One thing I’ve noticed while working with data/AI workflows is that the “acceptance criteria first” idea applies even more strongly once you move beyond code generation into data pipelines and analytics.<p>LLMs can generate queries, transformations, or even Spark jobs that look reasonable but if the underlying data contracts, schema expectations, or evaluation criteria aren’t defined, you end up with something that looks correct but is semantically wrong.<p>In practice, the teams that get the most value from AI-assisted development tend to have:
clearly defined datasets
reproducible data pipelines
well-defined outputs / metrics
Once those pieces are in place, AI becomes much more useful because it’s operating inside a structured system instead of guessing context.
That’s also why there’s been a lot of interest lately in lakehouse-style platforms that combine data engineering, analytics, and AI workflows in one place (e.g. platforms like IOMETE).<p>When the data layer is structured and reproducible, AI tooling becomes far more reliable.
Curious if others here have seen the same pattern when using LLMs for data engineering or analytics work.</p>
]]></description><pubDate>Sun, 15 Mar 2026 08:31:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=47385400</link><dc:creator>Shyaamal11</dc:creator><comments>https://news.ycombinator.com/item?id=47385400</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47385400</guid></item><item><title><![CDATA[New comment by Shyaamal11 in "Show HN: Forge, the NoSQL to SQL Compiler"]]></title><description><![CDATA[
<p>The cross warehouse portability problem you're solving is real. I've watched teams maintain four separate FLATTEN implementations for the same pipeline just because they were multi-cloud. The compiler framing makes sense.<p>Curious how you handle schema drift at the introspection phase specifically when an API starts returning a field as sometimes a string and sometimes an object depending on the endpoint response. Does Forge pick a winner or surface it as a conflict for the user to resolve?</p>
]]></description><pubDate>Sun, 15 Mar 2026 08:23:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=47385357</link><dc:creator>Shyaamal11</dc:creator><comments>https://news.ycombinator.com/item?id=47385357</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47385357</guid></item><item><title><![CDATA[New comment by Shyaamal11 in "Ask HN: Who is hiring? (March 2026)"]]></title><description><![CDATA[
<p>Amazing to see such a support</p>
]]></description><pubDate>Fri, 13 Mar 2026 19:23:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=47368551</link><dc:creator>Shyaamal11</dc:creator><comments>https://news.ycombinator.com/item?id=47368551</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47368551</guid></item><item><title><![CDATA[New comment by Shyaamal11 in "Show HN: Open-sourced an email QA lib 8 checks across 12 clients in 1 audit call"]]></title><description><![CDATA[
<p>The spam scoring caught my eye — 45+ heuristic signals is a lot. How do you handle false positives for transactional emails? A password reset or order confirmation might legitimately trigger some of those signals (no unsubscribe, image-heavy, urgent language) even though they're completely clean emails. Does the transactional exemption you mention cover most of those cases or is there still manual tuning needed?</p>
]]></description><pubDate>Tue, 03 Mar 2026 18:35:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47236703</link><dc:creator>Shyaamal11</dc:creator><comments>https://news.ycombinator.com/item?id=47236703</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47236703</guid></item><item><title><![CDATA[New comment by Shyaamal11 in "US orders diplomats to fight data sovereignty initiatives"]]></title><description><![CDATA[
<p>Worth separating two things the thread keeps conflating: data residency and data sovereignty are not the same, and the CLOUD Act is the clearest proof.
Residency = where data physically sits. Sovereignty = who legally controls it.
You can store data on AWS Frankfurt and still have zero sovereignty, the controlling entity is US-domiciled and fully subject to the CLOUD Act. Geographic residency without legal sovereignty is essentially compliance theater.
Real sovereignty requires the controlling entity, jurisdiction, and infrastructure layer to sit outside US reach which typically means infrastructure you actually operate, not just "hosted in your region" SaaS.<p>(Disclosure: I work at IOMETE where we think about this a lot. I am happy to go deeper if useful.)</p>
]]></description><pubDate>Tue, 03 Mar 2026 17:36:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=47235837</link><dc:creator>Shyaamal11</dc:creator><comments>https://news.ycombinator.com/item?id=47235837</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47235837</guid></item></channel></rss>