<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: olliem36</title><link>https://news.ycombinator.com/user?id=olliem36</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 14 May 2026 20:26:10 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=olliem36" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by olliem36 in "Claude for Small Business"]]></title><description><![CDATA[
<p>I agree and that's what i'm working on (for businesses) - an all-one-one consolidated AI application that's setup and ready for non-technical users.<p>It's called Zenning AI - we're a small team in London, testing it with a few companies at the moment!</p>
]]></description><pubDate>Thu, 14 May 2026 06:18:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=48131759</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=48131759</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48131759</guid></item><item><title><![CDATA[New comment by olliem36 in "Show HN: I built a small browser engine from scratch in C++"]]></title><description><![CDATA[
<p>Did you use GPT 5.2 Codex? lol</p>
]]></description><pubDate>Wed, 28 Jan 2026 19:18:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=46800229</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=46800229</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46800229</guid></item><item><title><![CDATA[New comment by olliem36 in "alpr.watch"]]></title><description><![CDATA[
<p>Surveillance of the surveillants to prevent the surveilled</p>
]]></description><pubDate>Wed, 17 Dec 2025 18:18:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=46303309</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=46303309</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46303309</guid></item><item><title><![CDATA[New comment by olliem36 in "Claude Advanced Tool Use"]]></title><description><![CDATA[
<p>Sounds good for tasks like the excel example in the article, but I wonder how this approach will hold up in other multi-step agentic flows. Let me explain:<p>I try to be defensive in agent architectures to make it easy for AI models to recover/fix workflows if something unexpected happens.<p>If something goes wrong halfway through the code execution of multiple 'tools' using Programmatic Tool Calling, it's significantly more complex for the AI model to fix that code and try again compared to a single tool usage - you're in trouble, especially if APIs/tools are not idempotent.<p>The sweet spot might be using this as a strategy to complete tasks that are idempotent/retryable (like a database 'transaction') if they fail half way through execution.</p>
]]></description><pubDate>Tue, 25 Nov 2025 11:43:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=46044877</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=46044877</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46044877</guid></item><item><title><![CDATA[New comment by olliem36 in "Which table format do LLMs understand best?"]]></title><description><![CDATA[
<p>We ended up making middleware for LLM 'tools/functions' that take common data/table formats like CSV, Excel and JSON.<p>The tool uses an LLM to write code to parse the data and conduct the analysis to return back to the LLM. Otherwise, we found pumping raw table data into a LLM is just not reliable, even if you go to the effort to conduct analysis on smaller chunks and merge the results.</p>
]]></description><pubDate>Sun, 05 Oct 2025 21:11:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=45485304</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=45485304</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45485304</guid></item><item><title><![CDATA[New comment by olliem36 in "LLM Observability in the Wild – Why OpenTelemetry Should Be the Standard"]]></title><description><![CDATA[
<p>I think the best way to explain this is to provide an example.<p>Scenario: A B2B fintech company processes chargebacks on behalf of merchants, this involves dozens of steps which depend on the type & history of the merchant, dispute cardholder. It also involves collection of evidence from the card holder.<p>There's a couple of key ways that LLMs make this different from manual workflows:<p>Firstly, the automation is built from a prompt. This is important as it means people who are non-technical and are not necessarily comfortable with non-code tools to pull data from multiple places into a sequence. This increases the adoption of automations as the effort to build & deploy them is lower. In this example, there was no automation in place despite the people who 'own' this process wanting to automate it. No doubt there's a number of reasons for this, one being they found todays workflow builders too hard to use.<p>Secondly, the collection of 'evidence' to counter a chargeback can be nuanced, which often requiring back and forth with people to explain what is needed and check the evidence is sufficient against a complicated set of guidelines. I'd say a manual submission form that guides people through evidence collection with hundreds of rules subject to the conditions of the dispute and the merchant could do this, but again, this is hard to build and deploy.<p>Lastly, LLMs monitors the success of the workflow once it's deployed, to help those who are responsible for it measure its impact and effectiveness.<p>The end result is that a business has successfully built and deployed an automation that they did not have before.<p>To answer your second question, dynamic routing describes the process of evaluating how complicated a prompt or task is, and then selecting an LLM that's 'best fit' to process it. For example, short & simple prompts should usually get routed to faster but less intelligent LLMs. This typically makes users happier as they get results more quickly. However, more complex prompts may require larger, slower and more intelligent LLMs and techniques such as 'reasoning'. The result will be slower to produce, but will be likely be far more accurate compared to a faster model. In the above example, a larger LLM with reasoning would probably be used.</p>
]]></description><pubDate>Sun, 05 Oct 2025 09:57:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=45480344</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=45480344</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45480344</guid></item><item><title><![CDATA[New comment by olliem36 in "Managing context on the Claude Developer Platform"]]></title><description><![CDATA[
<p>At Zenning AI, a generalist AI designed to replace entire jobs with just prompts. Our agents typically run autonomously for hours, so effective context management is critical. I'd say that we invest most of our engineering effort into what is ultimately context management, such as:<p>1. Multi-agent orchestration
2. Summarising and chunking large tool and agent responses
3. Passing large context objects by reference between agents and tools<p>Two things to note that might be interesting to the community:<p>Firstly, when managing context, I recommend adding some evals to our context management flow, so you can measure effectiveness as you add improvements and changes.<p>For example, our evals will measure the impact of using Anthropics memory over time. Thus allowing our team to make a better informed decisions on that tools to use with our agents.<p>Secondly, there's a tradeoff not mentioned in this article: speed vs. accuracy. Faster summarisation (or 'compaction') comes at a cost of accuracy. If you want good compaction, it can be slow. Depending on the use case, you should adjust your compaction strategy accordingly. For example, (forgive my major generalisation), for consumer facing products speed is usually preferred over a bump in accuracy. However, in business accuracy is generally preferred over speed.</p>
]]></description><pubDate>Sun, 05 Oct 2025 08:47:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=45480012</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=45480012</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45480012</guid></item><item><title><![CDATA[New comment by olliem36 in "LLM Observability in the Wild – Why OpenTelemetry Should Be the Standard"]]></title><description><![CDATA[
<p>We've built a multi-agent system, designed to run complex tasks and workflows with just a single prompt. Prompts are written by non-technical people, can be 10+ pages long...<p>We've invested heavily in observability having quickly found that observability + evals are the cornerstone to a successful agent.<p>For example, a few things measure:<p>1. Task complexity (assessed by another LLM)
2. Success metrics given the task(s) (Agin by other LLMS)
3. Speed of agent runs & tools
4. Errors of tools, inc time outs.
5. How much summarizaiton and chunking occurs between agents and tool results
6. tokens used, cost
7. reasoning, model selected by our dynamic routing..<p>Thank god its been relatively cheap to build this in house.. our metrics dashboard is essentially a vibe coded react admin site.. but proves absolutely invaluable!<p>All of this happed after a heavy investment in agent orchestration, context management... it's been quite a ride!</p>
]]></description><pubDate>Sat, 27 Sep 2025 22:55:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=45400037</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=45400037</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45400037</guid></item><item><title><![CDATA[New comment by olliem36 in "Stripe Launches L1 Blockchain: Tempo"]]></title><description><![CDATA[
<p>Co-founder of Lopay here, we're a small but heavy Stripe user with £1B+ processed across Connect, Terminal, Identity, Instant payouts, Issuing... you name it.<p>We're looking at stable coins for the following use cases:<p>1.   Instant clearing and settlement of 'floats' & liquidity - EG moving liquidity between our network to support instant/same day payouts or instant funding of a spend card.<p>2.     Instant cross border payments (lots of people doing this already in companies that operate multinationally). EG, our USD top-ups today take 3 days in fiat, which can cause operational issues.<p>3.    Offering our merchants (who are typically small businesses) optionality to hold USD in countries that have volatile currencies.<p>I'll also note that many people forget that the cost of a payment network isn't merely the movement of money, it's also KYC, dispute resolution, fraud prevention etc...<p>I wonder if the tempo team has looked at AI automating dispute resolution and fraud detection/prevention 'on chain'.. The network could fund the compute required for the AI to complete these tasks.</p>
]]></description><pubDate>Fri, 05 Sep 2025 10:01:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=45136843</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=45136843</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45136843</guid></item><item><title><![CDATA[New comment by olliem36 in "Zenobia Pay – A mission to build an alternative to high-fee card networks"]]></title><description><![CDATA[
<p>Cofounder of Lopay here - we have the same mission: offer free payments to businesses, but we're working with existing networks to do this.<p>QR code payments are particularly hard in countries like US and UK as you're trying to change consumer behaviour. I tried doing this in 2014 and again in 2019 - both failed to gain traction (aside from during COVID).<p>In the UK it's possible to accept card payments for 0% via Lopay, but only if you spend your earnings on our card (essentially, passing the fees onto the merchant/supplier you're paying). We're launching the same proposition in the US soon too.<p>If you don't use our card, our headline rate is 0.79%.<p>We're a lean team of just 36, supporting over 40k weekly transacting businesses with £1B+ in card processing. If anyone reading this is interested in this space, we're hiring and on the look out for driven people to join us!</p>
]]></description><pubDate>Thu, 14 Aug 2025 10:31:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=44898824</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=44898824</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44898824</guid></item><item><title><![CDATA[New comment by olliem36 in "I pulled data on 1378 restaurants from Google Maps to rank them in order"]]></title><description><![CDATA[
<p>Founder of Salamanca here, an app that aggregates every major restaurant booking platform into one app (OpenTable, SevenRooms, Tock, Resy, The Fork and others..)<p>Firstly, nice site - always love new tools to discover restaurants, thanks for posting, I’ve shared your blog post with friends, it was a brilliant read.<p>I have some recent experience working with restaurant reviews, I found that using only Google reviews can be unreliable, as some places that have top reviews may not be generally accepted as the ‘best’ restaurants.<p>We currently use a combination of Google reviews + Trip Advisor + Reviews from the booking platforms and we have web crawlers to check if the restaurant is featured on reputable restaurant guides or review sites.<p>We aggregate all of this review data and compute a “score”, so when users search for available tables in a city we can show available tables at the highest scoring restaurants first.<p>We apply Wilson score confidence intervals, to trust restaurant scores that have more reviews.<p>We are also applying an exponential decay when users list nearby restaurants, as you might be willing to travel a little further to go to a higher scoring restaurant.<p>Working with review data is fascinating.. we’re going to be launching an AI summary of recent reviews and our computed score in the coming weeks to help our users understand our ratings.<p>Our app went live on the App Store only a few days ago and we expect it to be live on Google play later this week.. so it’s an extremely busy time!<p>If you’re interested in what we’re doing please reach out, it would be great to connect, I really enjoyed your article!</p>
]]></description><pubDate>Mon, 17 Feb 2025 08:24:32 +0000</pubDate><link>https://news.ycombinator.com/item?id=43076557</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=43076557</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43076557</guid></item><item><title><![CDATA[New comment by olliem36 in "It's Time to Stop Taking Sam Altman at His Word"]]></title><description><![CDATA[
<p>Great analogy! I'll borrow this when explaining my thoughts on how LLMs pose to replace software engineers.</p>
]]></description><pubDate>Sat, 05 Oct 2024 14:09:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=41749986</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=41749986</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41749986</guid></item><item><title><![CDATA[New comment by olliem36 in "Ask HN: Who is hiring? (June 2024)"]]></title><description><![CDATA[
<p>AllTheTables | Founding Software Engineer | London, UK | 3 days in central London office, 2 days remote | Full time<p>AllTheTables is the first platform to aggregate booking platforms such as OpenTables, The Fork, Resy and many more. AllTheTables offers any website or mobile app the ability to add restaurant reservations to their platform and earn commission from each booking. Many millions of visitors to travel guides and restaurant review apps/websites can soon book directly without needing to leave the website they discover the restaurant on.<p>I have seen multiple exits as a first-time startup employee and co-founder. I’m looking for a Founding Software Engineer who is “mid-level” in experience. At least 4 years professional experience shipping API’s are required. Ideally this will be someone who wants their career to grow with a startup (CTO or Head of Engineering progression).<p>Tech stack: Typescript Node on the backend and React on the frontend. Postgres Database. Heroku hosting for simplicity & shipping fast.<p>Funding: £1m seed (June 2024)<p>Compensation: Between £60-70k + equity inline with founding engineer/CTO<p>Apply via email: ollie@allthetables.com  (No agencies please)
Website: <a href="https://allthetables.com" rel="nofollow">https://allthetables.com</a></p>
]]></description><pubDate>Mon, 03 Jun 2024 19:02:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=40566231</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=40566231</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40566231</guid></item><item><title><![CDATA[New comment by olliem36 in "What is it like to be a crab?"]]></title><description><![CDATA[
<p>If you like crabs, there’s a crab museum in Margate (Kent, UK)<p><a href="http://www.crabmuseum.org" rel="nofollow noreferrer">http://www.crabmuseum.org</a><p>According to their flyer, it’s the #1 crab museum in Europe!</p>
]]></description><pubDate>Sat, 25 Nov 2023 15:25:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=38414069</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=38414069</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38414069</guid></item><item><title><![CDATA[New comment by olliem36 in "Algolia has updated their free tier from 10k to 1M rows"]]></title><description><![CDATA[
<p>Previously, Algolia had a restriction of 10k rows on their free tier called "build". This has now been updated to 1 millions rows.<p>Depending on your project, I found their 10k row limit quite restrictive as my projects typically had significantly more rows which I wanted to index for search.<p>A jump from 10k to 1 million is substantial, so I have taken advantage of it this weekend. I hope others out there find this as useful as I have done!<p>Please note, this increase didn't apply to my existing Algolia accounts, so I had to create a new account to take advantage of this higher allowance.</p>
]]></description><pubDate>Sun, 16 Apr 2023 07:57:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=35588105</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=35588105</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=35588105</guid></item><item><title><![CDATA[Algolia has updated their free tier from 10k to 1M rows]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.algolia.com/pricing/">https://www.algolia.com/pricing/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=35588104">https://news.ycombinator.com/item?id=35588104</a></p>
<p>Points: 4</p>
<p># Comments: 1</p>
]]></description><pubDate>Sun, 16 Apr 2023 07:57:14 +0000</pubDate><link>https://www.algolia.com/pricing/</link><dc:creator>olliem36</dc:creator><comments>https://news.ycombinator.com/item?id=35588104</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=35588104</guid></item></channel></rss>