<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: iknownothow</title><link>https://news.ycombinator.com/user?id=iknownothow</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 10 Apr 2026 09:33:46 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=iknownothow" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by iknownothow in "Ramtrack.eu – RAM Price Intelligence"]]></title><description><![CDATA[
<p>I just checked again. 16GB DDR5 (291 Euro) to 32GB (519 Euro). Seems right to me and does not match what you reported. Maybe they updated the website by the time I checked.</p>
]]></description><pubDate>Fri, 20 Mar 2026 12:51:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47453828</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=47453828</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47453828</guid></item><item><title><![CDATA[New comment by iknownothow in "Ramtrack.eu – RAM Price Intelligence"]]></title><description><![CDATA[
<p>I checked the prices for 64GB DDR5. There's some variance based on brand/model but the average and trend seems more or less right. Did you happen to notice that it is about prices in the EU?</p>
]]></description><pubDate>Thu, 19 Mar 2026 16:28:57 +0000</pubDate><link>https://news.ycombinator.com/item?id=47442013</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=47442013</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47442013</guid></item><item><title><![CDATA[New comment by iknownothow in "Bucketsquatting is finally dead"]]></title><description><![CDATA[
<p>Thank you author Ian Mckay! This is one of those good hygiene conventions that save time by not having to think/worry each time buckets are named. As pointed out in the article, AWS seems to have made this part of their official naming conventions [1].<p>I'm excited for IaC code libraries like Terraform to incorporate this as their default behavior soon! The default behavior of Terraform and co is already to add a random hash suffix to the end of the bucket name to prevent such errors. This becoming standard practice in itself has saved me days in not having to convince others to use such strategies prior to automation.<p>[1] <a href="https://aws.amazon.com/blogs/aws/introducing-account-regional-namespaces-for-amazon-s3-general-purpose-buckets/" rel="nofollow">https://aws.amazon.com/blogs/aws/introducing-account-regiona...</a></p>
]]></description><pubDate>Fri, 13 Mar 2026 10:00:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=47362434</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=47362434</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47362434</guid></item><item><title><![CDATA[New comment by iknownothow in "Bucketsquatting is (finally) dead"]]></title><description><![CDATA[
<p>Potential reasons I can think of for why they don't disallow name reuse:<p>a) AWS will need to maintain a database of all historical bucket names to know what to disallow. This is hard per region and even harder globally. Its easier to know what is currently in use rather know what has been used historically.<p>b) Even if they maintained a database of all historically used bucket names, then the latency to query if something exists in it may be large enough to be annoying during bucket creation process. Knowing AWS, they'll charge you for every 1000 requests for "checking if bucket name exists" :p<p>c) AWS builds many of its own services on S3 (as indicated in the article) and I can imagine there may be many of their internal services that just rely on existing behaviour i.e. allowing for re-creating the same bucket name.</p>
]]></description><pubDate>Fri, 13 Mar 2026 09:50:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=47362377</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=47362377</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47362377</guid></item><item><title><![CDATA[New comment by iknownothow in "Bucketsquatting is (finally) dead"]]></title><description><![CDATA[
<p>I'd ask politely to refrain from such comments :)<p>This is not me criticising you. I totally understand the urge to say it. We're all thinking the thing you're thinking of. It takes effort not to give into it ;)<p>The reason I personally would refrain from making such comments is that they have the potential to end up as highest ranked comment. That would be a shame. Topic of S3 bucketsquatting is rather important and <i>very</i> interesting.</p>
]]></description><pubDate>Fri, 13 Mar 2026 09:33:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=47362266</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=47362266</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47362266</guid></item><item><title><![CDATA[New comment by iknownothow in "Will Amazon S3 Vectors kill vector databases or save them?"]]></title><description><![CDATA[
<p>I agree, Databricks is one of many in the space. If S3 makes Databricks redundant, then they also make others like Databricks redundant too.</p>
]]></description><pubDate>Tue, 09 Sep 2025 13:40:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=45181767</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=45181767</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45181767</guid></item><item><title><![CDATA[New comment by iknownothow in "Will Amazon S3 Vectors kill vector databases or save them?"]]></title><description><![CDATA[
<p>S3 has much bigger fish in its sight than the measely vector db space. If you see the subtle improvements in features of S3 in recent years, it is clear as day, at least to me, that they're going after the whale that is Databricks. And they're doing it the best way possible - slowly and silently eating away at their moat.<p>AWS Athena hasn't received as much love for some reason. In the next two years I expect major updates and/or improvements. They should kill off Redshift.</p>
]]></description><pubDate>Tue, 09 Sep 2025 11:01:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=45180347</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=45180347</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45180347</guid></item><item><title><![CDATA[New comment by iknownothow in "Incapacitating Google Tag Manager (2022)"]]></title><description><![CDATA[
<p>I just did a wget of the site and noticed the following line at the end.<p>> <script async src="<a href="https://www.googletagmanager.com/gtag/js?xxxxxxx"></script>" rel="nofollow">https://www.googletagmanager.com/gtag/js?xxxxxxx"></script></a><p>I am going to use this for sure, but it is a little ironic.</p>
]]></description><pubDate>Fri, 04 Jul 2025 20:05:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=44467459</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=44467459</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44467459</guid></item><item><title><![CDATA[New comment by iknownothow in "AbsenceBench: Language models can't tell what's missing"]]></title><description><![CDATA[
<p>I'd say that's where we're headed. A big model that's trained from the start to use tools and know when to use certain tools and how to use tools. Like us :)<p>I wouldn't be surprised if someone's building a dataset for tool use examples.<p>The newer gen reasoning models are especially good at knowing when to do web search. I imagine they'll slowly get better at other tools.<p>At current levels of performance, LLMs having the ability to get well curated information by themselves would increase their scores by a lot.</p>
]]></description><pubDate>Sat, 21 Jun 2025 18:47:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=44339792</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=44339792</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44339792</guid></item><item><title><![CDATA[New comment by iknownothow in "AbsenceBench: Language models can't tell what's missing"]]></title><description><![CDATA[
<p>To be fair, I'd put finding literal string diffs in the category of asking LLMs to do rote arithmetic.<p>The attention mechanism does far too much complex thinking for such a dumb task. This is precisely where you need to dumb down and focus and be disciplined rather than do high level next token prediction.<p>You'd benefit from actually asking the LLM to list the full document and compare, kind of like reasoning, and similar to how LLMs perform better when they break down arithmetic or algebra tasks into smaller steps.<p>Also my guess would be that the models that perform well are MoE models where there may be an Expert or two that does well on tasks that needs focus rather than intuition. So without knowing anything about Gemini Flash, my guess would be that it's an MoE model.</p>
]]></description><pubDate>Sat, 21 Jun 2025 18:40:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=44339743</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=44339743</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44339743</guid></item><item><title><![CDATA[New comment by iknownothow in "AbsenceBench: Language models can't tell what's missing"]]></title><description><![CDATA[
<p>As far as I can tell, the paper covers text documents only. Therefore your example doesn't quite apply.<p>It is well known that LLMs have a ways to go when it comes to processing images like they process text or audio.<p>I don't think there's any good performing multimodal model that accepts image pixels directly. Most  vision capabilities are hacks or engineered in. An image undergoes several processing steps and each processor's outputs are fed to the transformer as tokens. This may happen in one network but there's non-transformer networks involved. Examples of preprocessing:<p>* OCR
* CNNs (2D pattern recognizers) with different zooms, angles, slices etc
* Others maybe too?</p>
]]></description><pubDate>Sat, 21 Jun 2025 18:24:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=44339619</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=44339619</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44339619</guid></item><item><title><![CDATA[New comment by iknownothow in "Log-Linear Attention"]]></title><description><![CDATA[
<p>> Log-linear attention replaces the fixed-size hidden state with a logarithmically growing set of hidden states<p>Does this mean the models can be smaller too (on top of the primary benefit of being faster)?</p>
]]></description><pubDate>Sat, 07 Jun 2025 20:25:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=44212402</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=44212402</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44212402</guid></item><item><title><![CDATA[New comment by iknownothow in "The Darwin Gödel Machine: AI that improves itself by rewriting its own code"]]></title><description><![CDATA[
<p>That's fair. I care about the end result.<p>The problem is about taking information in 2D/3D space and solving the problem. Humans solve these things through vision. LLMs or AI can do it using another algorithm and internal representation that's way better.<p>I spent a long time thinking about how to solve the ARC AGI 2 puzzles "if I were an LLM" and I just couldn't think of a non-hacky way.<p>People who're blind use braille or touch to extract 2D/3D information. I don't know how blind people represent 2D/3D info once it's in their brain.</p>
]]></description><pubDate>Sat, 31 May 2025 09:21:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=44143051</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=44143051</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44143051</guid></item><item><title><![CDATA[New comment by iknownothow in "The Darwin Gödel Machine: AI that improves itself by rewriting its own code"]]></title><description><![CDATA[
<p>Don't take this the wrong way, your opinion is also vibes.<p>Let's ground that a bit.<p>Have a look at ARC AGI 1 challenge/benchmark. Solve a problem or two yourself. Know that ARC AGI 1 is practically solved by a few LLMs as of Q1 2025.<p>Then have a look at the ARC AGI 2 challenge. Solve a problem or two yourself. Note that as of today, it is unsolved by LLMs.<p>Then observe that the "difficulty" of ARC AGI 1 and 2 for a human are relatively the same but challenge 2 is much harder for LLMs than 1.<p>ARC AGI 2 is going to be solved *within* 12 months (my bet is on 6 months). If it's not, I'll never post about AI on HN again.<p>There's only one problem to solve, i.e. "how to make LLMs truly see like humans do". Right now, any vision based features that the models exhibit comes from maximizing the use of engineering (i.e. applying CNNs on image slices, chunks, maybe zooming and applying ocr, vector search etc), it isn't vision like ours and isn't a native feature for these models.<p>Once that's solved, then LLMs or new Algo will be able to use a computer perfectly by feeding it screen capture. End of white collar jobs 2-5 years after (as we know it).<p>Edit - added "(as we know it)". And fixed missing word.</p>
]]></description><pubDate>Fri, 30 May 2025 20:34:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=44139638</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=44139638</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44139638</guid></item><item><title><![CDATA[New comment by iknownothow in "Whenever – typed and DST-safe datetimes for Python"]]></title><description><![CDATA[
<p>Thanks for the reply and apologies for the general cynicism. It's not lost on me that it's people like you that build tools that make the work tick. I'm just a loud potential customer and I'm just forwarding the frustration that I have with my own customers onto you :)<p>Your customers are software devs like me. When we're in control of generating timestamps, we know we must use standard ISO formatting.<p>However, what do I do when my customers give me access to an S3 bucket with 1 billion timestamps in an arbitrary (yet decipherable) format?<p>In the GitHub issue you seem to have undergone an evolution from purity to pragmatism. I support this 100%.<p>What I've also noticed is that you seem to try to find grounding or motivation for "where to draw the line" from what's already been done in Temporal or Python stdlib etc. This is where I'd like to challenge your intuitions and ask you instead to open the flood gates and accept any format that is theoretically sensible under ISO format.<p>Why? The damage has already been done. Any format you can think of, already exists out there. You just haven't realized it yet.<p>You know who has accepted this? Pandas devs (I assume, I don't them). The following are legitimate timestamps under Pandas (22.2.x):<p>* 2025-03-30T (nope, not a typo)<p>* 2025-03-30T01 (HH)<p>* 2025-03-30 01 (same as above)<p>* 2025-03-30  01 (two or more spaces is also acceptable)<p>In my opinion Pandas doesn't go far enough. Here's an example from real customer data I've seen in the past that Pandas doesn't parse.<p>* 2025-03-30+00:00 (this is very sensible in my opinion. Unless there's a deeper theoretical regex pattern conflicts with other parts of the ISO format)<p>Here's an example that isn't decipherable under a flexible ISO interpretation and shouldn't be supported.<p>* 2025-30-03 (theoretically you can infer that 30 is a day, and 03 is month. BUT you shouldn't accept this. Pandas used to allow such things. I believe they no longer do)<p>I understand writing these flexible regexes or if-else statements will hurt your benchmarks and will be painful to maintain. Maybe release them under an new call like `parse_best_effort` (or even `youre_welcome`) and document pitfalls and performance degradation. Trust me, I'd rather use a reliable generic but slow parser than spend hours writing a write a god awful regex that I will only use once (I've spent literal weeks writing regexes and fixes in the last decade).<p>Pandas has been around since 2012 dealing with customer data. They have seen it all and you can learn a lot from them. ISOs and RFCs when it comes to timestamps don't mean squat. If possible try to make Whenever useful rather than fast or pure. I'd rather use a slimmer faster alternative to pandas for parsing Timestamps if one is available but there aren't any at the moment.<p>If time permits I'll try to compile a non exhaustive list of real world timestamp formats and post in the issue.<p>Thank you for your work!<p>P.S. seeing BurntSushi in the GitHub issue gives me imposter syndrome :)</p>
]]></description><pubDate>Sun, 13 Apr 2025 22:39:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=43676367</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=43676367</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43676367</guid></item><item><title><![CDATA[New comment by iknownothow in "Whenever: Typed and DST-safe datetimes for Python"]]></title><description><![CDATA[
<p>I've read the link and the GitHub readme page.<p>I'm sure I'm in the top 1% of software devs for the most number of timestamps parsed. [1]<p>DST is not a problem in Python. It's parsing string timestamps. All libraries are bad, including this one, except Pandas. Pandas does great at DST too btw.<p>And I'm not shilling for Pandas either. I'm a Polars user who helicopters Pandas in whenever there's a timestamp that needs to be parsed.<p>Pandas has great defaults. Here's string timestamps I expect to be paesed by default. I'm willing to pass timezone in case of naive timestamps:<p>* All ISO 8601 formats and all its weird mutant children that differ by a tiny bit.<p>* 2025-05-01 (parsed not as date, but as timestamp)<p>* 2025-05-01 00:00:00 (or 00.0 or 00.000 or 0.000000 etc)<p>* 2025-05-01 00:00:00z (or uppercase Z or 00.0z or 00.000z or 0.000000z)<p>* 2025-05-01 00:00:00+02:00 (I don't need this converted to some time zone. Store offset if you must or convert to UTC. It should be comparable to other non naive timestamps).<p>* 2025-03-30 02:30:00+02:00 (This is a non existent timestamp wrt European DST but a legitimate timestamp in timestamp representation, therefore it should be allowed unless I specify CET or Europe/Berlin whatever)<p>* There's other timestamps formats that are non standard but are obvious. Allow for a Boolean parameter called accept_sensible_string_parsing and then parse the following:<p><pre><code>  \* 2025-05-01 00:00 (HH:mm format)

  \* 2025-05-01 00:00+01:00 (HH:mm format)

</code></pre>
[1] It's not a real statistic, it's just that I work with a lot of time series and customer data.<p>Disclaimer: I'm on the phone and on the couch so I wasn't able to test the lib for its string parsing before posting this comment.</p>
]]></description><pubDate>Sun, 13 Apr 2025 18:10:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=43674670</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=43674670</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43674670</guid></item><item><title><![CDATA[New comment by iknownothow in "Reasoning models are just LLMs"]]></title><description><![CDATA[
<p>Is there a terminology battle happening in some circles? And if so, what are the consequences of being wrong and using the wrong terminology?<p>I follow the rnd and progress in this space and I haven't heard anyone make a fuss about it. They are all LLMs or transformers or neural nets but they can be trained or optimized to do different things. For sure, there's terms like Reasoning models or Chat models or Instruct models and yes they're all LLMs.<p>But you can now start combining them to have hybrid models too. Are Omni models that handle audio and visual data still "language" models? This question is interesting in its own right for many reasons, but not to justify or bemoan the use of term LLM.<p>LLM is a good term, it's a cultural term too. If you start getting pedantic, you'll miss the bigger picture and possibly even the singularity ;)</p>
]]></description><pubDate>Mon, 10 Feb 2025 09:19:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=42998406</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=42998406</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42998406</guid></item><item><title><![CDATA[New comment by iknownothow in "Show HN: S3HyperSync – Faster S3 sync tool – iterating with up to 100k files/s"]]></title><description><![CDATA[
<p>How does it compare with s5cmd [1]? s5cmd is my goto tool for fast s3 sync and they have the following at the top of their Github page:<p>> For uploads, s5cmd is 32x faster than s3cmd and 12x faster than aws-cli. For downloads, s5cmd can saturate a 40Gbps link (~4.3 GB/s), whereas s3cmd and aws-cli can only reach 85 MB/s and 375 MB/s respectively.<p>[1] <a href="https://github.com/peak/s5cmd">https://github.com/peak/s5cmd</a></p>
]]></description><pubDate>Tue, 09 Jul 2024 09:07:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=40913960</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=40913960</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40913960</guid></item><item><title><![CDATA[New comment by iknownothow in "Block AI bots, scrapers and crawlers with a single click"]]></title><description><![CDATA[
<p>I've been inadvertently working on this topic and I'd like to share some findings.<p>* Do not confuse bots with DDoS. While bot traffic may end up overwhelming your server, your DDoS SaaS will not stop that traffic unless you have some kind of bot protection enabled, for example the product described in post.<p>* A lot of bots announce themselves via user agents, some don't.<p>* If you're running an ecom shop with a lot of product pages, expect a large portion of traffic to be bots and scrapers. In our case it was upto 50%, which was surprising.<p>* Some bots accept cookies and these skew your product analytics.<p>* We enabled automatic bot protection and a of lot our third party integrations ended up being marked as bots and their traffic was blocked. We eventually turned that off.<p>* (EDIT) Any sophisticated self implemented bot protection isn't worth the effort for most companies out there. But I have to admit, it's very exciting to think about all the ways to block bots.<p>What's our current status? We've enabled monitoring to keep a look out for DDoS attempts but we're taking the hit on bot traffic. The data on our the website isn't really private info, except maybe pricing, and we're really unsure how to think about the new AI bots scraping this information. ChatGPT already gives a summary of what our company does. We don't know if that's a good thing or not. Would be happy to hear anyone's thoughts on how to think about this topic.</p>
]]></description><pubDate>Wed, 03 Jul 2024 14:01:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=40866041</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=40866041</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40866041</guid></item><item><title><![CDATA[New comment by iknownothow in "Can a law make social media less 'addictive'?"]]></title><description><![CDATA[
<p>Many parents don't know social media is bad. I've had to convince my partner that it is harmful over many months. She's only now stopped doom scrolling instagram. We finally came to an agreement that our child will not get a "smart" phone until 14. Our child will have access to old school phones and any other less-harmful communication tech. I intend to expose my child to smart phones and computers in all ways that is educational so that they don't fall behind.<p>EDIT: I would rather not have needed to convince my partner that social media FB, tiktok and instagram is obviously bad and I'd have liked to have some laws to give me some basic protection.</p>
]]></description><pubDate>Mon, 01 Jul 2024 08:16:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=40843631</link><dc:creator>iknownothow</dc:creator><comments>https://news.ycombinator.com/item?id=40843631</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40843631</guid></item></channel></rss>