<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: lolinder</title><link>https://news.ycombinator.com/user?id=lolinder</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sat, 18 Apr 2026 10:37:12 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=lolinder" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by lolinder in "I wrote my PhD Thesis in Typst"]]></title><description><![CDATA[
<p>Alternatively, they're people who write documents in a field where LaTeX is the standard, they're not computer savvy enough to try to even look for something new that might be acceptable or might compile to LaTeX, and at any rate they want to focus more on their research than they do on changing the typesetting norms in their field.<p>(No shade on people who <i>do</i> decide to use alternatives, and Typst is great!)</p>
]]></description><pubDate>Mon, 23 Jun 2025 12:29:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=44355103</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44355103</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44355103</guid></item><item><title><![CDATA[New comment by lolinder in "U.S. bombs Iranian nuclear sites"]]></title><description><![CDATA[
<p>Yes, that too. It was in <i>newspapers</i> hours before the attack, so there's no way Iran didn't know it was coming.</p>
]]></description><pubDate>Sun, 22 Jun 2025 15:28:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=44347706</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44347706</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44347706</guid></item><item><title><![CDATA[New comment by lolinder in "U.S. bombs Iranian nuclear sites"]]></title><description><![CDATA[
<p>Possibly. It could also be evidence that the bombs didn't really do their job. Either they missed or the bunkers were fortified enough and/or deep enough to avoid even the bunker busters.</p>
]]></description><pubDate>Sun, 22 Jun 2025 15:24:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=44347676</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44347676</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44347676</guid></item><item><title><![CDATA[New comment by lolinder in "Record DDoS pummels site with once-unimaginable 7.3Tbps of junk traffic"]]></title><description><![CDATA[
<p>Coordinating a botnet to launch a DDoS is commodity software at this point. You could argue that the engineering that went into the coordination software is good, which may or may not be true, but simply launching a botnet is well within the capabilities of a script kiddie and not something that shows sophistication on the part of the attacker.</p>
]]></description><pubDate>Sun, 22 Jun 2025 15:20:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=44347641</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44347641</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44347641</guid></item><item><title><![CDATA[New comment by lolinder in "Record DDoS pummels site with once-unimaginable 7.3Tbps of junk traffic"]]></title><description><![CDATA[
<p>Like I said: it broke records for data throughput. It doesn't hurt that Cloudflare has an interest in publicizing the size of the DDoS attacks it fights off.<p>> in spite of not having met your arbitrary and personal bar?<p>I'm not sure what you mean by this. I didn't establish any sort of bar for what sorts of DDoS should get headlines, I'm just agreeing with OP that that line in the article doesn't make any sense. There may be other reasons to believe this attack was well-engineered but the article doesn't get into them.</p>
]]></description><pubDate>Sat, 21 Jun 2025 16:08:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=44338610</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44338610</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44338610</guid></item><item><title><![CDATA[New comment by lolinder in "Record DDoS pummels site with once-unimaginable 7.3Tbps of junk traffic"]]></title><description><![CDATA[
<p>Not necessarily. It could be one for loop running on tens of thousands of compromised IoT devices, with the only thing distributed being the command that starts the loops.</p>
]]></description><pubDate>Sat, 21 Jun 2025 15:46:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=44338421</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44338421</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44338421</guid></item><item><title><![CDATA[New comment by lolinder in "Record DDoS pummels site with once-unimaginable 7.3Tbps of junk traffic"]]></title><description><![CDATA[
<p>Right. So why does the fact that they targeted 34,500 ports show it was a well-engineered attack? By itself it's just evidence that they know how to iterate over ports. Coupled with the data size (7.3Tbps) we know they had an enormous botnet. None of this points to a well-engineered attack, it just means that lousy IoT has made botnets incredibly cheap.<p>A well-engineered attack would not draw headlines for its scale because it would take down its target without breaking any records.</p>
]]></description><pubDate>Sat, 21 Jun 2025 15:43:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=44338406</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44338406</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44338406</guid></item><item><title><![CDATA[New comment by lolinder in "Sega mistakenly reveals sales numbers of popular games"]]></title><description><![CDATA[
<p>Video games I feel like reverse this general trend, though. Unless they have a major story component (and sometimes even if they do) many games get iteratively 'better' (better for the purposes of making sales if not of making original fans happy) for various reasons: improvements to the core game loop, polish that makes the game more appealing to new audiences, and most importantly graphics.<p>Story-based content is what struggles with sequels because it's really hard to both capture the feeling of the original sufficiently to satisfy existing fans while also telling a new story that's interesting in its own right. Being derivative without being too derivative.</p>
]]></description><pubDate>Sat, 21 Jun 2025 13:36:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=44337496</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44337496</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44337496</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>That would help, but who defines which sites are required for the task? If it's the LLM you haven't solved prompt injection because the LLM can be persuaded to open other sites that the user didn't intend.</p>
]]></description><pubDate>Sat, 21 Jun 2025 12:47:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=44337168</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44337168</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44337168</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>Alright, I think we can agree on that. I'll see you over in that new standardization discussion fighting fiercely for protections to make sure companies don't abuse it to compromise the open web.</p>
]]></description><pubDate>Fri, 20 Jun 2025 23:33:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=44333100</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44333100</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44333100</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>I draw the line where robotstxt.org (the semi-official home of robots.txt) draws the line [0]:<p>> A robot is a program that automatically traverses the Web's hypertext structure by retrieving a document, and recursively retrieving all documents that are referenced.<p>To me "recursive" is key—it transforms the traffic pattern from one that strongly resembles that of a human to one that touches every page on the site, breaks caching by visiting pages humans wouldn't typically, and produces not just a little bit more but orders of magnitude more traffic.<p>I was persuaded in another subthread that Nxtscape should respect robots.txt if a user issues a recursive request. I don't think it should if the request is "open these 5 subreddits and summarize the most popular links uploaded since yesterday", because the resulting traffic pattern is nearly identical to what I'd have done by hand (especially if the browser implements proper rate limiting, which I believe it should).<p>[0] <a href="https://www.robotstxt.org/faq/what.html" rel="nofollow">https://www.robotstxt.org/faq/what.html</a></p>
]]></description><pubDate>Fri, 20 Jun 2025 23:26:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=44333070</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44333070</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44333070</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p><a href="https://www.robotstxt.org/faq/what.html" rel="nofollow">https://www.robotstxt.org/faq/what.html</a><p>> A robot is a program that automatically traverses the Web's hypertext structure by retrieving a document, and recursively retrieving all documents that are referenced.<p>There's nothing recursive about "summarize all the cooking recipes linked on this page". That's a single-level iterative loop.<p>I will grant that I should alter my original statement: if OP wanted to respect robots.txt when it receives a request that should be interpreted as an instruction to recursively fetch pages, then I'd think that's an appropriate use of robots.txt, because that's not materially different than implementing a web crawler by hand in code.<p>But that represents a tiny subset of the queries that will go through a tool like this and respecting robots.txt for non-recursive requests would lead to silly outcomes like the browser refusing to load reddit.com [0].<p>[0] <a href="https://www.reddit.com/robots.txt" rel="nofollow">https://www.reddit.com/robots.txt</a></p>
]]></description><pubDate>Fri, 20 Jun 2025 23:01:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=44332910</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44332910</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44332910</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>> If you are using an ad-blocker by definition you are intentionally breaking the intended behavior by the creator of any given website (for personal gain), in that context any discussion about robots.txt or any other behavior that the creator expects is a moot point.<p>I'm not asking if you believe ad blocking is ethical, I got that you don't. I'm asking if it turns my browser into a scraper that should be treated as such, which is an orthogonal question to the ethics of the tool in the first place.<p>I strongly disagree that user agents of the sort shown in the demo should count as robots. Robots.txt is designed for bots that produce tons of traffic to discourage them from hitting expensive endpoints (or to politely ask them to not scrape at all). I've responded to incidents caused by scraper traffic and this tool will <i>never</i> produce traffic in the same order of magnitude as a problematic scraper.<p>If we count this as a robot for the purposes of robots.txt we're heading down a path that will end the user agent freedom we've hitherto enjoyed. I cannot endorse that path.<p>For me the line is simple, and it's the one defined by robotstxt.org [0]: "A robot is a program that automatically traverses the Web's hypertext structure by retrieving a document, and recursively retrieving all documents that are referenced. ... Normal Web browsers are not robots, because they are operated by a human, and don't automatically retrieve referenced documents (other than inline images)."<p>If the user agent is acting on  my instructions and accessing a specific and limited subset of the site that I asked it to, it's not a web scraper and should not be treated as such. The defining feature of a robot is amount of traffic produced, not what my user agent does with the information it pulls.<p>[0] <a href="https://www.robotstxt.org/faq/what.html" rel="nofollow">https://www.robotstxt.org/faq/what.html</a></p>
]]></description><pubDate>Fri, 20 Jun 2025 22:52:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=44332846</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44332846</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44332846</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>> If your browser behaves, it's not going to be excluded in robots.txt.<p>No, it's common practice to allow Googlebot and deny all other crawlers by default [0].<p>This is within their rights when it comes to true scrapers, but it's part of why I'm very uncomfortable with the idea of applying robots.txt to what are clearly user agents. It sets a precedent where it's not inconceivable that we have websites curating allowlists of user agents like they already do for scrapers, which would be very bad for the web.<p>[0] As just one example: <a href="https://www.404media.co/google-is-the-only-search-engine-that-works-on-reddit-now-thanks-to-ai-deal/" rel="nofollow">https://www.404media.co/google-is-the-only-search-engine-tha...</a></p>
]]></description><pubDate>Fri, 20 Jun 2025 22:42:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=44332789</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44332789</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44332789</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>robotstxt.org [0] is pretty specific in what constitutes a robot for the purposes of robots.txt:<p>> A robot is a program that automatically traverses the Web's hypertext structure by retrieving a document, and recursively retrieving all documents that are referenced.<p>This is absolutely <i>not</i> what you are doing, which means what you have here is not a robot. What you have here is a user agent, so you don't need to pay attention to robots.txt.<p>If what you are doing here counted as robotic traffic, then so would:<p>* Speculative loading (algorithm guesses what you're going to load next and grabs it for you in advance for faster load times).<p>* Reader mode (algorithm transforms the website to strip out tons of content that you don't want and present you only with the minimum set of content you wanted to read).<p>* Terminal-based browsers (do not render images or JavaScript, thus bypassing advertising and according to some justifications leading them to be considered a robot because they bypass monetization).<p>The fact is that the web is designed to be navigated by a diverse array of different user agents that behave differently. I'd seriously consider imposing rate limits on how frequently your browser acts so you don't knock over a server—that's just good citizenship—but robots.txt is not designed for you and if we act like it is then a <i>lot</i> of dominoes will fall.<p>[0] <a href="https://www.robotstxt.org/faq/what.html" rel="nofollow">https://www.robotstxt.org/faq/what.html</a></p>
]]></description><pubDate>Fri, 20 Jun 2025 22:30:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=44332715</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44332715</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44332715</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>> I know its not completely true, I know read-mode can help you bypass the ads _after_ you already had a peek at the cluttered version<p>What about reader mode that is auto-configured to turn on immediately on landing on specific domains? Is that a robot for the purposes of robots.txt?<p><a href="https://addons.mozilla.org/en-US/firefox/addon/automatic-reader-view/" rel="nofollow">https://addons.mozilla.org/en-US/firefox/addon/automatic-rea...</a><p>And also, just to confirm, I'm to understand that if I'm navigating the internet with an ad blocker then you believe that I should respect robots.txt because my user agent is now a robot by virtue of using an ad blocker?<p>Is that also true if I browse with a terminal-based browser that simply doesn't render JavaScript or images?</p>
]]></description><pubDate>Fri, 20 Jun 2025 22:26:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=44332694</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44332694</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44332694</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>> only reading the content the user would otherwise have gone through.<p>Why? My user agent is configured to make things easier for me and allow me to access content that I wouldn't otherwise choose to access. Dark mode allows me to read late at night. Reader mode allows me to read content that would otherwise be unbearably cluttered. I can zoom in on small text to better see it.<p>Should my reader mode or dark mode or zoom feature have to respect robots.txt because otherwise they'd allow me to access content that I would otherwise have chosen to leave alone?</p>
]]></description><pubDate>Fri, 20 Jun 2025 21:43:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=44332395</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44332395</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44332395</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>So is Chrome. Very artificial. It's still not a robot for the purposes of robots.txt.<p>What coherent definition of robot excludes Chrome but includes this?</p>
]]></description><pubDate>Fri, 20 Jun 2025 21:42:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=44332385</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44332385</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44332385</guid></item><item><title><![CDATA[New comment by lolinder in "Show HN: Nxtscape – an open-source agentic browser"]]></title><description><![CDATA[
<p>This is a user agent and I would be incredibly frustrated if they respected robots.txt. Robots.txt was designed to encourage recursive web crawlers to be respectful. It's specifically not meant to exclude agents that are acting on users' direct requests.<p>Website operators should not get a say in what kinds of user agents I used to access their sites. Terminal? Fine. Regular web browser? Okay. AI powered web browser? Who cares. The strength of the web lies in the fact that I can access it with many different kinds of tools depending on my use case, and we cannot sacrifice that strength on the altar of hatred of AI tools.<p>Down that road lies disaster, with the Play Integrity API being just the tip of the iceberg.<p><a href="https://www.robotstxt.org/faq/what.html" rel="nofollow">https://www.robotstxt.org/faq/what.html</a></p>
]]></description><pubDate>Fri, 20 Jun 2025 21:36:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=44332349</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44332349</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44332349</guid></item><item><title><![CDATA[New comment by lolinder in "Congestion pricing in Manhattan is a predictable success"]]></title><description><![CDATA[
<p>In the cases we're discussing rich people pay more <i>for increased access</i>, so they're still exchanging money for something of value, not just paying more into the system for the sake of it. They get opportunities and benefits out of the toll/fee system that are not available to people who cannot afford to pay.<p>If the taxes earned from these transactions are then spent on things that <i>also</i> benefit the wealthy just as much as they do the poor, then the rich are double-dipping and poor people still end up net <i>behind</i> the wealthy. They lose access to something that previously was paid for out of property taxes in exchange for more revenue funding services that the wealthy are just as likely to use.<p>This model at least doesn't further exacerbate the regressiveness of the tax by only funding things used by the rich, but it doesn't restore balance.<p>That's why I say that the only way that you flip the tax to be progressive is if the proceeds benefit the poor <i>disproportionately</i> rather than benefiting everyone equally.</p>
]]></description><pubDate>Fri, 20 Jun 2025 17:43:19 +0000</pubDate><link>https://news.ycombinator.com/item?id=44330069</link><dc:creator>lolinder</dc:creator><comments>https://news.ycombinator.com/item?id=44330069</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44330069</guid></item></channel></rss>