<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: alexsmirnov</title><link>https://news.ycombinator.com/user?id=alexsmirnov</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 17 Apr 2026 04:55:29 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=alexsmirnov" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by alexsmirnov in "OpenAI backs Illinois bill that would limit when AI labs can be held liable"]]></title><description><![CDATA[
<p>Much longer than that, and was available way before an internet. I graduated STEM high school in St. Petersburg in 1981, and I had several classmates who were big funs of chemistry. That they were able to create from textbooks, school lab ingredients, and understanding:<p>WWI era poison gas, tear gas, potassium cyanide, and bunch of explosives like acetone peroxide.<p>LLMs have all of that knowledge in training data</p>
]]></description><pubDate>Fri, 10 Apr 2026 17:31:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=47721265</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=47721265</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47721265</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Qwen3.6-Plus: Towards real world agents"]]></title><description><![CDATA[
<p>Exactly.<p>I did create my own MCP with custom agents that combine several tools into a single one. For example, all WebSearch, WebFetch, Context7 exposed as a single "web research" tool, backed by the cheapest model that passes evaluation. The same for a codebase research<p>Use it with both Claude and Opencode saves a lot of time and tokens.</p>
]]></description><pubDate>Thu, 02 Apr 2026 16:05:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47616303</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=47616303</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47616303</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Study: 'Security Fatigue' May Weaken Digital Defenses"]]></title><description><![CDATA[
<p>Almost instantly, compared to my experience working for a big health care provider... I waited 6 moths for IT department to allow me install development tools on work laptop.<p>And while security rules created enormous roadblocks for work, whey also left enough holes to be exploited. Before getting required permissions, I managed to create dual boot with linux and share files between 'approved' and 'illegal' systems</p>
]]></description><pubDate>Mon, 23 Mar 2026 17:36:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47492614</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=47492614</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47492614</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Ask HN: AI productivity gains – do you fire devs or build better products?"]]></title><description><![CDATA[
<p>> you start off checking every diff like a hawk, expecting it to break things, but honestly, soon you see it's not necessary most of the time.<p>I see it's necessary ALL the time. The AI generated code can be used as scaffolding, but it's newer get close to real production quality. The expierence from small startup with team of 5 developers. I do review and approve all PRs, and none ever able to pass AI code review from the first iteration.</p>
]]></description><pubDate>Mon, 23 Mar 2026 03:04:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47485079</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=47485079</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47485079</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Claudetop – htop for Claude Code sessions (see your AI spend in real-time)"]]></title><description><![CDATA[
<p>The calculation misses subagents tokens, that can be a significant differences.
Better to parse session jsonl files ~/.claude/projects/<project>/<session id>.jsonl and ~/.claude/projects/<project>/<session id>/subagents/<agent id>.jsonl</p>
]]></description><pubDate>Mon, 16 Mar 2026 00:01:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=47393455</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=47393455</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47393455</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Code Review for Claude Code"]]></title><description><![CDATA[
<p>This mostly matches my own estimates for pr-review command that I use. But it's pretty sophisticated: 6 specialized agents, best practices skills, CVE database, bunch of scripts. To reduce cost, most of agents use cheap open source models.</p>
]]></description><pubDate>Tue, 10 Mar 2026 02:06:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=47318322</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=47318322</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47318322</guid></item><item><title><![CDATA[New comment by alexsmirnov in "When AI writes the software, who verifies it?"]]></title><description><![CDATA[
<p>Actually, they extremely bad at that. All training data contains cod + tests, even if tests where created first. So far, all models that I tried failed to implement tests for interfaces, without access to actual code.</p>
]]></description><pubDate>Wed, 04 Mar 2026 07:45:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=47244384</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=47244384</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47244384</guid></item><item><title><![CDATA[New comment by alexsmirnov in "What Claude Code chooses"]]></title><description><![CDATA[
<p>Considering how little data needed to poison llm <a href="https://www.anthropic.com/research/small-samples-poison" rel="nofollow">https://www.anthropic.com/research/small-samples-poison</a> , this is a way to replace SEO by llm product placement:<p>1. create several hundreds github repos with projects that use your product ( may be clones or AI generated )<p>2. create website with similar instructions, connect to hundred domains<p>3. generate reddit, facebook, X posts, wikipedia pages with the same information<p>Wait half a year ? until scrappers collect it and use to train new models<p>Profit...</p>
]]></description><pubDate>Thu, 26 Feb 2026 22:18:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=47172794</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=47172794</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47172794</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Why Developers Keep Choosing Claude over Every Other AI"]]></title><description><![CDATA[
<p>I do in similar way, connect claude code to litellm router that dispatches model requests to different providers: bedrock, openai, gemini, openrouter and ollama for opensource models. I have special slash command and script that collect information about session, project and observed problems to evaluation dataset. I can re-evaluate prompts and find models that do a job in particular agent faster/cheaper, or use automated prompt optimization to eliminate problems.</p>
]]></description><pubDate>Thu, 26 Feb 2026 21:52:55 +0000</pubDate><link>https://news.ycombinator.com/item?id=47172454</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=47172454</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47172454</guid></item><item><title><![CDATA[New comment by alexsmirnov in "My AI Adoption Journey"]]></title><description><![CDATA[
<p>For me, AI is the best for code research and review<p>Since some team members started using AI without care, I did create bunch of agents/skills/commands and custom scripts for claude code. For each PR, it collects changes by git log/diff, read PR data and spin bunch of specialized agents to check code style, architecture, security, performance, and bugs. Each agent armed with necessary requirement documents, including security compliance files. False positives are rare, but it still misses some problems. No PR with ai generated code passes it. If AI did not find any problems, I do manual review.</p>
]]></description><pubDate>Fri, 06 Feb 2026 06:00:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=46909625</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=46909625</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46909625</guid></item><item><title><![CDATA[New comment by alexsmirnov in "LNAI – Define AI coding tool configs once, sync to Claude, Cursor, Codex, etc."]]></title><description><![CDATA[
<p>I did create and actively use a similar tool, but with different purpose: configure AI tools for each team member to use the same code style and architecture guides across projects. It includes:
- build docker images for claude code and opencode dev containers.
- creates custom MCP server that works as a proxy and combines several tools into a single one ( for example, web search, fetch, and context7 tools exposed as a single "web_research" that invokes custom code to answer question )
- copy code style, documentation, and best practice rules for technologies used in our projects
- deploys a bunch of helper scripts useful for development
- configure agents, skills, hooks, and commands to use those rules. Configuration changed per "mode" : documentation, onboarding, code review, and web development all have different settings. 
- run AI tools in docker container with limited permissions
- feedback tool to generate session report, that is used for automatic evaluation and prompt optimization.<p>This came out of necessity, as active using of AI assistants in uncontrollable way significantly degraded code quality. The goal is to enforce the same development workflow across team
This is internal tool. If someone interesting, I can create a public repo from it</p>
]]></description><pubDate>Tue, 03 Feb 2026 18:58:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=46875532</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=46875532</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46875532</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Ask HN: Do you also "hoard" notes/links but struggle to turn them into actions?"]]></title><description><![CDATA[
<p>I do use Obsidian on pair with Claude code and git.<p>I organize notes by tags, folders, and links from tree of "map of content" notes. Those documented as rules for AI. All notes came to "Inbox" folder, and from time  to time I run special script that checks inbox, formats notes, tags them, and put in the most appropriate place. "git diff" to check results and fix mistakes, reset if it went wrong.<p>As notes organized by the limited number of well defined rules, they became easy to search and navigate by AI. Claude Code easily finds requested notes, working as advanced search engine, and they became a starting point for "deep research"  : find relevant notes, follow links, detect gaps, search internet. Repeat until reach required confidence level.<p>The most advanced workflow so far is combination of TRIZ (Theory of Inventive Problem Solving) + First Principles Framework. Former generates ideas and hypotheses, later validates them and converge on final answer.</p>
]]></description><pubDate>Sat, 31 Jan 2026 07:17:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=46834250</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=46834250</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46834250</guid></item><item><title><![CDATA[New comment by alexsmirnov in "The Code-Only Agent"]]></title><description><![CDATA[
<p>This was implemented far ago, at least by huggingface "smolagents". <a href="https://huggingface.co/docs/smolagents/index" rel="nofollow">https://huggingface.co/docs/smolagents/index</a> . I did use them, with evaluations. For the most cases, modern models tool call outperforms code agent. They just trained to use tools, not a code</p>
]]></description><pubDate>Mon, 19 Jan 2026 06:27:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=46675615</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=46675615</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46675615</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Don't fall into the anti-AI hype"]]></title><description><![CDATA[
<p>This is exact the impression that I got. Every question or task given to LLM returns pretty reasonable, but flawed result. For the coding, those are hard to spot but dangerous mistakes. They all look good and perfectly reasonable, but just wrong. Anthropic compared Claude Code to a "slot machine", and I fell that AI coding now is something close to gambling addiction. As small wins keep gambler to make more bets, so correct results from AI keep developers to use it: "I see it made correct solution, let's try again!"
At a startup CTO, I review most of the pull requests from team members, and team uses AI tools actively. The overall picture strongly confirms your second conclusion.</p>
]]></description><pubDate>Mon, 12 Jan 2026 19:14:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=46592863</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=46592863</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46592863</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Why users cannot create Issues directly"]]></title><description><![CDATA[
<p>This is not about understanding the message, but switching user mental activity.
I go myself in the similar situations many times. One example: I tried to pay my bills in online bank application, but got into error. After several attempts, I did read message and it say "Header size exceed..." . It give me clue that app probably put too much history into cookies. Clear browser data, log in again, and all got works.<p>Even when error message was clearly understandable for my expertise, it took surprisingly long tome to switch from one mental activity - "Pay bills", to another - "Investigate technical problem". And you have to throw away all short memory to switch into another task. So all rumors about "stupid" users is direct consequence from how human mind works.</p>
]]></description><pubDate>Sat, 03 Jan 2026 20:09:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=46481015</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=46481015</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46481015</guid></item><item><title><![CDATA[New comment by alexsmirnov in "The Gorman Paradox: Where Are All the AI-Generated Apps?"]]></title><description><![CDATA[
<p>A lot of discussions around vibe AI coding flaws: awful architecture, performance problems, security holes, lack of maintainability, bugs, and low code quality. All correct, but none of those is matter if:<p>- you create small utility that covers only features needed only for you. As many researches show that any individual uses only less than 20% of software functionality, your tool covers only 10-20% that matters for you<p>- it only runs locally, on user computer or phone, and never has more than one customer. Performance, security, compliances do not matter<p>- the code lies next to application, and small enough to fix any bug instantly, in a single AI agent run<p>- as a single user, you don't care about design, UX, or marketing. Do the job is only matter<p>It means, majority of vibe coded applications run under radar, used only by a few individuals. I can see it myself: I have a bunch of vibe code utilities that never intended for a broad auditory . And, many of my friend and customers, mention the same: "I vibe coded utility that does ... for me". This means a big consequences for software development: the area for commercial development shrinks, nothing that can be replaced by the small local utility has a market value.</p>
]]></description><pubDate>Sun, 14 Dec 2025 19:48:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=46266159</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=46266159</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46266159</guid></item><item><title><![CDATA[New comment by alexsmirnov in "The past was not that cute"]]></title><description><![CDATA[
<p>I've read book written by captain Kocebu, that was on duty to protect Russian holdings in Alaska. They visited San Francisco in 1805 and 1815, and several chapters described life of native people in the mission.
He described harsh conditions, hard work, no freedom at all, and very high death rates. Shocking even for a early XIX century naval officer. Once a year, those people allowed to visit their tribes and relatives. And they always came back!
So, the real hunter gathers, who had first hand comparison for both nomadic and agrarian life, prefer near slavery in mission to life in the wild.</p>
]]></description><pubDate>Tue, 09 Dec 2025 01:39:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=46200297</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=46200297</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46200297</guid></item><item><title><![CDATA[New comment by alexsmirnov in "I hate screenshots of text"]]></title><description><![CDATA[
<p>I saw it more and more recent times, but worse is coming:
With my teenage daughter ( and her friends ), I see that they even do not bother with screenshots, and take pictures by phone...</p>
]]></description><pubDate>Tue, 11 Nov 2025 20:02:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=45892072</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=45892072</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45892072</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Eating stinging nettles"]]></title><description><![CDATA[
<p>It's also worth to mention that early spring sprouts do not sting at all. Cold "борщ" from them is delicious.</p>
]]></description><pubDate>Sat, 08 Nov 2025 03:08:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=45853771</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=45853771</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45853771</guid></item><item><title><![CDATA[New comment by alexsmirnov in "Kagi News"]]></title><description><![CDATA[
<p>I prefer to not.
LLM prompts and feeds selection clearly expose my political preferences, interests, and location.
And this is research project for more serious task. For 200 lines of code, there are 1000+ for evaluation, automatic prompt optimization.
But idea is simple, so there should be no problem to implement it</p>
]]></description><pubDate>Wed, 01 Oct 2025 01:37:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=45433407</link><dc:creator>alexsmirnov</dc:creator><comments>https://news.ycombinator.com/item?id=45433407</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45433407</guid></item></channel></rss>