<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: Jianghong94</title><link>https://news.ycombinator.com/user?id=Jianghong94</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 17 Apr 2026 07:03:18 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=Jianghong94" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by Jianghong94 in "Mercury 2: Fast reasoning LLM powered by diffusion"]]></title><description><![CDATA[
<p>Honestly I don't understand why they/any fast-and-error-prone model position themselves as coding agents; my experience tells me that I'd much rather working with a slow-but-correct model and let it run longer session than handholding a fast-but-wrong model.</p>
]]></description><pubDate>Wed, 25 Feb 2026 19:15:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47156353</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=47156353</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47156353</guid></item><item><title><![CDATA[New comment by Jianghong94 in "I'm not worried about AI job loss"]]></title><description><![CDATA[
<p>This. At this point AI/LLM/Claude Code is still a power user tool; the more you know about your domain + the more you're willing to reasonably use it, the more gain you have.<p>That being said the real danger is not coming from AI today, it's more C-suites believing AI can just zero shot any problem you throw at it.</p>
]]></description><pubDate>Fri, 13 Feb 2026 21:46:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=47008264</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=47008264</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47008264</guid></item><item><title><![CDATA[New comment by Jianghong94 in "I'm not worried about AI job loss"]]></title><description><![CDATA[
<p>Well, like I said, there're hidden incentives behind the scene; in my case, the hidden incentive is that, the requester/client is one of the company's subpar broker, and PM probably decided to just offer an average level of commitment, not going above and beyond. Hence the plan was to do exactly what the broker want even though that was messy and inferior. You can't just write down that kind of motivation on paper anywhere.<p>---
I said it because I did the analysis, and realized that if I implement the original version, which basically is a crazy way to iteratively solve the MIP problem, it's much harder to reason with internally, and much harder to code correctly. But obviously it keep the broker happy (the developer is doing exactly what I said)</p>
]]></description><pubDate>Fri, 13 Feb 2026 21:35:50 +0000</pubDate><link>https://news.ycombinator.com/item?id=47008162</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=47008162</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47008162</guid></item><item><title><![CDATA[New comment by Jianghong94 in "I'm not worried about AI job loss"]]></title><description><![CDATA[
<p>Maybe I'm being naive here, but for AI (heck, for any good algorithm) to work well, you need some at least loosely-clearly defined objectives. I assume it's much more straightforward in semi, but there're many industries, once you get into the details, all kinds of incentives start to disalign and I doubt AI could understand all kinds of nuances.<p>E.g. once I was tasked to build a new matching algorithm for a trading platform, and upon fully understanding of the specs I realized it can be interpreted as a mixed integer programming problem; the idea got shot down right away because PM don't understand it. There're all kinds of limiting factors once you get into the details.</p>
]]></description><pubDate>Fri, 13 Feb 2026 20:00:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=47007056</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=47007056</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47007056</guid></item><item><title><![CDATA[New comment by Jianghong94 in "Is 2026 next year?"]]></title><description><![CDATA[
<p>I believe so, see my result with Haiku extended thinking on. I think the weights are just too biased towards blurping out the majority of the training data of 'next year is xxx'. Interesting problem to solve indeed.</p>
]]></description><pubDate>Tue, 02 Dec 2025 20:17:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=46126251</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=46126251</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46126251</guid></item><item><title><![CDATA[New comment by Jianghong94 in "Is 2026 next year?"]]></title><description><![CDATA[
<p>I think the current trick for LLM API provider is to insert the today is $DATE into the system prompt, so maybe it's worthwhile to do that and see if that automatically fixes those OSS models?</p>
]]></description><pubDate>Tue, 02 Dec 2025 20:15:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=46126224</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=46126224</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46126224</guid></item><item><title><![CDATA[New comment by Jianghong94 in "Is 2026 next year?"]]></title><description><![CDATA[
<p>I did a similar test especially with the extended thinking on and off for Haiku, and once you have extended thinking on, the result is more or less the same as Sonnet.<p>Thought process:
The user is asking if 2026 is next year. According to the context, today's date is Tuesday, December 02, 2025. So the current year is 2025. That means next year would be 2026. So yes, 2026 is next year.Yes, 2026 is next year. 
Actual resp
Since we're currently in December 2025, 2026 is just about a month away.</p>
]]></description><pubDate>Tue, 02 Dec 2025 20:13:46 +0000</pubDate><link>https://news.ycombinator.com/item?id=46126207</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=46126207</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46126207</guid></item><item><title><![CDATA[New comment by Jianghong94 in "The lazy Git UI you didn't know you need"]]></title><description><![CDATA[
<p>I don't think JB UIs been changing that much, albeit I haven't been working in the industry long enough. I think the last major UI redo was like 2,3 years ago and most of it is to make UI more compact, and I definitely like it. That being said YMMV.<p>IMHO another good thing of using JB IDE git UI (especially in a corporate setting) instead of using another software is that everyone has the IDE so it's easier to collaborate. Imagine if you're helping a junior member debug their local branch and they don't have lazy git installed.</p>
]]></description><pubDate>Thu, 13 Nov 2025 00:39:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=45908991</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=45908991</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45908991</guid></item><item><title><![CDATA[New comment by Jianghong94 in "The lazy Git UI you didn't know you need"]]></title><description><![CDATA[
<p>Wait, no one mentions the default JetBrains IDE git UI? I mean, I get it if you're working from another IDE/text editor that doesn't have good git UI support out of the box, but JB's git UI is reasonably good enough that I don't want anything else.<p>Things that I use (and I like):
1. quick checkout to another branch and automatically stash and unstash your local changes; when I just need to inspect code elsewhere I find it really useful. My changes are small so I can always remember to stash them later;
2. compare branch/commit etc via UI; again I know you can do that in git diff, but then you would need to know the command and the commit SHA to compare; in UI it comes in really handy, just select the branch or commits you want to compare and that's it. I've seen my coworkers trying to come up with the command and I just say: use IDE and a couple of clicks they got it working.
3. filter commits by user and by folder.</p>
]]></description><pubDate>Tue, 11 Nov 2025 15:07:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=45888087</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=45888087</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45888087</guid></item><item><title><![CDATA[New comment by Jianghong94 in "What we talk about when we talk about sideloading"]]></title><description><![CDATA[
<p>An even more grotesque practice is to charge a stratosphere level premium for the product itself <i>AND</i> put its control behind a subscription e.g. 8sleep</p>
]]></description><pubDate>Tue, 28 Oct 2025 21:02:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=45739129</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=45739129</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45739129</guid></item><item><title><![CDATA[New comment by Jianghong94 in "Living Dangerously with Claude"]]></title><description><![CDATA[
<p>this seem solvable if the whitelisting just allows regex</p>
]]></description><pubDate>Fri, 24 Oct 2025 04:00:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=45690622</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=45690622</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45690622</guid></item><item><title><![CDATA[New comment by Jianghong94 in "OpenAI acquires Sky.app"]]></title><description><![CDATA[
<p>My take is that it's a standalone business consideration: Apple users are more inclined to pay for software (definitely the case for iPhone vs. Android, although I haven't found a source for Windows).</p>
]]></description><pubDate>Fri, 24 Oct 2025 03:52:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=45690578</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=45690578</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45690578</guid></item><item><title><![CDATA[New comment by Jianghong94 in ""Vibe code hell" has replaced "tutorial hell" in coding education"]]></title><description><![CDATA[
<p>OR problems are hard because whoever try to vibe coding it probably don't realize they fall into a specific algorithm and can prompt llm to do thatl; what's worse is that even if you tell them so they won't be able to understand the math behind it and would much prefer their vide coding solution.</p>
]]></description><pubDate>Fri, 10 Oct 2025 16:37:29 +0000</pubDate><link>https://news.ycombinator.com/item?id=45540848</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=45540848</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45540848</guid></item><item><title><![CDATA[New comment by Jianghong94 in "AI 2027"]]></title><description><![CDATA[
<p>Not only does the article claim that when we get to self-improving ai it becomes generally intelligent, it's assuming that AI is pretty close right now:<p>> OpenBrain focuses on AIs that can speed up AI research. They want to win the twin arms races against China (whose leading company we’ll call “DeepCent”)16 and their US competitors. The more of their research and development (R&D) cycle they can automate, the faster they can go. So when OpenBrain finishes training Agent-1, a new model under internal development, it’s good at many things but great at helping with AI research.<p>> It’s good at this due to a combination of explicit focus to prioritize these skills, their own extensive codebases they can draw on as particularly relevant and high-quality training data, and coding being an easy domain for procedural feedback.<p>> OpenBrain continues to deploy the iteratively improving Agent-1 internally for AI R&D. Overall, they are making algorithmic progress 50% faster than they would without AI assistants—and more importantly, faster than their competitors.<p>> what do we mean by 50% faster algorithmic progress? We mean that OpenBrain makes as much AI research progress in 1 week with AI as they would in 1.5 weeks without AI usage.<p>To me, claiming today's AI IS capable of such thing is too hand-wavy.  And I think that's the crux of the article.</p>
]]></description><pubDate>Fri, 04 Apr 2025 17:36:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=43585531</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=43585531</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43585531</guid></item><item><title><![CDATA[New comment by Jianghong94 in "AI 2027"]]></title><description><![CDATA[
<p>Putting the geopolitical discussion aside, I think the biggest question lies in how likely the *current paradigm LLM* (think of it as any SOTA stock LLM you get today, e.g., 3.7 sonnet, gemini 2.5, etc) + fine-tuning will be capable of directly contributing to LLM research in a major way.<p>To quote the original article,<p>> OpenBrain focuses on AIs that can speed up AI research. They want to win the twin arms races against China (whose leading company we’ll call “DeepCent”)16 and their US competitors. The more of their research and development (R&D) cycle they can automate, the faster they can go. So when OpenBrain finishes training Agent-1, a new model under internal development, it’s good at many things but great at helping with AI research. (footnote: It’s good at this due to a combination of explicit focus to prioritize these skills, their own extensive codebases they can draw on as particularly relevant and high-quality training data, and coding being an easy domain for procedural feedback.)<p>> OpenBrain continues to deploy the iteratively improving Agent-1 internally for AI R&D. Overall, they are making algorithmic progress 50% faster than they would without AI assistants—and more importantly, faster than their competitors.<p>> what do we mean by 50% faster algorithmic progress?
We mean that OpenBrain makes as much AI research progress in 1 week with AI as they would in 1.5 weeks without AI usage.<p>> AI progress can be broken down into 2 components:<p>> Increasing compute: More computational power is used to train or run an AI. This produces more powerful AIs, but they cost more.<p>> Improved algorithms: Better training methods are used to translate compute into performance. This produces more capable AIs without a corresponding increase in cost, or the same capabilities with decreased costs.<p>> This includes being able to achieve qualitatively and quantitatively new results. “Paradigm shifts” such as the switch from game-playing RL agents to large language models count as examples of algorithmic progress.<p>> Here we are only referring to (2), improved algorithms, which makes up about half of current AI progress.<p>---<p>Given that the article chose a pretty aggressive timeline (the algo needs to contribute late this year so that its research result can be contributed to the next gen LLM coming out early next year), the AI that can contribute significantly to research has to be a current SOTA LLM.<p>Now, using LLM in day-to-day engineering task is no secret in major AI labs, but we're talking about something different, something that gives you 2 extra days of output per week. I have no evidence to either acknowledge or deny whether such AI exists, and it would be outright ignorant to think no one ever came up with such an idea or is trying such an idea. So I think it goes down into two possibilities:<p>1. This claim is made by a top-down approach, that is, if AI reaches superhuman in 2027, what would be the most likely starting condition to that? And the author picks this as the most likely starting point, since the authors don't work in major AI lab (even if they do they can't just leak such trade secret), the authors just assume it's likely to happen anyway (and you can't dismiss that).
2. This claim is made by a bottom-up approach, that is the author did witness such AI exists to a certain extent and start to extrapolate from there.</p>
]]></description><pubDate>Fri, 04 Apr 2025 17:26:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=43585417</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=43585417</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43585417</guid></item><item><title><![CDATA[New comment by Jianghong94 in "AI 2027"]]></title><description><![CDATA[
<p>Well based on what I'm reading, the OP's intent is that, not all (hence 'fully') validation, if not most of, can be done in-silico. I think we all agree that and that's the major bottleneck making agents useful - you have to have human-in-the-loop to closely guardrail the whole process.<p>Of course you can get a lot of mileage via synthetically generated CoT but does that lead to LLM speed up developing LLM is a big IF.</p>
]]></description><pubDate>Fri, 04 Apr 2025 16:54:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=43585030</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=43585030</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43585030</guid></item><item><title><![CDATA[New comment by Jianghong94 in "AI agents: Less capability, more reliability, please"]]></title><description><![CDATA[
<p>Yep that's what I've been thinking. This shouldn't be that hard, at this point LLMs should already have all the 'rules' (e.g. credit card A buys flight X give you m point which can be converted into n miles) in their params or can easily query the web to get it out. Dev need to encode the whole thing into a decision mechanism and once executed ask LLM to chase down the specific path (e.g. bombard ticket office with emails).</p>
]]></description><pubDate>Mon, 31 Mar 2025 16:16:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=43536703</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=43536703</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43536703</guid></item><item><title><![CDATA[New comment by Jianghong94 in "AI agents: Less capability, more reliability, please"]]></title><description><![CDATA[
<p>Now THAT's the workflow I'd like to see AI agent automate, streamline and democratize for everybody.</p>
]]></description><pubDate>Mon, 31 Mar 2025 15:52:12 +0000</pubDate><link>https://news.ycombinator.com/item?id=43536409</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=43536409</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43536409</guid></item><item><title><![CDATA[New comment by Jianghong94 in "AI Agents: Less Capability, More Reliability, Please"]]></title><description><![CDATA[
<p>Superhuman results 1/10 are, in fact, a very strong reliability guarantee (maybe not up to today's nth 9 decimal standard that we are accustomed to, but probably much higher than any agent in real-world workflow).</p>
]]></description><pubDate>Mon, 31 Mar 2025 15:50:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=43536391</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=43536391</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43536391</guid></item><item><title><![CDATA[New comment by Jianghong94 in "FrontierMath: A benchmark for evaluating advanced mathematical reasoning in AI"]]></title><description><![CDATA[
<p>I guess the primary reason is that the answers must be numbers that can be verified easily. Otherwise, you just flood the validator with long LLM reasoning that's hard to verify. People have been proposing using LEAN as a medium for answers but AFAIK even LEAN is not mainstream in the general math community, so there's always trade-offs.<p>Also, coming up with good problems is an art in its own right; the Soviets was famous for institutionalizing anti-Semitism via special math puzzles for Jews in Moscow Univerisity entrance exams. The questions are constructed as such that are hard to solve but have some elementary solutions to divert criticism.</p>
]]></description><pubDate>Mon, 11 Nov 2024 00:14:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=42103558</link><dc:creator>Jianghong94</dc:creator><comments>https://news.ycombinator.com/item?id=42103558</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42103558</guid></item></channel></rss>