<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: ismailmaj</title><link>https://news.ycombinator.com/user?id=ismailmaj</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Thu, 09 Apr 2026 06:43:55 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=ismailmaj" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by ismailmaj in "MegaTrain: Full Precision Training of 100B+ Parameter LLMs on a Single GPU"]]></title><description><![CDATA[
<p>Obsolete because of what? Because with limited hardware you’re never aiming for state of the art, and for fine-tuning, you don’t steer for too long anyway.</p>
]]></description><pubDate>Wed, 08 Apr 2026 13:55:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=47690278</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47690278</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47690278</guid></item><item><title><![CDATA[New comment by ismailmaj in "25 Years of Eggs"]]></title><description><![CDATA[
<p>I don't know why people mess with tesseract in 2026, attention-based OCRs (and more recently VLMs) outperformed any LSTM-based approach since at least 2020.<p>My guess is that it's the entry-point to OCR and the internet is flooded by that, just like pandas for data processing.</p>
]]></description><pubDate>Sun, 22 Mar 2026 12:55:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=47477019</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47477019</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47477019</guid></item><item><title><![CDATA[New comment by ismailmaj in "BitNet: Inference framework for 1-bit LLMs"]]></title><description><![CDATA[
<p>Assuming 2 bit per values (first bit is sign and second bit is value).<p>actv = A[_:1] & B[_:1]<p>sign = A[_:0] ^ B[_:0]<p>dot = pop_count(actv & !sign) - pop_count(actv & sign)<p>It can probably be made more efficient by taking a column-first format.<p>Since we are in CPU land, we mostly deal with dot products that match the cache size, I don't assume we have a tiled matmul instruction which is unlikely to support this weird 1-bit format.</p>
]]></description><pubDate>Wed, 11 Mar 2026 18:09:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=47339074</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47339074</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47339074</guid></item><item><title><![CDATA[New comment by ismailmaj in "Surpassing vLLM with a Generated Inference Stack"]]></title><description><![CDATA[
<p>Confusing, since this is specific to an architecture that no one making money will use (8B is consumer space, not enterprise).
The produced code shouldn't hold much interesting IP?</p>
]]></description><pubDate>Wed, 11 Mar 2026 17:43:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=47338732</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47338732</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47338732</guid></item><item><title><![CDATA[New comment by ismailmaj in "Why the global elite gave up on spelling and grammar"]]></title><description><![CDATA[
<p>Oh no, cortisol spike in my text-only forum.</p>
]]></description><pubDate>Wed, 11 Mar 2026 17:24:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=47338468</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47338468</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47338468</guid></item><item><title><![CDATA[New comment by ismailmaj in "BitNet: 100B Param 1-Bit model for local CPUs"]]></title><description><![CDATA[
<p>The title and the repo uses 1-bit when it means 1.58 bits tertiary values, it doesn't change any of my arguments (still xors and pop_counts).</p>
]]></description><pubDate>Wed, 11 Mar 2026 17:02:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=47338172</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47338172</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47338172</guid></item><item><title><![CDATA[New comment by ismailmaj in "BitNet: 100B Param 1-Bit model for local CPUs"]]></title><description><![CDATA[
<p>You drop the memory throughput requirements because of the packed representation of bits so an FMA can become the bottleneck, and you bypass the problem of needing to upscale the bits to whatever FP the FMA instruction needs.<p>typically for 1-bit matmul, you can get away with xors and pop_counts which should have a better throughput profile than FMA when taking into account the SIMD nature of the inputs/outputs.</p>
]]></description><pubDate>Wed, 11 Mar 2026 14:37:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47336142</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47336142</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47336142</guid></item><item><title><![CDATA[New comment by ismailmaj in "Yann LeCun raises $1B to build AI that understands the physical world"]]></title><description><![CDATA[
<p>Ah yes, OpenAI the puppet of Microsoft that is currently declaring war against GitHub, sounds logical.</p>
]]></description><pubDate>Wed, 11 Mar 2026 11:35:56 +0000</pubDate><link>https://news.ycombinator.com/item?id=47334283</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47334283</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47334283</guid></item><item><title><![CDATA[New comment by ismailmaj in "Surpassing vLLM with a Generated Inference Stack"]]></title><description><![CDATA[
<p>Any place we can find the code?</p>
]]></description><pubDate>Wed, 11 Mar 2026 11:01:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=47334028</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47334028</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47334028</guid></item><item><title><![CDATA[New comment by ismailmaj in "Yann LeCun raises $1B to build AI that understands the physical world"]]></title><description><![CDATA[
<p>Don’t think that’s a fair interpretation of what I said.<p>Liquid money rich? No.<p>Can get pulled for big tech packages? Also no, for most of the employees.<p>AFAIK, big tech didn’t aggressively poach OpenAI-like talent, they did spend 10M+ pay packages but it was for a select few research scientists. Some folks left and came but it boiled down to culture mostly.</p>
]]></description><pubDate>Tue, 10 Mar 2026 19:39:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47327864</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47327864</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47327864</guid></item><item><title><![CDATA[New comment by ismailmaj in "Yann LeCun raises $1B to build AI that understands the physical world"]]></title><description><![CDATA[
<p>I don't see a world where they become threatening and the employees don't become rich from investors flooding in.</p>
]]></description><pubDate>Tue, 10 Mar 2026 18:01:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=47326708</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47326708</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47326708</guid></item><item><title><![CDATA[New comment by ismailmaj in "I put my whole life into a single database"]]></title><description><![CDATA[
<p>In my experience, tracking objective things like "nutrition" and "sleep hours" is immensely useful to reflect on what went wrong, and tracking subjective things like "mood" or "stress" is useless given hedonic adaptation or heavy swings that make problems obvious, and not need tracking.<p>What's key is be able to visualize metrics easily on the data and frictionless data entry, I've got a decent setup with iPhone Action + Obsidian + QuickAdd scripts on Obsidian Sync (mobile + laptop). for visualization I use Obsidian Bases and Obsidian notes that run Dataview code blocks and Chart.js, couldn't be happier.<p>I could track things that are not interesting to reflect on like vitamin D supplementation for accountability but I've never bothered, especially if it's taken ~daily.</p>
]]></description><pubDate>Tue, 10 Mar 2026 11:32:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47321841</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47321841</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47321841</guid></item><item><title><![CDATA[New comment by ismailmaj in "My “grand vision” for Rust"]]></title><description><![CDATA[
<p>Surprised with the push back of the comments, getting effects on Rust would be a dream.<p>Could even enable some stuff like passing loggers around not by parameters but by effect.</p>
]]></description><pubDate>Mon, 09 Mar 2026 10:16:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=47307079</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47307079</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47307079</guid></item><item><title><![CDATA[New comment by ismailmaj in "Tech employment now significantly worse than the 2008 or 2020 recessions"]]></title><description><![CDATA[
<p>Don’t need the EU for that, they are hitting everything in 2026 including unemployment, though nothing passed yet.</p>
]]></description><pubDate>Sun, 08 Mar 2026 09:45:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=47295936</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47295936</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47295936</guid></item><item><title><![CDATA[New comment by ismailmaj in "MacBook Air with M5"]]></title><description><![CDATA[
<p>I bought the M4 air given that the consensus was also that it was the best value for 1k, but I ended up returning it and went for pro base model for a few reasons (still valid for M5 AFAIK from a bit of research):<p>- pro motion (120hz screen).<p>- better display brightness which is important when there is a bright sun outside.<p>- 1 more USB-C port and HDMI port (no dongle hell).<p>- 20% more battery life.<p>- This is more personal, but 13" is too small and 15" is too big, so 14" MBP worked best for me (~25 HFOV with a stand + KBM).<p>It's hard to justify saving 400 bucks given the gap between the models, but the decision is closer since the air has 16GB memory by default since M4 AFAIK.</p>
]]></description><pubDate>Tue, 03 Mar 2026 19:18:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=47237375</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47237375</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47237375</guid></item><item><title><![CDATA[New comment by ismailmaj in "Show HN: I built a sub-500ms latency voice agent from scratch"]]></title><description><![CDATA[
<p>Semantic end of turn being 300ms of silence is horrible because I ended up intentionally um-ing to finish my thoughts before getting answer.<p>It was difficult to detrain and that made me stop using voice chat with LLMs all together.</p>
]]></description><pubDate>Tue, 03 Mar 2026 08:50:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=47229895</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47229895</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47229895</guid></item><item><title><![CDATA[New comment by ismailmaj in "Gemini 3.1 Pro"]]></title><description><![CDATA[
<p>3.1 feels to me like 3.0 but that takes a long time to think, it didn't feel like a leap in raw intelligence like 2.5 pro was.</p>
]]></description><pubDate>Thu, 19 Feb 2026 21:37:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=47079826</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47079826</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47079826</guid></item><item><title><![CDATA[New comment by ismailmaj in "YouTube Is Down"]]></title><description><![CDATA[
<p>Cannot trust those damn Americans during on call.</p>
]]></description><pubDate>Wed, 18 Feb 2026 13:32:42 +0000</pubDate><link>https://news.ycombinator.com/item?id=47060746</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47060746</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47060746</guid></item><item><title><![CDATA[New comment by ismailmaj in "Async/Await on the GPU"]]></title><description><![CDATA[
<p>Warp specialization is an abomination that should be killed and I'm glad this could be an alternative.<p>I hope they can minimize the bookkeeping costs because I don't see it gain traction in AI if it hurts big kernels performance.</p>
]]></description><pubDate>Tue, 17 Feb 2026 20:44:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=47053053</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=47053053</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47053053</guid></item><item><title><![CDATA[New comment by ismailmaj in "Gemini 3 Deep Think"]]></title><description><![CDATA[
<p>top 10 elo in codeforces is pretty absurd</p>
]]></description><pubDate>Thu, 12 Feb 2026 18:37:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=46993043</link><dc:creator>ismailmaj</dc:creator><comments>https://news.ycombinator.com/item?id=46993043</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46993043</guid></item></channel></rss>