<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: philipkglass</title><link>https://news.ycombinator.com/user?id=philipkglass</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 17 Apr 2026 13:10:49 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=philipkglass" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by philipkglass in "The noise we make is hurting animals. Can we learn to shut up?"]]></title><description><![CDATA[
<p>This was interesting enough to prompt an investigation. It looks like the idea comes from this letter that he wrote late in life:<p><a href="https://philosophics.blog/2021/08/01/the-futility-of-words/" rel="nofollow">https://philosophics.blog/2021/08/01/the-futility-of-words/</a><p>I don't know German so I can't search for the original with any facility.</p>
]]></description><pubDate>Thu, 16 Apr 2026 21:16:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=47799602</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47799602</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47799602</guid></item><item><title><![CDATA[New comment by philipkglass in "CPUs Aren't Dead. Gemma2B Out Scored GPT-3.5 Turbo on Test That Made It Famous"]]></title><description><![CDATA[
<p>I think that we're getting there. I put together a workstation in early 2023 with a single 4090 GPU. I did it to run things like BERT and YOLO image classifiers. At that point the only "open weights" LLM was the original Llama from Meta, and even that was open-weights only because it was leaked. It was a very weak model by today's standards.<p>With the same hardware I now get genuine utility out of models like Qwen 3.5 for categorizing and extracting unstructured data sources. I don't use small local models for coding since frontier models are so much stronger, but if I had to go back to small models for coding too they would be more useful than anything commercially available as recently as 4 years ago.</p>
]]></description><pubDate>Wed, 15 Apr 2026 18:37:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47783304</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47783304</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47783304</guid></item><item><title><![CDATA[New comment by philipkglass in "The Future of Everything Is Lies, I Guess: Safety"]]></title><description><![CDATA[
<p><i>In short, the ML industry is creating the conditions under which anyone with sufficient funds can train an unaligned model. Rather than raise the bar against malicious AI, ML companies have lowered it.</i><p>This is true, and I believe that the "sufficient funds" threshold will keep dropping too. It's a relief more than a concern, because I don't trust that big models from American or Chinese labs will always be aligned with what I need. There are probably a lot of people in the world whose interests are not especially aligned with the interests of the current AI research leaders.<p>"Don't turn the visible universe into paperclips" is a practically universal "good alignment" but the models we have can't do that anyhow. The actual refusal-guards that frontier models come with are a lot more culturally/historically contingent and less universal. Lumping them all under "safety" presupposes the outcome of a debate that has been philosophically unresolved forever. If we get hundreds of strong models from different groups all over the world, I think that it will improve the net utility of AI and disarm the possibility of one lab or a small cartel using it to control the rest of us.</p>
]]></description><pubDate>Mon, 13 Apr 2026 17:56:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=47755646</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47755646</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47755646</guid></item><item><title><![CDATA[New comment by philipkglass in "Seven countries now generate nearly all their electricity from renewables (2024)"]]></title><description><![CDATA[
<p>At present, those tandem cells are still experimental. Nobody is manufacturing them on gigawatt scale like for other solar cell technologies.</p>
]]></description><pubDate>Sun, 12 Apr 2026 14:51:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47740418</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47740418</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47740418</guid></item><item><title><![CDATA[First criticality for Indian fast breeder reactor]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.world-nuclear-news.org/articles/first-criticality-for-indian-fast-breeder-reactor">https://www.world-nuclear-news.org/articles/first-criticality-for-indian-fast-breeder-reactor</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=47679228">https://news.ycombinator.com/item?id=47679228</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 07 Apr 2026 18:16:45 +0000</pubDate><link>https://www.world-nuclear-news.org/articles/first-criticality-for-indian-fast-breeder-reactor</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47679228</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47679228</guid></item><item><title><![CDATA[New comment by philipkglass in "Show HN: I Built Paul Graham's Intellectual Captcha Idea"]]></title><description><![CDATA[
<p>At least one of the test questions was just a screen shot from a tweet. It was difficult to read. I'd suggest extracting text from screen shots with OCR. Apple has built-in functionality for this on their operating systems with Live Text. There are strong open source systems based on small vision language models for this, too. The one I have been recommending lately is GLM-OCR:<p><a href="https://github.com/zai-org/GLM-OCR" rel="nofollow">https://github.com/zai-org/GLM-OCR</a><p>It's fast and can run even on low-resource computers.<p>---<p>Does this CAPTCHA actually resist computers? I didn't try feeding the questions I got to an LLM, but my sense is that current frontier models could probably pass all of these too. Making generated text pass the pangram test is simple enough for someone actually writing a bot to spin up automated accounts.</p>
]]></description><pubDate>Mon, 06 Apr 2026 16:53:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=47663485</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47663485</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47663485</guid></item><item><title><![CDATA[New comment by philipkglass in "Solar and batteries can power the world"]]></title><description><![CDATA[
<p>Unlike the US, Vietnam is a net importer of fuel. It imports over 40 million tons of coal per year:<p><a href="https://statbase.org/data/vnm-coal-imports/" rel="nofollow">https://statbase.org/data/vnm-coal-imports/</a><p>It also started importing liquid natural gas in 2023.<p>But it has abundant sunlight, access to low cost Chinese solar panels that will produce electricity for decades instead of being burned once, and a substantial domestic photovoltaic manufacturing industry of its own.<p>"Renewable Energy Investments in Vietnam in 2024 – Asia’s Next Clean Energy Powerhouse" (June 2024)<p><a href="https://energytracker.asia/renewable-energy-investments-in-vietnam-asias-next-clean-energy-powerhouse/" rel="nofollow">https://energytracker.asia/renewable-energy-investments-in-v...</a><p><i>In 2014, the share of renewable energy in Vietnam was just 0.32%. In 2015, only 4 megawatts (MW) of installed solar capacity for power generation was available. However, within five years, investment in solar energy, for example, soared.</i><p><i>As of 2020, Vietnam had over 7.4 gigawatts (GW) of rooftop solar power connected to the national grid. These renewable energy numbers surpassed all expectations. It marked a 25-fold increase in installed capacity compared to 2019’s figures.</i><p><i>In 2021, the data showed that Vietnam now has 16.5 GW of solar power. This was accompanied by its green energy counterpart wind at 11.8 GW. A further 6.6 GW is expected in late 2021 or 2022. Ambitiously, the government plans to further bolster this by adding 12 GW of onshore and offshore wind by 2025.</i><p>These growth rates are actually much faster than growth rates in the US.</p>
]]></description><pubDate>Fri, 03 Apr 2026 17:47:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=47629725</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47629725</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47629725</guid></item><item><title><![CDATA[New comment by philipkglass in "Google releases Gemma 4 open models"]]></title><description><![CDATA[
<p>When you self-host, there's still a client/server relationship between your self-hosted inference server and the client that manages the processing of individual pages. You can get timeouts depending on the configured timeouts, the speed of your inference server, and the complexity of the pages you're processing. But you can let the client retry and/or raise the initial timeout limit if you keep running into timeouts.<p>That said, this is already a small and fast model when hosted via MLX on macOS. If you run the inference server with a recent NVidia GPU and vLLM on Linux it should be significantly faster. The big advantage with vLLM for OCR models is its continuous batching capability. Using other OCR models that I couldn't self-host on macOS, like DeepSeek 2 OCR or Chandra 2, vLLM gave dramatic throughput improvements on big documents via continuous batching if I process 8-10 pages at a time. This is with a single 4090 GPU.</p>
]]></description><pubDate>Fri, 03 Apr 2026 14:54:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=47627353</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47627353</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47627353</guid></item><item><title><![CDATA[New comment by philipkglass in "Google releases Gemma 4 open models"]]></title><description><![CDATA[
<p>I also get connection timeouts on larger documents, but it automatically retries and completes. All the pages are processed when I'm done. However, I'm using the Python client SDK for larger documents rather than the basic glmocr command line tool. I'm not sure if that makes a difference.</p>
]]></description><pubDate>Thu, 02 Apr 2026 22:04:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=47620775</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47620775</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47620775</guid></item><item><title><![CDATA[New comment by philipkglass in "Google releases Gemma 4 open models"]]></title><description><![CDATA[
<p>If you have a basic ARM MacBook, GLM-OCR is the best single model I have found for OCR with good table extraction/formatting. It's a compact 0.9b parameter model, so it'll run on systems with only 8 GB of RAM.<p><a href="https://github.com/zai-org/GLM-OCR" rel="nofollow">https://github.com/zai-org/GLM-OCR</a><p>Use mlx-vlm for inference:<p><a href="https://github.com/zai-org/GLM-OCR/blob/main/examples/mlx-deploy/README.md" rel="nofollow">https://github.com/zai-org/GLM-OCR/blob/main/examples/mlx-de...</a><p>Then you can run a single command to process your PDF:<p><pre><code>  glmocr parse example.pdf

  Loading images: example.pdf
  Found 1 file(s)
  Starting Pipeline...
  Pipeline started!
  GLM-OCR initialized in self-hosted mode
  Using Pipeline (enable_layout=true)...

  === Parsing: example.pdf (1/1) ===
</code></pre>
My test document contains scanned pages from a law textbook. It's two columns of text with a lot of footnotes. It took 60 seconds to process 5 pages on a MBP with M4 Max chip.<p>After it's done, you'll have a directory output/example/ that contains .md and .json files. The .md file will contain a markdown rendition of the complete document. The .json file will contain individual labeled regions from the document along with their transcriptions. If you get all the JSON objects with<p><pre><code>  "label": "table"
</code></pre>
from the JSON file, you can get an HTML-formatted table from each "content" section of these objects.<p>It might still be inaccurate -- I don't know how challenging your original tables are -- but it shouldn't be terribly slow. The tables it produced for me were good.<p>I have also built more complex work flows that use a mixture of OCR-specialized models and general purpose VLM models like Qwen 3.5, along with software to coordinate and reconcile operations, but GLM-OCR by itself is the best first thing to try locally.</p>
]]></description><pubDate>Thu, 02 Apr 2026 20:47:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=47619948</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47619948</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47619948</guid></item><item><title><![CDATA[New comment by philipkglass in "Renewables reached nearly 50% of global electricity capacity last year"]]></title><description><![CDATA[
<p>It takes time for statistical agencies to compile reports. I haven't yet found reports covering the growth in renewable <i>generation</i> (actual terawatt hours) for all of 2025. But this covers 3 quarters of the year:<p><a href="https://ember-energy.org/latest-updates/solar-and-wind-growth-meets-all-new-electricity-demand-in-the-first-three-quarters-of-2025/" rel="nofollow">https://ember-energy.org/latest-updates/solar-and-wind-growt...</a><p><i>In the first three quarters of 2025, solar generation rose by 498 TWh (+31%) and already surpassed the total solar output in all of 2024. Wind generation grew by 137 TWh (+7.6%). Together, they added 635 TWh, outpacing the rise in global electricity demand of 603 TWh (+2.7%).</i></p>
]]></description><pubDate>Thu, 02 Apr 2026 17:44:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=47617666</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47617666</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47617666</guid></item><item><title><![CDATA[New comment by philipkglass in "Google releases Gemma 4 open models"]]></title><description><![CDATA[
<p>Do you have plans to do a follow-up model release with quantization aware training as was done for Gemma 3?<p><a href="https://developers.googleblog.com/en/gemma-3-quantized-aware-trained-state-of-the-art-ai-to-consumer-gpus/" rel="nofollow">https://developers.googleblog.com/en/gemma-3-quantized-aware...</a><p>Having 4 bit QAT versions of the larger models would be great for people who only have 16 or 24 GB of VRAM.</p>
]]></description><pubDate>Thu, 02 Apr 2026 17:25:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=47617356</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47617356</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47617356</guid></item><item><title><![CDATA[New comment by philipkglass in "Slovenia becomes first EU country to introduce fuel rationing"]]></title><description><![CDATA[
<p>"Lithium mining commences in Finland"<p><a href="https://www.electrive.com/2026/02/12/lithium-mining-commences-in-finland/" rel="nofollow">https://www.electrive.com/2026/02/12/lithium-mining-commence...</a><p><i>This week, the first spodumene vein was blasted from the rock at the open-pit mine in western Finland, marking the occasion with a ceremonial event attended by invited guests and media.</i></p>
]]></description><pubDate>Fri, 27 Mar 2026 21:56:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=47548847</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47548847</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47548847</guid></item><item><title><![CDATA[New comment by philipkglass in "Last gasps of the rent seeking class?"]]></title><description><![CDATA[
<p>The small models that I can run at home are becoming more capable, and I have replaced some API-based tasks with local inference as they improve, but large open weights models are still a lot stronger. The nice thing with larger open weights models is that competing providers serve them at modest margins and prices. I don't have the hardware to run the largest Qwen models, but I can get API access at low cost. Since there are only modest barriers to new commercial inference providers for these models I'm not worried that API access to them will become drastically more expensive at some future time.</p>
]]></description><pubDate>Fri, 27 Mar 2026 16:44:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=47545078</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47545078</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47545078</guid></item><item><title><![CDATA[New comment by philipkglass in "Data centers are transitioning from AC to DC"]]></title><description><![CDATA[
<p>One more exception: uranium. It actually splits into smaller atoms when it's used as fuel.</p>
]]></description><pubDate>Wed, 25 Mar 2026 16:17:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=47519435</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47519435</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47519435</guid></item><item><title><![CDATA[New comment by philipkglass in "Blocking Internet Archive Won't Stop AI, but Will Erase Web's Historical Record"]]></title><description><![CDATA[
<p>I don't think that it is karma-gated, but you can only edit the comment for two hours after it's posted. Then it's frozen.</p>
]]></description><pubDate>Tue, 24 Mar 2026 18:44:38 +0000</pubDate><link>https://news.ycombinator.com/item?id=47507307</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47507307</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47507307</guid></item><item><title><![CDATA[New comment by philipkglass in "The bridge to wealth is being pulled up with AI"]]></title><description><![CDATA[
<p>I think that the parent means Wall-E: <a href="https://en.wikipedia.org/wiki/WALL-E#Plot" rel="nofollow">https://en.wikipedia.org/wiki/WALL-E#Plot</a></p>
]]></description><pubDate>Tue, 24 Mar 2026 16:07:30 +0000</pubDate><link>https://news.ycombinator.com/item?id=47504797</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47504797</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47504797</guid></item><item><title><![CDATA[New comment by philipkglass in "Nanopositioning Metrology, Gödel, and Bootstraps"]]></title><description><![CDATA[
<p>If you found this interesting and are a non-expert, I recommend browsing <i>Foundations of Mechanical Accuracy</i> by Wayne Moore (1970):<p><a href="https://archive.org/details/Foundations_of_Mechanical_Accuracy_by_Wayne_R_Moore_1970" rel="nofollow">https://archive.org/details/Foundations_of_Mechanical_Accura...</a><p>Manufacturing got down to nanometer precision starting with macroscopic precision visible to the naked eyeball.</p>
]]></description><pubDate>Mon, 23 Mar 2026 16:40:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=47491859</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47491859</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47491859</guid></item><item><title><![CDATA[New comment by philipkglass in "FDA links raw cheese to outbreak; Makers "100% disagree," refuse recall"]]></title><description><![CDATA[
<p>A local small grocery chain started stocking raw milk (with many warnings) and I decided to risk consuming it to see for myself. I couldn't tell the difference between it and ordinary full fat milk. I wondered if it was a fraud (commercial milk falsely advertised as Local Forbidden Delicacy Milk), but maybe there's not much difference. Or maybe I am not a subtle taster. I also can't taste the superiority of an $80 bottle of wine when it's pitted against an $18 bottle.</p>
]]></description><pubDate>Wed, 18 Mar 2026 15:27:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=47426970</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47426970</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47426970</guid></item><item><title><![CDATA[New comment by philipkglass in "GPT‑5.4 Mini and Nano"]]></title><description><![CDATA[
<p>The crazy cheap models may be adequate for a task, and low cost matters with volume. I need to label millions of images to determine if they're sexually suggestive (this includes but is not limited to nudity). The Gemini 2.0 Flash Lite model is inexpensive and performs well. Gemini 2.5 Flash Lite is also good, but not noticeably better, and it costs more. When 2.0 gets retired this June my costs are going up.</p>
]]></description><pubDate>Tue, 17 Mar 2026 18:18:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=47416281</link><dc:creator>philipkglass</dc:creator><comments>https://news.ycombinator.com/item?id=47416281</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47416281</guid></item></channel></rss>