<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: evmaki</title><link>https://news.ycombinator.com/user?id=evmaki</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 26 Apr 2026 08:27:12 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=evmaki" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by evmaki in "Claude Code to be removed from Anthropic's Pro plan?"]]></title><description><![CDATA[
<p>I would love someone to play devil's advocate against this perspective:<p>While these tools stand to enable the democratization of productive capability in software engineering and other tasks (creating a renaissance for solopreneurs, let's say), what seems more likely to actually happen is that entrenched capital will become the only player with real access to this "knowledge as a utility" (was it Altman who called it that?).<p>We already see this playing out in two fronts: 1) the gradual reduction of services and 2) the DRAM market, where local-first tools (i.e., potential disruptors of the emerging "knowledge monopoly" created by the big AI firms) are being stifled by supply shortages. How many promising small-to-medium-sized competitors are being snuffed out of existence (or never starting) due to the insanity of the DRAM/storage/CPU (soon) markets?<p>The currently-subsidized access that we have to the big Opus-like models will, in parallel, be gradually be taken away until only the big players can afford it. And in the end what we will have is hyper-productive skeleton crews at a few consolidated firms performing (or selling expensive access to) basically all of the knowledge labor for society, with very little potential for disruption due to the hardware and "knowledge" scarcity engineered (in part, maybe) by this monopoly.<p>Not necessarily a closely held belief – just a hunch – which is why I want to see what parts of the picture I might be missing.</p>
]]></description><pubDate>Wed, 22 Apr 2026 00:13:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=47856763</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=47856763</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47856763</guid></item><item><title><![CDATA[New comment by evmaki in "Tell HN: Fiverr left customer files public and searchable"]]></title><description><![CDATA[
<p>Extremely bad stuff here. Can't believe it's been 7 hours now and you can still pull up people's complete prepared tax returns right from a Google search. This should be a business-ending breach of trust and good practices, but I worry there's probably a lack of regulatory might or will to make anything happen.</p>
]]></description><pubDate>Wed, 15 Apr 2026 02:25:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=47773942</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=47773942</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47773942</guid></item><item><title><![CDATA[New comment by evmaki in "University of Texas limits on teaching of "unnecessary controversial subjects""]]></title><description><![CDATA[
<p>Disappointing to see my alma mater gradually self-sabotage into irrelevance.</p>
]]></description><pubDate>Thu, 19 Feb 2026 20:51:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=47079131</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=47079131</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47079131</guid></item><item><title><![CDATA[New comment by evmaki in "Warren Buffett dumps $1.7B of Amazon stock"]]></title><description><![CDATA[
<p>> My Echo, that I use solely to voice activating lights and switches, is now an ad machine<p>I've been wondering if it is even possible for a publicly-traded company to deliver a voice assistant product without these incentives involved. I have to imagine the UX of these devices would be much different if they were built by a private company without the same market pressures. It would need to be self-contained and local, so that the infrastructure burden (e.g., data and AI in the cloud) wouldn't create a need for subscription service or data collection revenue to cover the cost.</p>
]]></description><pubDate>Wed, 18 Feb 2026 18:12:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=47064147</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=47064147</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47064147</guid></item><item><title><![CDATA[New comment by evmaki in "Claims of disability are highest at elite universities"]]></title><description><![CDATA[
<p>Some of that is probably because students at elite universities come from families who understand exactly how to navigate meritocratic systems. Parents who know you need a tutor to ace the ACT and an essay editor to polish your application essays likely also understand that, in some situations, you should get a psychiatric evaluation and disclose any diagnoses so you can receive accommodations.<p>I'm fairly convinced a big part of clearing the barrier to entry of these elite institutions is having a deep understanding of exactly the things you need to do to succeed given the structure of the system and the nature of the competition. Students at "non-elite" institutions are more likely to come from backgrounds where even if you DO have a disability, maybe nobody ever tells you that you can go to the doctor for it, or that something like "accommodations" exist to help you.</p>
]]></description><pubDate>Tue, 17 Feb 2026 23:46:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=47055128</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=47055128</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47055128</guid></item><item><title><![CDATA[New comment by evmaki in ""Token anxiety", a slot machine by any other name"]]></title><description><![CDATA[
<p>The LLM is not the slot machine. The LLM is the lever of the slot machine, and the slot machine itself is capitalism. Pull the lever, see if it generates a marketable product or moment of virality, get rich if you hit the jackpot. If not, pull again.</p>
]]></description><pubDate>Mon, 16 Feb 2026 22:43:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=47041366</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=47041366</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47041366</guid></item><item><title><![CDATA[New comment by evmaki in "SkillsBench: Benchmarking how well agent skills work across diverse tasks"]]></title><description><![CDATA[
<p>> I ask it to reflect on why, and update the Skill to clarify, adding or removing detail as necessary.<p>We are probably undervaluing the human part of the feedback loop in this discussion. Claude is able to solve the problem given the appropriate human feedback — many then jump to the conclusion that well, if Claude is capable of doing it under some circumstances, we just need to figure out how to remove the human part so that Claude can eventually figure it out itself.<p>Humans are still serving a very crucial role in disambiguation, and in centering the most salient information. We do this based on our situational context, which comes from hands-on knowledge of the problem space. I'm hesitant to assume that because Claude CAN bootstrap skills (which is damn impressive!), it would somehow eventually do so entirely on its own, devoid of any situational context beyond a natural language spec.</p>
]]></description><pubDate>Mon, 16 Feb 2026 21:54:27 +0000</pubDate><link>https://news.ycombinator.com/item?id=47040849</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=47040849</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=47040849</guid></item><item><title><![CDATA[Mirage: A new breed of sampler, powered by tiny generative audio models]]></title><description><![CDATA[
<p>Article URL: <a href="https://evanking.io/posts/mirage/">https://evanking.io/posts/mirage/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46965386">https://news.ycombinator.com/item?id=46965386</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 10 Feb 2026 19:22:17 +0000</pubDate><link>https://evanking.io/posts/mirage/</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=46965386</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46965386</guid></item><item><title><![CDATA[New comment by evmaki in "America isn't exceptional – it's the exception"]]></title><description><![CDATA[
<p>US incarceration rates increased 500% in the decades following the enactment of war on drugs legislation, starting with the Controlled Substances Act in 1970.<p><a href="https://en.wikipedia.org/wiki/Incarceration_in_the_United_States" rel="nofollow">https://en.wikipedia.org/wiki/Incarceration_in_the_United_St...</a><p><a href="https://en.wikipedia.org/wiki/Incarceration_in_the_United_States#/media/File:US_incarceration_timeline-clean.svg" rel="nofollow">https://en.wikipedia.org/wiki/Incarceration_in_the_United_St...</a><p>More speculatively, I think the prison system has also taken over the role of the mental health institutions that were wound down under Reagan. Over half of the incarcerated population has a mental health condition, and likely are not receiving adequate mental health care while incarcerated.</p>
]]></description><pubDate>Tue, 10 Feb 2026 04:15:47 +0000</pubDate><link>https://news.ycombinator.com/item?id=46955325</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=46955325</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46955325</guid></item><item><title><![CDATA[New comment by evmaki in "Interview Coder Just Leaked Full Names and Companies of All SWEs Who Cheated [video]"]]></title><description><![CDATA[
<p>Dishonesty layered on dishonesty, marketed with an arrogant smirk on top. I feel like tech culture has fully internalized the ethos of "no attention is bad attention", so having a lack of scruples and a talent for rage baiting is now seen as an advantage. It might pay off well for some people in the short term, but it's not a sustainable way to run a society.</p>
]]></description><pubDate>Tue, 13 Jan 2026 01:24:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=46596336</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=46596336</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46596336</guid></item><item><title><![CDATA[New comment by evmaki in "Show HN: I built a synth for my daughter"]]></title><description><![CDATA[
<p>What a beautiful and charming project. Kudos for taking it all the way from zero to one with such a polished design. That's no small feat. I've built prototypes for eurorack and even with some simplifying constraints it's a lot of work.<p>Best of luck with your Kickstarter!</p>
]]></description><pubDate>Mon, 17 Nov 2025 19:10:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=45956893</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=45956893</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45956893</guid></item><item><title><![CDATA[New comment by evmaki in "Show HN: Aris – a free AI-powered answer engine for kids"]]></title><description><![CDATA[
<p>The underlying idea tracks. The next generation of kids is going to interact with AI, and we should anticipate that and try to build systems that are healthy and safe for them to interact with.<p>On the other hand, I wonder if this doesn't just further alienate children from their parents. Kids are already given access to unlimited supernormal stimuli via iPads so that parents don't have to parent. This just seems like more of that: now parents don't even need to have basic conversations with their kids because the AI can do it.<p>Anecdotally, some of the most formative interactions I had as a child started by asking my parents questions. These were things that not only shaped me as a person, but deepened my relationship with my parents. These interactions are important, and I wonder if Aris doesn't just abstract it away into another "service" that further deepens social decay. I would not be the person I am today if I hadn't had the chance to ask my dad as an angsty pre-teen what the point of life is, and for him to tell me it is to learn and create so that we can make a better world for humanity. I guarantee a smoothed-over LLM would not have offered something so personally impactful.<p>My two cents is that you should ponder that deeper point a little bit, and think about how it informs the way you market your idea, and scope the service it provides.</p>
]]></description><pubDate>Fri, 12 Sep 2025 23:42:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=45227968</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=45227968</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45227968</guid></item><item><title><![CDATA[New comment by evmaki in "OpenAI says it's scanning users' conversations and reporting content to police"]]></title><description><![CDATA[
<p>So dropping out of CS to start selling something was more important to him than 2 more years of CS education. Maybe he realized that continuing his engineering education was unnecessary because he preferred selling things. Sounds like a salesman.</p>
]]></description><pubDate>Tue, 02 Sep 2025 21:37:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=45109405</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=45109405</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45109405</guid></item><item><title><![CDATA[Mishearings: Hacking a tiny ASR model to write Dadaist poetry]]></title><description><![CDATA[
<p>Article URL: <a href="https://evanking.io/posts/mishearings/">https://evanking.io/posts/mishearings/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=44856410">https://news.ycombinator.com/item?id=44856410</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Sun, 10 Aug 2025 16:48:38 +0000</pubDate><link>https://evanking.io/posts/mishearings/</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=44856410</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=44856410</guid></item><item><title><![CDATA[New comment by evmaki in "Leaving Google"]]></title><description><![CDATA[
<p>> AI is nothing but a tool for wealth transfer from what remains in the middle class to the top ultra wealthy.<p>Is that inherent to the technology, or is that just inherent to the way we've chosen to organize society? Really, any technological paradigm shift going back to the industrial revolution has mainly served to enrich a small few people and families, but that's not some immutable property of technology. Like you say, it's a tool. We've chosen (or allowed) it to be wielded toward one end rather than another. I can smash my neighbor's head in with a hammer, or I can build a home with it.<p>At one point in the United States, there was political will to update our social structures so that the fruits of technological and economic progress did not go disproportionately to one class of society (think early 20th century trust busting, or the New Deal coming out of the Great Depression). I'm afraid we find ourselves with a similar set of problems, yet the political will to create some other reality beyond further wealth concentration seems to be limited.</p>
]]></description><pubDate>Sun, 11 May 2025 19:25:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=43956340</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=43956340</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43956340</guid></item><item><title><![CDATA[New comment by evmaki in "Building a fully local LLM voice assistant to control my smart home"]]></title><description><![CDATA[
<p>Awesome write-up - especially the fact that you've gotten it working with good performance locally. It certainly requires a little bit more hardware than your typical home assistant, but I think this will change over time :)<p>I've been working on this problem in an academic setting for the past year or so [1]. We built a very similar system in a lab at UT Austin and did a user study (demo here <a href="https://youtu.be/ZX_sc_EloKU" rel="nofollow">https://youtu.be/ZX_sc_EloKU</a>). We brought a bunch of different people in and had them interact with the LLM home assistant without any constraints on their command structure. We wanted to see how these systems might choke in a more general setting when deployed to a broader base of users (beyond the hobbyist/hacker community currently playing with them).<p>Big takeaways there: we need a way to do long-term user and context personalization. This is both a matter of knowing an individual's preferences better, but also having a system that can reason with better sensitivity to the limitations of different devices. To give an example, the system might turn on a cleaning robot if you say "the dog made a mess in the living room" -- impressive, but in practice this will hurt more than it helps because the robot can't actually clean up that type of mess.<p>[1] <a href="https://arxiv.org/abs/2305.09802" rel="nofollow">https://arxiv.org/abs/2305.09802</a></p>
]]></description><pubDate>Sun, 14 Jan 2024 15:54:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=38991525</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=38991525</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38991525</guid></item><item><title><![CDATA[New comment by evmaki in "Using ChatGPT for home automation"]]></title><description><![CDATA[
<p>Thanks! That is part of the challenge as this idea scales imo - once you've increased the number of plugins or "levers" available to the model, you start to increase the likelihood that it will pull some of them indiscriminately.<p>To your point about turning random commands into API calls: if you give it the raw JSON from a Philips Hue bridge and ask it to manipulate it in response to commands, it can even do oddly specific things like triggering Hue-specific lighting effects [0] without any description in the plugin yaml. I'm assuming some part of the corpus contains info about the Hue API.<p>[0] <a href="https://evanking.io/posts/homegpt/" rel="nofollow">https://evanking.io/posts/homegpt/</a></p>
]]></description><pubDate>Sat, 20 May 2023 19:41:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=36014929</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=36014929</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36014929</guid></item><item><title><![CDATA[New comment by evmaki in "Using ChatGPT for home automation"]]></title><description><![CDATA[
<p>It certainly makes logical sense. I think if you have the ability to control the light in the first place via an API, it's probably an LED smart bulb and thus doesn't produce much heat. At least, I'm not aware of any incandescent smart bulbs.</p>
]]></description><pubDate>Sat, 20 May 2023 19:36:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=36014889</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=36014889</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36014889</guid></item><item><title><![CDATA[New comment by evmaki in "Using ChatGPT for home automation"]]></title><description><![CDATA[
<p>Super exciting to see the work happening in this area! I can especially appreciate the use of ChatGPT to orchestrate the necessary API calls, rather than relying on some kind of middleware to do it.<p>I have been working in this area (LLMs for ubiquitous computing, more generally) for my PhD dissertation and have discovered some interesting quality issues when you dig deeper [0]. If you only have lights in your house, for instance, the GPT models will always use them in response to just about any command you give, then post-rationalize the answer. If I say "it's too chilly in here" in a house with only lights, it will turn them on as a way of "warming things up". Kind of like a smart home form of hallucination. I think these sorts of quality issues will be the big hurdle to product integration.<p>[0] <a href="https://arxiv.org/abs/2305.09802" rel="nofollow">https://arxiv.org/abs/2305.09802</a></p>
]]></description><pubDate>Sat, 20 May 2023 18:03:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=36014135</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=36014135</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=36014135</guid></item><item><title><![CDATA[Using ChatGPT to control a smart home]]></title><description><![CDATA[
<p>Article URL: <a href="https://evanking.io/posts/homegpt/">https://evanking.io/posts/homegpt/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=35212768">https://news.ycombinator.com/item?id=35212768</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 18 Mar 2023 20:24:05 +0000</pubDate><link>https://evanking.io/posts/homegpt/</link><dc:creator>evmaki</dc:creator><comments>https://news.ycombinator.com/item?id=35212768</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=35212768</guid></item></channel></rss>