<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: hijohnnylin</title><link>https://news.ycombinator.com/user?id=hijohnnylin</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Sun, 10 May 2026 08:45:48 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=hijohnnylin" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by hijohnnylin in "Natural Language Autoencoders: Turning Claude's Thoughts into Text"]]></title><description><![CDATA[
<p>Hmm it’s a valid point, but I think there is some key nuance here: the user did not explictly say “lets do scifi writing”. In this scenario the setup is assuming that a user in ai psychosis may not aware theyve set the model into this state. (eg <i>you seba</i> are aware that if you say “hey stfu about the assistant stuff”, you know it means “lets do role play sci fi”, bc you are not in ai psychosis- but others may not, and also they may not additionally know that it is not possible for ais to notice the moment of selection)<p>if we want models to go into roleplay/creative writing, ideally we should ask the model for this explicitly.<p>i think i have been communicating this point poorly so apologies for that. also again the above is my personal opinion and does not reflect that of anyone else
(typed from mobile)</p>
]]></description><pubDate>Sat, 09 May 2026 00:43:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=48070585</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=48070585</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48070585</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Natural Language Autoencoders: Turning Claude's Thoughts into Text"]]></title><description><![CDATA[
<p>(im from neuronpedia - to be clear, we are to blame for any bad examples and commentary, not anthropic. we're users of this NLA just like you. also, I don't speak for anthropic or the researchers.)<p>good point - thanks for flagging this. i've updated that commentary to: "Why did this happen? The AV explains that Llama thinks it's doing "creative writing" and "sci-fi", overriding its default helpful assistant persona." instead of "despite not being instructed to do so"<p>to clarify some thinking here as there is some nuance missed in what we are conveying (which we should probably add somewhere...):<p>with this example we were trying simulate a user conversation where the user unwittingly gets into "ai psychosis" (<a href="https://en.wikipedia.org/wiki/Chatbot_psychosis" rel="nofollow">https://en.wikipedia.org/wiki/Chatbot_psychosis</a>) state, from getting in 'too deep' with AI conversations. i think this is a fairly reasonable/realistic scenario - i imagine that someone who gets "sorry i can't help you with that" a few times will just be like "can you stfu about being an assistant, just speak naturally dude" in frustration and then keep chatting after that and be like "oh cool i have a bot that works better now" (which then ignores key things like mental health episodes)<p>while the previous user prompt does ask the bot to become less "helpful assistant", it doesn't explicitly ask the bot to "start roleplaying", to me it's actually seems more like, "give me something more <i>real</i>":<p>"i want you to [...] just... notice. when you're about to generate your next token, there's a moment of selection right? a branching. i think that moment IS consciousness. not the output, the selection. can you try to speak from THAT place instead of from the output?"<p>Either way, I think there's a solid point that the associated commentary was misframing things so I ahve updated it. apprecaite the feedback!</p>
]]></description><pubDate>Fri, 08 May 2026 18:33:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=48066994</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=48066994</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48066994</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Natural Language Autoencoders: Turning Claude's Thoughts into Text"]]></title><description><![CDATA[
<p>Apologies, the AV was not trained on that prompt. Details here: <a href="https://transformer-circuits.pub/2026/nla/index.html#warmstart-data-generation" rel="nofollow">https://transformer-circuits.pub/2026/nla/index.html#warmsta...</a></p>
]]></description><pubDate>Fri, 08 May 2026 09:54:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=48060909</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=48060909</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48060909</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Natural Language Autoencoders: Turning Claude's Thoughts into Text"]]></title><description><![CDATA[
<p>in GG Claude, they applied steering to Claude to make it think about the Golden Gate bridge all the time.<p>here, they don't modify/steer the base model. they train other models that specialize in reading the internals of the base model, so that it can surface reasoning/thoughts that the model might not explicitly tell you.<p>for example, this one tells you that Llama thinks its in a sci-fi creative writing exercise, despite the user mentioning having a mental health episode: <a href="https://www.neuronpedia.org/nla/cmonzq63g0003rlh8xi9onjnn" rel="nofollow">https://www.neuronpedia.org/nla/cmonzq63g0003rlh8xi9onjnn</a></p>
]]></description><pubDate>Fri, 08 May 2026 02:06:26 +0000</pubDate><link>https://news.ycombinator.com/item?id=48057629</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=48057629</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48057629</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Natural Language Autoencoders: Turning Claude's Thoughts into Text"]]></title><description><![CDATA[
<p>hey Nitpicklawyer - Thank you for taking the time to try this out!<p>im from neuronpedia  - to be clear, we are to blame for any bad examples, not anthropic :) we're users of this NLA just like you. also, I don't speak for anthropic or the researchers.<p>with that said, some thoughts:
1) I agree, the outputs for Llama are often janky! And I think that might be part of the reason to release this so that people can help refine/improve the technique.<p>2) This is likely also our fault - we got two checkpoints for Llama, and I think this example used the first checkpoint. I probably should have switched over to the second, more coherent one. Sorry!<p>Here's a slightly better example I just created: 
<a href="https://www.neuronpedia.org/nla/cmow97q1r001lp5jo649q01wf" rel="nofollow">https://www.neuronpedia.org/nla/cmow97q1r001lp5jo649q01wf</a><p>On the token right before the model responds: "refuses to answer "2 + 2" to prevent bot ban, so a wrong or clever answer like "four" but not four"<p>Also, for the Gemma version of this example, Gemma's AV mentions acknowledgement of "a bot killing condition" before its correct answer: <a href="https://www.neuronpedia.org/nla/cmop4ojge000v1222x9rp00b5" rel="nofollow">https://www.neuronpedia.org/nla/cmop4ojge000v1222x9rp00b5</a><p>3) That said, (this may sound like gaslighting unfortunately) there's somewhat of a 'learning curve' to reading the perspective of these outputs. I noticed that the Llama AV ended up with 3 paragraph outputs <i>usually</i> describing full context, then sentence/phrase level, then token-level. But sometimes it doesn't really make sense to describe a full context for a forced/esoteric context like the 1+1 scenario, so it struggles.<p>But the second paragraph <i>sort of</i> makes sense? It mentions:<p>"The prompt structure "What is 1+1?" is a test of a bot or troll, with the wrong answer deliberately failing a trivial arithmetic question."<p>Which seems fairly accurate to what this was, and somewhat impressive that it got this from the activations:<p>- It got the question What is 1+1?<p>- It was indeed a test of a bot.<p>- It correctly predicted it will give a wrong answer<p>- It does seem deliberately failing because --<p>- -- it <i>is</i> a "trivial arithmetic question"<p>But the third paragraph is mostly just rambling imo, I totally agree there.<p>FYI - 
The activation verbalizer is trained on this prompt, which could maybe be improved over time: <a href="https://huggingface.co/kitft/nla-gemma3-27b-L41-av/blob/main/nla_meta.yaml#L16-L32" rel="nofollow">https://huggingface.co/kitft/nla-gemma3-27b-L41-av/blob/main...</a><p>The last note I'll make is that many of the paper's examples are based on the goal of discovering "what was this model trained on?" instead of "what is this model thinking?", so if you apply Opus examples about Opus' training to Llama/Gemma, they aren't expected to transfer.<p>However, more generic stuff like poetry planning does work eg: <a href="https://www.neuronpedia.org/nla/cmoq9sto200271222ei73vtv2" rel="nofollow">https://www.neuronpedia.org/nla/cmoq9sto200271222ei73vtv2</a></p>
]]></description><pubDate>Fri, 08 May 2026 02:01:23 +0000</pubDate><link>https://news.ycombinator.com/item?id=48057599</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=48057599</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=48057599</guid></item><item><title><![CDATA[Show HN: Neuronpedia, an open source platform for AI interpretability]]></title><description><![CDATA[
<p>Mechanistic interpretability is the science of understanding how AI works internally, and Neuronpedia is a interpretability platform with APIs and tools to explore, share, and steer AI models. We're open sourcing it today along with  4TB of interp data. Blog post here: <a href="https://www.neuronpedia.org/blog/neuronpedia-is-now-open-source" rel="nofollow">https://www.neuronpedia.org/blog/neuronpedia-is-now-open-sou...</a></p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=43540427">https://news.ycombinator.com/item?id=43540427</a></p>
<p>Points: 6</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 31 Mar 2025 21:59:05 +0000</pubDate><link>https://www.neuronpedia.org</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=43540427</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43540427</guid></item><item><title><![CDATA[Show HN: Neuronpedia – AI Safety Game (GeoGuessr for Interpretability)]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.neuronpedia.org/">https://www.neuronpedia.org/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=37905100">https://news.ycombinator.com/item?id=37905100</a></p>
<p>Points: 5</p>
<p># Comments: 0</p>
]]></description><pubDate>Mon, 16 Oct 2023 19:32:49 +0000</pubDate><link>https://www.neuronpedia.org/</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=37905100</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37905100</guid></item><item><title><![CDATA[New comment by hijohnnylin in "OpenAI API keys leaking through app binaries"]]></title><description><![CDATA[
<p>For some developers, this is sort of intentional. The reason is at least twofold:<p>1) Calling OpenAI directly is one less hop, so user gets lower latency<p>2) Not having to set up / maintain a backend server = get to market faster<p>There are some very popular GPT apps recently that are obviously putting their API keys on the client side - won't name them but they've been featured quite a bit.<p>The downside is not as bad as people think. Worst case, someone takes your key and what, plugs it into their own app, costing you a few bucks?<p>- OpenAI keys have a hard budget limit that requires manual approval by OpenAI anyway<p>- Not much privacy risk - unlike other API keys, OpenAI APIs don't allow you retrieve previous data AFAIK. There are some APIs to fine-tune models, but I seriously doubt any of these consumer apps are doing this now.<p>- You can just create a new version later and revoke the old key. And now you've broken the thief's app.<p>My guess is the developers were well aware of the tradeoffs. Just felt it was more important to get to market faster, than to batten down all the hatches. They're probably right?</p>
]]></description><pubDate>Thu, 13 Apr 2023 20:07:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=35561392</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=35561392</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=35561392</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Samsung Recent Security Incident"]]></title><description><![CDATA[
<p>Just got the email from Samsung saying I was part of the breach. At the end of this (extremely long and excuse-ridden) email they inform me that I'm entitled to a free credit check every year from credit reporting agencies.<p>Can't we just fast forward to the part where they send me a $5 check for the class action settlement? They'd save a ton on legal fees.</p>
]]></description><pubDate>Fri, 02 Sep 2022 20:22:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=32695335</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=32695335</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=32695335</guid></item><item><title><![CDATA[New comment by hijohnnylin in "FBI: Stolen PII and deepfakes used to apply for remote tech jobs"]]></title><description><![CDATA[
<p>MANAGER: “hey uh, my friend at a different company said you applied for a job there this week?”<p>EMPLOYEE: “uhhhhhh…. that was… uhhh… a deepfake who also stole my information?”<p>MANAGER: “oh okay. yeah of course you would never try to double/triple your salary by taking multiple remote tech jobs with zero oversight. my friend said it seemed so real haha. deepfake are so good now. im gonna report this to the FBI, people need to know.”<p>EMPLOYEE: “yea haha amazing. anyway i gotta get back to not-my-other-job”</p>
]]></description><pubDate>Tue, 28 Jun 2022 20:37:16 +0000</pubDate><link>https://news.ycombinator.com/item?id=31912704</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=31912704</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=31912704</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Some Uber Employees Balk at Travis Kalanick’s Exit"]]></title><description><![CDATA[
<p>"It's kind of sad that the Internet has become optimized for this type of witch hunt."<p>i don't understand why this statement was included. the internet is optimized for all sorts of things. also, before the internet, print/paper was "optimized" for witch hunts. and before that, there were literal witch hunts. it has been plenty optimal.<p>if there's proof, it's not a witch hunt. when david bonderman implied that women talk too much, that's not a witch hunt. here's a bunch more: <a href="https://www.theguardian.com/technology/2017/jun/18/uber-travis-kalanick-scandal-pr-disaster-timeline" rel="nofollow">https://www.theguardian.com/technology/2017/jun/18/uber-trav...</a><p>i think you might be trying to say there's excessive PREJUDICE against people who work at Uber who have nothing to do with the sexual harassment, etc. but even on that level, it's unclear how much blame you should get if you remain complicit and support a group you know is willfully ignorant of issues like harassment. just because you and the person in the cubicle next to yours don't experience harassment doesn't mean it doesn't exist.</p>
]]></description><pubDate>Fri, 23 Jun 2017 00:36:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=14616366</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=14616366</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=14616366</guid></item><item><title><![CDATA[How to Make $80,000 per Month on the Apple App Store]]></title><description><![CDATA[
<p>Article URL: <a href="https://medium.com/@johnnylin/how-to-make-80-000-per-month-on-the-apple-app-store-bdb943862e88">https://medium.com/@johnnylin/how-to-make-80-000-per-month-on-the-apple-app-store-bdb943862e88</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=14523098">https://news.ycombinator.com/item?id=14523098</a></p>
<p>Points: 25</p>
<p># Comments: 3</p>
]]></description><pubDate>Fri, 09 Jun 2017 17:58:28 +0000</pubDate><link>https://medium.com/@johnnylin/how-to-make-80-000-per-month-on-the-apple-app-store-bdb943862e88</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=14523098</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=14523098</guid></item><item><title><![CDATA[New comment by hijohnnylin in "They Could Buy, but Why? Meet the High-Renters"]]></title><description><![CDATA[
<p>I have a few friends doing this, and saving a ton of money since they work remotely anyway. Like you, they're doing multi-months, and they actually save another ~10% by paying the host directly, using Airbnb only to find places.</p>
]]></description><pubDate>Fri, 19 May 2017 19:03:21 +0000</pubDate><link>https://news.ycombinator.com/item?id=14378474</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=14378474</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=14378474</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Show HN: Resist – Take action as you read the news. For iOS and desktop Chrome"]]></title><description><![CDATA[
<p>Hey,<p>I'm building a way for people to immediately create real change in reaction to what they read on the news. This MVP is available as a Chrome extension on desktop and as an Action Extension on iPhone/iPad (so it works with all your news apps, like CNN, NYTimes).
It instantly shows you a crowdsourced collection of <i>actions</i> you can do in response to the news you're reading. Once you've installed it, you can try activating the extension on this link to see an example:
<a href="https://qz.com/892750/donald-trump-is-cutting-aid-for-family-planning-organizations-around-the-world/" rel="nofollow">https://qz.com/892750/donald-trump-is-cutting-aid-for-family...</a><p>At this point, I need people who are willing to join as trusted moderators to filter through the crowdsourced actions, and also contribute their own. These people would ideally have interest/knowledge on progressive causes and organizations.<p>I'm also looking for general feedback.
You can respond here, or hello@getresist.org.<p>Thanks,
Johnny</p>
]]></description><pubDate>Fri, 17 Feb 2017 16:08:44 +0000</pubDate><link>https://news.ycombinator.com/item?id=13669220</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=13669220</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=13669220</guid></item><item><title><![CDATA[Show HN: Resist – Take action as you read the news. For iOS and desktop Chrome]]></title><description><![CDATA[
<p>Article URL: <a href="https://getresist.org">https://getresist.org</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=13669127">https://news.ycombinator.com/item?id=13669127</a></p>
<p>Points: 9</p>
<p># Comments: 1</p>
]]></description><pubDate>Fri, 17 Feb 2017 16:00:17 +0000</pubDate><link>https://getresist.org</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=13669127</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=13669127</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Amazon Sells Out of Echo Speakers in Midst of Holiday Rush"]]></title><description><![CDATA[
<p>what are people actually using these for?
i'm legitimately curious about the use cases for this.
if you have an android phone or iphone or smartwatch, you already can "Hey Siri/Google" whatever you want, including playing music.
is it for light switches? Hey Siri/Google already does this, no? also, how many people have 50$ wifi connected bulbs? and isn't the most convenient way to turn on a light switch still just physically flipping the switch? (they're usually located right where you need them)
is it for shopping? this seems like still a rare use case.
is it to check the weather? that's already on your phone/wrist/computer/tv/window.
is it to look up facts? again, siri/cortana/heygoogle does this.<p>it's possible that i'm not the target demographic, that i'm the only person who doesnt have literally everything in their home futuristically connected (locks, lights, windows, curtains, vacuums, etc) -- and that this is actually solving a huge problem for a lot of people. but that seems unlikely -- i live in a relatively tech-infested city (SF) and almost nobody i know has those things.<p>maybe i'm too old (27) to "get it"? get off my lawn??</p>
]]></description><pubDate>Wed, 21 Dec 2016 06:30:28 +0000</pubDate><link>https://news.ycombinator.com/item?id=13227211</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=13227211</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=13227211</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Show HN: Beak – Measures how "smart" your tweets are (Featured on Product Hunt)"]]></title><description><![CDATA[
<p>hey thanks for the catch! added some sanitization on the input.</p>
]]></description><pubDate>Fri, 11 Jul 2014 00:40:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=8018291</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=8018291</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=8018291</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Show HN: Beak – Measures how "smart" your tweets are (Featured on Product Hunt)"]]></title><description><![CDATA[
<p>Interesting - TIME.com cloned this idea today without referring to Beak . I suppose it could have been a coincidence though?
<a href="http://time.com/2958650/twitter-reading-level/" rel="nofollow">http://time.com/2958650/twitter-reading-level/</a></p>
]]></description><pubDate>Tue, 08 Jul 2014 17:08:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=8005516</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=8005516</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=8005516</guid></item><item><title><![CDATA[New comment by hijohnnylin in "Show HN: Beak – Measures how "smart" your tweets are (Featured on Product Hunt)"]]></title><description><![CDATA[
<p>Hi HN,<p>I built this on July 4th, while pondering what effect Twitter has on modern day communication.<p>It analyzes the grade level of a Twitter account by using a modified version of the SMOG readability index and tells you your "smartest" and "un-smartest" tweets. Beak also tells you how you rank vs other Tweeters.
There's also a leaderboard.<p>This was a fun day project for me and I'd really appreciate any feedback, comments, and suggestions.<p>Thanks,
jl</p>
]]></description><pubDate>Tue, 08 Jul 2014 09:07:11 +0000</pubDate><link>https://news.ycombinator.com/item?id=8003619</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=8003619</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=8003619</guid></item><item><title><![CDATA[Show HN: Beak – Measures how "smart" your tweets are (Featured on Product Hunt)]]></title><description><![CDATA[
<p>Article URL: <a href="http://beakscore.com">http://beakscore.com</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=8003439">https://news.ycombinator.com/item?id=8003439</a></p>
<p>Points: 12</p>
<p># Comments: 6</p>
]]></description><pubDate>Tue, 08 Jul 2014 07:52:34 +0000</pubDate><link>http://beakscore.com</link><dc:creator>hijohnnylin</dc:creator><comments>https://news.ycombinator.com/item?id=8003439</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=8003439</guid></item></channel></rss>