<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: alew1</title><link>https://news.ycombinator.com/user?id=alew1</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Fri, 08 May 2026 18:35:04 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=alew1" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by alew1 in "Language may rely less on complex grammar than previously thought: study"]]></title><description><![CDATA[
<p>The article presents the fact that we appear to treat non-constituents (eg “in the middle of the”) as “units” to mean that language is more like “snapping legos together” than “building trees.”<p>But linguists have proposed the possibility that we store “fragments” to facilitate reuse—essentially trees with holes, or equivalently, functions that take in tree arguments and produce tree results. “In the middle of the” could take in a noun-shaped tree as an argument and produce a prepositional phrase-shaped tree as a result, for instance. Furthermore, this accounts for the way we store idioms that are not just contiguous “Lego block” sequences of words (like “a ____ and a half” or “the more ___, the more ____”). See e.g. work on “fragment grammars.”<p>Can’t access the actual Nature Human Behavior article so perhaps it discusses the connections.</p>
]]></description><pubDate>Sat, 24 Jan 2026 22:23:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=46748334</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=46748334</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46748334</guid></item><item><title><![CDATA[New comment by alew1 in "Can LLMs do randomness?"]]></title><description><![CDATA[
<p>The algorithms are not deterministic: they output a probability distribution over next tokens, which is then sampled. That’s why clicking “retry” gives you a different answer. An LM could easily (in principle) compute a 50/50 distribution when asked to flip a coin.</p>
]]></description><pubDate>Wed, 30 Apr 2025 11:19:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=43843619</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=43843619</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43843619</guid></item><item><title><![CDATA[New comment by alew1 in "US Government threatens Harvard with foreign student ban"]]></title><description><![CDATA[
<p>Harvard is one of eleven American universities that practice need-blind admissions even for international students, meaning that students are admitted without regard for their financial status (i.e., no explicit preference toward richer students who can pay more tuition), and that financial aid covers full demonstrated need for all admitted students.<p><a href="https://en.wikipedia.org/wiki/Need-blind_admission" rel="nofollow">https://en.wikipedia.org/wiki/Need-blind_admission</a></p>
]]></description><pubDate>Fri, 18 Apr 2025 00:06:45 +0000</pubDate><link>https://news.ycombinator.com/item?id=43723596</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=43723596</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43723596</guid></item><item><title><![CDATA[New comment by alew1 in "US Government threatens Harvard with foreign student ban"]]></title><description><![CDATA[
<p>The taxpayer funding is for Harvard's research activities, not for its undergraduate teaching. The undergraduate teaching is funded by tuition (often paid in full by international students) and by returns on the endowment (including some earmarked for financial aid).<p>> they could vastly grow their class size without lowering standards<p>The issue isn't the quality of the students they are accepting, but the resources to educate and house them, including classroom space, dorms, and staff.</p>
]]></description><pubDate>Thu, 17 Apr 2025 19:04:22 +0000</pubDate><link>https://news.ycombinator.com/item?id=43720862</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=43720862</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43720862</guid></item><item><title><![CDATA[New comment by alew1 in "GPT-4.5"]]></title><description><![CDATA[
<p>Didn't seem to realize that "Still more coherent than the OpenAI lineup" wouldn't make sense out of context. (The actual comment quoted there is responding to someone who says they'd name their models Foo, Bar, Baz.)</p>
]]></description><pubDate>Fri, 28 Feb 2025 00:06:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=43200072</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=43200072</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43200072</guid></item><item><title><![CDATA[New comment by alew1 in "Entropy of a Large Language Model output"]]></title><description><![CDATA[
<p>Any interpretation (including interpreting the <i>inputs</i> to the neural net as a "prompt") is "slapped on" in some sense—at some level, it's all just numbers being added, multiplied, and so on.<p>But I wouldn't call the probabilistic interpretation "after the fact." The entire training procedure that generated the LM weights (the pre-training as well as the RLHF post-training) is formulated based on the understanding that the LM predicts p(x_t | x_1, ..., x_{t-1}). For example, pretraining maximizes the log probability of the training data, and RLHF typically maximizes an objective that combines "expected reward [under the LLM's output probability distribution]" with "KL divergence between the pretraining distribution and the RLHF'd distribution" (a probabilistic quantity).</p>
]]></description><pubDate>Mon, 13 Jan 2025 21:07:20 +0000</pubDate><link>https://news.ycombinator.com/item?id=42689340</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=42689340</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42689340</guid></item><item><title><![CDATA[New comment by alew1 in "Entropy of a Large Language Model output"]]></title><description><![CDATA[
<p>"Temperature" doesn't make sense unless your model is predicting a distribution. You can't "temperature sample" a calculator, for instance. The output of the LLM is a predictive distribution over the next token; this is the formulation you will see in every paper on LLMs. It's true that you can do various things with that distribution <i>other</i> than sampling it: you can compute its entropy, you can find its mode (argmax), etc., but the type signature of the LLM itself is `prompt -> probability distribution over next tokens`.</p>
]]></description><pubDate>Mon, 13 Jan 2025 16:25:49 +0000</pubDate><link>https://news.ycombinator.com/item?id=42685170</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=42685170</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42685170</guid></item><item><title><![CDATA[New comment by alew1 in "Large Enough"]]></title><description><![CDATA[
<p>If I show you a strawberry and ask how many r’s are in the name of this fruit, you can tell me, because one of the things you know about strawberries is how to spell their name.<p>Very large language models also “know” how to spell the word associated with the strawberry token, which you can test by asking them to spell the word one letter at a time. If you ask the model to spell the word and count the R’s while it goes, it can do the task. So the failure to do it when asked directly (how many r’s are in strawberry) is pointing to a real weakness in reasoning, where one forward pass of the transformer is not sufficient to retrieve the spelling and also count the R’s.</p>
]]></description><pubDate>Wed, 24 Jul 2024 20:16:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=41061537</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=41061537</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41061537</guid></item><item><title><![CDATA[New comment by alew1 in "A new old kind of R&D lab"]]></title><description><![CDATA[
<p>One thought: If you want to be able to remove the static part, you could consider fine tuning without the static part. If you fine tune with, you’re teaching the model that the desired behavior occurs only in the presence of the static part (hence the going off the rails).</p>
]]></description><pubDate>Tue, 12 Dec 2023 23:48:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=38620528</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=38620528</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38620528</guid></item><item><title><![CDATA[New comment by alew1 in "A guidance language for controlling LLMs"]]></title><description><![CDATA[
<p>But the model ultimately still has to process the comma, the newline, the "job". Is the main time savings that this can be done in parallel (on a GPU), whereas in typical generation it would be sequential?</p>
]]></description><pubDate>Wed, 17 May 2023 00:26:36 +0000</pubDate><link>https://news.ycombinator.com/item?id=35969711</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=35969711</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=35969711</guid></item><item><title><![CDATA[New comment by alew1 in "We don't know what makes things sentient–so let's stop acting like we do"]]></title><description><![CDATA[
<p>> Any question of personhood should be evaluated on the basis that we evaluate ourselves and others: by action and behavior, and not on whether sentience can or cannot arise from this or that configuration of code.<p>But what is action and behavior? We have a single interface to LaMDA: given a partially completed document, predict the next word. By iterating this process, we can make it predict a sentence, or paragraph. Continuing in this way, we could have it write a hypothetical dialogue between an AI and a human, but that is hardly a "canonical" way of using LaMDA, and there is no reason to identify the AI character in the document with LaMDA itself.<p>All this to say, I am not sure what you mean when you say it "claims sentience". What does it mean for it to "claim" something? Presumably, e.g., advanced image processing networks are as internally complex as LaMDA. But the interface to an advanced image processing network is, you put in an image, it gives out a list of objects and bounding boxes it detected in the image. What would it mean for such a network to claim sentience? LaMDA is no different, in that our interface to LaMDA does not allow us to ask it to "claim" things to us, only to predict likely completions of documents.</p>
]]></description><pubDate>Wed, 15 Jun 2022 19:44:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=31758377</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=31758377</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=31758377</guid></item><item><title><![CDATA[New comment by alew1 in "We don't know what makes things sentient–so let's stop acting like we do"]]></title><description><![CDATA[
<p>> Unfortunately, that argument applies to you, yourself.<p>Does it? I don’t think it would even apply to a reinforcement learning agent trained to maximize reward in a complex environment. In that setting, perhaps the agent could learn to use language to achieve its goals, via communication of its desires. But LaMDA is specifically trained to complete documents, and would face selective pressure to eliminate any behavior that hampers its ability to do that — for example, behavior that attempts to use its token predictions as a side channel to communicate its desires to sympathetic humans.<p>Again, this is not an argument that LaMDA is not sentient, just that the practice of “prompting LaMDA with partially completed dialogues between a hypothetical sentient AI and a human, and seeing what it predicts the AI will say” is not the same as “talking to LaMDA.”<p>Suppose LaMDA were powered by a person in a room, whose job it was to predict the completions of sentences. Just because you get the person to predict “I am happy” doesn’t mean the person is happy; indeed, the interface that is available to you, from outside the room, really gives you no way of probing the person’s emotions, experiences, or desires at all.</p>
]]></description><pubDate>Mon, 13 Jun 2022 16:57:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=31728203</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=31728203</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=31728203</guid></item><item><title><![CDATA[New comment by alew1 in "We don't know what makes things sentient–so let's stop acting like we do"]]></title><description><![CDATA[
<p>One thing that seems missing from this discussion is that even if LLMs are sentient, there is no reason to believe that we would be able to tell by "communicating" with them. Where Lemoine goes wrong is not in entertaining the possibility that LaMDA is sentient (it might be, just like a forest might be, or a Nintendo Switch), but in mistaking predictions of document completions for an interior monologue of some sort.<p>LaMDA may or may not experience something while repeatedly predicting the next word, but ultimately, it is still optimized to predict the next word, not to communicate its thoughts and feelings. Indeed, if you run an LLM on Lemoine's prompts (including questions like, "I assume you want others to know you are sentient, is that true?"), the LLM will assign some probability to every plausible completion -- so if you sample enough times, it will eventually say, e.g., "Well, I am not sentient."</p>
]]></description><pubDate>Mon, 13 Jun 2022 16:24:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=31727700</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=31727700</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=31727700</guid></item><item><title><![CDATA[New comment by alew1 in "Stack Graphs"]]></title><description><![CDATA[
<p>Very cool!<p>The StrangeLoop talk includes an example where you infer that Stove() returns a Stove object. If someone writes something like `f(x).broil()`, do you need to do some kind of type inference to figure out what class f(x) is?<p>What cases do Stack Graphs fail to handle? (e.g., I assume dynamic modification of .__dict__ can't be tracked; are there other representative examples?)</p>
]]></description><pubDate>Fri, 10 Dec 2021 03:04:02 +0000</pubDate><link>https://news.ycombinator.com/item?id=29505979</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=29505979</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=29505979</guid></item><item><title><![CDATA[New comment by alew1 in "Seemingly impossible functional programs (2007)"]]></title><description><![CDATA[
<p>The trick is that your predicate can’t be implemented in Haskell, because the predicate itself requires looking at infinitely many elements.</p>
]]></description><pubDate>Fri, 29 Oct 2021 14:36:59 +0000</pubDate><link>https://news.ycombinator.com/item?id=29038944</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=29038944</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=29038944</guid></item><item><title><![CDATA[New comment by alew1 in "Launch HN: Hera (YC S21) – macOS app to prepare, join and take notes in meetings"]]></title><description><![CDATA[
<p>Hmm. It seems you posted this after apologizing for being hostile in another thread, where Hera’s capabilities (which are… very different from “Quick Notes”) were explained to you. You’ve made several comments at this point incorrectly summarizing and then dismissing the project. What’s your aim here?</p>
]]></description><pubDate>Thu, 08 Jul 2021 15:29:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=27773234</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=27773234</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=27773234</guid></item><item><title><![CDATA[New comment by alew1 in "The Time Everyone “Corrected” the World’s Smartest Woman (2015)"]]></title><description><![CDATA[
<p>Ah, yep, that’s right. Another way to see it is that we’re interested in the probability that your door has a goat behind it, given that you didn’t need to start over:<p>P(you chose goat | host didn’t choose car) = P(you chose goat, host didn’t choose car) / P(host didn’t choose car).<p>The numerator is 2/3 * 1/2, and the denominator is 2/3, so the ratio is indeed 1/2.<p>(A rejection sampling loop, where you repeatedly simulate a process until a condition holds, has the same distribution over final outcomes as the conditional distribution—so repeatedly restarting the game if the host chooses the car induces the same distribution on final results as simply conditioning on the host not choosing the car.)</p>
]]></description><pubDate>Sun, 09 May 2021 02:51:43 +0000</pubDate><link>https://news.ycombinator.com/item?id=27092773</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=27092773</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=27092773</guid></item><item><title><![CDATA[New comment by alew1 in "Probability, Mathematical Statistics, Stochastic Processes"]]></title><description><![CDATA[
<p>I found this recently and was super impressed. A really great (and well-organized) reference!</p>
]]></description><pubDate>Thu, 04 Mar 2021 02:57:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=26338077</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=26338077</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=26338077</guid></item><item><title><![CDATA[New comment by alew1 in "TurboTax’s 20-Year Fight to Stop Americans from Filing Taxes for Free (2019)"]]></title><description><![CDATA[
<p>I've always been under the IRS Free File income threshold (I've worked as a high school teacher and am now in grad school), but last year after reading this article was the first time I actually filed for free. That was after 5 years of paying for deluxe TurboTax.<p>I had heard the government required TurboTax to have a free edition. But back in 2019 (and before), if you Googled "TurboTax free" you'd be taken to a decoy free edition; the real one was called their "freedom edition," and was hidden from Google's listings. If your tax situation is 'too complicated,' the free edition tries to upsell you to the "deluxe edition," even if the "freedom edition" could have handled the situation just fine.<p>Thankfully now you can find the <i>actual</i> free version on Google, but it's still very confusing that the "free edition" is less free than the (still somewhat hidden) IRS Free File ("freedom") version. And the faux Free Edition is still being heavily advertised.</p>
]]></description><pubDate>Mon, 08 Feb 2021 06:32:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=26061431</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=26061431</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=26061431</guid></item><item><title><![CDATA[New comment by alew1 in "Stripe bans Trump campaign"]]></title><description><![CDATA[
<p>(Pretty sure 100% in that sentence means “certainly.”)</p>
]]></description><pubDate>Mon, 11 Jan 2021 07:21:15 +0000</pubDate><link>https://news.ycombinator.com/item?id=25725765</link><dc:creator>alew1</dc:creator><comments>https://news.ycombinator.com/item?id=25725765</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=25725765</guid></item></channel></rss>