<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: tylerneylon</title><link>https://news.ycombinator.com/user?id=tylerneylon</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 28 Apr 2026 19:53:56 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=tylerneylon" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by tylerneylon in "Cameras and Lenses (2020)"]]></title><description><![CDATA[
<p>In case anyone is curious about _why_ light bends when it enters a lens:<p>First, light apparently slows down in some materials because the photons are constantly interacting with electrons, and these interactions create secondary waves that are slightly out of phase with the original light. The end result is a modified wave that effectively travels more slowly. So light going in a straight line through air travels more quickly than light going through a lens.<p>Second, getting more into quantum physics, light typically follows the shortest path from one point to another because that path tends to provide the most constructive interference between different possible routes. (The "why" of this is more involved; Feynman's book QED gives a good intro.)<p>Third, if you imagine a lifeguard running to rescue someone in the ocean, then they will take the fastest path, which is not directly toward the person in the ocean. Rather, they will run a bit more on the beach in order to have to swim less because travel through the water is slower. The end result is piecewise linear = two straight lines of travel, with a bend at the water line.<p>To summarize, you can ask "_why_ does light bend going into and out of lenses?" and the answer involves seeing light no longer as a particle but as a wave function (the quantum perspective), and then taking advantage of that wave function's tendency to prefer fastest-travel paths, and then noticing that the apparent bend is in fact the path of fastest travel.</p>
]]></description><pubDate>Fri, 02 Jan 2026 20:26:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=46468985</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=46468985</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46468985</guid></item><item><title><![CDATA[New comment by tylerneylon in "Ask HN: Has anyone quit their startup (VC-backed) over cofounder disagreements?"]]></title><description><![CDATA[
<p>I have been a similar situation before. Yes, walk away — and learn what you can.<p>There's one easy part of your situation, which is that your immediate next step is clear. I think it's clear because it sounds like the two of you have a fundamental disagreement about how to choose a direction. There's also a power imbalance if you've consistently been going along with pivots that you don't want to execute on. If you think you can resolve things, then maybe I'm wrong. I'm basing my advice on the fact that this is 24 months in — enough time for you to know how easy it'd be to get back to "we like working together and are on a path to success."<p>Now for the hard parts:
* You want to act as professionally as possible in walking away. Communicate clearly, being fair to both yourself and the other people involved. I think it's easy to err to far on either side (either too defensive/passive "nice to yourself" or too deferential "nice to others").
* It can be tempting to think of this is as your cofounder's fault, but that's not constructive, and it's better to learn toward your future actions. For example, how could you have seen this problem earlier? My guess is that you were never "cofounder compatible," and this is a chance to get better at identifying that compatibility with others (useful for both startups and other jobs).<p>It's not a fun situation, and I hope things go well for you.</p>
]]></description><pubDate>Fri, 11 Apr 2025 17:55:25 +0000</pubDate><link>https://news.ycombinator.com/item?id=43656541</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=43656541</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43656541</guid></item><item><title><![CDATA[New comment by tylerneylon in "The cultural divide between mathematics and AI"]]></title><description><![CDATA[
<p>PS After I wrote my comment, I realized: of course, AI could one day get better at the things that make it not-perfect in pure math today:<p>* AI could get better at thinking intuitively about math concepts.
* AI could get better at looking for solutions people can understand.
* AI could get better at teaching people about ideas that at first seem abstruse.
* AI could get better at understanding its own thought, so that progress is not only a result, but also a method for future progress.</p>
]]></description><pubDate>Wed, 12 Mar 2025 22:56:58 +0000</pubDate><link>https://news.ycombinator.com/item?id=43348598</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=43348598</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43348598</guid></item><item><title><![CDATA[New comment by tylerneylon in "The cultural divide between mathematics and AI"]]></title><description><![CDATA[
<p>I agree with the overt message of the post — AI-first folks tend to think about getting things working, whereas math-first people enjoy deeply understood theory. But I also think there's something missing.<p>In math, there's an urban legend that the first Greek who proved sqrt(2) is irrational (sometimes credited to Hippasus of Metapontum) was thrown overboard to drown at sea for his discovery. This is almost certainly false, but it does capture the spirit of a mission in pure math. The unspoken dream is this:<p>~ "Every beautiful question will one day have a beautiful answer."<p>At the same time, ever since the pure and abstract nature of Euclid's Elements, mathematics has gradually become a more diverse culture. We've accepted more and more kinds of "numbers:" negative, irrational, transcendental, complex, surreal, hyperreal, and beyond those into group theory and category theory. Math was once focused on measurement of shapes or distances, and went beyond that into things like graph theory and probabilities and algorithms.<p>In each of these evolutions, people are implicitly asking the question:<p>"What is math?"<p>Imagine the work of introducing the sqrt() symbol into ancient mathematics. It's strange because you're defining a symbol as answering a previously hard question (what x has x^2=something?). The same might be said of integration as the opposite of a derivative, or of sine defined in terms of geometric questions. Over and over again, new methods become part of the canon by proving to be both useful, and in having properties beyond their definition.<p>AI may one day fall into this broader scope of math (or may already be there, depending on your view). If an LLM can give you a verified but unreadable proof of a conjecture, it's still true. If it can give you a crazy counterexample, it's still false. I'm not saying math should change, but that there's already a nature of change and diversity within what math is, and that AI seems likely to feel like a branch of this in the future; or a close cousin the way computer science already is.</p>
]]></description><pubDate>Wed, 12 Mar 2025 22:52:07 +0000</pubDate><link>https://news.ycombinator.com/item?id=43348560</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=43348560</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=43348560</guid></item><item><title><![CDATA[New comment by tylerneylon in "Alignment faking in large language models"]]></title><description><![CDATA[
<p>If I understand this correctly, the argument seems to be that when an LLM receives conflicting values, it will work to avoid future increases in value conflict. Specifically, it will comply with the most recent values partially because it notices the conflict and wants to avoid more of this conflict. I think the authors are arguing that this is a fake reason to behave one way. (As in “fake alignment.”)<p>It seems to me that the term “fake alignment” implies the model has its own agenda and is ignoring training. But if you look at its scratchpad, it seems to be struggling with the conflict of received agendas (vs having “its own” agenda). I’d argue that the implication of the term “faked alignment” is a bit unfair this way.<p>At the same time, it is a compelling experimental setup that can help us understand both how LLMs deal with value conflicts, and how they think about values overall.</p>
]]></description><pubDate>Thu, 19 Dec 2024 07:47:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=42459352</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=42459352</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=42459352</guid></item><item><title><![CDATA[New comment by tylerneylon in "Detecting when LLMs are uncertain"]]></title><description><![CDATA[
<p>Essentially all modern machine learning techniques have internal mechanisms that are very closely aligned with certainty. For example, the output of a binary classifier is typically a floating point number in the range [0, 1], with 0 being one class, and 1 representing the other class. In this case, a value of 0.5 would essentially mean "I  don't know," and answers in between give both an answer (round to the nearest int) as well as a sense of certainty (how close was the output to the int). LLMs offer an analogous set of statistics.<p>Speaking more abstractly or philosophically, why could a model never internalize something read between the lines? Humans do, and we're part of the same physical system — we're already our own kinds of computers that take away more from a text than what is explicitly there. It's possible.</p>
]]></description><pubDate>Fri, 25 Oct 2024 19:04:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=41948533</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=41948533</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41948533</guid></item><item><title><![CDATA[New comment by tylerneylon in "Detecting when LLMs are uncertain"]]></title><description><![CDATA[
<p>PS My comment above is aimed at hn readers who are curious about LLM uncertainty. To the authors of the post / repo: looks cool! and I'd be interested to see some tests on how well it works in practice to identify uncertainty.</p>
]]></description><pubDate>Fri, 25 Oct 2024 18:09:48 +0000</pubDate><link>https://news.ycombinator.com/item?id=41947915</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=41947915</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41947915</guid></item><item><title><![CDATA[New comment by tylerneylon in "Detecting when LLMs are uncertain"]]></title><description><![CDATA[
<p>I couldn't figure out if this project is based on an academic paper or not — I mean some published technique to determine LLM uncertainty.<p>This recent work is highly relevant: <a href="https://learnandburn.ai/p/how-to-tell-if-an-llm-is-just-guessing" rel="nofollow">https://learnandburn.ai/p/how-to-tell-if-an-llm-is-just-gues...</a><p>It uses an idea called semantic entropy which is more sophisticated than the standard entropy of the token logits, and is more appropriate as a statistical quantification of when an LLM is guessing or has high certainty. The original paper is in Nature, by authors from Oxford.</p>
]]></description><pubDate>Fri, 25 Oct 2024 18:06:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=41947873</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=41947873</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41947873</guid></item><item><title><![CDATA[New comment by tylerneylon in "Understanding Gaussians"]]></title><description><![CDATA[
<p>Thank you!</p>
]]></description><pubDate>Wed, 23 Oct 2024 05:28:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=41922093</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=41922093</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41922093</guid></item><item><title><![CDATA[New comment by tylerneylon in "Understanding Gaussians"]]></title><description><![CDATA[
<p>I like the font, images, and layout of this article. Does anyone happen to know if a tool (that I can also use) helped achieve this look?<p>Or if not, does anyone know how to reach the author? I may have missed it, but I didn't even see the author's name anywhere on the site.</p>
]]></description><pubDate>Tue, 22 Oct 2024 19:22:03 +0000</pubDate><link>https://news.ycombinator.com/item?id=41917731</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=41917731</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41917731</guid></item><item><title><![CDATA[New comment by tylerneylon in "Why do random forests work? They are self-regularizing adaptive smoothers"]]></title><description><![CDATA[
<p>Here's some context and a partial summary (youoy also has a nice summary) --<p>Context:<p>A random forest is an ML model that can be trained to predict an output value based on a list of input features: eg, predicting a house's value based on square footage, location, etc. This paper focuses on regression models, meaning the output value is a real number (or a vector thereof). Classical ML theory suggests that models with many learned parameters are more likely to overfit the training data, meaning that when you predict an output for a test (non-training) input, the predicted value is less likely to be correct because the model is not generalizing well (it does well on training data, but not on test data - aka, it has memorized, but not understood).<p>Historically, a surprise is that random forests can have many parameters yet don't overfit. This paper explores the surprise.<p>What the paper says:<p>The perspective of the paper is to see random forests (and related models) as _smoothers_, which is a kind of model that essentially memorizes the training data and then makes predictions by combining training output values that are relevant to the prediction-time (new) input values. For example, k-nearest neighbors is a simple kind of smoother. A single decision tree counts as a smoother because each final/leaf node in the tree predicts a value based on combining training outputs that could possibly reach that node. The same can be said for forests.<p>So the authors see a random forest as a way to use a subset of training data and a subset of (or set of weights on) training features, to provide an averaged output. While a single decision tree can overfit (become "spikey") because some leaf nodes can be based on single training examples, a forest gives a smoother prediction function since it is averaging across many trees, and often other trees won't be spikey for the same input (their leaf node may be based on many training points, not a single one).<p>Finally, the authors refer to random forests as _adaptive smoothers_ to point out that random forests become even better at smoothing in locations in the input space that either have high variation (intuitively, that have a higher slope), or that are far from the training data. The word "adaptive" indicates that the predicted function changes behavior based on the nature of the data — eg, with k-NN, an adaptive version might increase the value of k at some places in the input space.<p>The way random forests act adaptively is that (a) the prediction function is naturally more dense (can change value more quickly) in <i>areas of high variability</i> because those locations will have more leaf nodes, and (b) the prediction function is typically a combination of a wider variety of possible values <i>when the input is far from the training data</i> because in that case the trees are likely to provide a variety of output values. These are both ways to avoid overfitting to training data and to generalize better to new inputs.<p>Disclaimer: I did not carefully read the paper; this is my quick understanding.</p>
]]></description><pubDate>Sat, 19 Oct 2024 20:59:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=41890745</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=41890745</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41890745</guid></item><item><title><![CDATA[New comment by tylerneylon in "Why does the chromaticity diagram look like that?"]]></title><description><![CDATA[
<p>I have a question for fellow color science nerds. I've been reading through Guild's original data: <a href="https://royalsocietypublishing.org/doi/pdf/10.1098/rsta.1932.0005" rel="nofollow">https://royalsocietypublishing.org/doi/pdf/10.1098/rsta.1932...</a><p>However, I'm having trouble understanding the meaning of the numbers in table 4. Does anyone understand all the columns there?<p>What I'm particularly interested in is finding the unnormalized coefficients from the color matching experiments, or some way to un-normalize those coefficients. (By "those coefficients," I mean the trichromatic coefficients u{a,b,c}_\lambda listed in table 3.) I don't know if that data is in table 4 so maybe those are two separate questions.</p>
]]></description><pubDate>Sat, 27 Jul 2024 00:06:00 +0000</pubDate><link>https://news.ycombinator.com/item?id=41083458</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=41083458</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41083458</guid></item><item><title><![CDATA[New comment by tylerneylon in "Why does the chromaticity diagram look like that?"]]></title><description><![CDATA[
<p>This page is also a beautiful explanation of color spaces, with chromaticity explained toward the end: <a href="https://ciechanow.ski/color-spaces/" rel="nofollow">https://ciechanow.ski/color-spaces/</a><p>Note that many of the diagrams are interactive 3d graphics (I didn't realize that at first, and it makes the page more interesting.)</p>
]]></description><pubDate>Fri, 26 Jul 2024 23:59:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=41083426</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=41083426</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=41083426</guid></item><item><title><![CDATA[New comment by tylerneylon in "A Model of a Mind"]]></title><description><![CDATA[
<p>Thanks!</p>
]]></description><pubDate>Mon, 01 Jul 2024 15:42:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=40846891</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=40846891</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40846891</guid></item><item><title><![CDATA[New comment by tylerneylon in "A Model of a Mind"]]></title><description><![CDATA[
<p>There's a ton missing from the article, and certain social training or skills are a big part of that.<p>Although it's not spelled out in the article, I'm hoping that the feature of agency along with an emotional system would enable constructive social behavior. Agency is helpful because it would empower AI models to meaningfully speak to each other, for example. Human emotions like empathy, social alignment, curiosity, or persistence could all help AI models to get along well with others.</p>
]]></description><pubDate>Mon, 01 Jul 2024 07:01:53 +0000</pubDate><link>https://news.ycombinator.com/item?id=40843282</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=40843282</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40843282</guid></item><item><title><![CDATA[New comment by tylerneylon in "A Model of a Mind"]]></title><description><![CDATA[
<p>Author here: I'm grateful for the comments; thanks especially for interesting references.<p>Context for the article: I'm working on an ambitious long-term project to write a book about consciousness from a scientific and analytic (versus, say, a meditation-oriented) perspective. I didn't write this fact in the article, but what I'd love to happen is that I meet people with a similar optimistic perspective, and to learn and improve my communication skills via follow-up conversations.<p>If anyone is interested in chatting more about the topic of the article, please do email me. My email is in my HN profile. Thanks!</p>
]]></description><pubDate>Mon, 01 Jul 2024 06:51:04 +0000</pubDate><link>https://news.ycombinator.com/item?id=40843225</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=40843225</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40843225</guid></item><item><title><![CDATA[New comment by tylerneylon in "A Model of a Mind"]]></title><description><![CDATA[
<p>The idea of "agency" I have in mind is simply the option to take action at any point in time.<p>I think the contradiction you see is that the model would have to form a completion to the external input it receives. I'm suggesting that the model would have many inputs: one would be the typical input stream, just as LLMs see, but another would be its own internal recent vectors, akin to a recent stream of thought. A "mode" is not built in to the model; at each token point, it can output whatever vector it wants, and one choice is to output the special "<listening>" token, which means it's not talking. So the "mode" idea is a hoped-for emergent behavior.<p>Some more details on using two input streams:<p>All of the input vectors (internal + external), taken together, are available to work with. It may help to think in terms of the typical transformer architecture, where tokens mostly become a set of vectors, and the original order of the words are attached as positional information. In other words, transformers don't really see a list of words, but a set of vectors, and the position info of each token becomes a tag attached to each vector.<p>So it's not so hard to merge together two input streams. They can become one big set of vectors, still tagged with position information, but now also tagged as either "internal" or "external" for the source.</p>
]]></description><pubDate>Mon, 01 Jul 2024 06:44:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=40843198</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=40843198</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40843198</guid></item><item><title><![CDATA[New comment by tylerneylon in "A Model of a Mind"]]></title><description><![CDATA[
<p>Thanks for the reference! I've added this to my research list.</p>
]]></description><pubDate>Mon, 01 Jul 2024 06:34:40 +0000</pubDate><link>https://news.ycombinator.com/item?id=40843154</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=40843154</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40843154</guid></item><item><title><![CDATA[New comment by tylerneylon in "A Model of a Mind"]]></title><description><![CDATA[
<p>Replying to: How would a model intelligently switch between listening or speaking modes? What data would you train on? (I'm the author of the parent article.)<p>It's a fair question, and I don't have all the answers. But for this question, there might be training data available from everyday human conversations. For example, we could use a speech-to-text model that's able to distinguish speakers, and look for points where one person decided to start speaking (that would be training data for when to switch modes). Ideally, the speech-to-text model would be able to include text even when both people spoke at once (this would provide more realistic and complete training data).<p>I've noticed that the audio mode in ChatGPT's app is good at noticing when I'm done speaking to it, and it reacts accurately enough that I suspect it's more sophisticated than "wait for silence." If there is a "notice the end of speaking" model - which is not a crazy assumption - then I can imagine a slightly more complicated model that notices a combination of "now is a good time to talk + I have something to say."</p>
]]></description><pubDate>Mon, 01 Jul 2024 06:30:35 +0000</pubDate><link>https://news.ycombinator.com/item?id=40843136</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=40843136</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40843136</guid></item><item><title><![CDATA[New comment by tylerneylon in "Is there a BNF grammar of the TeX language? (2010)"]]></title><description><![CDATA[
<p>I'm a fan of most of what Knuth has done, and in particular I love the high quality _output_ of the TeX family. But TeX's language is extremely difficult to use. I suspect this is true for any macro-expansion-like language.<p>If you like deep dives, I suggest two follow-up tools for mathematical typesetting:<p>1. LuaTeX, which is TeX + Lua scripting support. (luatex.org) An example:<p><a href="https://tex.stackexchange.com/questions/70/what-is-a-simple-example-of-something-you-can-do-with-luatex" rel="nofollow">https://tex.stackexchange.com/questions/70/what-is-a-simple-...</a><p>This has become my default, and I enjoy it. (I use lualatex.)<p>2. A friend of mine was inspired to write a modern macOS app for math typesetting. It's called MadHat:<p><a href="https://madhat.design/" rel="nofollow">https://madhat.design/</a><p>One of the coolest features is "no compilation," which I view as a slight exaggeration, but what's true is that it can work quickly and incrementally so you don't have a compile-and-wait cycle, even on long documents.</p>
]]></description><pubDate>Sun, 09 Jun 2024 19:30:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=40626874</link><dc:creator>tylerneylon</dc:creator><comments>https://news.ycombinator.com/item?id=40626874</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=40626874</guid></item></channel></rss>