<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: gajomi</title><link>https://news.ycombinator.com/user?id=gajomi</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Tue, 14 Apr 2026 22:13:32 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=gajomi" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[New comment by gajomi in "Ask HN: Who is hiring? (December 2023)"]]></title><description><![CDATA[
<p>I didn't see any of the specific job titles you mentioned on the career site. Perhaps this is part of the "antedisciplinary" method you refer to? Practically speaking though, should one just use the "Open Apply" option for Machine Learning Engineer/Scientist roles?</p>
]]></description><pubDate>Mon, 04 Dec 2023 18:16:01 +0000</pubDate><link>https://news.ycombinator.com/item?id=38520812</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=38520812</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=38520812</guid></item><item><title><![CDATA[New comment by gajomi in "The worst programmer I know"]]></title><description><![CDATA[
<p>PSA... check out this guy's LinkedIn profile. It's a great read.</p>
]]></description><pubDate>Sun, 03 Sep 2023 05:27:39 +0000</pubDate><link>https://news.ycombinator.com/item?id=37367946</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=37367946</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=37367946</guid></item><item><title><![CDATA[New comment by gajomi in "The inflated promise of science education"]]></title><description><![CDATA[
<p>I think that is why he asked the rhetorical "which masks?".</p>
]]></description><pubDate>Tue, 13 Sep 2022 03:19:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=32820313</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=32820313</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=32820313</guid></item><item><title><![CDATA[New comment by gajomi in "Learning how dictionaries work"]]></title><description><![CDATA[
<p>There are a couple starting points you could take. I spent a weekend hacking out a program that generates fake word/definition pairs with a transformer model set against a dictionary: <a href="https://youtu.be/XnJ2TKAn-Vk?t=1547" rel="nofollow">https://youtu.be/XnJ2TKAn-Vk?t=1547</a>. If you substitute fake words for real words and have a sufficiently accurate model you could quickly generate reasonable and novel definitions.<p>There are more complete versions of this kind of thing publicly available: <a href="https://github.com/turtlesoupy/this-word-does-not-exist" rel="nofollow">https://github.com/turtlesoupy/this-word-does-not-exist</a><p>> This would be amazing, for example, to run on a large corpus, generate the dictionary, and then run it again to find words that are used but not defined - not just in the original corpus but in the definitions too.<p>I think this would be how you would gauge success of the model. That is to say, you would evaluate model accuracy on a set of held-out words with definitions that never appeared in your dictionary training set but appeared in context in your corpus. You would have to manually annotate whether or not the generated definition of these held out words was acceptable.</p>
]]></description><pubDate>Thu, 11 Nov 2021 16:43:18 +0000</pubDate><link>https://news.ycombinator.com/item?id=29190004</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=29190004</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=29190004</guid></item><item><title><![CDATA[New comment by gajomi in "The Nobel Prize in Physics 2021"]]></title><description><![CDATA[
<p>Let me take a stab at this (I'll maybe take it halfway there). First of all we want to know what kind of matrix we are talking about.<p>Imagine that you have a whole bunch generative models (its best if you imagine a fully connected Boltzmann machine in particular, whose states you can think of as a binary vector consisting only of zeros and ones) that have the same form but different random realizations of their parameters. This is a typical example of what a toy model of a so-called "spin glass" looks like in statistical physics (the spins are either up down down, usually represented as +1/-1). Each of these models, having been initialized randomly will have their up particular frequency of a particular location (also called site) of the boolean vector being either a one or a zero.<p>If the tendency of a site to be either or one or a zero was independent of every other site the analysis of such a model would be pretty straightforward: every model would just have a vector of N frequencies and we could compare how close the statistical behavior of each model was to the other by comparing how closely the N frequencies at each site matched one another. But in the general case there will be some interaction or correlation between sites in a given model. If the interaction strength is strong enough this can result in a model tending to generate groups of patterns in its sequence of zeros and ones that are close to one another. Furthermore if we compare the overlap of the apparent patterns between two such models, each with their own random parameters, we will find that some of them overlap more than others.<p>What we can then do is to ask the question of how much, on average do the patterns of these random models overlap with on another in the full set of all models. This leads us to the concept of an "overlap matrix". This matrix will have one set of values along the diagonal (corresponding to how much a models patterns tend to overlap with themselves) and off diagonal values capturing the overlap between. You can find through simulation or with some carefully constructed calculations that when the interaction strength between sites is small that the off diagonal elements don't tend to zero, but rather a single number different from the diagonal value. This is perhaps intuitive: these models were randomly initialized but they are going to overlap in their behavior in some places.<p>Where things get interesting though is when you increase the interaction strength you find that the overlap matrix starts to take on a block diagonal form, wherein clusters of models overlap with one another at a certain level and at a lower but constant level with out-of-cluster models. This is called  one replica symmetry breaking (1RSB). These different clusters of models can be thought of as having learned different overall patterns with the similarity quantified by their overalp. If you keep increasing the interaction strength you will find that this happens again and again, with a k-fold replica symmetry braking (kRSB) with a sort of self similar block structure emerging in the overlap matrix (picture is worth a thousand words [1]).<p>Now the real wild part that Parisi figured out is what happens when you take this process to the regime of full replica symmetry breaking. You can't really do this with simulations and the calculations are very tricky (you have a bunch of terms either going to infinity or zero that need to balance out correctly) but Parisi ending up coming up with an expression for the distribution of overlaps for the infinitely sized matrix with full interaction strength in play. The expression is actually a partial differential equation that itself needs to be solved (I told you the calculations were tricky right), but amazingly, it seems to capture the behavior of these kinds of models correctly.<p>Whereas mathematicians have a pretty good idea of how to understand the 1RSB process rigorously, the Parisi full replica symmetry breaking scheme is very much not understood and remains of interest both to complex systems researches trying to understand their models and applied mathematicians (probability people in particular) trying to lay the foundations needed to explore the ideas being explored by theorists.<p>Hope that helps a bit!<p>[1] <a href="https://www.semanticscholar.org/paper/Spin-Glasses%2C-Boolean-Satisfiability%2C-and-Survey-Landau/f2983762266175a1775113f0864718c8200f05b1/figure/2" rel="nofollow">https://www.semanticscholar.org/paper/Spin-Glasses%2C-Boolea...</a></p>
]]></description><pubDate>Tue, 05 Oct 2021 16:32:10 +0000</pubDate><link>https://news.ycombinator.com/item?id=28761709</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=28761709</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=28761709</guid></item><item><title><![CDATA[New comment by gajomi in "The Nobel Prize in Physics 2021"]]></title><description><![CDATA[
<p>+1 on the whole references as a nice introduction. I think the authors overstate the preparation of their hypothetical "pedestrian" (either that or they need to get away from the physics department a bit more often), but a great reference nevertheless. I also got a lot out of sections of Nishimori's textbook [1]. In particular it helps motivate problems outside of physics and provides some references to start digging into more rigorous approaches via cavity methods (which I think, incidentally, are also more intuitive). I am a novice in this area but am sort of crossing my fingers that some of the ideas in this area will make their way into algorithms for inferring latent variables in some of the upcoming modern associative neural networks. What I mean here is that it would be cool not just to have an understanding of total capacity of the network but also correct variational approximations to use during training.<p>[1] <a href="https://neurophys.biomedicale.parisdescartes.fr/wp-content/uploads/2014/10/Nishimori-2001-Statistical-Physics-of-Spin-Glasses-and-Information-Processing-An-Introduction.pdf" rel="nofollow">https://neurophys.biomedicale.parisdescartes.fr/wp-content/u...</a>
[2] <a href="https://ml-jku.github.io/hopfield-layers/" rel="nofollow">https://ml-jku.github.io/hopfield-layers/</a></p>
]]></description><pubDate>Tue, 05 Oct 2021 15:22:52 +0000</pubDate><link>https://news.ycombinator.com/item?id=28760689</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=28760689</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=28760689</guid></item><item><title><![CDATA[New comment by gajomi in "Software Dev Can't Be Automated – It’s a Creative Process with an Unknown Goal"]]></title><description><![CDATA[
<p>I have been using the phrase "Artificial Stupidity" as well, but with the opposite meaning. Specifically I like to think of human-like artificial stupidity as a challenge for machine intelligence, in which an algorithm is able to replicate the rather sophisticated and incredibly entangled logic, intuitions and calculus of humans at the height of their stupidity. This seems to me a much greater challenge than the standard sort of supervised learning problems in that a truly stupid AI must be able to imagine latent variables that allow it to explain away real world observations in a way that is both statistically implausible but casually serendipitous to their stupid peers. This seems to me to be a requirement for any kind of useful AGI.</p>
]]></description><pubDate>Sun, 15 Aug 2021 06:43:08 +0000</pubDate><link>https://news.ycombinator.com/item?id=28186700</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=28186700</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=28186700</guid></item><item><title><![CDATA[New comment by gajomi in "Fun and dystopia with AI-based code generation using GPT-J-6B"]]></title><description><![CDATA[
<p>This would be the opposite of a Turing test though, since most people wouldn't be able to do this.</p>
]]></description><pubDate>Fri, 25 Jun 2021 00:43:41 +0000</pubDate><link>https://news.ycombinator.com/item?id=27625812</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=27625812</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=27625812</guid></item><item><title><![CDATA[Space station detectors found the source of weird ‘blue jet’]]></title><description><![CDATA[
<p>Article URL: <a href="https://www.sciencenews.org/article/space-station-detectors-found-source-weird-blue-jet-lightning/">https://www.sciencenews.org/article/space-station-detectors-found-source-weird-blue-jet-lightning/</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=25873006">https://news.ycombinator.com/item?id=25873006</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Fri, 22 Jan 2021 16:24:10 +0000</pubDate><link>https://www.sciencenews.org/article/space-station-detectors-found-source-weird-blue-jet-lightning/</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=25873006</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=25873006</guid></item><item><title><![CDATA[New comment by gajomi in "TreeCard: The wooden debit card that plants trees"]]></title><description><![CDATA[
<p>I can imagine a variant of this card that tries to tackle the point you bring up. Not all purchases are equal. Some of them have negative carbon footprints. The tree planting offsets the footprint for every purchase. These things are not easy to estimate exactly but since there are probably order of magnitude differences in carbon footprints for different products there should be useful information about the relative impact of a purchase to display to the consumer and to demonstrate the overall impact of the project.</p>
]]></description><pubDate>Sun, 03 Jan 2021 17:19:13 +0000</pubDate><link>https://news.ycombinator.com/item?id=25623129</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=25623129</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=25623129</guid></item><item><title><![CDATA[New comment by gajomi in "Reinforcement learning is supervised learning on optimized data"]]></title><description><![CDATA[
<p>It seems to me that they are basically describing a variational formulation of the "optimization perspective" of reinforcement learning, which is cool, but I am confused... where is the supervised learning? Like what is the input and what is the output?</p>
]]></description><pubDate>Wed, 14 Oct 2020 04:43:54 +0000</pubDate><link>https://news.ycombinator.com/item?id=24773150</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=24773150</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=24773150</guid></item><item><title><![CDATA[New comment by gajomi in "Julia 1.5 Highlights"]]></title><description><![CDATA[
<p>FYI this is the _first_ call which is dominated by the JIT process. So not really "this time to plot" so much as the "time to compile". Calls to the resulting jitted function will be on the order of what you quote (I have not timed myself for the specific result they quote here but have done for similar in the past).</p>
]]></description><pubDate>Mon, 03 Aug 2020 17:37:05 +0000</pubDate><link>https://news.ycombinator.com/item?id=24040809</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=24040809</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=24040809</guid></item><item><title><![CDATA[New comment by gajomi in "Teaching physics to neural networks removes 'chaos blindness'"]]></title><description><![CDATA[
<p>> it sounds like they've increased the accuracy of a neural network model of a system, notably for edge cases, by training it on complete a complete model of said system.<p>Not quite. It's really just that they require the dynamics to be Hamiltonian, which would be highly atypical of the kind of dynamics an otherwise unconstrained neural network would learn. This is reflected in their loss function, the first of which learn an arbitrary second order differential equation, the second of which enforces Hamiltonian dynamics.<p>I don't understand how this was considered novel enough to warrant at PRE paper.<p>Here is a link to the paper:<p><a href="https://journals.aps.org/pre/pdf/10.1103/PhysRevE.101.062207" rel="nofollow">https://journals.aps.org/pre/pdf/10.1103/PhysRevE.101.062207</a></p>
]]></description><pubDate>Tue, 23 Jun 2020 22:15:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=23621042</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=23621042</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=23621042</guid></item><item><title><![CDATA[New comment by gajomi in "Malaria 'Completely Stopped' by Microbe"]]></title><description><![CDATA[
<p>> we shouldn’t just be ranking people by age, if we want to come to moral outcomes?<p>Agreed. But I think quantifications can still be useful.<p>Person-years probably aren't the right sort of quantification though, in that they blend together too many kinds of human experience. But I imagine most people would agree that statements like "X% of grandparents, lost 10 years earlier than expected" or "Y% of children under 10" allow for comparing the relative impact on families.<p>Quantifying losses doesn't mean that we consider the quantified as fungible.</p>
]]></description><pubDate>Mon, 04 May 2020 17:29:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=23070852</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=23070852</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=23070852</guid></item><item><title><![CDATA[On the first sequence without triple in arithmetic progression]]></title><description><![CDATA[
<p>Article URL: <a href="https://mathoverflow.net/questions/338415/on-the-first-sequence-without-triple-in-arithmetic-progression">https://mathoverflow.net/questions/338415/on-the-first-sequence-without-triple-in-arithmetic-progression</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=22542596">https://news.ycombinator.com/item?id=22542596</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Wed, 11 Mar 2020 03:10:13 +0000</pubDate><link>https://mathoverflow.net/questions/338415/on-the-first-sequence-without-triple-in-arithmetic-progression</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=22542596</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=22542596</guid></item><item><title><![CDATA[New comment by gajomi in "StyleGAN2"]]></title><description><![CDATA[
<p>How is it speculative? The background does not always allow for the discrimination to be made but in the random sample of ~20 faces I looked it was the main factor in maybe 1/4 of the cases (I was %100 accurate on this random sample). Of course, my random sample is not your random sample. We could probably do a controlled experiment to get at these kinds of attributions systematically.<p>I might also say that a second major discriminating factor is skin texture, especially at boundaries.</p>
]]></description><pubDate>Fri, 13 Dec 2019 17:07:31 +0000</pubDate><link>https://news.ycombinator.com/item?id=21783318</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=21783318</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=21783318</guid></item><item><title><![CDATA[New comment by gajomi in "Eigenvectors from eigenvalues"]]></title><description><![CDATA[
<p>Came here to pretty much write these same notes. The fact that you need to know the spectrum of all the minors really limits the usefulness for solving general eigen* problems. But I can imagine that this could make a great starting point for all kinds of approximate calculations/asymptotic analyses.<p>Regarding the first caveat that you bring up though, whereas the problem statement says you need a Hermitian matrix I think the results should generalize to non-hermitian matrices. In particular take a look at the third proof of the theorem at the end of the article. The only assumption required here is that the eigenvalues are simple (which does not preclude them being complex/coming in complex pairs).<p>Protip: I had to read the second step in the proof a few times before I could see what was going on. Explicitly what you do here is to (i) multiple from the right by the elementary unit vector e_j (ii) set up the matrix vector inverse using Cramer's rule (iii) notice that this matrix can be permuted to have a block diagonal element equal to the minor M_i with the other block diagonal entry equal to 1 and the corresponding column equal filled with zeros (iv) Write the determinant in the block form, which simplifies to the expression involving M_i by itself then finally (v) multiply by e_j from the left to get scalar equation.</p>
]]></description><pubDate>Fri, 15 Nov 2019 18:30:34 +0000</pubDate><link>https://news.ycombinator.com/item?id=21547374</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=21547374</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=21547374</guid></item><item><title><![CDATA[New comment by gajomi in "Stegasuras: Neural Linguistic Steganography"]]></title><description><![CDATA[
<p>FYI you need to type something into the "Secrete message" and ask to "Encrypt" using the prepopulated "LM context".<p>Both the context and secrete message change the encrypted message</p>
]]></description><pubDate>Thu, 05 Sep 2019 20:03:17 +0000</pubDate><link>https://news.ycombinator.com/item?id=20890202</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=20890202</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=20890202</guid></item><item><title><![CDATA[New comment by gajomi in "Ask HN: Best resources to gain math intuition?"]]></title><description><![CDATA[
<p>It's probably too hard to answer the general question about "how to gain math intuition". Mathematics is just too vast.<p>However, you have written<p>> I think when I took physics it really brought out these flaws and lack of intuition<p>which suggests you have some good practical experience with physics problem solving which has precipitated a certain feeling that you need to learn more about some kind of math. I would advise that you try to exploit this. In the same breath I want to recognize (as someone who did their BS and MS in physics) that physicists are not always so careful or explicit in how they are doing their mathematics. So learning means eventually going beyond physics sources and into a much wider world of mathematical thought. The particular things that mathematicians care about may or may not be relevant to the problem you are trying to solve in physics, and a good part of developing that intuition is to figure out which particular caveats that a mathematician expounds upon (more often than not, some esoterica about the space(s) that they are working in or the class of isomorphisms under which their results are invariant) matter physically. As you develop and intuition about these things a bonus is that you will be able to skim through mathematics resources much faster.</p>
]]></description><pubDate>Tue, 27 Aug 2019 00:27:06 +0000</pubDate><link>https://news.ycombinator.com/item?id=20805095</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=20805095</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=20805095</guid></item><item><title><![CDATA[New comment by gajomi in "Modern SAT solvers: fast, neat and underused"]]></title><description><![CDATA[
<p>Stupid question... Are there equivalent (in the sense of being fast and having nice interfaces) solvers for #SAT problems? A whole pile of inference in machine learning can be framed in this way which makes me curious.</p>
]]></description><pubDate>Sun, 19 May 2019 16:25:33 +0000</pubDate><link>https://news.ycombinator.com/item?id=19954064</link><dc:creator>gajomi</dc:creator><comments>https://news.ycombinator.com/item?id=19954064</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=19954064</guid></item></channel></rss>