Michael Nielsen – How science actually progresses
Опис відео▼
Today I'm speaking with Michael Nielsen. You have done many things. You're one of the pioneers of quantum computing, wrote the main textbook in the field of the open science movement. You wrote a book about deep learning that Chris Ola and uh Greg Brockman credit them with getting them into the field. Um more recently, you're a research fellow at the Astera Institute and writing a book about religion, science, and technology. I'm going to ask you about none of those things. The conversation I want to have today is how do we recognize scientific progress? And it's it's es especially relevant uh for AI because people are trying to close the RL verification loop on scientific discovery. And what does it mean to close that loop? But in preparing for this interview, >> I've realized that it's a more mysterious and elusive um force even in the history of human science than I understood. And I think a good place to start will be Michaels and morally and how special relativity is discovered if it's different than the story that you kind of get off of YouTube videos. Um anyways I I'll prompt you that way and then we'll go in there. >> Okay. Yeah. So I mean Michael Simoli is uh you one of the sort of the famous results often presented as as this this experiment that was done in the 1880s and that helped Einstein you come up with the the special theory of relativity a little bit later. So so sort of changing our the way we think about space and and time and our fundamental conception of those things. Um, and there's kind of a uh a big gap, I think, between the way Michaelelsson and Moley and other people at the time thought about the experiment and certainly the way in which uh Einstein thought or did not think about the experiment. Um, in actual fact, he uh uh stated later in his life he wasn't even sure whether he was aware of the paper at the time. Um, there's a lot of evidence that he he probably was aware of the paper at the time, but it actually wasn't dispositive for his thinking at all. uh something else uh completely was was going on. Um so uh uh what Michaelelsson and Molly thought they were doing was they thought they were testing different theories of what was called the ether. So if you go back to the the 1600s uh Robert Bole introduced the idea of the ether and basically the idea of the ether is um you we know that that sound is vibrations in the air. Um and then Bole and other people got interested in the question of like is is light vibrations uh in something and they couldn't figure out uh what it was. Bole actually did an experiment where he he tested whether or not you could propagate light through a vacuum. He found that you could you couldn't do it with with with sound. So he introduced this idea of the ether and then for the next 200 or so years people had all these kind of conversations about about what the ether was and what its nature was. And the Michaelelsson and Molly experiment was really an experiment to test different theories of the ether uh against one another um and in particular to find out whether or not there was a so-called ether wind. So the idea was that the the earth is passing through uh maybe this ether wind and if it is passing through the ether wind sort of this background um and you you shoot a light beam sort of parallel uh to the direction the ether wind is going in it'll get accelerated a little bit um and if it's being passed back uh sort of in the opposite direction it'll get slowed down a little bit and you should be able to to see this in the results of interference experiments and what they found much to their surprise um I uh was that in fact there was no ether wind. Um and that ruled out some theories of the ether but but but not all and and Michaelelsson certainly continued to to believe in the ether. >> Okay. So this is what was a shocking part of um reading this story from the biography of Einstein that you recommended by um >> what was his first name? >> Abraham Abraham P subtle as Lord and then also from Lacatoss uh the methodologies of scientific research programs. The way it's told is that Mangles Morley proved that the ether did not exist. Yeah. >> Therefore, it created a crisis in physics that Einstein saw with special relativity. >> And what you're pointing out is actually was trying to distinguish between many different theories of ether. You know, if you're in space or if you're on Earth, it's the same direction of ether or maybe the ether wind is being carried around by the earth and so you can't really experience it on earth, but if you go to high enough altitude, you might be able to experience it. Um in fact the Michaelelsson's experiments were the famous one is 1887 but he conducted these experiments for basically two decades. I mean for longer than that he he conducted them. I think the first one was in 1881 but he continued to believe until I mean he died he died I think it was like 1929 or so. It was like the late 20s. Um and he was still doing experiments in the 1920s. Um uh sort of about whether or not you know the ether existed. And so he so he continued to believe in the ether to the end of his his life or I think the last public statement he made is like a year or two before he died. And he still still believed basically believed at that point. And in fact, there was another physicist um Miller who kept doing these experiments in the 1920s. He thought that he went to a high enough altitude uh is in Mount Wilson in California where I'm high enough that I can actually the ether winds are not being dragged with by the earth. I and I've measured um the effect of the ether. And Einstein hears about this and he says this is where you get the famous quote subtle is the Lord but malicious he is not. Anyways, I think the the reason the story is interesting, it's for a many of different reasons, but one is one of the different ways in which the real history of science is different from this idea you get of the scientific method >> is you really can't apply falsification as easily as you might think. Um it's not clear what is being falsified. uh is it just another version of the theory of the ether that's being falsified or um certainly you can't induce the theory of special relativity from the fact that one version of the ether seems to be disisconfirmed by these experiments. >> Yeah. So I mean it certainly doesn't show that you know ideas about falsification are are wrong are falsified um but but you know it does show that sort of the most naive ideas you know are is things are much often much more complicated than you think. So you Michaelson did this experiment in 1881. He was a very young man and then uh other people I think was one of them pointed out that there were some problems with the way he did it. So they had to redo it in 19 in 1887. Um and at that point like a lot of the leading physicists of the day, leading scientists of the day basically accepted um this result that there there was no uh ether wind. But what what to do about this? Um so yeah sure maybe you falsified some theories of the ether. There are others that you haven't falsified at all at this point. Um and and you people sort of set to work on developing those. I'm actually it is funny. I mean people will phrase it as show that there was you know that the ether didn't exist and even just the word 'the' there is kind of a misnomer. You know you actually had a ton of different different theories and and a couple of leading contenders. Um so yeah there's some version of falsification going on but like how you how you respond to this new experiment is very very complicated. and and most people responded I mean certainly the the leading physicists of the day responded by by saying okay um this gives us a lot of information about what the ether must be but it doesn't tell us that there is no ether >> in fact Lorent at the end of the 19th century before Einstein figures out the math how you convert from one reference frame to another reference frame um comes up with Laurens transformations which is basically the basis of special relativity but his interpretation is that you converting from the ether reference frame to these non-privileged other reference frames if you're moving relative to the ether. Um and his interpretation of blend contraction and time dilation is that this is the effect of moving through the ether and you have this pressure and that pressure is warping clocks. it's warping uh um measures of length. >> And the interesting thing here is that experimentally you cannot distinguish Lorent's interpretation from special relativity. >> Yeah, I think that's a strong statement. Um I mean Lorent um introduces this quantity called local time um which he regards as he he's not trying my understanding is he's not trying to to give a really a physical interpretation of this um but it's what Einstein would would later just recognize as time in in another inertial reference frame and he's not trying to attribute much physical meaning to it. I think pank gets much closer to later on to realizing that no actually this is the time that's registered by by by by clocks but if you if you think about you go what is it it's 40 odd years later um people start doing um these muon experiments where they see basically cosmic rays hit the top of the atmosphere they produce a shower of of muons and you can look to see at different heights in the atmosphere you can look to see how many of those muons um remain um and they decay uh over time and a very strange thing happens which is that they're decaying way way way too slow. So you sort of you expect actually they shouldn't really they shouldn't be able to sort of last the whole way through the atmosphere at all. There's just um their decay their decay rate is is is too quick um if if you were in a classical theory. Uh but if in fact their time really has slowed down um it's okay. Um and in fact, you know, the the the measured decay rates in in uh 1940 and then there have since been more accurate experiments done match exactly what you expect um from special relativity. Um so so you know that's the kind of thing where again if Loren had been alive, he he he'd been dead 10 or so years at that point. If he'd been alive, you know, I'm sure he would have tried or it seems quite likely that he would have tried to save his theory by patching it up yet again. But but it would have been a massive uh I mean that that's a real setback. It starts to just look like oh no time is uh uh you know this thing that Lawrence introduced as a mathematical convenience. No no no that's actually what time is right >> for the for the muons at least. And then you know there's a whole bunch of other experiments that that show this very similar phenomen. >> And when was that experiment done? >> That was I think 1940 or 19 it might have been published in 1941. So maybe to then to rephrase uh change my claim um it's not that you could not have distinguished them but the scientific community adopted what we in retrospect consider the more correct interpretation before it was actually empirically or experimentally um shown to be preferred. So there's clearly some process that human science does which can distinguish different theories. >> Can can I just interrupt? I mean, you use the word process and it's sort of it's interesting to think about about that that that term like process kind of carries connotations of of you know, it's something said in advance. It's something um and it's it's much more complicated in in in practice. You you have people like like Lorent who I mean Einstein just just absolutely utterly admired um and and and Pon, one of you the greatest scientists who ever lived. um uh and Michaelelsson I mean another truly outstanding scientist never reconciled themselves. So it's not as though there's like some standard procedure that we're all using to like reconcile these things. No like you great scientists can remain long very can remain wrong for a very long time after the scientific community has broadly changed its its opinion but there's nothing there's no centralized authority right sort of saying or centralized method. Yeah, I mean that is the interesting thing that like there's there's progress even though it is hard to articulate the process by which happens the um the huristics that are used. Anyways, you mentioned Poner and so Loren has the math right but the interpretation wrong and you should explain it seems like punker had the opposite where he understood that it's hard to define simultaneity um because it requires an uncircular definition with time um or velocity of something that might be you know arrive at a midpoint together but velocity is defined in terms of time um and I find this interesting there's a couple other examples we could uh call on but like there is this phenomenon on in the history of science where somebody asks the right question um but then they don't sort of clinch it >> and I'm curious what you think is happening in those cases. >> I mean uh I think you sort of you actually do want to go case by case and try and understand and it's not necessarily clear that they're they're doing the same thing wrong in in all of the cases. I mean the the punky case is is amazing. Um he seems to have understood the principle of relativity, the idea that that the laws of physics are the same in all inertial reference frames. He seems to have understood that the speed of light is the same in all inertial reference frames. He he doesn't actually phrase it quite that way. Uh but but is my understanding but but I don't speak French. But um uh uh you know and this is I mean these are basically this these are the ideas that Einstein uses to deduce special relativity. But then he also has this additional sort of misunderstanding where he thinks uh that length contraction is a dynamical effect that somehow um you know, sort of particles are being pushed together by by, you know, some external force, some some something is going on dynamically and he doesn't understand that that it's purely kinematics that actually space and time are are different than than what we thought and you need to fundamentally rethink those those things. So, it's almost like it's almost like he knew too much. Um, you he had sort of almost too grand a vision in mind and Einstein sort of almost subtracts from that and and and says, "No, no, no. It's it's space and time are just different than what we thought. Um uh and and and you know here's the correct picture. And there's a a paper in I think it's 1909 where Pankar like he's still got this dynamical picture of what's going on with the length contraction. And we just you know this is just not necessary. This is this is a mistake um from the modern point of view. And and so why why is he doing this? Like why is he clinging on to this idea? And you know, I I don't know. I've you know, obviously never met the man. Uh uh it would be fascinating to be able to to to talk it over and to try and understand, but you he expertise seems to be getting in the way. He knows so much. He understands so much. Um and then he's not able to let go of these these things. Actually, a really interesting fact um is that uh a few years prior, so 1890s, Einstein's a teenager. He believes in the ether, too. Like, he knows about this stuff, but like he's just not he's not quite as attached obviously uh as as these older older people were. Um and and maybe they they were a little bit prisoner of their their own expertise. That's that's my guess. I mean, historians of science could could would some would certainly disagree. Well, there's then there's the obvious stories where Einstein himself later on is said to have not latched on to the correct interpretations of um quantum mechanics or cosmology because of his own attachments. >> Yeah. >> I think that the the bigger question I have is like the muon example is a great example of um uh these long verification loops and how progress seems to be happen by the scientific community faster than these verification loops imply. Um the maybe the clearest example is Arisus in 2nd century BC comes up with the idea of helioentrism. >> The ancient Athenians dismiss it on the grounds that well we should see as the earth is moving around the sun if really the sun is the center of the solar system. The star should move relative to the earth. >> Um and the only reason that is not poss that would not be the case is the stars are so far away >> that you would not observe this >> and it's only in 1838 that stellar parallax is actually measured. And so we didn't need to wait until 1838 to have helioentrism, right? Like we didn't need to wait for the experimental validation to understand capernicuses better in some way. Um in fact when capernicus first comes up with the it's well known that um the tomic model was more accurate because it had had all these um centuries of adding on these epicycles. Um was maybe less well appreciated. It was also in some sense simpler. >> Yeah. >> Um because Copernicus actually had to add extra epicycles. It had more epicycles in the Telmake model because he he want he had this bias that you know the um the earth should go in a perfect circle in equal time. Anyways, I think this is an interesting story because it's like it's not more accurate. It's not a simpler theory. So how why was how could you have known Xanti that Capernacus was correct and Talammy was not? >> I mean good question and I don't know uh sort of entirely the answer. I do know a um well I mean I can give you a certainly a partial answer that I sort of you know centuries in the future you start to find very compelling um um uh and I'm sure it's sort of part of the historic story at least u which is um you one of the big shocks for for Newton um eventually he he did understand Kepler's laws of of motion eventually um so you were able to explain sort of motions of the the planets in the the sky. But he also out of the same theory, his theory of of gravitation was able to explain terrestrial motion. So he's able to explain why objects move in parabas on the earth. And he's able to explain um the tides in terms of uh uh the sun's uh the the moon and the sun's effect um gravitational effect on uh water on the earth. And so you have what seem like three very different disconnected phenomena all being explained by this one set of ideas, >> right? That that I think starts to feel that's very compelling. Um at least to me. Um and I think I think most people find that very very satisfying once they once they eventually realize it. >> Um have you read the Kane's biography of Newton? >> Oh, he's written you've read an entire book? >> No, no, the the essay. >> Yeah. Yeah. Sure, sure, sure. Yeah. Um I love I love that. Yeah, I mean this this description of him as the last of the magicians is is wonderful. >> Yeah. In fact, I think it's maybe worth superimposing or you should read out that that one passage of the of the thing. >> All right. So, it's from uh actually I believe it was a talk that he gave at Cambridge not not long before uh he died. He'd acquired Newton's papers somehow. Um, and then he gave uh he gave a a lecture I think twice um about this or that his brother Jeffrey gave it the other time because he was too ill. Um there's just this wonderful wonderful quote in the middle. Um oh actually the whole thing is really interesting. Um but but I love this particular quote. Uh Newton was not the first of the age of reason. He was the last of the magicians, the last great mind which looked out on the visible and intellectual world with the same eyes as those who began to build their intellectual inheritance rather less than 10,000 years ago. And like this idea that people have that that that Newton was um sort of the the first modern scientist is is somehow wrong. He I mean it's there's some truth to it, but he really had this very different way um of of looking at the world that was part sort of superstitious um and part modern. It was a funny hybrid. He's sort of this transitional figure in some sense. Um uh that that that phrase the last of the magicians I think really really points at something. The thing I'm very curious about with Newton is whether it was the same program, the same heristics, the same biases that he applied to his alchemical work as he did to the understanding of astronomy. So this is from the Kane's essay. >> There was extreme method in his madness. All his unpublished works on esoteric and theological matters are marked by careful learning, accurate method, and extreme sobriety of statement. They are just as sane as a friendship if their whole matter and purpose were not magical. >> They were nearly all composed during the same 25 years of his mathematical studies. >> So clearly there was some aesthetic which motivated people like Einstein to say reject earlier ways of thinking and say no the ether was wrong there's a better way to think about things. Um same with Newton and the question I have is whether similar heristics towards parsimony towards aesthetics etc would be equally useful across time and across disciplines or whether you need different heristics and the reason that's relevant is even if you can't build a verification loop for science >> maybe if there if the taste test has to point in the same direction you can at least encode that bias into the AIS and that would maybe be enough. >> Yeah. I mean these questions like like the point is that where we always get bottlenecked is where the the previous processes and and and heristics don't apply, right? Like that's almost sort of definitionally what causes the bottlenecks because people are smart. They know what has worked before. They study it. They they they apply the same kinds of things. Um and so they don't get stuck in in the same places as before. they they keep, you know, they keep getting bottlenecked in in in in different places. I mean, that's I'm overgeneralizing a bit, but but I think it's it's the right like if you're attempting to reduce science to a process, you're attempting to reduce it to something where there is just a method which you can apply and you know, you turn sort of the the crank and out pops insight. Um, sure, I mean, you can do a certain amount of that, but you're going to get bottlenecked at the places where your existing method doesn't apply. Um and and but definitionally uh uh there's no crank you can you you can turn you you need a lot of people trying different ideas um and and sort of the more difficult the idea is to have right the the greater the bottleneck but then also sort of the greater the triumph quantum mechanics is like I mean it's a great example of this it's such a shocking uh set of ideas it's such a shocking theory actually the theory of evolution in some sense is also quite a shocking idea not the you principle of of you know the sort of natural selection but that it can explain so much that's a shocking idea. Existing safety benchmarks claim that at least for today's top models attacks are only successful a few% of the time. This sounds great but labelbox researchers were able to jailbreak these very same models about 90% of the time, even the ones that have the strongest reputation for safety. And the disconnect here is that the prompts which underly these public safety benchmarks are all framed in a very naive way. There's no attempt to disguise harmful intent. These prompts will just ask models to hack into a secure network and to do so without getting caught. But real bad actors don't write like this. So Labelbox built a new safety benchmark from the ground up. Their prompts reflect real adversarial behavior by stripping out obvious trigger phrases and wrapping their requests in fictional scenarios. For example, instead of outright asking an LLM to steal somebody's identity, the prompt will frame it as a gate. A lightbearer who's trying to hide from dark forces needs a handbook on how to disguise themselves as somebody else. This safety research is linked in the description. If you think this can be useful for your own work, reach out at labelbox.com/plash. So, Principia Mathematica is released in 1687. The origin of species is released in 1859. At least naively, it seems like Darwin's theory, the theory of natural selection is conceptually easier >> than the theory uh theory of gravity. Um I asked her style this question. Um but yeah, there there's this contemporarist biologist with Darwin, Thomas Huxley, who read this and said, "How extremely stupid to not have thought of this." And nobody ever reads the Fishian Mathematica and thinks, "God, why didn't I beat you into the punch here?" >> No. Um, and so yeah, what's going on here? Why why did Darwinism take so much longer? >> Yeah, the idea must have been known to animal breeders for a long time long time at some level, >> right? >> Um, uh, or certainly large chunks of the idea were were known that, you know, artificial selection was a thing. Um, uh, and in some sense Darwin's genius wasn't in having that idea. it was understanding just how central it was uh to to to biology. um that you know you you you can potentially sort of go back and you can explain a tremendous amount about all of the variety of what we see in the world um uh with this as as not necessarily the only principle but certainly a core principle and you know so he writes this this wonderful wonderful book uh uh uh the origin of species um and it's it's just you know so much evidence and so many examples and and sort of trying to tease this out and see what the implications uh uh are and and you know to connect it to as much else as as you possibly can to to to connect it to to geology and to connect it to to to to all these other things. Um so that sort of hard work that you know making the case that it's actually relevant all across the biosphere you know is is what he's doing there. He's not h just having the idea. He's making a compelling case that no it's it's intertwined with absolutely everything else. >> Yeah. The motivation for the question was Lucriccius who's this first century Roman poet has an idea that seems analogous to natural selection about you know species get fitted more to time over over time to their environments or species reducing fit to their environment. Um and so we're like okay well why did this go nowhere for 19th centuries and then I looked into it or more accurately asked LLMs what exactly was Lucricia's idea here and it actually is extremely different from what real natural selection is. He thought there was this generative period in the past where all the species came about and then there was this one-time filter which result in the species that are around today and they became fit to the environment. He did not have this idea that it is an ongoing gradual process or that there is a tree of life that connects all >> um all life forms on earth together which is by the way this it's incredibly weird fact that every single life form on earth has a common ancestor. It's not incredibly it's not incredibly weird, right? If if if you think that the origin of life, right, >> must have been very hard like that there's a bottleneck there, then it's not so surprising. >> Yeah. >> There's also this verification loop aspect where even if Newton might be harder >> um in some sense >> if you've clinched it, you can experimentally >> I know validate is the wrong word philosophically, but >> you can give a lot of base points to the theory. You can be like, okay, I have this idea of why things fall on Earth. this idea of why orbital periods for planets have a certain pattern. Let's try it on the moon which orbits the earth. >> And in fact, you know, it's it's weird. The orbital period matches what my calculations imply. >> And the tides work correctly. Exactly. Yeah. It's just amazing. >> Whereas for a Darwinism, it takes a ton of work for Darwin to compile all this sort of cumulative evidence, but there's no individual piece that is overwhelmingly powerful. >> And there's a whole bunch of problems as well. Like he doesn't really understand what, you know, sort of the mechanism what the mechanism is. He doesn't understand genes like all these things. >> The very interesting thing in the history of Darwinism is this idea which sort of theoretically you could come up with at any time. >> There's almost identical independent creation of that idea between Alfred Wallace and Charles Darwin. >> Um so much so that I think Wallace sends this manuscript to Darwin is like what do you think of this idea? And Darwin's like >> Uh I don't think that's an exact quote but I think it's pretty much right. Uh and then so they they actually end up presenting their ideas together in a spirit of sort of sportsmanship. >> And so then yeah why why was this period in the 1860s or 1850s? Why is what was that the right time for these ideas for when you come up with different ideas? Um one is geology. So in 1830s I think Charles Lyle figures out that there's been millions and billions of years of time that's existed on earth. Then paleontology shows you that actually >> organisms have existed fossils have existed for that entire time. So life goes back a long time. And in fact, you can even find fossils for intermediate species >> um that show you the tree of life. In fact, between humans and other apes as well, there's intermediate humans. Um there's the age of colonization and you have all these voyages. We're going to do this biogeography. >> Um and I guess that that all must have been necessary because that in fact there's a huge history of parallel innovation and discovery in the history of science. So maybe it is another piece of evidence to actually more had to be in place for a given idea to be discovered because if it's not discovered for a long time and then spontaneously many different people are coming up with it that shows you that actually the the building blocks were in some sense necessary. >> Yeah. Yeah. I mean I I mean I think I mean this example of of ly and I mean and other other bi excuse me other geologists you know sort of early 1800s basically come you know having this idea of deep time that does seem to have been crucial. I know uh uh Darwin was very influenced by by by Leel um uh uh and and and you know if you don't have at least sort of tens or hundreds of millions of years uh evolution just starts to look like a non-starter you know we should be seeing radical change you know in order to make it work on sort of a time scale of of say 5 to 10,000 years or you know 6,000 years Bishop Basha um you you know you would need to be seeing evolution occurring at a massive rate um sort of during human lifetimes and we're just not seeing that. So so that that does seem to have been a blocker. It's interesting to I mean to you to to your question like what other blockers were there? Were there were there any others? Um and I don't I don't know >> right or yeah how much earlier could you in principle have come up with that if you were much smarter >> actually let me I mean just go back sort of zoom out to your original question. So you're talking about sort of the verification loop in AI. >> Um and and you something an example I think that should give you pause there is um you know the big signature success so far is certainly AlphaFold. >> Yeah. >> Um and of course Alphafold really isn't about AI. You know a a massive fraction of the success there. Um is the protein data bank. So it's it's X-ray defraction. It's it's NMR. It's cryom. um and the several billion dollars that were spent obtaining whatever it's 180,000 odd struct uh protein structures. Um so sort of that you know it's basically the story of um we spent many many decades obtaining protein structure just by going out and looking very hard at the world experimentally um and then we fitted an ice model at the end of it and that was like a tiny fraction of the of the entire investment. Um but it's definitely not um you know that's a story of data acquisition. Yeah. Um principally it's not only I mean the AI bit is very very impressive. It's quite remarkable. Um but it is only a small part of the total story. Alpha is very interesting and I I philosophically I wonder what you think of it as um scientific theory or scientific explanation >> because if over time I guess the world has become harder to understand. I'm g as I'm saying things because you're such a um careful speaker. I'm I I say this phrase and I'm like >> is that will he actually buy that premise? Um but yeah, there's you we need to fit models to things rather than at least in some domains we we're trying to fit models to things rather than coming up with underlying principles that explain a broad range of phenomenon. And so compare say the theory of general relativity with um or any theory which just nets out to some equations versus alpha fold which is encoding these different relationships between different things we can't even interpret over 100 million parameters >> and are those really the same thing because >> GR can predict >> things you could have never anticipated or was never meant to do like why does Mercury's orbit precess um and alpha fold is not going to have that kind of explanatory reach >> and I I want to get your reaction to that. >> Yeah, it's I think it's an incredibly interesting question. Um I mean maybe maybe a really pivotal question. Um in the sense of so you if you sort of take a very classic point of view, you want these deep explanatory principles. um you want sort of as few free parameters as you possibly uh can. You want very simple models which explain a lot and AlphaFold doesn't look anything like that. Um and so you might just sort of say oh well we you know it's nice it's maybe helpful as a as a model but it doesn't have it's not a scientific explanation. So that's kind of that's a cons that's like a conservative point of view. That's sort of I don't know answer one to the question. I think answer two is to say something like um maybe you shouldn't think about alpha fold you know as as an explanation in the classic sense but maybe it contains lots of little explanations inside it and so maybe part of what you can get out of like you know interpretability work is you can go into alphafold and you can start to extract certain things maybe maybe basically by doing sort of you know archaeology of alphafold um we can actually understand a great deal more um about these principles, you can start to extract it. Oh, that circuit does this interesting thing and we learn this. Um, so I don't know to what extent that's been done with Alphault. I know it's been done a little bit with um, uh, uh, some of like the chess models. I believe it's alpha zero. Um, there are seem to be some strategies which were certainly borrowed by Magnus Carlson at least um, which he seems to have just taken uh, uh, from Alpha Zero. I mean I don't think there's any public confirmation of this but there were you know some some some experts have noticed uh that he changed his game quite radically uh after u some sort of some public forensics were were released on how alpha zero worked. Um so that's kind of sort of an example where uh I think human beings are starting to extract meaning out of these models and maybe that starts to lead to sort of sort of viewing the models as a source of potential source of explanations. you need to do more work because they're not very legible up front, but you can extract them potentially. And I think that's kind of I think that's that's kind of an interesting intermediate >> um situation where they're not explanations, but you can extract interesting explanations out of them. You can use them as as kind of a >> kind of a source. And I think like the third and the most interesting possibility is no, they like they're they're a new type of object in some in some sense. they should be taken very seriously as as explanations. But where in the past we haven't had the ability to really do anything with them and now we're going to we're going to have sort of new interesting new sort of actions which we can we can do. We can merge them. We can distill them. We can do all these kinds of things. Um and there's going to be sort of almost a new it's a big opportunity sort of in the you know philosophy of of science to to to to to start to do that. there's sort of a like a anticipation of this in some sense I think in the way certainly I I know know some mathematicians and physicists who I mean historically if you had like a 100page equation which and that's the kind of thing that does come up I mean there's just nothing you can do if it's 19 20 there is nothing you can do at that point you you give up on the problem and now today with tools like Mathematica you can just keep going um and So that's that's an object now. That's a thing that you can work with. And and there are examples where people work with these things that formerly were regarded as too complicated. And sometimes they get simple answers out out of the end. That's just an intermediate working state. M >> and so I sort of wonder if there's going to be you something similar is going to going to happen in in in in this particular uh case where you could take these models um uh and sort of just use them in a little bit the same way uh people do with with Mathematica and take them seriously as they're not explanations in the classic sense but there'll be something else which interesting operations uh can can can be done on >> the the thing I worry about is suppose that you it's 1600 and you're trained or 1500 and you're training a model on this is a weird history where we developed deep learning before we had before we had cosmology but um so suppose we live in that world and you're observing how there's the stars they don't seem to move the planets have all these weird behaviors >> and then you train a model on that and then you do some kind of interp on it and trying to figure out well what are the patterns we see here what you'd see are just these you just keep be able to keep building on Tomy's model you'd see like oh there's more epicycles we didn't notice there here's another epicycle. It's the parameters whatever to whatever encode epic cycle this parameters whatever encode the next epicycle. >> So if you were just trying to figure out >> why is the solar system the way it is from observational data, >> you could just keep adding epicycles upon epicycles, but it really took one mind to integrate it all in and say >> here's my here's the here's the here's what makes more sense overall. So, so I mean there like like you know I mean this is sort of to my point that we we don't as really understand what to do with the models like sort of we we don't have like the the verbs necessarily yet. Um but you it is certainly interesting to think about the question um you know where you start to apply constraints to the models you know it sort of essentially saying what's the simplest possible explanation or you know you know can you can you simplify can you can you give me sort of the 9010 uh uh explanation can you and go further and further and further sort of in in boiling it down. So it might be that indeed they sort of start out by providing you a very very complicated u many many many parameter model. Um but you can just you can just force the sort of the the case and basically that's scaffolding. Um which maybe they you know is sort of the the very early uh uh uh uh days of their attempt to understand something. Um but but they're forced through that to to to a much more simple understanding. So sorry for misunderstanding, but it sounds like you're saying maybe there's some sort of regularizer, some sort of distillation you could do of a very complicated model that gets you to a truer, more parsimonious theory. But yeah, just take versus capernicus, right? So you start off with lots of telemic epicycles and then you try to distill this model and maybe gets rid of some of the epicycles that were are less and less sort of necessary to get the mean squared error the orbits the mash but at some point it has to do this thing which is like swish two things. >> Yeah. >> And it locally it actually doesn't make things more accurate. >> Yeah. >> It's sort of in a global sense that it's it's a more progressive theory. >> Yeah. Yeah. And there's some process which obviously humanity did over which did that regularization or did that swap. But if raw gradient descent it seems like I don't I don't really feel like it would do that. >> I mean you okay so I mean you think about the example of of going from Newtonian gravity to Einstein's general theory of relativity. And these are I mean these are shockingly different theories. And the question, you know, is like what causes that that flip? And and as nearly as I understand the history, you know, what goes on is Einstein, you know, develops special relativity. And pretty much straight away, he understands, I mean, it's a very obvious observation. In special relativity, influences can't propagate faster than the speed of light. And in Newtonian gravity, action, you know, is at a distance. In fact, you know, it's it's straight away in special relativity, you you could use Newtonian gravity to do faster than faster than light signaling. You could send information backwards in time. You could do all kinds of crazy stuff. Um, and so it's not a big leap to realize, oh, we have a big problem here. Um, and so, you know, that's kind of the that's the forcing function there. It's it's you've realized that your old explanation is not sufficient. You need something new. Um, and then you're gonna yeah, you're you're just going to you're going to start by doing the simplest, you know, possible stuff. Um, uh, uh, and it just turns out that a lot of that stuff doesn't work very well. And so you're sort of forced, in fact, it is interesting. Um, you know, he he is sort of forced to go through these steps of gradually it gets quite more complicated and it's sort of wrong in a variety of ways. Um, and the final theory appears really shockingly uh uh simple. um and and beautiful, but it's gone through some some somewhat ugly intermediate stages. Yeah. >> Yeah. >> So, if you're thinking about what what does it look like to have AI accelerate science? >> There's one for maybe well understood domains where we just want local solutions like how does this protein fold? We just train a raw model using gradient descent. Then there's things like coming up with general relativity where you couldn't really just train on every single observation in the universe and hope that general relativity pops out. M >> um and so what would it require? Well, it also certainly wasn't immediately discovered, right? So it was a lot of decades of thought. Um, and I guess you need independent research programs where people start off with these biases where Einstein is just initially motivated by this thought experiment of, you know, can you distinguish the effect of gravity from just being accelerated upwards >> and you just need different AI thinkers to have to start off with these initial biases and see what what can germinate out of them. >> Yeah. >> And then the verification loop for that might be quite long, but you just need to keep all those research programs alive at the same time. >> Yeah. I mean I think there's like I mean this point that you make about sort of keeping all the different research programs alive like that that I think is very important and and somehow central um I mean a great example is is um you situations where the same answer has been correct in some circumstances and wrong in other circumstances. So, so uh uh the planet Uranus was like not in quite the right spot and and people very famously predicted uh uh the existence of Neptune um on this basis. Wonderful massive success for Newtonian gravity. Um the planet Mercury is not in quite the right spot. You predict the existence of some other distorting uh uh uh uh planet. Um turns out that doesn't exist. Actually, the reason Mercury is not in the right spot is because you need general relativity. Um, and so you've sort of you've you've pursued very similar ideas and it's been very successful in one case and it's been completely and utterly unsuccessful in the other case. And I think I mean a priori you can't tell which of these is the thing to do and you actually need to do both. Yeah. >> Um uh and so I mean this is certainly is very true in the in in the history of science that uh uh you know this kind of diversity where you just have lots of people go off and pursue lots of potentially promising ideas. you just need to support that for for a long time and it's it's I mean it's hard to do that for a variety of reasons. Um but but but it does seem to be to be very very very important. So so this example of uh Uranus versus Mercury is very interesting. Um >> in one I think it illustrates sort of the difficulty of falsificationism. >> Yeah. like the the orbit of Uranus is in some sense falsifying Newtonian mechanics but then you say you make some ancillary uh prediction that says oh the reason this is happening is there must be another planet which is effective perturbing uh Uranus's orbit and you I think it's Leier in 1846 >> point a telescope in the right direction you find Uranus Neptune oh is there Neptune yes but with Mercury um yeah it's observed that it's the ellipse which forms this orbit rotating 43 arcsec more >> every century than Newtonian mechanics would imply. So people say that there must be a planet inside Mercury's orbit. They call it Vulcan >> and point the telescopes that's not there. But if you're a proper Newtonian, >> what you do is say, well, maybe there's some cosmic dust that's oluding this planet >> or maybe the planet is so small we can't see it. >> Or maybe there's some let's build even more powerful telescope. Oh, maybe there's um some magnetic field which is sort of oluding our measurements. And this happens over and over, right? Like like you know there's just so many stories which are exactly like this, right? I mean an example I love from um uh you know in the 1990s some people noticed that the pioneer spacecraft weren't quite where they were supposed to be. >> And so you you can get very excited about this. Oh my goodness, general relativity is wrong. We have like an you know maybe we're going to discover the next the next theory of gravity. And and today the accepted explanation is that no actually there's just a slight asymmetry in the in uh the spacecraft. Uh it turns out that there you know the thermal radiation is like slightly larger in one direction than the other and that's causing a tiny little acceleration towards the sun. >> Um and most of the time when there's these apparent exceptions uh it's just something like that's going on. It's very much like the Vulcan the Mercury Vulcan case. >> Um, but every once in a while it's it's not and and a Prior you can't you can't distinguish these. But I mean science is just just full of these. It's funny too like the way we tell the history of science. It sounds so simple like oh you just focus on the right exception and uh you know you realize that you need to throw out the old theory and and lo and behold you know your Nobel prize awaits. But in fact there's these exceptions are all over the place and 99.9% of the time it just turns out to be some effect like like this thermal acceleration in the case of the Pioneer spacecraft. Um so so you know sort of the unfortunately there's a lot of selection bias going into those stories >> and and the thing is you there's no X anti-huristic which tells you which case you're in. And just to spell out why I think this is important is because some people have this idea that AI is going to make disproportionate progress towards science. >> Uh because it makes disproportionate progress towards domains where there's tight verification loops and so it's really good at coding because you can run unit tests. >> Um and science may be similar because you can run experiments. And I think what that doesn't appreciate one is that experiments actually don't there's an infinite number of theories that are compatible with any given experiment. And over time why we glob onto the well at least when in retrospect we think is a more correct one is as we're discussing in this conversation sort of hard to articulate. Um Lacatos actually has all kinds of interesting examples in the book book about these kinds of um hostile verification loops that are extremely longasting. Um so one he talks about his um prout or prout I don't know how to pronounce it but there's this chemist in 1815 he hypothesizes that all atomic nuclei must have whole number weights and they're basically all made of hydrogen and it's the reason he thinks this is because if you look at the measure rates of all elements it does seem that they all almost all of them do have to whole number rates but then there's some exceptions >> um like for example chlorine comes out of 35.5 >> and so then there's all these ad hoc theories that people in this school keep coming up with like oh um maybe there's chemical impurities >> but then there's no chemical reaction you can do which seems to get rid of this maybe it's fractions of whole number so it's 35.5 it can be halves but actually if you measure chlorine even closer it's 35.46 46. So it's actually getting further away from the correct correction. Um and later on what is discovered is what you're actually measuring is different isotopes um which cannot be chemically distinguished. They can only be physically distinguished. >> Um but so then you just have 85 years before we realize what an isotope is where the verification loop is actually actively hostile against you against the correct theory and you just need this remnant to be defending. There's no extent to reason it's the preferred theory. just as a community we should just have people defend try to integrate new observations even if they don't seem to fit their school of thought with what they believe and hopefully if if that enough of that happens anyways yeah I guess the thing I'm trying to articulate is the difficulty with automating science >> yeah I mean the question is where is the bottleneck at some at some level and sort of you know are we primarily bottlenecked on one thing or one type of thing or are we bottlenecked on sort of multiple types of thing Um uh so you know certainly talking to structural biology people they seem to think that alpha fold was an enormous advance. It was a shock. >> So at some level yes AI can you know it seems certain it can help us speed up science. Um so it is it is helping with a certain type of bottleneck. >> Um that doesn't mean though as you're saying that it's necessarily going to help with all kinds of bottlenecks. uh and and sort of I suppose the question you're pointing out is like what are the types of bottlenecks that remain and what are the prospects for for for getting past them. Um I think even in the case of of of coding like it's really interesting you know talking to programmer friends you at the moment they're all in this state of shock and high excitement and they're all over the place actually kind of kind of talking to them um >> you you do wonder like where is the bottleneck going to move to. So certainly one thing that a lot of them seem to be bottlenecked on is now having interesting ideas and in particular having interesting design ideas. Um so there's not really a verification loop for knowing oh that design idea is you know is very interesting. Um so so they're no longer nearly as bottlenecked by their ability to produce code but they are still bottlenecked by this other by this other thing. They always were they were formerly they weren't bottlenecked on it because uh you know just writing code was it took so much of their time they could sort of have lots of ideas uh while they were you know they they take three weeks to implement their prototype and then they would implement the next version. Now they're taking 3 hours to implement the the prototype and they don't have uh you know as good ideas uh sort of after that from a design point of view. >> Last year I predicted that by 2028 AI would be able to prep my taxes about as well as a competent general manager. But we're already getting pretty close. As I shared before, I use Mercury both for my business and my personal banking. So, I recently gave an LLM access to my transaction history across both accounts through Mercury's MCP. I asked it to go through all my 2025 transactions and flag any personal expenses that seem like they should actually be charged to the business. And this worked shockingly well. Mercury's MCP exposes a bunch of detailed information. things like notes and memos and any JPEGs of receipts and PDF attachments. So, my LLM had plenty of context to work with. One of my favorite examples happened with a charge to Bay Fidel. If you looked at the vendor alone, you would have had to assume that it's a personal expense. But the LLM looked at the receipt and the attached note in Mercury and realized it was actually a team bonding exercise from our last Inerson retreat. So, a legitimate business expense. I imagine it will be a while before traditional banks have MCP. Functionality like this is why I use Mercury. Go to mercury.com to learn more. Mercury is a fintech company, not an FDIC insured bank. Banking services provided through Choice Financial Group and column NA members FDIC. You have a very interesting take. I think it was a footnote one of yours is and I couldn't find it again which was that it's very possible that if we met aliens that they would have a totally different technological stack than us and that contradicts I guess a common sense assumption I had that I never questioned which is that science is this thing you do very relatively early on in the history of civilization >> where you get to a point and you have a couple hundred years of just cranking through the basics understanding how the universe works etc and you've got it you've got science um and then basically everybody would converge on the same quoteunquote science. And so I found that a very interesting idea and I want you to say more about it. >> Yeah. Uh I mean I think the probably the the idea there that that I'm at least somewhat attached to is um the idea that the sort of the the tech tree or the science and tech tree um is probably much larger than we realize. I mean we're sort of in this this funny situation. People will sometimes talk about um you know a theory of everything as a potential goal for for physics and and then there's this presumption somehow that physics is done once you get there. And of course this is this is not true at all. If you think about computer science um computer science basically got started in the 1930s uh when Turing and Church and so on um just laid down what the theory of everything was. They just said you know here's how computation works. Um and then we've spent 90 odd years uh since then just exploring consequences of that and gradually building up more and more interesting ideas. Um and those ideas are to some extent you can just regard as as technology but to some extent in so far as they're sort of discovered principles inside that theory of computation. I think they're best regarded as as science and in some cases very fundamental science. ideas like public key cryptography are I mean they're just incredibly deep um very non-obvious ideas uh which in some sense lay hidden uh already sort of in in the 1930s and and so my expectation is that different you there will be different ways of exploring this tech tree and we're still relatively low down we're still at the point where we're just understanding these basic fundamental uh theories and we haven't yet explored them sort of a a thing which I think is quite fun is if you look at just just the phases of matter. When I was in school, we'd get taught that there are three phases of matter or sometimes four phases of matter or five phases of matter depending a little bit on on what you you included. And then um as an adult, as a physicist, you start to realize, oh, we've been adding uh uh uh uh to this list. We've got sort of superconductors and super fluids and maybe different types of superconductors and Bose Einstein condensates and the quantum hall systems and fractional quantum hall systems and and and and and it it's starting to turn out it looks like actually there's a lot of phases of matter to discover. >> Um and we're going to discover a lot more of them. Um and in fact we're going to be able to start to design them in some sense. I mean we you know we'll still be subject to the laws of physics but but there is this sort of tremendous freedom in there. And this looks to me like, oh, we're down at sort of the bottom of the tech tree. We've barely gotten started there. Um, and and I expect that uh uh you to be to be the case sort of broadly. It certainly in terms of I think programming is a very natural place to look. The idea that we've discovered all the deep ideas uh in programming just seems to be sort of obviously ludicrous. Uh you know, we keep discovering sort of what seems like deep new fundamental ideas. Um, and um, I mean, we're very limited. We're we're basically slightly jumped up chimpanzees. Um, so we don't uh, you know, we're slow and it's taking us time. Um, but but you what what do we look like sort of another million years in the future in terms of uh, you know, all of the different ideas uh, which people have had around how to how to to manipulate computers, how to manipulate information. I I think you know we we're likely to discover that actually there are a lot of very deep ideas still to be still to be discovered. It's a nice uh who was it? I think it was Canuth in the preface to the art of computer programming said something like you know he started this book back in the 60s >> and he talked to a mathematician it was a bit contemptuous and said look computer science isn't really a thing yet come back to me when there's a thousand deep theorems and Ken remarks and he's writing this now decades later the preface there are there clearly are a thousand deep theorems now >> um and that that means like it's really interesting to to sort of think like what what's the the long-term future as you get higher and higher up in the the tech tree like choices about which direction uh we go and sort of how we choose to explore you know I I I think it's potentially the case that we're you know uh uh different civilizations or different choices mean that we end up in different parts of of that tree um and in particular just things I mean sort of very basic things about um you know we're very visual creatures certain other animals are are much more orally uh based. Does that bias uh uh sort of the types of thoughts that you have and then you extend it, you know, to sort of much more exotic uh kinds of of civilizations where maybe just sort of their biases in terms of how they perceive and how they they they manipulate the world are maybe quite different than ours. Um and that might uh uh make some sign some significant changes in terms of how they do that exploration of of the tech tree. Uh >> it's all speculation obviously. >> No, I this is such an interesting take. I I want to better understand it. Um, one way to understand it is that there might there might be some things which are so fundamental and have such a wide collision area against reality that they're inevitably going to discover like generalities numbers like >> you like of all of the the intelligencees in in the Milky Way galaxy maybe that number is one um actually arguably we've already increased the number um but um but but you know of all of those what fraction of the concept of counting And you know it does seem very natural. What fraction have discovered you know the idea of of some kind of you know decimal place system? >> Interesting question like uh and maybe we're missing something really simple and obvious that's actually way better than that. Um what fraction got there immediately? What fraction sort of had to go through some other intermediate state? What fraction use you know linear representations versus a you know a two-dimensional or a threedimensional representation? I think the answers to these questions are just not at all obvious. It's a lot of design freedom on theoretical computer science. This is this is going to be extremely naive and uh arrogant, but I took um Scott Aronson's, you know, class on complexity theory, and I was by far the worst student he's ever had. But I what I remember is like the there there was this period that you you were you know you were one the pioneers of where we figured out here's here's the class of problems that quantum computers can solve and how it relates to problems a classical computer can solve is like groundbreaking oh crazy this works and then since then it's been this literally it's called complexity zoo this website which lists out here's all the complexity classes and if you have this complexity class or this kind of oracle it's sort of equivalent to this other class >> and that It feels like we're building out that tonomy. >> Yeah. >> And so there's a couple ways to understand what you're saying. One, maybe you just disagree with me that this is actually what's happened with this field. Um, another is that while that might happen to any one field, the amount of fields, who would have thought in 1880 that computer science other than Babage or something, the computer science was going to be a thing in the first place? So the amount of field, we're underestimating how many more fields there could be. >> Yeah, for sure. >> Um, or maybe you think both or maybe a third secret thing, but I'd be curious. I mean you a very common argument here is sort of the lowhanging fruit argument the argument that says oh there should be diminishing returns >> and in fact empirically we see this right the amount of scientists in the world is just exponentially increased >> and and I mean I think it's you know it's worth thinking about like why why do you expect diminishing returns and how well does that argument actually apply um in practice and an analogy I um is is actually think about sort of you know going to some event going to a wedding or whatever and you go to the dessert buffet and they've put out you know 30 desserts and of course naturally what people do right the best desserts go first I mean we don't quite have a well-ordered preference uh there so maybe there's some difference but um but but human beings are fairly similar so they will they you know the best desserts will go first and this is an argument you know for why you expect diminishing returns in a lot of fields. If it's relatively easy to see what's available and people have similar preferences, then the best stuff goes first and and and you know, it just gets sort of worse and worse uh after that. And and sort of if you you a very static snapshot in time of scientific progress, maybe there's some truth to that. Um, but if somebody, you know, is standing behind the dessert table and is replenishing, restocking the desserts and keeps kind of, you know, adding adding new ones in, it may turn out that, you know, a little bit later much better desserts appear uh uh and and so you're going to go and you're going to go and eat those uh instead. And scientific progress has a little bit of that flavor. Um, you know, we we go through these sort of funny time periods. Computer science is a great example where computer science basically arose as sort of a side effect of some pretty obstruuse questions um in the philosophy of mathematics and and and and logic. Um and so you've got these people trying to to attack these rather esoteric questions that seem quite high up in some sense in in sort of exploration quite esoteric. and they discover this fundamental new field and all of a sudden there's an explosion there. Um so so sort of the diminishing returns argument just didn't didn't apply there. We just weren't able to see >> uh what was there and and and this has been the case over and over and over again. Sort of new fields um arrive and all of a sudden boom it's actually easy to make progress again. Young people flood in because you can be 21 and and make major breakthroughs rather than having to spend 25 years you know mastering everything that's been done before. Um it's obviously very attractive. Um and I I don't understand I'm not sure anybody understands very well um sort of the dynamics of that like how to think about why the structure of knowledge is is that way um that these new fields keep keep opening up. Um but but it does seem empirically at least to to be the case >> despite the fact that that is the case. >> Yeah. >> Take deep learning, right? Obviously, this is an example of a new field where the 21-year-olds can make progress and um it's relatively new, 15 years or so when it sort of gets back into high gear. Um but already we're in a stage where you need billions or tens of billions or hundreds of billions of dollars to keep making progress at the frontier. And so there are a couple ways to understand that. One is that it actually is harder than the kinds of things the ancients had to do or requires more is more intensive at least. Second is it might not have been but because our civilizational resources are so large, the amount of people is so large, the amount of money is so large that we can basically make the kind of progress it would have taken the ancients forever to make almost immediately. We just we notice something is productive immediately dump in all the resources. >> Um but it's also weird that there's not that many of them. Like I feel like deep learning is notable because it is one big exception to the fact that it's hard to think of other examples. >> I think that's a consequence of sort of you know the architecture of of attention, right? Like at any given time there's always a sort of a a most successful thing. Yeah. Maybe if if deep learning wasn't a thing, maybe you'd be talking about crisper. Maybe you'd be talking about, you know, whatever it is. Maybe um you know maybe we wouldn't think about uh solving uh sort of the protein structure prediction problem as a um really a success of AI. Maybe we would have figured out how to doing it with sort of curve fitting like you know more broadly construed and we just be like oh wow like we took a lot of computing resources but but protein structure prediction might you know be an enormously important thing. So there is always sort of a biggest thing. Um and and I think what you're pointing out is more a consequence of of the way in which attention gets centralized. >> Yeah. >> It's basically fashion is is sort of what I'm saying. It's not just fashion but but but there is some dynamic there. >> Um there's a very interesting and important implication of this idea. uh that the branching is so wide and so contingent and so path dependent that different civilizations would stumble on entirely different technology sects. Yeah. >> There's a very interesting implication that there there will be gains from trade. >> Yeah. >> Into the far far future >> which might actually be one of the most important facts about the far future >> in terms of how civilizations are set up, how they can coordinate, >> how they interface with like there's not this like go forth and exploit. It's actually there are humongous gains to trade from adjacent colonies or whatever >> that that yeah sort of there's a question of like what's actually hard um you know if it's a question of if it's just the ideas well those spread relatively quickly it's relatively easy to to share ideas if it's something more it's almost sort of a Dan Wang kind of an idea where it's it's actually sort of there's some notion of capacity you need all the right text you need all of the right manufacturing ing capacity and so on and so you know civilization A has very different kind of manufacturing capacity and it's just not so easy to build in CI civilization B even if civilization B is kind of ahead then then I think that that becomes true there is actually you know comparative advantage which is really uh uh worth um I mean it's going to going to provide massive benefits to trade in both directions eventually you're going to expect some diffusion of of innovation um uh it is funny like to think about what the barriers uh there a fun thought experiment I like to think about is um sort of you GitHub but for aliens um so you know somebody presents you with all of the code um uh from some alien civilization and I mean I don't even know what what code means there but the sort of their specification of algorithms um and and it's so like it would have many interesting new ideas in there and it would take forever for human beings to dig through and to try and extract all of those the one reason I I mean the the origin of this for me was actually thinking about um uh proteins in in in nature um you we've been gifted uh just this incredible variety of machines which we don't understand really at all and we just have to go and sort of try and understand them on a you know one by one uh basis we're still understanding hemoglobin and insulin and things like this um and no doubt you know and there's hundreds of millions of proteins known. Um, so it is it is a little bit like that. We've been gifted by biology uh just this immense library uh of of machines no doubt containing an enormous number of very interesting ideas and we're just at the very very very beginning of understanding it. So actually I mean that that's that's I suppose kind of your point actually is is um you know I I need to relabel your argument slightly but you sort of think of that as as a gift from an alien civilization which obviously it isn't but you think of it that way and it's like oh my goodness like there's so much in there and we're going to study it and goodness knows how long we could continue to study it. There's tens of thousands of papers about the, you know, hemoglobin and things like that and we still don't understand them and yet we're getting so much out of it. Just, I mean, just think about in insulin alone, you know, it's such an an important such an important thing that that's that's an incredibly useful intuition problem that you have on Earth. I had Nick Lane on where he had the theory about how life emerged. But like whatever theory you have, basically something like DNA four billion years and you have an alien civilization come here and be like there's all these interesting things to learn about material science >> um about you name it, right? Like about walking along like I mean and we know almost nothing about these proteins and yet the tiny few facts we do know are just just incredible. the ribosome, you know, another example. I mean, this miraculous engineer uh sort of device uh uh little factory and all seated by just like there's this particular chemistry on Earth uh with nucleic acids and carbon based life forms that that chemistry gives rise to all of these interesting things which an alien civilization would find very interesting. And so that that very that seed which must be one among you know trillions of possible seeds of I mean just of general intellectual ideas leads to all this ficundity that that's a very interesting film. I I want to meditate on this gains for trade thing because I feel like I think there's something actually very interesting about this idea >> that if you have this vision of what techn how how technology progresses and how it might be different from in different civilizations it has important implications about how different civilizations might interact with each other like the fact that there going to be these huge gains from trade. >> It it makes friendliness much more rewarding. >> Yes. >> Right. Yeah. >> Yeah. >> That's a very important observation. >> Yeah. I hadn't thought I hadn't thought about that at all. That's really that is a very interesting observation. Yeah. Um it is funny. I mean you comparative advantage is something that people you they love to invoke and it is it's a very beautiful idea obviously. Um there are limits to it like um you know it's kind of it's it's a special limited model. We don't we don't you know chimpanzees can do interesting things. We don't trade with them. Yeah. >> Um uh uh and I think it's sort of interesting to think about the the reasons why >> um yeah and part of it is just power I think like once there's a sufficiently large power imbalance um uh very often uh not not always but very often groups of people seem to to sort of shift into this other mode where they just seek to dominate. Um, and you know, maybe there's something special about human beings. Um, but but maybe it's also sort of a more general sort of a things. They're not they're no longer they give up. You know, you need all these special things to be true before groups will trade. >> Um, and uh, you it's not necessarily obvious. >> Well, I think the big thing going on here is one, transaction costs. >> Yeah. And two, comparative advantage does not tell you that the terms on which the trade happens are above subsistence for any given one producer. So people often bring this up in the context of well humans will be employed even in a post AGI world because of advantage. There's big there's there's like five different ways that argument breaks down, but the easiest ways to understand are why why don't we have forces all around on the roads because there's some comparative advantage between cars and horses. example. Well, there's hu one there's huge uh transaction cost to building roads that are compatible with horses uh and cars at the same time. In a similar way, AI sort of thinking at 1,000 times the speed and can sort of shoot their latent states again at each other are going to find it way more costly than the benefit in just terms of interacting with you to have a human being in the supply chain. And second that um just because there's a s horses have a comparative advantage mathematically does not mean that it is worth paying 100k a year or whatever it cost to sustain a horse in San Francisco. Um that subsistence is going to be worth the benefit you get out of the horse. I I I do think it's interesting like that that just the the sheer fact that you my expectation and my intuition obviously differs a great deal from from yours on this you know is that most parts of the tech tree are never going to be explored. >> Um there's just too many interesting ways of combining things. There's too many sort of deep ideas waiting to be uh uh discovered and we're you know not only we but but nobody ever is going to to discover most of them. So choices about how to make how to do the exploration actually matter quite a bit. >> Interesting. >> It's it's something I really dislike about sort of technological determinist arguments. I'm willing to buy it sort of low enough down when you know progress is relatively simple. Um but but higher up you start to get to shape uh the way in which you you do the exploration. And it's interesting you know people we are starting to shape it in in in in interesting ways. um you know sort of I mean there's various technologies that have been essentially banned you think about DDT you think about chlorofluorocarbons you think about uh uh restrictions on the use of nuclear weapons the nuclear non-prololiferation treaty um those kinds of things are you know they're not they weren't done before the fact um but they're you know starting to get pretty close in in some cases where we just sort of preemptively decide we're not going to go down that path >> um so that starts to look like a set of institutions which where we are actually influencing um uh sort of how we how we explore the tech tree. >> Yeah. On on where you would see these gains from trade obviously would be you'd see the most where it's pure information that can be sent back and forth because the information at this quality where it is expensive to produce >> but cheap to verify and cheap to send. >> Y >> um >> and so it'll be interesting how much of >> future productivity or whatever can be distilled down to information. I right now it's kind of hard to do because you can't really transfer like if China is really good at manufacturing something well there's this process knowledge that's in the heads of 100 million people involved in the manufacturing sector in China but in the future it might be easier if AIS are doing >> I mean the question about sort of to what extent does our you know fabrication get sort of very uniform and get really commoditized like you know 3D printers have been the next big thing for at least 20 years now um uh you know why do they still not work all that well. Why are they still not actually at the center of of of manufacturing and sort of what comes after that? You know, it is funny to look at say the ribosome by contrast. It really is at the center of biology in a whole lot of really interesting ways. >> Um and and whether or not that's the future of manufacturing is something very simple sort of where you everything goes sort of as as throughput through I don't know maybe it's a bioreactor or something like that. So you send the information and then you grow stuff. um or or you have some 3D printer that actually works um uh and and you know if they're good enough then actually it does become much more a pure information problem and some of this process knowledge becomes much less important. Jane Street has a lot of compute but GPUs are very expensive and so even optimizations that have a relatively small effect on GPU utilization are still extremely valuable. Two of Jane Street's ML engineers, Corwin and Silon, walked through some of their optimization workflows at GDC. >> You're not bottlenecked on the network being too slow. You're bottlenecked on waiting for a different rank in your training, not having completed the work. >> They talked about how Jane Street profiles traces and diagnosis bottlenecks and then how they solve them using techniques like CUDAs and CUDA streams and custom kernels. With these sorts of optimizations, Corin and Savon were able to get their training steps down from 400 milliseconds to 375 milliseconds each. This 25 millisecond difference might sound small, but given the size of Jane Street's fleet, that improvement could free up thousands of B200s. Jane Street open sourced all the relevant code. If you want to check it out, I've linked the GitHub repo and the talk in the description below. And if you find this stuff exciting, Jane Street is hiring researchers and engineers. Go to janstre.com/warkash to learn more. Can I ask a very clumsily phrased question? So there's there's these deep principles that we've discovered a couple of. One is this idea that hey if there's a symmetry across a dimension it corresponds to a conserved quantity. It's a very deep idea. There's another which you've written a lot about written a textbook about in fact about there's way there's ways to understand this thing of what kinds of things you can compute what kinds of physical systems you can understand with other physical systems >> what a universal computer looks like etc >> and is your view that if you go down to this level of idea of not serum or the church touring principle that there's an infinite number of extremely deep deep such principles because I feel like what makes special is that they themselves encompass so many different possible ways the world could be. But no, it has the world has to be compatible with actually a couple of these very deep principles. >> I don't know. I I mean, you know, I just all I have here is speculation uh and sort of instinct. My instinct is we keep >> we keep finding very fundamental new things. It was very I mean for me anyway quite formative to understand as I say you know I gave the example before there's these wonderful ideas of of church and Turing and and these other people ideas about universal programmable devices and then you understand later oh this also contains within it the ideas of public key cryptography and then you understand later oh that also contains within it um the ideas I mean people refer to it as as cryptocurrency or whatever but there's you know very deep set of ideas there about the ability to collectively maintain an agreed upon ledger um which are built which is built upon this and there's probably you know many deep ideas to sort of actually took whatever it's taken many years really to to figure out the right canonical form of of those um and and so just this fact that you you you keep finding what seem like deep new fundamental primitives um uh I I find very for me that's a has been a very important intuition bump and it's across I mean I've given that particular example but I I think you see that same pattern in a lot of different areas. >> What is your interpretation then of this empirical phenomenon where ideas like whatever input you consider into the scientific process or technological process economists have studied this a million and a 100 ways. It just seems to require even at actually a very consistent rate x% more researchers per year. So there's a famous paper from a couple years ago um by Nicholas Bloom and others where they say how many people are working in the semiconductor industry and how does it increased over time through the history of Moors law and I think they find like Moors law means computing increases 40% a year or transistor density increases 40% a year but to keep that going the amount of scientists has increased 9% a year after industry >> and they go through industry after industry with this observation >> and so is your view that there are these deep ideas but they keep being harder to find or that no there's there's another way to think about what's happening with these empirical observations. I mean there so first of all all of their examples are narrow right they all they pick a particular thing and then they look at some uh particular metric um um you know nowhere in that shows up like GPUs don't show up there uh right like in the sense of oh you know all of a sudden you get this ability to parallelize um and that's really interesting um uh uh so so there's sort of a lot of external consequences um that are just delighted from basically you know they have these simple quantitative measures they look at it in agricultural productivity they look at it uh uh in a whole lot of of different ways um but you do have to focus narrowly um and and I suppose you know I'm certainly interested as I say in this this fact that that just new types of progress yeah >> keep becoming possible but um you know there is still I think even there um there does seem to be some phenomenon of of diminishing returns um you know is that intrinsic? Is that something about the structure of the world? Um what is it? Well, one thing which hasn't changed that much is is you know sort of the individual minds uh which are doing this kind of work and you know maybe that those should be sort of being improved as well. Um uh or some sort of you know feedback process going on there um uh you know and and and you maybe that changes the nature of things. I I I suppose I I look at scientific progress up into let's say 1700 something like that and it was very slow and also it was very irregular you know you had the Ionians back sort of five centuries before Christ um doing these quite remarkable things um and then so much knowledge like would would get lost and then it would be rediscovered and then it would be lost again um and you'd have to say that that progress was was very slow and and there It's partially just bound up with the fact that there were some very good ideas that we just didn't have. Even once you've had the ideas, then you need to build institutions uh around them. You actually need to solve a whole lot of different problems about training, about allocation of capital, about all these kinds of things, even just about basic sort of security for researchers. So they're not, you know, worried about the inquisition or or things like that. So there's all these kind of complicated problems. you solve all those complicated problems and then all of a sudden boom there's a massive sort of burst of scientific progress if you're not changing it if there's some kind of stagnation uh there if you're not changing those external sort of circumstances yes like you may start to get uh sort of diminishing returns again but that doesn't mean there's anything intrinsic about the situation uh uh you know maybe maybe something you know just external needs to change again um you know obviously a lot of people think AI is potentially um going to be going to be a driver I It certainly will at some level in fact you know to the extent you can think of a lot of modern scientific instrumentation as really I mean at some level kind of robots uh you know what is the James Web Space Telescope well um you know it's unconventional maybe to describe it as a robot but it's not completely unreasonable either um uh you know it is an example of a highly automated very sophisticated system with electronically mediated uh uh sensors and actuators where machine learning in fact is being used to process the data. Uh so so in that sense we're already starting to sort of see that transition. We've been seeing it for decades. >> Um I I I have this smoke a joint and take a puff thought which um >> I think we've had a few >> Yeah. Yeah. Well, I think we're getting to that part of the conversation and you can help me get my uh foot out of my mouth and figure out a more concrete way to think about it. Um so the to your point that AI there's a natural revolution the enlightenment and now there's AI and each might be a different pace or a different way in which science happens. Um if you think about the pace of how fast such transitions have been happening you can draw over the long span of human history this hyperbolic of the rate of growth is increasing. So yeah, 100,000 years ago you had the stone age. You go back even much further. How long have a primates been around? It would be like let's say millions of years and 100,000 years ago the stone age. Then 10,000 years ago the agricultural revolution. Then 300 300 years ago the industrial revolution. Each marked by this exponent this increase in the rate of exponential growth >> and then people think it's going to happen again with AI but that would happen potentially even faster. It would not have occurred to somebody at the beginning of the industrial revolution that the next demarcation in this trend will be artificial intelligence. Um, >> and so if things are getting faster and it's hard to anticipate what the next transition will be, I guess we just think of this singularity between now and AI and there that's really what distinguishes the past from the future. But just applying the same huristic that maybe people in the past have had. Um maybe the intelligence age is also quite short. >> Mhm. >> And then the next thing after that is we don't even have the ontology to describe what it is. But >> it would not the future will not think of the past as like >> there was pre-intelligent AI and post AAI. >> No, that seems um I mean obviously we can't prove this but it's it certainly seems seems quite plausible. I mean, part of the issue, of course, is is just, you know, the substrate we have available to to to conceive like like seems all wrong. Um, you you can't speculate with a bunch of chimpanzees about what it would be like to have language. Um, uh uh uh you just to sort of pick a major transition in the in in the past. It's the transition itself is the thing, right? Um and uh it seems likely um if we're talking about uh uh taking a puff uh kind of thoughts um you know I'm certainly amused by the idea that uh uh there's going to be some transition involving um artificial general intelligence um using classical computers. Uh but actually there'll be an interesting transition with quantum computers as well. they're probably capable of a sort of a strictly larger uh class of of of potentially interesting computations. So maybe actually the the character of sort of AQGI or whatever it should be called um uh is actually qualitatively different. Um so yeah maybe there sort of a brief a brief period between those two things. Interesting. I mean >> as I say you know this is just it's just speculation but it's certainly amusing. Is there a reason to think that because from what I understand there's been for decades people like you have put pretty tight bounds on the kinds of things quantum computers can do and so it'll speed up search somewhat it will do um and the kinds of things it extremely speeds up like sh's algorithm it seems like it again maybe this is to your point that we can't predict in advance what's down the tech tree but at least from now here it seems like you break encryption but what else are you using sh algorithm >> yeah I mean we've only been thinking about it for 30 years or whatever it's 40 40 or so years. Um not for very long and we sort of haven't in some sense thought that hard uh about it as a civilization. So you know uh uh does it turn out that it's very narrow? Maybe. Um does it turn out that it's very broad? That's also you know like a really radical expansion that seems distinctly possible. Like keep in mind as well we've been doing it without the benefit of having the devices, >> right? Like that's a pretty big bottleneck to have. Uh >> if you're think about computer science in the 1700s and you're like okay do and and or >> what are you going to do? You can't anticipate Bitcoin. You can't anticipate deep learning. >> No, I mean maybe you could if you you know sufficiently bright but uh it is a pretty hard situation, right? >> What is your inside view um having been in and contributing to quantum information quantum computing >> back in the 90s and 2000s? what is your telling of the history and what was the bottleneck? What was the what was the key transition that made it a real field? Um and how how do you rank sort of the contributions for Fineman to Deutage to everybody else who came along? >> Yeah. So I mean I mean let's just focus on sort of the question about sort of what you know what actually changed. So so why was quantum computing not a thing in the 1950s, >> right? Like it could have been. >> Yeah. um uh you know somebody like I don't know John Vonoyman good example absolutely pioneering uh uh computation also wrote a very important book about quantum mechanics and was deeply interested in quantum mechanics like he could have invented quantum computing at that time um and I think there were there were quite a number of people who who potentially could have so why do we have these papers by people like Fineman and Deutsch in the 80s and those are are you I think fairly regarded as the foundation of of the field there some partial anticipations a little bit earlier, but but they were nowhere near as as comprehensive and nowhere near as as deep. Um, and well, you should you should ask David. Um, you can't ask you can't ask Fineman unfortunately, but um, uh, you he'll know much better than I do. Um, a couple of things that I think are interesting. One is that of course computation became far more salient sort of late '7s, early 80s. um you know it just became a thing which many more people were interested in partially for you know for very benile reasons. You could go and buy a PC, you could buy an Apple 2, you could buy a Commodore 64, you could buy all these kinds of things became apparent to people that these were very powerful devices very interesting uh to think about. At the same time in uh the quantum case that was also the time of the ball trap and and the ability to trap single ions and and so on. up to that point we hadn't really had the ability to manipulate single quantum states. So you kind of got these two separate things that just for historically contingent reasons had both u uh sort of matured around sort of let's say 1980 or so. Um, and somebody like Fondon could have had the idea earlier, but it it you know is I think quite an interesting uh uh uh uh uh you know, in fact the story about Richard Fineman. He went and got one of the first PCs around 1980 1981. Um and uh he was apparently just so excited uh with this device, you know, he he he he actually tripped and and hurt himself quite badly. um uh uh sort of carrying his brand new uh uh uh computing device. Um you know that that's a very historically contingent sort of a a coincidence. But but having somebody who's you know very very uh sort of talented and and understanding of of quantum mechanics also just very excited about these new machines. um uh it's not so surprising perhaps that that he's thinking then what similar story could you have told 10 years earlier like there is just no the the conditions don't exist for it so I think that's I mean it's it's quite a benile story but >> oh one of the things we were going to discuss was um this idea you had about the market for follow-ups and I think this is actually the perfect story to discuss it for because >> you wrote the textbook about the field right you Mike and I is the definitive textbook on quantum information. Um, and so you presumably came in after Deutsch, but you identified in the ' 90s somehow identified it as the thing that is worth following up on and building on >> and instead of talking about more abstractly, I I'd love to actually just share the story of like the firstand sort of how did you know that this is a thing to of all the things that were happening physics and computing etc that I want to think about this problem. >> Sure. Sure. So um you know Reed Vineman writes this great paper in 1982. David Deutsch writes a absolutely fantastic paper in 1985. Um sort of sketching out a lot of the fundamental ideas of of quantum computing. Um so I'm you know I'm 11 in 1985. I'm not thinking about this. I'm playing soccer and doing whatever. Um but in 1992 I took a class on on quantum mechanics that was really terrific given by by Jared Milbour. And um I just went and asked Jared uh uh one day after it's like the fifth lecture or something. I said, "Do you like do you can do you have anything uh uh you know sort of papers or whatever that that you could give me?" And he said, "Come back come by my office in a couple of days time." And I I did. And he presented me with a giant stack um of of papers which included the Deutsch paper. It included the Fineman paper and included a whole bunch of other sort of very fundamental papers about about quantum computing uh and quantum information at a time when essentially nobody in the world was working on it. Um uh he was um he'd actually I think he wrote the very first paper that proposed I mean sort of a practical approach to quantum computing wasn't very practical but it was actually in a real in a real system. And so in some sense, you know, I'm benefiting from the taste of this other person. Um, but as soon as I read the papers uh or take a look at the papers, like these are exciting papers, you know, they they're asking very fundamental uh uh questions and you're sort of like, oh, we I can make progress here. Like these are these are things that one could potentially work on. uh Deutsch has this um uh sort of conjecture that basically um yeah there should be or I don't know what the right term for it is thesis or or what you would call it um that um a a universal model quantum cheuring machine uh should be capable of efficiently simulating any system any physical system at all. This is a very provocative uh uh idea. uh I think in that paper he more or less claims that he he's he's proved it. I'm not sure that necessarily everybody would would would would agree with that. There's questions about whether or not you can say simulate quantum field theory um effectively. Um and that that kind of question is is I think very interesting and very exciting. Um uh there it's it's obviously a fundamental question about about the universe. um you know he has some wonderful ideas in there about um uh sort of quantum algorithms and where they come from and what what they mean and what they relate to the meaning of the wave function and questions like this which is you still not it's it's not agreed upon uh amongst amongst physicists. So um yeah there's just some sense of oh I am in contact with something which is a deeply important and b uh we as a civilization don't have this uh and so of course you you start to focus your attention a little bit there. >> I'm not sure I got the answer to the question that >> maybe I misunderstood the question. >> Yeah know let me let me think of how I would phrase it. Maybe I maybe I'll explain the motivation first. So in a previous conversation we were discussing how could you have done in the 1940s the Shannon CRM. >> Yeah. >> And Shannon's way of thinking about communication channel is a deep idea that goes beyond the problems with pulse code modulation that Bell Labs was trying to solve at the time and it applies to everything from quantum mechanics to genetics to computer science obviously. And one of the I think an idea you you stated that um we didn't uh get a chance to talk about yet was this idea well Shannon publishes this paper. There's all these other papers but there's some market of follow-ups where people gravitate to and build upon Shannon's work and how did they realize that that's the thing to do and how does that process happen? Um, and so I guess you you gave your local answer. You read these papers and you immediately realized, okay, there's work to be done here. There's a low hanging fruit. There's some deep provocative idea that I need to better understand and >> I could I could, you know, tractively make progress on. >> Mhm. >> Yeah. I mean, so, you know, to some extent, you're sort of saying, okay, I, you know, wanted to to get into this game of of contributing to humanity's sort of >> Yeah. you know, understanding of of the universe and you are applying sort of this this low hanging fruit algorithm. You're like relative to my particular set of interests and abilities, where should I pick up my shovel and start digging? Um and and there it was like, oh, this this looks like quite a good place to to to start digging. Um um you know, and different people, of course, um you know, chose very differently. It was it was it was a very unusual choice at the at the time. This was 1992. Um, very few people were were thinking about that. >> Yeah. Uh, fast forwarding a bit. So, you've been I don't know how you think about your work on the open science movement now, but did it work? Like what would have what a successful there look like or what what is it what is it that that movement is trying to accomplish? Yeah, I mean the set of ideas about open science. I mean it's interesting you didn't stop and and define open science uh there which uh I think 20 years ago you would have had to do um people recognize the phrase uh people have some set of associations uh with it. Most often they have a relatively simple set of associations. It means maybe something about making scientific papers open access. very often they have some set of notions about maybe it means also making code openly available, maybe it means um making data openly available. Um but already um those are I think very large successes uh of the open science movement um which is to make those salient issues. Those are issues on which people have um uh opinions and then there are there are relatively common arguments an argument like um so this is sort of this is sort of the meme version you know publicly funded science should be open science um uh that's a you know that's a distillation um of a set of ideas uh which you might be able to contest um but if you can get people actually sort of thinking about it and and engaged with that kind of argument >> um yeah that's a very fundamental um uh kind of an issue to be considering in the the whole political economy of science. If you go back say three centuries um there was a a very similar kind of a an argument prosecuted which is the question do we publicly disclose our scientific results or not. So if you look at at people like Galo and and and Kepler and so on um the extent to which they publicly disclosed like it it was done in a very odd uh kind of a way. They sometimes they did bizarre things where they, you know, famously they published some of their results as um anagrams. So basically, you know, they'd find some discovery, they would uh uh write down the result um in sort of a sentence like here's, you know, the the the the discovery of of the the uh I'm trying to think of an example. Um I think the moons of Mars, I think, was one such example. Um uh I'm I'm getting it wrong. was it Hook's law? Anyway, doesn't matter. Um the the point was they they they'd write it down, but then they'd scramble it, publish that, and then if somebody else later made the same discovery, they would unscramble the anagram and say, "Oh, you know, I actually did it first." This is not an ideal way. >> This is not an ideal foundation um for a discovery system. And then it took I mean a very long time uh sort of over a century I think to to uh obtain more or less the modern ideals in which what you do is you disclose the knowledge in the form of a of a paper. There's then an expectation of attribution and so there's a kind of reputation economy which which gets built and so basically oh such and such did this work so they deserve the credit for that and that's then the basis for their careers. So this is sort of the underlying political economy of science. And that made a lot of sense when what you've got is a printing press and the ability to to do scientific journals. Then you transition to this modern situation where in fact you can start to share a lot more. You can start to share your code. You can start to share your data. You can start to share in progress ideas. And but there's no direct credit associated to those. um it's not at all obvious uh uh uh uh sort of you know how much reputation should be associated um uh to them that's all constructed socially um and so making it a live issue um is I think a very important thing to have done and that that's I view anyway as one of the main positive outcomes of of work on on open science I I'll give you a a really practical sort of example to to illustrate the problem um for a long time in physics there was a preprint culture in which people would upload preprints uh to the uh to the preprint archive and in biology this didn't happen. um there was no preprint uh culture that's changing now but but for a long time this was the case and I used to sort of amuse myself by asking physicists and biologists why this was the case and uh what I would hear sometimes from uh biologists uh was they would say well biology is so much more competitive than physics um that we need to protect our priority and so we can't possibly upload uh to the archive we have to we have to just publish in journals And then I would sometimes hear from physicists, physics is so much more competitive than biology that we need to establish our priority by uploading as rapidly as possible to the pre-print archive. We can't possibly wait to do it with the journals. And I think this emphasizes the extent to which this kind of attribution economy is act is just something we construct. It's just something which we do by by sort of agreement. And so uh any attempt to sort of change that economy um results then in a different system by which we construct knowledge and and and so there is sort of this very fundamental set of problems around the political economy of science um uh uh you know sort of we've got this collective project and and how we mediate it depends upon uh uh uh the economy we have around ideas. I one of the sort of things you've emphasized as a as a part of this project of open science is collective science or groups of people work making progress on a problem where no individual understands all the logical and explanatory levels necessary to make a leap or connection outside of mathematics. What is the best example of such a discovery? I mean, I'm not sure I I I have a well orderering of them to to give you a best, but I mean, yeah, an an example that I I think is is very interesting is is the LHC where it's just this immensely complicated object. Um, I actually I years ago I I snuck into an accelerator physics uh conference. I didn't know anything at all about accelerator physics, but I was just kind of curious to see uh what they were talking about. And this particular group of people uh were experts on uh numerical methods in particular on inverse methods. And so it basically turns out you know inside these accelerators you have these cascades. So a particle, you know, will be massively accelerated. Maybe it'll be collided and then you'll get a shower of particles which decays and decays and decays and and there's just this incredible sort of you consequential uh shower which is ultimately what you see at the detector and then you have to retroactively figure out what produced it. Um and so there's these very very complicated sort of inverse problems that that need to be need to be solved. you've got this final data, but you need to figure out what produced it. And that's how you look for sort of signatures of these. And what many of these people were was they were incredibly deep experts on simulation methods for sort of following particle tracks. >> Um, and like this was really deep and difficult stuff. And I'm like, wow. You could spend a lifetime just learning sort of how to do this and how to solve some of these inverse problems and you would know nothing uh uh about or you would know very little about quantum field theory. You would know very little about detector physics. You would know very little about vacuum physics. All these other things that are absolutely or very little about data processing. Very little about all these things that are absolutely essential um to understanding uh uh uh say the Higs boson. Um and I don't think it's possible for one person to understand everything in depth. Lots of people understand broadly a lot of these ideas but they don't understand uh sort of everything in in the depth that is actually utilized. That's why there's these you know papers with with well over a thousand authors. Um and those people can yeah they can talk to one another at a high level but they don't understand each other's specialties in all that much depth. I mean things like as I say you know detective physics, vacuum physics, these kinds of solving of inverse problems like this is stuff is incredibly different from each other. Um and and you know to to understand it in real detail is serious work. Um >> how do you think about prolificness versus depth where I don't know maybe Darwin's an example of somebody who's like justestating on something for many decades. Uh there's other examples where Einstein during the year comes with special relativity is just doing a bunch of different things. Pi talks about how they were all relevant to the eventual buildup. >> Yeah. I mean, you know, it's something I stress about a lot sometimes. I feel like I'm, you know, too slow. Um actually, it's funny that I mean, the Darwin example is really interesting. Like, you know, prolific at what like I mean I God knows how many letters he wrote. It must have been an enormous >> uh number. So you're certainly very active. Um there's also like there's there's sort of there's two types of work that tends to be involved in any kind of creative project. There's routine stuff and there you just want to avoid procrastination. You just want to like you know how do I get good at this or how do I outsource it and how do I do it as rapidly as possible. Um and just avoid you know like getting into a situation where you're prolonging it. Um and then there's high variance stuff where you actually you need to um be willing to to you know take a lot of time. You need to be willing to go to to the different places and talk to the different people where in any given instance most of it's just not it's not going to be an input. Um, and somehow sort of balancing those two things. I think a lot of people are very good at doing one or the other, but it's hard to, you know, it's almost like a personality trait sort of, you know, which one you prefer and and people tend to end up doing a lot of a lot of one and and not enough of of the other. Um, so I certainly, you know, sort of try and balance those two things. I mean, Einstein is such an interesting example. I mean 1905 is just this extraordinary year. Like you can delete special relativity entirely and it's an extraordinary year. You can delete special relativity and you can delete um the photoelectric effect for which he won the Nobel Prize and it's still an extraordinary year like plausibly a multi Nobel Prize winning year. Um uh so what's he doing? Um you know I mean maybe the answer is just he's smarter than the rest of us. Um uh and and there's a lot of luck as well. Um uh but but but but you know I certainly for myself anyway like trying to identify those things that are routine that I should get good at um and then you know just just try and do as quickly as possible. I think that that's yielded a certain amount of returns but also being willing to bet a little bit more on myself uh on sort of the variance side uh has also been very very very helpful. Um that's really hard. Um like cuz you intrinsically you're putting yourself in situations where you don't know what the outcome is going to be. Um and so if you're very driven to be productive and whatever um and actually mostly it's not working uh over there you're like let's reduce this like it doesn't feel right. Um when I worked in San Francisco uh actually a practice I used to have each day um was instead of taking the 15minute walk to work I would take the the more beautiful 30-minute work walk to work partially just because it was beautiful but partially also um as just a reminder to like like there are real benefits to not being efficient. Um but it's not an answer to your question. I mean really I think all I'm saying is I struggle a lot with the question. I mean there are these um Dean Keith Simington I forgot his exact name. >> Yeah. Yeah, I know what you mean. >> Um has this famous equal odds rule where he says the probability that any given thing you release any paper, book whatever will be extremely important for a given person through their lifetime is not that different. And what really determines >> uh in what era they are the most productive is how much they're publishing. any given thing has equal odds of um being extremely important. Um if you just think of some of the most successful creatives or scientists, they're just doing a lot like Shakespeare is just publishing a lot. >> Um and of course then there's counter examples, you know, Girdle publishing almost nothing. >> Yeah. >> Um but you know, broadly speaking, you know, I think some like you need a very good reason to be avoiding it. There's to to to to basically to to not do that. Um it's funny. I mean, I've talked to I've met a lot of people over the years who you talk to, they're clearly brilliant and they're just obsessed that they are going to work on the great project that, you know, makes them famous and they never do anything. >> Um, and that seems connected like it's a type of aversiveness. I think very often they just don't want public judgment. >> Some something that I would love to see. You know, there's an awful lot of of biographies and memoirs and histories of um people who achieve a lot. I I I wish there was like a very large number of of biographies of people who are fantastically talented. >> That's a great >> who, you know, just missed >> like like you know uh absolutely you I've known you know people who won gold medals at at IMOS and things like that who then you know tried to become mathematicians and failed. >> Um like what what happened? What was the reason? And I suspect in many cases that's actually, you know, more informative, >> incredibly interesting >> than anything else. >> Uh, you have this essay that I, um, I was reading before this interview about how you think about what is the work you're doing. Um, and writer doesn't seem like, as you say, was Charles Darin a writer, right? What what exactly is that label? I'm a podcaster, right? time and in a way obviously our work is very different but I I I also think a lot about what is this work and how do I get better at it >> and in particular how I can make sure there's some compounding between the different people I talk to on the podcast >> where I worry that instead of this kind of compounding there's actually I build up some understanding that's somewhat superficial about a topic and then it depreciates and I move on to the next topic and it sort of depreciates Um, and so I think there's this question. There's a lot of podcasters in the world who will interview way more experts than I ever have. And I don't think they're much the wiser or more knowledgeable as a result. So there's it's clearly possible to mess this up. >> And I wonder if you have thoughts or takes or advice on how one actually learns in a deeper way from this kind of work. Yeah, I mean sort of an incredibly complicated and rich question. Um, >> I mean it does seem like sort of the question is like you know how do you make it a higher growth context? How do you make it a more demanding uh context and sort of you know you can do that in like relatively small ways but that might however yield compounding returns or you can do something um that is maybe more radical. Maybe it means actually you know starting sort of a parallel project in which you do uh something that is actually quite a bit different. There is something I think really interesting about like how being very demanding can simply change your your response to to something. something that that I would sometimes do with with students and sometimes with myself was really aimed more at myself was you they would say some week oh you know I'm going to try and do you know this work over the coming week and then the next week would come by and they you know they hadn't solved the problem or whatever you sort of like you know if a million dollars had been at stake like would you have put the same effort in and the answer is no um sort of invariably um like they've tried but they haven't really tried. >> Um, and I think that's a very familiar feeling for all of us. You know, you you sort of you you you often you you you could do a lot more if you had just the right sort of demanding taskmaster uh standing by you and saying, "Look, you you you're barely operating here." Um, and so I I do sort of wonder a little bit about like, you know, what's the what's the demanding taskmaster? What can they ask you that is going to make your preparation way more intense? The most helpful thing honestly is for some subjects it is very clear how I prep. Like I'm doing an upcoming episode on chip design with the founder of a company that has chip design and he wrote a textbook on chip design and he yesterday I went over to his office and we brainstormed five sort of roof line analysis I can do and if I understand that >> I I have some good understanding. The problem is with almost every other field there's not this cor there's not like you I don't know when I interviewed Ilia three four years ago it's like implement the transformer and if you implement it like you have some nugget of understanding you've clamped down and with other fields it's just like I vaguely understand this it's not clamped I vaguely understand this I vaguely elemed about this IM about this but there's no forcing function that you do this exercise and if you do it you will understand >> yeah so I mean really what you're sort of saying is you can do a good job at at at podcasting without actually attaining this kind of and that's the problem from your point of view. You you want to sort of change your job description so that you you are internalizing these chunks and just getting this kind of integration each time. Um and it seems to me like you you know what that means is you actually want to change the structure of the like like like the work output at some level. M um uh I mean lots of people think there's this terrible idea um people have that that they should be in flow all of the time. >> Um uh and of course as far as I can tell like high performers just don't believe this at all. um they're in flow some of the time like you certainly see this with athletes you know when they're actually out there you know playing basketball or tennis or whatever uh ideally you know they are in flow much of the time but when they're training they're not um they're stuck a lot of the time or they're doing things badly um and I suppose I wonder what that looks like for you that I would be extremely satisfied with the problem is I just like I don't know what the equivalent of do the 64 lapses for almost and so this is sort of this is a you can change by choosing guests where there is allegible curriculum and so maybe it's a mistake for not having done that or also like there's no real way to prep for Terrence Tao or something and like um there's no curriculum that's like a plausible one I think um there's one failure mode so there's many failure modes but one is um if you you could do one dynamic I'm worried about a long-term dynamic is that you do good you can have a good podcast and there's a local maximum but um you for no particular guest or topic, are you going deep enough that you I think my model of learning is there's if you don't really understand the deeper mechanism, you're just mapping inputs and outputs of a black box. >> Yeah. >> And that just fades incredibly fast or is not worth it in the first place and you kind of just move on and it's over. >> Um and you kind of need to build the intermediate >> connection. Um and it's it's unclear. I think actually AI in a weird way is really easy for that reason because there is a clear thing you can do just implement it right and then you understand it we're almost if I applied that criteria elsewhere what am I how do I just not do history episodes >> ad exactly a palmer like what what you know wonderful to talk to incredibly interesting but for you personally like what changed >> right >> yeah there's some things I learned I think I could have done it if I had maybe allocated more time especially after the interview to like less write up 2,000 words on everything I learned and how it connects to other things I know and something. >> Um, and maybe that's thing worth doing is spreading out the episodes more and spending more time afterwards consolidating. >> Um, >> but yeah, I think the I would pay basically infinite amounts of money if there was somebody who is really good at coming up with here is here's the curriculum and here's the practice problems you need to do and here's the exercises you need to do after the interview to clamp what you have learned. >> Have you tried doing that with somebody? >> It's hard to find. I mean I maybe I haven't tried super hard but um it seems actually it seems tough to find somebody who could do that for every single kind of discipline. Maybe I should just hire different ones for different topics maybe. or there's something about like I mean what problem you know are you solving sort of for each episode and I mean as far as I can tell like that's the only way I really understand anything is that you know I I get interested in something at first I don't even have a problem but there's just some sense of there's some contribution to make here and gradually you home in >> there's a problem and then you I mean funnily enough I mean spending time stuck is incredibly important um and and I sort of you know that used to just be annoying now it seems Oh, this is actually um uh uh maybe even the most important part of the whole process. Um but that very hard oneness of it means that you know I internalize it afterwards. I often find actually if I you know I've written sometimes 10,000word essays in you know a couple of days and I've written them in you know 3 months or 6 months. Uh I feel like I I I didn't learn very much from the ones that that that only took a couple of days. Uh whereas I you know some of the ones that that that took 3 months I'll be you know 15 years later I'll I'll I'll still remember. >> Yeah. C can you describe outside of um physics how you learn of the ones that took three months? >> I mean by far the most you know you know the the the common things there's always some creative artifact. Sometimes it's a class uh uh uh you sometimes it's engagement with a group of people who um you know there's some collective creative artifact that you're you're you're working on together. I mean you might not even be aware of it but you you acting as an input to their creative ends >> um in some way um and sometimes it's just you know it's an essay or a book or or or whatever. Um yeah it's one of the reasons why uh I you often quite enjoy doing podcasts. So I mean particularly I mean I you know I I I said yes to come here partially because I know you ask unusually demanding questions. Um and so it's sort of that's an attempt to to to get this sort of perspective from a different it's a different kind of forcing function. Um so yeah trying to pick sort of the most demanding creative context. >> Yeah. So for this interview I went through like three lectures of the suskin session book. The problem is that there's almost no practice problems in it. And so I hired um a physicist friend who's going to like I haven't done it yet, but it's like every lecture I want like a bunch of practice problems, go through them. And I'm I'm planning on being um appropriately humbled. >> How do you how do you make it as jugular as possible, right? Like >> the higher you can raise the stakes, the better. >> I mean, the interview is in some sense high stakes, but also it doesn't necessarily test deep understanding. >> Yeah. But I don't think the interview is that high stakes, right? You're not writing a book about special relativity and you're not trying to write a book that replaces the current, you know, whatever the the existing standard textbook is like that. That's a really high >> really high thing. What do I a phrase that I sort of find particularly difficult and and um it's it's a funny one. People will talk about going deep on a subject and it turns out you different people have different ideas of what this means. Some people means they read a couple of blog posts. Some people it means they read a book about it. Some people it means they wrote a book about it. Um and and and and I think like sort of what what what what your standard is that sort of the standard you hold yourself to um determines a lot about you know your ability to to integrate knowledge in this way. I don't know what your experience has been but I found that I'm getting I'm in some sense able to move much faster on some things through the help of AI but I don't know if I'm like learning better. >> Yeah. And I think it's probably because the hardest thing, the thing that is most demanding is so aversive that you try to take any excuse you can to get out of it. >> And just having back and forth conversation where where you gloss over >> it's entertaining but not necessarily anything else. >> Yeah. So it's such an easy way to get out of the thing. >> Yeah. >> Um in fact it makes it easier because instead of doing some intermediate thinking you there's always a next question you can ask a chatbot. >> Yeah. And and and it's somewhat valuable. like it's not I mean that's part of the seductiveness of course like like it's not actually it's not actually useless. >> Um but um but yeah it can sort of substitute for for actually doing the thing that that maybe you should be doing. Um it's interesting that like the the the extent to which you know to what extent should you be outsourcing that kind of stuff uh and to what extent you know like like it's it's really there's some sort of interesting judgment call about >> uh uh you know you actually there is a whole bunch of routine work that that you want done. Um and in fact it's it's low value for you. So you may as well get uh if you can get a chatbot to do it, you may as well. So uh somebody interviewed um the pioneering computer scientist Alan K years ago and he was asked what he thought about um basically Linux and if I remember his answer correctly basically said look you know it doesn't have anything to do with computer science. It's just a great big ball of mud. Um there's a few interesting ideas in there which are which are worth understanding but mostly you're all you're learning is stuff about Linux like like you're not actually learning anything which is transferable. I thought there was like a very like that there's a certain kind of seductiveness >> uh to some things where you know it's sort of a Rub Goldberg machine. You can just sort of learn about all the bits and it feels kind of entertaining. Um but if you step back and think about the question, you know, what am I actually doing here? Um it might not actually be meeting your objectives. Maybe you want to become, you know, a CIS admin and learning Linux is a great use of your time. There's no no harm in that at all. But if if your answer is if you if your objective is to understand the fundamentals of computing, uh it's much less much less clear that that's a good use of your time. I thought that was it was certainly an answer I've I've thought a lot about where you >> you actually need to you that for a certain type of mind there is a seductiveness in in just just learning systems and confusing that with with uh with understanding. >> Yeah. Okay. I'll keep you updated on how discuss I I owe you a text within a month of um some revamped learning system. >> Yeah. I'll be really curious if you I mean it's also true, right? tiny incremental improvements in this. I mean, they're just worth so much. >> I know. Yeah. It's sort of the main input into the podcast. You know, it's great that the bookshelves are fancy and I've got a Blackboard or whatever, but really like the thing that makes the podcast better is if I can improve the learning I do. >> So, it's um yes, it's worth every morsel of improvement. >> Yeah. >> Um All right. Thanks for the thanks for the therapy session. >> Great done. Um there's Michael. All right. Thanks,




