| Episode | Status |
|---|---|
| Episode | Status |
|---|---|
All but the last 20 minutes of this episode should be comprehensible to non-physicists.Steve explains where frontier AI models are in understanding frontier theoretical physics. The best analogy is to...
You have a brilliant but unreliable genius colleague. Right? So you go down the hall to your colleague's office, and he's some kind of strange brain. His brain is clearly not like yours, but he has an encyclopedic mastery of all the literature, and he can do lightning calculations at this Blackboard. So every now and then you go and talk to him.
You say, hey. I had this idea. Do you have any thoughts on this? Has anybody already done this? And he just leaps to the board and writes out, like, you know, 100 lines of response to your query.
That's kind of what it's like trying to do theoretical physics with these LLMs. But the key part of the analogy is this human this genius whose office you go to with these requests is capable of deep insights, but also capable of very simple and also very profound mistakes. Right? So you can't just assume, oh, if this genius thinks there's some good analogy between a and b, It could be fruitful to think of a in terms of b or reformulate a in terms of b. With a human when a human genius says that to you, I think you could have a higher degree of of confidence, that suggestion, that analogy or analogy of analogies has something there.
But with the models, it's not always true. Okay? So so these are these intelligences are not like human intelligences. They don't understand physics quite the way we understand physics, but my take home message is they can be used in a very fruitful way. Welcome to Manifold.
This is going to be a solo episode where I talk about my recent work applying AI to theoretical physics. That work culminated in this research paper, which was published in physics letters b. It's on a fairly technical subject, so I won't really go into the physics until the end of this episode. For those of you listening on audio, I will try to make the whole episode as accessible as possible. But this particular episode is probably best to watch on video so you can see some of the slides or some of the excerpts from the papers that I'm gonna discuss.
In addition to this physics paper that's on the screen right now, I also wrote a companion paper which describes how the research was conducted in collaboration with AI. And let me just switch windows to show you that paper. That paper is called theoretical physics with generative AI. And, so, again, for audio listeners, I'm gonna try to make this as accessible as possible. I think at least the first half hour will be perfectly fine for you.
Later on when I get into the actual physics and the physics equations, Again, probably better to have the ability to look at what's on the screen. But for this per first part of the episode, which is gonna be mostly narrative, I think you'll be fine as an audio only listener. So for the last, easily, I think, six months, I've been testing all the major frontier models to see how good they are at physics and math. And, you may remember that I did an episode just a few episodes back with a UCLA CS professor who had written a paper using a particular kind of generator verifier architecture, which I'll discuss in more detail, which using that architecture, he was able to use off the shelf frontier models and get gold medal level performance on international math Olympiad problems. So that's an a previous episode that I'll link to in the show notes.
I've been conducting that similar kind of research, but more directly in just trying to figure out ways to use the models productively for physics and math, and also trying to trying to get a good feeling of their understanding of frontier research topics in, in mostly in physics, but a little bit in math as well. And as part of that work, I actually started collaborating with DeepMind. DeepMind there are some DeepMind researchers who are acknowledged in this paper that's on the screen right now who have built something called coscientist, and they've actually written a paper on coscientist. And coscientist is also a sort of souped up version of the frontier models that gen the general public has access to, but it's been improved in a certain way. It has a large token budget, and it is meant to be used by research scientists as a collaborator.
It is meant to both perform calculations, literature searches, etcetera, but also even produce new ideas for science research. So this whole field is very, very active right now. You may have also read a blog post by Terence Tau where he talks about the use of AI in his research. I think there is now more than there are now more than one Erdos problems that have been open for a long time that have been solved by the models or by scientists in collaboration with models. The paper that I wrote is, I believe, the first published theoretical physics article in which the main idea from the article actually originated from an AI, in this case, from GPT five.
That's something that I'll discuss in detail. So let me now go through what I say in this paper. The title of this paper is theoretical physics with generative AI. As I said, it's it's mainly about the current situation, how to productively use generative AI in theoretical physics research. And as the sort of specific example that I dis that I explore in the paper, it's the quantum field theory paper that I showed on the screen when this episode first started.
It's a topic of the paper. Tomanaga Schwinger Integrability Conditions Applied to State Dependent Modifications of Quantum Mechanics. Now that that's quite an esoteric topic. Only a small number of theoretical physicists really, I think, would be familiar with that terminology, but I will go over that. I have some slides to just to kind of briefly explain the paper in case some of you are some of you viewers slash listeners are physicists.
I have some slides where I go through the main results of the paper, but I'll do that at the end of this episode. So people who are not experts on that topic can just drop off. Before that happens, what I'm gonna discuss right now is I think of of broader interest. And you can see on the screen here, I say, remarkably, the main idea in the paper originated de novo from GPT five. I also used models like GPT five, Gemini, and Quinn Max extensively in the research.
I found those three to be extremely good at math and theoretical physics, and I use them in a kind of generator verifier pipeline or architecture, which I'll explain. K? Now Frontier LLMs have been trained extensively on textbook materials in physics and math, but also all the research articles on archive and probably in most of the main journals have been used and have been subjected to forward and backward next token prediction, I would suspect, by all of the major labs and all of the models, you know, at the frontier are actually pretty good at physics. So for example, undergraduate and graduate level courses in physics, in those courses, students can readily get the AIs to solve the homework problems that they're set. So we were in a situation where we we have to figure out how to deal with students learning physics, physics majors, because what used to be a very good exercise for them learn and the learning exercise for them, which is to be assigned some problems and then have to work out those problems.
In the old days, it was it was already possible to get solution sets. Like, there were lots of Chinese produced solution sets that you could find online, and they could help you solve the problems. But now the AIs can just solve them in in a few seconds or a few minutes. And so, that's had a huge impact on how we organize our courses at the university. A little bit harder than solving well known homework style problems that you would find in, at the back of the chapter in a textbook is actually trying to understand frontier research papers and and trying to productively help researchers push the frontier forward.
And, mostly, what I'm gonna talk about is that. But given that the models can solve problems that would appear, say, in a graduate level textbook, that suggests that they might be useful to researchers, in in trying to actually push the frontier for. Now the main problem that the models have, which I'm sure you're familiar with, is mistakes, which sometimes are called hallucinations. Sometimes another term for it is confabulations. And I've talked about that in previous Manifold episodes where AI is the subject.
In the context of physics research, the real problem is the models can both make simple mistakes. Like, they just make some simple calculation mistakes, although that's becoming rarer and rarer. The more troublesome mistake they make is they may propose some idea, for example, that a technique from some distant area of physics of theoretical physics can be applied to the specific problem or research question that you're interested in, that you're asking them about. And unlike humans, the models have read everything. So they can make what sound like plausible suggestions that, oh, this distant technique b could be applied to your problem a, and here's some reasoning for how that how that would work out, or here's a sketch of how that would work out.
And the problem is unless you are expert at both a and b, usually, person involved in the situation is an expert in a, but not in b, because humans are very finite in our in our, deep knowledge. If you know a, but you don't know b, that suggestion can sound extremely exciting, and then you can waste many, many hours trying to figure out whether that suggestion is actually gonna be fruitful, whether the the things the model is telling you in detail are correct. In my case, for this particular paper that I'm talking about, I actually had to go away and learn a bunch of stuff about b. In this case, b was something called axiomatic quantum field theory, which is not my area of research. But I had to go and learn a bunch of stuff about axiomatic quantum field theory to figure out that what the model suggested to me turned out not to be a very fruitful direction, turned out to be actually kind of completely wrong.
But that can waste a huge amount of researcher time. And if the AIs do tend to lead you on a kind of wild goose chase every now and then, that makes the ROI or the cost benefit analysis of using the models much worse because, okay, they help you most of the time, but then they waste ten hours of your time. Occasionally, very bad. Right? So this generate verify protocol that I came up with and that these the UCLA team that I one of whom I interviewed on a previous podcast, what they came up with as well is to have different instances of the model or even instances of different models playing the role of AI which generates a response to the research question, but then others which play the role of verifiers that are trying to find problems in the earlier response of the other model.
And by cycling through, generate, verify, generate, verify that, or generate, verify, modify, verify that kind of process, I've found that you can suppress to a large degree many of the problems, both simple errors and then also deep conceptual confabulations. You can suppress those to quite an extent. And I'll I'll go deeper into some examples, as we go on in this talk. Now I give an example in the paper of a verifier step, and this is the simplest possible thing. You you just have some output from model one, and you show it to model two, but model two has been prompted with the following.
You are a world class theoretical physicist. Check the following for errors, and then you you put in the the output from model one. Review each equation and each reasoning step. Identify all problems and summarize your findings. So very simple prompt.
You could actually make more elaborate ones. The UCLA prompts that were used to solve IMO problems are longer and more detailed. They're a few pages long, I think. This one is very, very short. But even this one will generate a good analysis by model two of what had been proposed by model one.
And if you cycle through this, it really is quite helpful. K? As I mentioned already, the main models I used were GPT five, Gemini, Quinmax. I also found that Deepsea, Grok, and some others were quite strong. Kim e k two were quite strong in their physics capabilities.
I think to some extent, just chaining together different models in this generate verify protocol is enough to take any of these top models and make them into a very useful research tool for scientists. K. So the the the broader take home message, whether you're a physicist or a mathematician or an engineer or even biologist, is that each individual model is smart but can make mistakes. And chaining these things together suppresses the probability of the mistakes. K?
I just wanna add that I I characterized the at least for physics, the the process of doing research with an LLM as analogous to the following. You have a brilliant but unreliable genius colleague. Right? So you go down the hall to your colleague's office, and he's some kind of strange brain. His brain is clearly not like yours, but he has an encyclopedic mastery of all the literature, and he can do lightning calculations at this blackboard.
So every now and then you go and talk to him. You say, hey. I had this idea. Do you have any thoughts on this? Has anybody already done this?
And he just leaps to the board and writes out, like, you know, 100 lines of response to your query. That's kind of what it's like trying to do theoretical physics with these LLMs. But the key part of the analogy is this human this genius whose office you go to with these requests is capable of deep insights, but also capable of very simple and also very profound mistakes. Right? So you can't just assume, oh, if this genius thinks there's some good analogy between a and b, it could be fruitful to think of a in terms of b or reformulate a in terms of b.
With a human when a human genius says that to you, I think you could have a higher degree of of confidence, that suggestion, that analogy or analogy of analogies has something there. But with the models, it's not always true. Okay? So so these are these intelligences are not like human intelligences. They don't understand physics quite the way we understand physics, but my take home message is they can be used in a very fruitful way.
I have to say a little bit about the research context that I was working in because I'm gonna go through the example of the the actual research project that led to this published paper and for which remarkably the main idea of the paper came from GPT five. I was asking GPT five. I was testing the capabilities of GPT five and some of the other models by doing the following. I uploaded or pointed the model at one of my own papers where I know all the details or, you know, maybe with the passing of time, I forgot some of the details, but I I still have some kind of deep knowledge in that area. I upload that to the model, and then I just start asking the model questions to to see how well it understands understands the paper.
Can it summarize the conclusions? Can it actually, as a useful thing, I can say, oh, what, you know, what follow-up work has been done in this area that I should be aware of? And it it'll immediately return, like, some interesting results from some papers that cited my paper, things that. So models are fully capable of doing that. I mean, again, slight error rate because sometimes it will confabulate some new result, but which came from some paper that doesn't exist that extends my earlier work.
It it could do that, but most of the time when it does that, it's actually true what it's saying, but some of the time, it's not true. Again, running it through a generator verifier pipeline suppresses that. And so if I had an automated generator verifier pipeline, which I'll talk about that maybe later in the episode, the building of such an automated thing that would make this whole process just that much easier rather than having to do this kind of thing manually. Okay. So the the old paper that I asked it about had to do with something called nonlinear modifications to the Schrodinger equation.
So the Schrodinger equation in quantum mechanics and quantum field theory tells you how the state of the universe evolves in time or the state of some group of particles or something or quantum fields evolves in time. And, interestingly, the Schrodinger equation is completely linear. Okay? So the the the state itself only appears to the first power in some sense on both sides of the equation. That's equation one on the screen if you can see this.
And so the a consequence of that is if you have two different solutions to the Schrodinger evolution equation, the sum or the difference of those two solutions is also a solution. K. That's what linearity means. And that's quite interesting. That's deeply related to things like quantum superposition and the the persistence of quantum superpositions in nature.
So very deep property of quantum mechanics linearity. I think scientists who are, you know, who aren't just slavishly absorbing ideas from textbooks have a right to ask, gee, is this linearity property exact? Could there could there really be something in nature which is exactly linear, or is it just some kind of useful approximation that eventually we'll discover corrections to the Schrodinger equation, modifications to the Schrodinger equation, which maybe are small. There's they have a coefficient which is small, but include nonlinearities. K?
It's a super important fundamental question. In physics, we don't know the answer. This was studied by Steve Weinberg in the nineteen eighties. He's one of the great physicists of the sort of second half of the twentieth century, Nobel Prize winner, one of the inventors of the, quote, standard model particle physics. And he, for some time, was studying, gee, is there any way to modify quantum mechanics in a meaningful way and test those modifications?
And he was thinking about nonlinear corrections to quantum evolution. And my old paper on this, which I guess I'll just flash it on the screen briefly, is this paper from 2015, I guess, as the the most I guess that was the last version that we uploaded. I think it published in late twenty fourteen. The title is locality and nonlinear quantum mechanics. And I was asking the model to see how well it understood this paper because, actually, most of my colleagues wouldn't actually understand this paper because it's a little bit out of the box what we're doing here.
We're allowing quantum mechanics to have nonlinear corrections, and we're showing that those nonlinear corrections allow distant objects to instantaneously become entangled. So it it allows entanglement, which is, in a sense, faster than light, which is highly problematic. Entanglement can build up in a faster than light way we show. And, this is related to what was considered the main problem with Weinberg's proposals or the proposal of adding nonlinearity to the Schrodinger revolution is that it seems to violate all kinds of properties like special relativity. It also allows this sounds kind of fanciful.
It's true. It allows people who are in different Everett branches of the multiverse to signal to each other, to send messages to each other. So there's all kinds of wacky stuff that happens if you allow this nonlinearity. Some physicists would just say, don't think about nonlinearity. It's so obvious that that can't be part of nature the way nature works, and that quantum mechanics is just exactly linear, which which could be true.
I mean, I I suppose I would bet in that direction, but it's an amazing conclusion, and it deserves to be tested. Right? You you need to test this conclusion. Okay. So here in the paper on the screen, I have a even though this companion paper, it's meant to be readable by AI researchers, I do have a little bit about the physics in here.
So I I sort of review little bit about the structure of quantum mechanics and then this linear versus nonlinear question that I've just described. Here on the screen, you can see the prompt that I gave to GPT five. I was testing a variety of models on a variety of my old papers. I've been doing that for quite some time just to see how well they understand them. And sometimes they understand them better than my colleagues, just to be totally honest.
Okay. Now what's the prompt here? The prompt is compare and contrast studies of nonlinearity in nonrelativistic quantum mechanics. That's the old work by Weinberg and some famous follow ups by two physicists called Giesen and Polchinsky, compare and contrast studies of nonlinearity and nonrelativistic quantum mechanics with treatments that incorporate what is known from quantum field theory, that the nonlinearity must involve the entire wave functional, which describes a potentially infinite number of degrees number of space like separated modes. Okay.
Actually, I would probably write that a little bit differently now, but that is literally what I wrote to the model on that fateful day. Okay? And, so that is what it is. I just included in the paper here for historical reasons. I just wrote that prompt in the paper because the paper that I showed on the screen that I wrote ten years ago is about extension of this nonladiarity analysis to quantum field theory, which is different from just the quantum mechanics of some few degrees of freedom.
It's the quantum mechanics of entire fields that occupy all of space time. And so there's a qualitative difference in in in what's going on. K? So the model responded by with I was a little bit surprised. The model responded with a beautiful, very technically correct, and also conceptually correct summary of our paper, which, again, like, we went through multiple rounds with referees who clearly didn't understand what we're doing.
But but but, ultimately, the paper was published. But the models understood it right away, and then the model at the end of its response, summary response, trying to be helpful. I think GPT is reinforcement learning trained to be to try to be helpful and make helpful suggestions. It said, if you want, I can sketch the Tomanaga Schwinger version of this evolution by space time space like hypersurfaces to show explicitly why a hypersurface local generator that depends nonlinearly on the global state cannot remain foliation independent without collapsing back to linear dynamics. Okay.
So it's offering to show me a calculation. It's doing this in a tech technique framework or, yeah, technique framework, Toman Agassinger formulation of quantum field theory, which is a little bit obscure, And it's sort of giving the answer that, oh, there's a problem with something called foliation independence if you include non if you allow nonlinear terms. Okay? So I was quite intrigued by that. Of course, it could have been one of these crazy hallucinations, but I was willing to go further and explore it.
And so my next response to my prompt next prompt in response to its response was, yes. Sketch the Tomanager Tomanager Schwinger version of this. So now if you're again, if you can see my screen, you'll see there's a long response. It's act actually, this response is much longer than even what I excerpt in the paper here. But the model within a few seconds gives it summarizes the Tomanager Schwinger formulation of field theory.
It introduces something called foliation independence of the evolution or something also called the integrate integrability condition, which allows you to take the state of the wave function on one slice, time slice, and evolve it into the future, to other time slices. And then there's a certain condition that you want to be true. You want small perturbations to the the way you slice space time to not matter. In other words, if I if I slice these are just coordinate systems, arbitrary coordinate systems that I can use to describe the what's going on in the universe to to to describe the location and space and time of events. If I just modify slightly the coordinate system I use, it shouldn't affect the physics.
And so that that's foliation dependence. And it then derives some nice conditions for foliation independence in the presence of a nonlinear correction to Schrodinger evolution. K? And if, again, if you're watching my screen, you can see me. I'm scrolling.
These equations are getting a little bit long. If you're not a mathematician or theoretical physicist, these look like probably pretty formidable equations. I actually stop once it's introduced the key equation. It's it it actually did a bunch of other stuff as well, but I didn't put that in this companion paper. And what's remarkable here is that this equation, the start equation that I have on the screen here turns out to be correct.
The further analysis of this equation is quite fascinating or at least it was to me. It led to an interesting paper, which was published has now been published in physics letters b. And I claim I don't really claim any credit for this this particular, equation. The the model just generated this equation. K?
So I thought that was kind of interesting. I thought that was worth first pursuing to see if the equation and the idea here is productive and makes sense. It does. Whether it's publishable in a journal, it was. And then I wrote the companion paper just to give this as an example of the model coming up with some innovative idea.
I mean, if a colleague came into my office, if a grad student or postdoc or colleague came into my office and say, hey, Steve. I read this old paper you wrote ten years ago, and it occurs to me we could apply the Tomanaga Schringer formalism to this blah blah blah, and I get this equation. And then he my colleague, she and I or he and I worked for the next month to, like, extend to to further explore the consequences and then wrote a paper. That's more or less what happened, except it happened much faster, and the colleague here was an AI. Was not my flesh and blood human colleague from down the hall.
K? So that's what happened. Some people are mad about this. People who don't like OpenAI are mad about this. I I don't know.
GPT five is OpenAI's model. Some people who think AIs can only generate slop are mad about this, but I'm just being completely transparent about what happened. I didn't set out to write an AI generated paper. I actually was just testing the models to see whether they could understand my earlier work, but the model then generated something interesting, which I started exploring, and it was worth writing up as a paper. K?
So, by the way, a sociological comment, The reactions to this work have been really strange and all over the map. So some people are excited by it, like some of the, like, pro AI, EAC type people retweeted it. I think I think the tweet that I made a week ago about this work, after the paper had been accepted at the journal and all this other stuff, I I tweeted out something about it. I think that was that post on x was viewed by, like, half a million people worldwide. So so a lot of people were interested in it.
But I also got, like, angry attacks over over this. Like, how dare you publish AI slop? Updated journal. Okay. Whatever.
I mean, the peep some of the people who were saying it was AI slop can't even understand understand the paper, let alone my old paper from ten years ago. But, hey, they're they're qualified to say it's AI slop. I don't know. Okay. In this section, I talk a little bit more about false directions that can come from the model, this mode of confabulation, deep confabulation, I would say, that is probably the most time costly for expert scientists, which is that the model proposes something which superficially looks very plausible, and it's meant to be plausible.
Remember, these models are trained to generate plausible text. That's all in a sense in a sense, all they are. They generate plausible text, which looks like the 15,000,000,000,000 or 30,000,000,000,000 tokens that they've been trained on, right, which includes all physics papers, right, and all math papers. Right? So there was a case of this in in in following up on this Tomahawk Schrodinger stuff.
One of the things the model suggested was, oh, we can use these results, something called the Rieschleider theorem or the split property. We can use these profitably in further analyzing, the Tomanaga Schwinger integrability conditions. And because I'm not an expert in that area, I had to go away and think about this for a while, and eventually, I realized what the models were saying to me was, I think, pretty pretty much incorrect. And the problem being that some of the assumptions from axiomatic quantum field theory are themselves probably invalidated by the state dependent modification. So one one can't really prove properties of state dependent quantum mechanics from those standard axioms of quantum field theory if in fact those are invalidated by the modification of the of the theory of the dynamics of the theory.
Okay. So so this was probably the most time costly wild goose chase that the models led me on. I will say that the generator verifier procedure that I described earlier did actually suggest that this was a wild goose chase because I could not get although as if if when I fed these axiomatic quantum field theory outputs into other verifier instances of the same or different models, I generally I didn't I did not find convergence. So one model would say, yeah. This is fruitful.
The other model would say, no. This is raw, you know, or in a more detailed way. But it would sort of oscillate like that and never converged. So one rule of thumb, I think, is it if you're very lazy and you're not gonna go and learn this other area b, act in this case, axiomatic field theory, if you're not gonna go and learn it to the point where you can judge yourself whether what the model is offering you is good or bad or has value, You could just cycle through a diverse set of models and just look to see whether there's convergence. And if there's not convergence, it doesn't automatically mean the proposal's wrong, but I think it there's a good chance that proposal's wrong.
Conversely, in most things, when you put them through the generator generator or for many things, you put them through the generator verifier pipeline. There is convergence where you get, like, at the end, you get multiple verifications, many models agreeing this is is probably actually okay. Now even that is not a guarantee that what the model the the result that the model is telling you is correct. It still could be wrong, but I think probabilistically, these are one, lack of convergence is a good kind of a strong negative signal, and convergence is a some at least somewhat strong positive signal of correctness. And I'm not saying you should use the models to be the final judge of any of this, but using them in this chained way does increase the probability that you're getting good stuff out of them and not wasting your time.
So it just it it I think it makes the value proposition strongly positive for expert researchers. Now if you're not an expert researcher, I think you're still kind of in trouble because you occasionally, the models will just lead you on a completely crazy wild goose chase or even make some very simple error that you don't catch. And so I as I as I comment in this companion paper, if you like, honestly, like, someone who's a PhD if you're if you're not in academia, if you're not an academic scientist, you might think, oh, someone who's, you know, in his fourth year or third year of a PhD program, they must already know a lot about their research area. That person is already an expert. From our perspective of frontier research, someone who's, you know, a year or two away from finishing their PhD, they're still kind of a novice.
And I would honestly say, like, those kind the people at that level using the models could easily generate tons and tons of subtly subtly wrong slop, which is kind of a bad outcome. Now some critics of my paper and of what I'm talking about would even say, like, oh, Steve Shue, an experienced physicist who's, you know, been publishing papers for thirty five, forty years now or something. Even he is taken in by these AIs and has written a slop paper. I don't think that's actually true, but but the point is I think there's some high level of expertise still required from the human in the loop to get good stuff out of the models. So I think that's true.
But it's also true that for the population, there are thousands of, say, physics or math, there are thousands of people who are real experts who have published many papers and really know their sub discipline. Those people, if using the models judiciously, perhaps in this generator verifier way, can experience a very big productivity boost in their research. Using the models to do computations, literature searches, maybe sometimes proposing new ideas, generating the LaTeX output, proofreading papers, putting in, you know, some really tedious stuff like putting in the citations and the bibliography. There are lots of things that models can do to save you time. And so my message is, students, be extra careful, but start spending some time with the models just to learn how to do it.
Experts, this could you you try it. You may find this to be extremely useful and productive. But, again, you should also be careful not to be led on wild goose chases by these models. K? Let me go on in the companion paper.
Here, I sort of summarize the results. I will go into this in more detail after I complete this part of the episode, which is for nonexperts. I'll I'll actually go through some slides describing the paper, so I'll come back to this. Here is a prompt that I used, and I did this with many different models. When the paper was nearing completion, I I asked or when we're well into the research, I asked the models, you know, do the t s the Tomahawk Schrodinger conditions for state dependent quantum mechanics exist elsewhere in the literature, question mark.
Do a careful search. K? So I kept pushing them just to make sure. Like, one of the concerns when you get something that looks novel from an AI is that it actually read that somewhere else. So it isn't actually a novel.
This is something AI researchers care a lot about. That result isn't actually novel. It read it somewhere, and it's just kind of reproducing it for you here. So I I check repeatedly for that. I've also now published a paper, and some people have other physicists have looked at it.
No one's come forward and said, like, no. This is actually derivative. This has actually been done before. There's actually a paper with this these exact this exact equation. So that doesn't seem to be the case.
And I think models are useful for that too to just do literature search and find, like, papers where, you know, the the result could have been anticipated. In this case, what I think actually happened is the model understood we were studying we were interested in nonlinear evolute nonlinear modification of the Schrodinger equation. We were interested in quantum field theory, and it it itself made this combinatorial connection between the Tomanaga Schirmer Schwinger work and this sort of Weinberg like work from the nineteen eighties or the work from my old paper from ten years ago, and it combined them to to produce these new ideas and these new equations. So so I think that, you know, like, again, is this, the deep kind of scientific innovation that, say, Einstein produced or, you know, the Max plot? No.
It's more like ordinary research where, you know, you combine two ideas, you get something interesting, you push it forward, and you write a paper. So I think that's how I would characterize what happened here, but it's still interesting because of the the AI did it, not my colleague down the hall. The AI did it. Okay. I submitted the paper to the journal.
The referee wrote a report. One of the things the referee wanted me to do was actually use these new equations to test a recent set of models, which had been proposed by two physicists, very good physicists, David Kaplan, and I don't know if his colleague is a student or a postdoc or or another professor, Rajendran. Him, I don't know. It's called the KR model, and the KR model proposes a particular form of nonlinearity for quantum mechanics and quantum field theory, and they've they've they've written extensive papers on this. So there doesn't seem to be any paper in literature that says, hey.
This this model has some problems special relativity, etcetera, etcetera, or or non local signaling, etcetera. So they they were trying to deliberately trying to avoid the original problems that the Weinberg work from the eighties had that was pointed out, as I said, by Giesen and Polchinski. So so KR were aware of this and trying to deal with this problem in their particular model. When I ran their model through my new TS equations, it doesn't satisfy the TS equation. K.
So seems like their model doesn't actually necessarily have the the properties that they thought it had. Again, this is probably something we're still gonna have to work out amongst ourselves, the the the handful of physicists that are looking deeply into this. But in any case, to respond to the referee report, I ran the KR model. I had not looked at the KR model in detail before that, and I I ran the KR model, physics model through the AI and just asked it. You can see the prompt here.
I asked it to calculate the TS integrability conditions in the KR model and discuss physical meaning of the results. Now what's interesting about this result is that to to do this calculation requires, in a way, a kind of geometric analysis of the past light cone, the the the set of events in space time that assuming a finite speed of light could have affected a particular point in space time. And so the calculation involves the overlap between two different past light cones. So you have point x, and it's past light cone, another point y, and it's past light cone. So some geometry here, and the models in a way are not, I think, not very good at geometry.
They don't really see things. At least the the the multimodal models might see things, but these model language models don't generally. So I was a little bit surprised that in this case, g p d five was able to actually correctly figure out what the overlap between these two past light cones was. It developed a very nice notation and answered the question as to whether there was a nonzero contribution to these integrability conditions coming from the KS model the KR model. And it it even like, I like this notation very much.
It defines the set j, which is the past light cone and talks about the intersection between j of x and j of y, etcetera. Now this is probably standard notation for some subset set of papers, but I if I'd invented it myself, I would have probably done it slightly differently and maybe not as economically. I think the moth did a great job. I mean, after all, this is, a few seconds of work. And the model did it right.
Gemini got it wrong. Quinn Max also got this wrong. But then I I cycled through the verifier stuff multiple times, and they they agreed that this was actually right. I checked it myself. I think it's right.
K. But just another example of the use of models in the research. K. So let me conclude on this sort of general discussion, which is aimed for nonexperts. I think this research article and how it was produced is a decent case study for where the models are right now and where how human experts and in this case, again, as I said, high experts, like people who are served well beyond PhD level in a particular field, how they can use the models profitably and to to accelerate discovery and and and scientific research.
I think it's pretty clear that if you know what you're doing, you you you you start working with the models, you gain some experience, maybe have the models, look at some of your old papers, and you discuss it with them. You'll get a sense of where they are, and then you use in a disciplined way this generate verify procedure to beat down the the errors, you you get a very useful tool. And, actually, I'm I'm I've actually built a prototype version that sort of automates the sort of something that operates in your browser that automates this pipeline to some degree. So I've been experimenting with that. Thanks to the people at x AI for inviting me to visit for a week and and build this in collaboration with them using the Grok model as the the engine that powers it.
But but I've set it up in such a way you can use any model to power it. I'm still playing with this, and I'm I I think I think by chaining together lots of models, you can you can have a supermodel that's much better. That's also quite similar to what the DeepMind people are doing it with Coscientists, and I continue to work with them on Coscientists. But, anyway, stay tuned. This is gonna get a lot better, I think.
It's but but I think it it's good for scientists of all types to spend some time with the models and understand how well or poorly do they perform in your particular sphere of expertise, and then how can you use them profitably to improve your research productivity? Okay. All of this is just gonna get better because, as I said, the the models can pretty much solve any textbook type problem. They're getting to the point where they can solve, you know, what are considered really hard international math Olympiad or Putnam problems. And I think they just lack the deepest level of understanding that an area expert, someone who had been doing, say, work in axiomatic field theory for twenty years, they don't have that level of comprehension, but I think we can easily generate the training data that will close that gap.
If if we get a lot of scientists using the models to actually do research, not just students using it to cheat on homework or something, but actual frontier, the best brains on the planet, human brains interacting strongly with the models on a day to day basis and generating training data, I think we can we can we can make another step forward. So stay tuned for that. I think that that's the project that I'm most excited about this about at the moment. In my acknowledgment, I I acknowledge, Yuraj Gottweiss and Vivek Natharjan, who are both at DeepMind. They're cocreators of this AI coscientist tool.
I didn't actually use coscientist in this research that I'm talking about, although I did use coscientist at the end to check, and it gave a pretty good report on on the work. He agreed with the results, etcetera. However, a lot of my intuition for how to use the models, what can go wrong, how you can get led on wild goose chases. I was led on some wild goose chases by coscientists. But as with all models, I'm not singling out coscientists for this.
But a lot of that intuition was built up in collaboration with these guys. I continue to work with these guys. I hope we can produce an even better research result using coscientists. It just happened that this work was done. It it it just spontaneously happened because of something that GPT five responded with.
Okay. So I think that's mostly what I wanted to talk about for this episode that's of a, shall we say, it's appropriate for nonexperts, for for people who are interested in AI but don't necessarily know that much about physics. Actually, let me add one more thing. The models don't really understand physics the way we do. I think the most pessimistic okay.
The most pessimistic interpretation of what general large language models are doing is that they're only stochastic parents. They they can only basically recombine stuff they've seen before in some kind of quasi stochastic way that sort of looks right, and that's the extent of their capabilities. I I think that's too pessimistic. I think that they're they're they're definitely better than that. I don't think we're really that close to AGI.
I think that's somewhat further away than people would have said a year or two ago, and I think gradually that's becoming a consensus among a lot of AI researchers. Here's a a sort of intermediate pessimistic take on what Gen AI is doing. They are able to generate text that resembles human reasoning, but they do not themselves reason the way that we we do or that a true symbolic manipulation program does. Okay? They're not manipulating objects with hard rules, concepts with hard rules, the way that Lean does, for example.
They're not doing that. But even if they only generate text that resembles human reasoning without actual reasoning ability of their own, those outputs are still useful to us. So so if it if it generates some stuff that, like, it didn't really reason to get to this result, but it nevertheless somehow generated it symbolically where it says, oh, you know, these conditions of Tomanaga Schwinder are probably violated by a lot of these nonlinear modifications of Schrodinger. Like, maybe it didn't reason the way that an ordinary human brain would reason to get or mister Spock would reason to get to that conclusion. Maybe it just sort of, like, blurted that out, and it blurted it out because it sort of resembles stuff that it's seen before.
It still could be correct just as the proofs for these IMO problems are actually correct. Right? Even though it may not be reasoning the way Lean does, which is a symbolic actual, like, axiomatic kind of reasoning program. Even though it's not doing that, it still might generate stuff that is correct and that is of value to a human expert using the system. So that that that's my position on where we are.
And welcome your feedback, by the way. If you if you're an AI researcher or a theoretical physicist or mathematician, you have and you have thoughts on this. I'm sure a lot of people are thinking about this. Feel just reach out and contact me because I would love to discuss it more. In fact, if you're doing research in this area, I'm I I would consider even doing, like, another future episode.
Like, if you have interesting results that you wanna talk about in using AI in your own work, I would love to maybe do an episode about that in the future. I'm going to go through some slides now, which, again, if you have access to my screen, you can you can see these. They're meant to explain the the actual physics content of the research paper that was published in physics letters. So if you aren't familiar with terms like relativistic covariance and Thomann Hage Swinger analysis, this might be hard for you to follow. Feel free to sign off.
But for the few physicists that wanna hear this, I think this is just a good place to cover the the results. It's not gonna be a full blown talk. I'll just try to go through this briefly. So title of the paper is relativistic covariance and nonlinear quantum mechanics, Tomanaga Schringer analysis. On the screen right now, I have some slides.
Here's the actual paper that was accepted at Fiske's letters. It's relatively brief paper. I think it's, like, just the the typical length of a letter. I'm just scrolling through it on the screen here so you can you can glance at it. But you you can find this on archive or I think even the physics letters version is open access, so you can you can grab this and look at it.
I encourage you to. You can email me and yell at me as some people already have. But, anyway, let me come back to the slides. So these are slides that I by the way, I had the AI prepare these slides. I literally took the LaTeX from the paper, and I said, make some slides so I can quickly give an overview of the result.
What it gave me was not perfect, but I was able to refine those slides. It took maybe another half hour to refine the slides. And so that's probably a relatively small fraction of what normally it would take me to make up slides for the a short talk. K. Yeah.
Probably, like, it took a third the normal amount of time or less. Okay. Alright. So on this slide now here, I'm gonna kind of assume you can see the screen. So if you're riding your bike and listening to this podcast, sorry.
I'm gonna kind of assume you can see my screen. Okay. So equation one, I have the Schrodinger equation, size, state vector. The thing which generates time evolution for psi is the Hamiltonian operator. And as you can see, the state psi only occurs linearly in this equation.
So, consequently, if I write a superposition a times psi one plus b times psi two, if psi one and psi two are solutions, then the resulting superposition state is also a valid solution. This has all kinds of consequences, and maybe the most colorful one is the many worlds description of quantum mechanics where you could have different branches where radically different things have happened, but they they coexist comfortably in the superposition state. And there could be a version of me which gets entangled with one set of outcomes and a version another version of me that whose memory records are entangled with another set of outcomes. And linearity means those those two branches of the wave function can just keep evolving in the Hilbert space forever. We could talk more about decoherence, etcetera, etcetera, but that's not not the not the point of this discussion.
Okay. But a fundamental question is, is the linearity in this equation fundamental or is it just an approximation? And this is something only the, like, most people most interested in foundations of physics focus on. Because if you talk to some guy in fluid mechanics and he he's looks at the Navier Stokes equation, he knows at some point Navier Stokes stops applying and you have to deal with, like, the individual motions of water molecules or something like this. He knows that.
But here, we're talking about supposedly the fundamental laws of physics. And it is a very a question worthy of a lot of attention whether there is any kind of structure in the the true fundamental description of our universe, which is exactly linear. I know we're all conditioned from our elementary quantum mechanics classes to just accept the form of this equation and that it's linear. But if you think about it carefully, this is a very deep ask. For someone to say that this is exactly here is a very deep ask.
Okay? At least for a fit for a physicist. Maybe a mathematician is totally happy with that because they don't care, but we we care about physical reality and, gee, did it really have to be that way? Did it really have to be exactly linear? Okay.
Next slide. K. This is sort of repeating a little bit what I said, but, you could imagine a modification of Schrodinger revolution, which contains some evolution in which the evolution operator, here it's written as f in equation two, is itself a function of the state. K? So this, as I mentioned earlier in the episode, this was proposed there was a very thorough set of proposals by Weinberg to study this kind of possibility, specific examples, specific tests using atomic physics.
Here, we're interested in the field theoretic generalization of this. So not just psi describing the location of a particle or something, but the quantum state that describes an entire field. So that the the electric and magnetic fields everywhere in the universe or the Higgs field everywhere in the universe. There is actually a state in the Hilbert space that corresponds to that whole thing. And quantum field theory describes the evolution of all of that, not just the wave function of a few electrons, but the the wave function of, in effect, all the degrees of freedom in the universe are evolving in time.
And there there is a Schrodinger like there may be a Schrodinger like description of that. At least there is one in quantum field theory. Okay. So amazingly, if you go all the way back, Tomanaga work was done, I believe, during World War two or in the wake of World War two. And they were interested in this question of, well, how do I unify ordinary quantum mechanics with special relativity?
The result of that is something called quantum field theory. And what conditions should what what equation is the analog of the Schrodinger equation for a quantum field? For that, you get this Tomanaga Schwinger equation. Because you're talking about special relativity, the things have to be coordinated invariant or covariant with the coordinate transformations, boosts, Poincare transformations. And the equivalent of pushing something forward in time.
So the original Schrodinger equation tells you how to push something forward in time. K? In field theory, you need a way to push something forward in time, but Tomanaga and Schringer had something which they called many fingered time because the normal vector to the Cauchy slice or time slice could be slightly different depending on where you are, depending on your chord and system. This functional derivative with respect to sigma of x is the moral equivalent of pushing something forward in time. Sigma of x is this kind of normal fluctuation away from one of the Cauchy surfaces.
And in order that space like separated places on a Cauchy surface can't talk to each other, okay, to so to preserve relativistic causality, it should be the case that pushing something forward a little or modifying the Cauchy surface slightly at x should not should should commute. Should not it should not matter whether you do something at y and then something at x or something at x or something at y because by causality, there hasn't been enough time for any causal influence to go from x to y. So there's a condition original to Tomanaga and Schwinger's formulation. I think Schwinger was probably the first one to actually write it this way, but I'm not sure, in which this commutator is is required to vanish. Now in the ordinary linear theory of quantum mechanics, this automatically vanishes if you have this microcausality condition that quantum operators commute at space like separation.
This is the old Tomanaga Schringer in linear quantum mechanics. K? And what the model proposed to me was to say, well, if you're gonna impose if you're gonna include state dependent modifications to Schrodinger revolution, then you have to modify the Tomanaga Schringer equations. This n hat is the nonlinear functional of the state psi, which is introduced. And then you can start asking questions about what happens to these commutation relations, these integrability conditions in the presence of n hat.
And you need to introduce some technical machinery. You need to be able to take a functional derivative of the n hat nonlinear term with respect to variations in this quantum state or in the coordinate system. If you vary if you push the Cauchy surface slightly forward with the sigma of x. Okay? So you have to develop a little bit of mathematical technology to deal with that.
The models just did that. You know, I didn't do it. I the models did it. I went and checked it. It seems to be correct.
K. So you generalize the this integrate integrability condition that the commutator of these two functional derivatives has to vanish on any state that has to be true if this if the theory is satisfactory. And you end up when you when you work this all out, you get this operator constraint. This is the starred equation that I mentioned in the earlier part of this episode. If the start equation is satisfied so the start equation is a necessary but not sufficient condition for you to be able to integrate forward in time from Cauchy surface to Cauchy surface in a Lorentz covariant way, in a relativistically covariant way.
K. It's necessary, but it's not sufficient. There could be some global obstacle to doing it, but if this fails, you're really in trouble because even on on some arbitrary slice, you you can't quite go forward in a relativistically covariant way. So I think it's a reasonable test or criterion that you would wanna subject in a model to. And if the model doesn't satisfy this condition, of course, doesn't mean the model can't describe reality, but it means that reality isn't fully relativistically covariant.
So that that's the underlying physics, the underlying cons conceptual basis for this. On this slide, I just have a straightforward composition check. People who work in this area, Toma Naga, Swinger type stuff, start talking about these slight variations in where how the Cauchy surface looks at x and y as little bubbles. And so you you can check the conditions that were derived by looking at the effect of two different bubble perturbations, and you you get the same equation. K?
Now here's an example. It's the original nonlinearity introduced by Weinberg. You have an operator, some arbitrary operator o, and the expectation value of o times the operator o is added to the Schrodinger the right hand side of the Schrodinger equation. So this this particular term, which is shown here, affects the quantum evolution. But the the c number coefficient of the operator o is that expectation value in the first equation on the slide, and that expectation value depends on the state.
So this is a state dependent modification to the time evolution of the state itself. So it's a nonlinearity. And so you can stuff this into the equations that we've derived, the Tomodaga Schwinger integrability conditions. You can grind it out. The models do this pretty much effortlessly.
I think maybe there were some slight errors in various models when they tried to do this and but they were all, like, suppressed out by doing the verifier generator process, and and I checked it myself. So, anyway so what's interesting is that this set of results for the specific Weinberg type term turns out the the whole thing when you when you simplify it ends up being proportional to a a commutator between two operators. And if x and y are space like separated, then conventionally, in ordinary quantum mechanics, we would say, oh, this commutator is zero because of causality. And so for this particular Weinberg term, you do get foliation independence or integrability condition satisfied, but modulo this assumption that these two operators, the operators that appear here, the the original one o, but also the Hamiltonian operator appear in these commentators. Under the provision that this microcausality still holds that the commutator between two any two operators that are space like separated is zero.
If that continues to hold, then you're okay. You've satisfied the integrability conditions. Now the subtle part of this paper is, well, is is that really reasonable? Right? So if I've modified quantum mechanics to have the state dependence or nonlinearity, can I still maintain microcausality?
Okay. So that's a question. And so let me just review what we usually say, like, in textbook treatments about microcausality. So, normally, we we we have an initial Cauchy slice or sorry. We can just say initial time, and we impose what are called equal time commutators.
So if you've ever studied quantum field theory, if you took an elementary course in quantum field theory, you would have seen this in a textbook where at some initial time, we oppose these so called canonical commutation relations between the field operators and their conjugate momenta. Because phi x and x prime at equal time so x does not equal x prime. Because they commute at equal time, if you go into a different Lorentz frame, instead of being at equal time, they could be at slightly different times, but they're still space like separated. K? So it's these commutation equal time commutation relations which give you micro causality because of general of because of relativistic covariance.
K? Now once you have that on your initial Cauchy Cauchy slice, if you define the operators at different times, say Heisenberg picture operators, if you define them at different times by using this time evolution operator, the standard time evolution operator, you can start with the canonical commutation relations on one Cauchy slice, and then you can show that because the time evolution operator has this specific unitary form where you've in you integrate over this Hamiltonian operator to get it, then this microcausality condition holds at all times, not just or on all Cauchy slices, not just on the one where you first impose the canonical commutation relation. So this is a standard textbook treatment. Now is this textbook treatment, does it still apply in the presence of these state dependent modifications of quantum mechanics? And what's I would say it's questionable.
Okay. So so the time evolution operator in standard quantum field theory is this integral over the Hamiltonian operator. It's the the first equation here. And, of course, this is totally linear. It's just some fixed operator and which acts on the state and then evolves the state or an operator forward in time slightly.
This nonlinear case is much more complicated where the actual time evolution is state dependent. You can't even actually speak of an operator by itself. You can talk about how the state evolves, but when you talk about the operator at some future time, that evolution itself is state dependent. So so the whole standard formalism for justifying microcausality so you start to justify microcausality, you start with the con canonical commutation relation. And then because the the dynamics of the theory are relativistically invariant, you can then conclude that on future you can evolve the operators forward.
And then on these future Cauchy slices, the the space like commutation relations continue to be zero. That whole framework is gone now because you can't really talk about the operators, Heisen Heisenberg picture. Everything is state dependent in the theory. So I would say that it's questionable even though the Tomonaga Schwinger equations reduced to commutators, which look like the micro they should be zero by the microcausality condition. We can't be sure of the microcausality condition because we've modified the underlying theory.
K? So that that's a point that's made in the paper. One of the reasons that I'm dubious about whether that microcausality condition continues to apply in the fully modified nonlinear state dependent version of quantum mechanics that we're in now is because in the old work, the the original work that I asked the model to look at, which I wrote, this is a paper that's more than 10 old now, We actually looked at some physical things that can happen when you have these nonlinear terms in the Schrodinger equation. And one of the things that you can show is that if you start out with two a quantum field and you have some state that you make at a and some state that you make far away at b, and a and b could be light years apart. If you introduce a thing like the Weinberg nonlinear state dependent term to the time evolution of these two states, like, could be coherent state wave packets at a and b.
As soon as you try to evolve them forward in time, the state forward in time, even though a and b were unentangled, like, they started in a product state, but they're totally unentangled and they're light years apart. Upon any amount of time evolution, they generically become entangled, which is a which is a causal process. Right? So normally, things the only way that two particles that are initially unentangled could become entangled is if they interact, and usually those interactions can only happen they can happen no faster than the speed of light. So, normally, in in ordinary quantum mechanics five two, objects or wave packets, which are some distance apart, if the if they evolve forward in time and the amount of time I evolve them forward is small in units where c equals one relative to their physic their space like separation, then, of course, they they cannot become entangled because there's no influence or physical thing that can happen between them in in such a short amount of time.
K? But that's violated the moment you introduce this nonlinear term because the nonlinear term is state dependent. I mean, you have to work out the details, but that that turns out to be the case. So this suggests to me that, actually, this microcausality condition is more subtle than than you think because definitely that instantaneous entanglement violates microcausality. And so it may not be the case that this formal, space like the the the commutator of two space like operators must vanish in the quantum field theory, that that that property may be violated once you introduce these nonlinear terms.
K? And and and in some examples that I looked at in the paper, the integrability conditions reduced to things which are proportional to these commutation relation type things. In other cases, you can find just explicit violations of the integrability conditions that come from the way that the the nonlinear term is constructed. Okay. So it's it's a little bit complicated.
You have to read the whole paper. But I think one of the most interesting physical questions is what happens to microcausality if you introduce any kind of state dependence or nonlinearity to quantum evolution. Again, the old work that I did with Chumann Ho, who who's now a head of an AI research group, by the way. The the work I did with Truman and Ho ten years ago already suggested that, yeah, microcausality cannot actually hold in some sense once you include such terms. The the very old Giesen Polchitsky thought experiment also allowed the same thing where sort of Alex and Bob could communicate superluminally with having some entangled cubic resource or something.
They could communicate superluminally if there were state dependent evolution in the Schrodinger equation. So so, again, like that slightly different physical concept there, but also, you know, seems to violate, relativistic causality. Okay. So I think there's still some open questions. It could be that just we need a completely radically different way of thinking about these theories once we allow state dependence in the evolution.
And I I think that is true. Let me finish by saying that on this really deep fundamental question of is quantum mechanics linear? And if so, you better have a good reason to justify it's exactly linear because that that would be an extremely strange result. Everything else in physics has some nonlinearities in it. Right?
It could be the case that any amount of nonlinearity, if you are adopting a quantum description of nature, which we are forced to deal with, empirically, know nature is quantum mechanical, the moment you allow some nonlinearity, any nonlinearity in the quantum evolution of the state, you get immediate violent nonlocality breakdown of microcausality or Giesen Polchinsky supraluminal signaling, signaling across Everett branches, or this Ho, Shu instantaneous entanglement of space like separated wave packets. I think those are all different manifestations of this problem that if you allow any amount of nonlinearity in the state evolution in quantum mechanics, you have to deal with this. You you don't have locality anymore. And it could be that it's a fundamental requirement of the physics of our universe that we have some degree of locality. Like, I I don't I don't want my or it's perhaps not a desirable property of, a physical theory that the evolution of an atom that's in my brain is somehow affected by the state of some atom that's on Jupiter.
Right? That that that's not a doesn't seem to be the case in our universe and maybe not a desirable property of a theory. And it's maybe that restriction that if you want some level of locality in a quantum field theory, you cannot tolerate any amount of nonlinearity in the state evolution. K. So I think that's that's the deep question, which I think still deserves a lot more research.
I think most physicists are just sort of sleeping about this. They either just sort of accept linearity. They're not thinking about it. Or, well, I think that's the case for most businesses. But it is a, I think, a deeper question and deserves more attention.
That is, of course, completely independent of whether you use AIs to investigate. This question, I think I think well, for me, myself, I I continue to actually investigate this, and I I am finding the AIs to be quite useful, in that process. Thanks a lot for your attention. Apologies to manifold listeners who aren't interested in crazy physics ideas. If you didn't like this episode, hopefully, next one would be better for you.
Take care.
Theoretical Physics With Generative AI – #101
Ask me anything about this podcast episode...
Try asking: