Wow, I didn't think this would HN. I actually planned to do the advertisement rounds only after the final ICLR submission.
This is our attempt at creating a model which understands multiple physics, which is in contrast to PINNs and Neural Operators, which focus on much more narrow systems.
Obviously, the biggest issue is still data (3D and real-world problems), but I think we and a few other groups make significant progress here.
For folks wondering whether to read or not, here is the conclusion from the paper verbatim
> We have demonstrated that a single transformer-based model can effectively learn and predict the dynamics of diverse physical systems without explicit physics-specific features, marking a significant step toward true Physics Foundation Models. GPhyT not only outperforms specialized architectures on known physics by up to an order of magnitude but, more importantly, exhibits emergent in-context learning capabilities—inferring new boundary conditions and even entirely novel physical phenomena from input prompts alone.
Anyone remember that one time, a year or so ago, when some company teased a physics based generative model which showcased a drop of water sliding down a beer bottle and the model could display the forces acting on it?
How do they prove their model preserves conservation principles? I looked in the paper & didn't find any evidence of how they verify that whatever their "trained" model is doing is actually physically plausible & maintains the relevant invariants like mass, energy, momentum, etc.
I guess it can be implemented in the 'sampler' part. When solving an actual PDE, project the output of the AI onto a space that preserves the invariants.
Author here: we do NOT do conservation of energy/momentum. We are currently trying to incorporate that in a follow up paper, but for now, the models that try that (e.g. PINNs (soft constraint) or hard constraint models, all perform badly when modeling multiple systems.
Perhaps, we will encounter the bitter lesson again and a well trained model will solve this. But as I said, we are also looking at hybrid models
I think very few of these "replace numerical solver with ML model" papers do anything to verify invariants are satisfied (they often are not well preserved).
They basically all just check that the model approximately reproduces some dynamics on a test data of PDEs, that's often sampled from the same distribution as the training dataset...
Author here: we do NOT do conservation of energy/momentum. We are currently trying to incorporate that in a follow up paper, but for now, the models that try that (e.g. PINNs (soft constraint) or hard constraint models, all perform bad when modeling multiple systems.
Perhaps, we will encounter the bitter lesson again and a well trained model will solve this. But as I said, we are also looking at hybrid models
From a quick scan, I do not think they explicitly encode that. They want "the model to predict the evolution of diverse physical systems governed by partial differential equations". It looks like a more sophisticated sibling of time series forecasting models rather than a physics-informed nonparametric symbolic regression model.
Yeah, It’s true that PDEs are the "top-tier tool" for describing physical phenomena—from the laws of motion in classical mechanics and electromagnetic waves in electromagnetism to the evolution of wave functions in quantum mechanics, they accurately model most macroscopic, classical scenarios. However, when it comes to covering all physical phenomena, they really "fall short": in quantum gravity, spacetime may be discontinuous, making the concept of differentiation meaningless; for complex systems like turbulence, PDEs cannot be solved nor can they capture macroscopic laws; even for the randomness of quantum measurements, PDEs can only predict probability distributions and fail to explain the underlying nature. In short, they are a "top-tier auxiliary," but by no means a "one-size-fits-all key."
Author here,
Wow, I didn't think this would HN. I actually planned to do the advertisement rounds only after the final ICLR submission.
This is our attempt at creating a model which understands multiple physics, which is in contrast to PINNs and Neural Operators, which focus on much more narrow systems.
Obviously, the biggest issue is still data (3D and real-world problems), but I think we and a few other groups make significant progress here.
For folks wondering whether to read or not, here is the conclusion from the paper verbatim
> We have demonstrated that a single transformer-based model can effectively learn and predict the dynamics of diverse physical systems without explicit physics-specific features, marking a significant step toward true Physics Foundation Models. GPhyT not only outperforms specialized architectures on known physics by up to an order of magnitude but, more importantly, exhibits emergent in-context learning capabilities—inferring new boundary conditions and even entirely novel physical phenomena from input prompts alone.
Not the "foundational model" of physics I was expecting, but this is still great to see!
Anyone remember that one time, a year or so ago, when some company teased a physics based generative model which showcased a drop of water sliding down a beer bottle and the model could display the forces acting on it?
Whatever happened to that? Vapourware?
I think you mean this: https://genesis-embodied-ai.github.io/ It seems that this is much more focused on robotics, but interesting nonetheless.
How do they prove their model preserves conservation principles? I looked in the paper & didn't find any evidence of how they verify that whatever their "trained" model is doing is actually physically plausible & maintains the relevant invariants like mass, energy, momentum, etc.
I guess it can be implemented in the 'sampler' part. When solving an actual PDE, project the output of the AI onto a space that preserves the invariants.
Author here: we do NOT do conservation of energy/momentum. We are currently trying to incorporate that in a follow up paper, but for now, the models that try that (e.g. PINNs (soft constraint) or hard constraint models, all perform badly when modeling multiple systems.
Perhaps, we will encounter the bitter lesson again and a well trained model will solve this. But as I said, we are also looking at hybrid models
I think very few of these "replace numerical solver with ML model" papers do anything to verify invariants are satisfied (they often are not well preserved). They basically all just check that the model approximately reproduces some dynamics on a test data of PDEs, that's often sampled from the same distribution as the training dataset...
Why? Is this important as a sanity check in the absence of any independent verifications?
Author here: we do NOT do conservation of energy/momentum. We are currently trying to incorporate that in a follow up paper, but for now, the models that try that (e.g. PINNs (soft constraint) or hard constraint models, all perform bad when modeling multiple systems.
Perhaps, we will encounter the bitter lesson again and a well trained model will solve this. But as I said, we are also looking at hybrid models
From a quick scan, I do not think they explicitly encode that. They want "the model to predict the evolution of diverse physical systems governed by partial differential equations". It looks like a more sophisticated sibling of time series forecasting models rather than a physics-informed nonparametric symbolic regression model.
Yeah, It’s true that PDEs are the "top-tier tool" for describing physical phenomena—from the laws of motion in classical mechanics and electromagnetic waves in electromagnetism to the evolution of wave functions in quantum mechanics, they accurately model most macroscopic, classical scenarios. However, when it comes to covering all physical phenomena, they really "fall short": in quantum gravity, spacetime may be discontinuous, making the concept of differentiation meaningless; for complex systems like turbulence, PDEs cannot be solved nor can they capture macroscopic laws; even for the randomness of quantum measurements, PDEs can only predict probability distributions and fail to explain the underlying nature. In short, they are a "top-tier auxiliary," but by no means a "one-size-fits-all key."
> in quantum gravity
GP was asking about conservation laws but in gravity you don't even have energy-momentum conservation.
ITAR will be quite a surprise to some when it suddenly makes an appearance.