Lagrangian Neural Network (LNN) [Physics Informed Machine Learning]

  Рет қаралды 29,410

Steve Brunton

Steve Brunton

Күн бұрын

Пікірлер: 29
@psychii678
@psychii678 4 ай бұрын
this in combination with some of the modern neural operator (fourier, wavelet) methods are really going to be the norm for most computational physics in industry that use continuum models I think
@zanubiadepasquale
@zanubiadepasquale 4 ай бұрын
So cool, thank you, professor! I have a possibly naive question: does this mean that MLPs are inherently unable to fully model such systems, no matter the complexity or depth of their architecture, because they will always lose a system's symmetry relations?
@maxbaugh9372
@maxbaugh9372 3 ай бұрын
So we have Lagrangian & Hamiltonian Neural Networks, I think the question is obvious: do we have Hamilton-Jacobi Neural Networks?
@as-qh1qq
@as-qh1qq 4 ай бұрын
Integrating chaotic systems: when Runge-Kute can be called naive
@muhammadfurqan8556
@muhammadfurqan8556 Ай бұрын
Do we use single (or same) optimizer for both grandient descent and ascent in these Lagrangian NN or separate one ??
@ingolifs
@ingolifs 4 ай бұрын
Can I clarify something? Does the NN just give the updated position and velocity at the next time step? And then you repeatedly use the NN to integrate the system to find its full time evolution? You can't use this sort of architecture (at least for simple problems) to find the state at an arbitrary point in time in a single NN calculation?
@tassiedevil2200
@tassiedevil2200 4 ай бұрын
This is a good question. I interpreted that you get back the accelerations i.e. enough to make a timestep from your (input) initial values. In either case (Baseline or Lagrangian NN) it seems the accelerations are the training data - it is just how they are used. For predictions, the difference is how the accelerations for the next step are generated - the autodifferentiated L version being superior, presumably being more constrained by the Lagrangian rather than simply being some sort of ML interpolator of the training observations. While I can see this is potentially interesting for deducing underlying dynamics from observations, I am curious how it useful it is for chaotic systems. Consider the training data as samples of positions and velocities in the phase space (e.g. 4 dimensional for the double pendulum), then of course velocities (given) and accelerations indicate the tangent to a trajectory at each of these phase space points. However, given that trajectories of initially close points diverge in chaotic systems, how realistic is this for marching forwards?
@MDNQ-ud1ty
@MDNQ-ud1ty 4 ай бұрын
In his older videos he used a PINN that was was an integrator to show that they are much better at long term predictability. I imagine it is exactly the same. These are local methods, not global. Since they are discrete methods they can't do any global derivations which would be symbolic. It is likely an impossible problem to have a global solver. Effectively you are then asking to be able to compute an exact timestep to get where you want with zero loss. This would, at the very least, require one to have the symbolic description of the system rather than just discrete sample points.
@esti445
@esti445 3 ай бұрын
I suggest a "transcendental neural network". Can I publish?
@dadsonworldwide3238
@dadsonworldwide3238 4 ай бұрын
I do long for an agent that can serf all models tuned and weighted although I'm sure it will be a while before we the people really get tomorrow's access today like that .
@mootytootyfrooty
@mootytootyfrooty 4 ай бұрын
least action seems like the only way you can actually ground a neural net if you want it stay in reality, even abstract from physics since at some point it needs to come back to reality where there is thermodynamics governing always. Seems like a necessary core for neural nets in general to adopt.
@johnwaczak8028
@johnwaczak8028 4 ай бұрын
Excellent video! About the intrinsic coordinates problem for HNNs, can't you use an auto-encoder to "discover" the correct (q,p) pairs from your input data like Greydanus et al do for the pendulum video example in the HNN paper? It seems like the added cost of computing the Hessian could be a significant bottleneck for more realistic, high-dimensional datasets.
@looper6394
@looper6394 4 ай бұрын
any proof that it will find the right lagrangian? in my case the qpp values fit, however the lagrangian is completly off. seems like there is no 1:1 relationship.
@FredericMbouleNgolle
@FredericMbouleNgolle 4 ай бұрын
Good eveny sir and tank you for yours videos Please Can WE use or those méthodes(all that you have présent) in a epidemylogycal model ? ( Driving by ODE or PDE system)
@kannan.j7867
@kannan.j7867 4 ай бұрын
Great content
@theekshanabandara9293
@theekshanabandara9293 4 ай бұрын
Very interesting! ❤️
@jaikumar848
@jaikumar848 4 ай бұрын
Hello sir ! Is it possible to make mathematical model/transfer function of Diode /Thyristor so that we can predict output of diode just by convolution of diode and input sine wave ?
@vinitsingh5546
@vinitsingh5546 4 ай бұрын
Could you please do a video explanation one for Implicit Neural Representations with Periodic Activation Functions? Thank you!
@799usman
@799usman 4 ай бұрын
To all those who read my comment: I want to apply a Lagrangian Neural Network (NN) to approximate or model a temporal signal, such as temporal traffic flow. However, I don't know where to start. Could you guide me on whether it is possible and where I can find related python-code? I would also be happy to learn if anyone has applied LNN to the MNIST features as an embedded layer in a neural network.
@BillTubbs
@BillTubbs 3 ай бұрын
An easy introduction/refresher on Newtonian, Hamiltonian, and Lagrangian mechanics here: kzbin.info/www/bejne/ZnWrf3qqqNKgfpo
@arafathasan-ec5cj
@arafathasan-ec5cj 4 ай бұрын
sir..can u make a video on tensor for a physics major.??????we will be grateful if u make one....
@hyperduality2838
@hyperduality2838 4 ай бұрын
Problem (input), reaction (hidden), solution (output) -- the Hegelian dialectic! Your mind (concepts) is the reaction or anti-thesis to the outside world (perception). Input vectors are dual -- contravariant is dual to covariant -- dual bases, Riemann geometry. Concepts are dual to percepts -- the mind duality of Immanuel Kant. "Always two there are" -- Yoda. Neural networks are based upon the Hegelian dialectic! Lagrangians are dual to Hamiltonians.
@Jaylooker
@Jaylooker 4 ай бұрын
I think this Lagrangian neural network would be good at most classical physical simulations and has applications like being used in a physics engine. There is the Lagrangian of the standard model so it should be possible to also replicate particle physics excluding gravitational effects.
@edisonj5335
@edisonj5335 4 ай бұрын
excellent
4 ай бұрын
@raaedalmayali3685
@raaedalmayali3685 4 ай бұрын
@AABB-px8lc
@AABB-px8lc 4 ай бұрын
I know, no one care, but please tell me when that Perceptron BS ended on this channel, and I can finnaly enjoy real math as half year or so ago.
@rudypieplenbosch6752
@rudypieplenbosch6752 4 ай бұрын
The videos, went from being really instructive, towards, just skimming the surface unfortunately. The content has changed and not for the better..
@SuperSuperGenius
@SuperSuperGenius 4 ай бұрын
I so want to dub a version of this video to ZZTops 'LaGrange.'
Neural Implicit Flow (NIF) [Physics Informed Machine Learning]
13:43
Hamiltonian Neural Networks (HNN) [Physics Informed Machine Learning]
19:24
PIZZA or CHICKEN // Left or Right Challenge
00:18
Hungry FAM
Рет қаралды 16 МЛН
Generative Model That Won 2024 Nobel Prize
33:04
Artem Kirsanov
Рет қаралды 211 М.
Do Complex Numbers Exist?
11:26
Sabine Hossenfelder
Рет қаралды 791 М.
Watching Neural Networks Learn
25:28
Emergent Garden
Рет қаралды 1,3 МЛН
The moment we stopped understanding AI [AlexNet]
17:38
Welch Labs
Рет қаралды 1,3 МЛН
Official PyTorch Documentary: Powering the AI Revolution
35:53
Neural ODEs (NODEs) [Physics Informed Machine Learning]
24:37
Steve Brunton
Рет қаралды 66 М.
This is why Deep Learning is really weird.
2:06:38
Machine Learning Street Talk
Рет қаралды 397 М.
A Brain-Inspired Algorithm For Memory
26:52
Artem Kirsanov
Рет қаралды 158 М.
The Most Important Algorithm in Machine Learning
40:08
Artem Kirsanov
Рет қаралды 511 М.