Neural differential equation
Neural differential equations are a class of models in machine learning that combine neural networks with the mathematical framework of differential equations.[1] These models provide an alternative approach to neural network design, particularly for systems that evolve over time or through continuous transformations.
The most common type, a neural ordinary differential equation (neural ODE), defines the evolution of a system's state using an ordinary differential equation whose dynamics are governed by a neural network: In this formulation, the neural network parameters θ determine how the state changes at each point in time.[1] This approach contrasts with conventional neural networks, where information flows through discrete layers indexed by natural numbers. Neural ODEs instead use continuous layers indexed by positive real numbers, where the function represents the network's state at any given layer depth t.
Neural ODEs can be understood as continuous-time control systems, where their ability to interpolate data can be interpreted in terms of controllability.[2] They have found applications in time series analysis, generative modeling, and the study of complex dynamical systems.
Connection with residual neural networks
[edit]Neural ODEs can be interpreted as a residual neural network with a continuum of layers rather than a discrete number of layers.[3] Applying the Euler method with a unit time step to a neural ODE yields the forward propagation equation of a residual neural network:
with ℓ being the ℓ-th layer of this residual neural network. While the forward propagation of a residual neural network is done by applying a sequence of transformations starting at the input layer, the forward propagation computation of a neural ODE is done by solving a differential equation. More precisely, the output associated to the input of the neural ODE is obtained by solving the initial value problem
and assigning the value to .
Universal differential equations
[edit]In physics-informed contexts where additional information is known, neural ODEs can be combined with an existing first-principles model to build a physics-informed neural network model called universal differential equations (UDE).[4][5][6][7] For instance, an UDE version of the Lotka-Volterra model can be written as[8]
where the terms and are correction terms parametrized by neural networks.
References
[edit]- ^ a b Chen, Ricky T. Q.; Rubanova, Yulia; Bettencourt, Jesse; Duvenaud, David K. (2018). "Neural Ordinary Differential Equations" (PDF). In Bengio, S.; Wallach, H.; Larochelle, H.; Grauman, K.; Cesa-Bianchi, N.; Garnett, R. (eds.). Advances in Neural Information Processing Systems. Vol. 31. Curran Associates, Inc. arXiv:1806.07366.
- ^ Ruiz-Balet, Domènec; Zuazua, Enrique (2023). "Neural ODE Control for Classification, Approximation, and Transport". SIAM Review. 65 (3): 735–773. arXiv:2104.05278. doi:10.1137/21M1411433. ISSN 0036-1445.
- ^ Chen, Ricky T. Q.; Rubanova, Yulia; Bettencourt, Jesse; Duvenaud, David K. (2018). "Neural Ordinary Differential Equations" (PDF). In Bengio, S.; Wallach, H.; Larochelle, H.; Grauman, K.; Cesa-Bianchi, N.; Garnett, R. (eds.). Advances in Neural Information Processing Systems. Vol. 31. Curran Associates, Inc. arXiv:1806.07366.
- ^ Christopher Rackauckas; Yingbo Ma; Julius Martensen; Collin Warner; Kirill Zubov; Rohit Supekar; Dominic Skinner; Ali Ramadhan; Alan Edelman (2024). "Universal Differential Equations for Scientific Machine Learning". arXiv:2001.04385 [cs.LG].
- ^ Xiao, Tianbai; Frank, Martin (2023). "RelaxNet: A structure-preserving neural network to approximate the Boltzmann collision operator". Journal of Computational Physics. 490: 112317. arXiv:2211.08149. Bibcode:2023JCoPh.49012317X. doi:10.1016/j.jcp.2023.112317.
- ^ Silvestri, Mattia; Baldo, Federico; Misino, Eleonora; Lombardi, Michele (2023), Mikyška, Jiří; de Mulatier, Clélia; Paszynski, Maciej; Krzhizhanovskaya, Valeria V. (eds.), "An Analysis of Universal Differential Equations for Data-Driven Discovery of Ordinary Differential Equations", Computational Science – ICCS 2023, vol. 10476, Cham: Springer Nature Switzerland, pp. 353–366, doi:10.1007/978-3-031-36027-5_27, ISBN 978-3-031-36026-8, retrieved 2024-08-18
- ^ Christoph Plate; Carl Julius Martensen; Sebastian Sager (2024). "Optimal Experimental Design for Universal Differential Equations". arXiv:2408.07143 [math.OC].
- ^ Patrick Kidger (2021). On Neural Differential Equations (PhD). Oxford, United Kingdom: University of Oxford, Mathematical Institute.