APS Logo

Optimizing Computationally-Intensive Simulations Using a Biologically-Inspired Acquisition Function and a Fourier Neural Operator Surrogate

POSTER

Abstract

Computational modeling of physical phenomena has enabled researchers to acquire insight that was previously only observable from costly real-world experiments. The adoption of physics simulations has resulted in numerous advancements in fusion energy, seismic inversion/monitoring, and national defense. However, optimizing simulation studies requires many realizations of the intensive simulations. Manually tweaking control parameters and searching for optimal results can be tedious and inefficient. To tackle such obstacles in simulation studies, we found that differential evolution combined with the covariance matrix adaptation strategy could effectively optimize simulations while simultaneously behaving as an acquisition function to collect samples. The samples collected may be used to train intelligent surrogate models such as a Fourier neural operator (FNO). Once a surrogate is constructed, it could be used to accelerate the sampling of the optimization search space further. This methodology effectively optimized a hydrodynamic simulation modeled by systems of partial differential equations; it may also be extended to simulation optimization in other disciplines.

Publication: [1] Andrew W Cook. Artificial fluid properties for large-eddy simulation of compressible turbulent mixing.<br>Physics of fluids, 19(5), 2007.<br>[2] Andrew W Cook. Enthalpy diffusion in multicomponent flows. Physics of Fluids, 21(5), 2009.<br><br>[3] Swagatam Das and Ponnuthurai Nagaratnam Suganthan. Differential evolution: A survey of the state-of-the-<br>art. IEEE transactions on evolutionary computation, 15(1):4–31, 2010.<br><br>[4] David Gottlieb and Chi-Wang Shu. On the gibbs phenomenon and its resolution. SIAM review, 39(4):<br>644–668, 1997.<br>[5] Kurt Hornik, Maxwell Stinchcombe, and Halbert White. Multilayer feedforward networks are universal<br>approximators. Neural networks, 2(5):359–366, 1989.<br>[6] Christian Igel, Nikolaus Hansen, and Stefan Roth. Covariance matrix adaptation for multi-objective<br>optimization. Evolutionary computation, 15(1):1–28, 2007.<br>[7] Nikola B. Kovachki, Zongyi Li, Burigede Liu, Kamyar Azizzadenesheli, Kaushik Bhattacharya, Andrew M.<br>Stuart, and Anima Anandkumar. Neural operator: Learning maps between function spaces. CoRR,<br>abs/2108.08481, 2021.<br>[8] Zongyi Li, Nikola Kovachki, Kamyar Azizzadenesheli, Burigede Liu, Kaushik Bhattacharya, Andrew Stuart,<br>and Anima Anandkumar. Fourier neural operator for parametric partial differential equations. arXiv preprint<br>arXiv:2010.08895, 2020.

Presenters

  • John P Lins

    Lawrence Livermore National Laboratory

Authors

  • John P Lins

    Lawrence Livermore National Laboratory

  • Wei Liu

    Lawrence Livermore National Laboratory