Reinforcement learning and approximate Bayesian computation for model selection and parameter calibration applied to a nonlinear dynamical system

T. G. Ritto*, S. Beregi, D. A.W. Barton

*Corresponding author for this work

Research output: Contribution to journalArticle (Academic Journal)peer-review

15 Citations (Scopus)
65 Downloads (Pure)

Abstract

In the context of digital twins and integration of physics-based models with machine learning tools, this paper proposes a new methodology for model selection and parameter identification. It combines (i) reinforcement learning (RL) for model selection through a Thompson-like sampling with (ii) approximate Bayesian computation (ABC) for parameter identification and uncertainty quantification. These two methods are applied together to a nonlinear mechanical oscillator with periodic forcing. Experimental data are used in the analysis and two different nonlinear models are tested. The initial Beta distribution that represents the likelihood of the model is updated depending on how successful the model is at reproducing the reference data (reinforcement learning strategy). At the same time, the prior distribution of the model parameters is updated using a likelihood-free strategy (ABC). In the end, the rewards and the posterior distribution of the parameters of each model are obtained. The results show that the combined methodology (RL-ABC) is promising for model selection from bifurcation diagrams. Prior parameter distribution was successfully updated, correlations between parameters were found, probabilistic envelopes of the posterior model are consistent with the available data, the most rewarded model was selected, and the reinforcing strategy allows to speed up the selection process.

Original languageEnglish
Article number109485
Number of pages15
JournalMechanical Systems and Signal Processing
Volume181
Early online date1 Jul 2022
DOIs
Publication statusPublished - 1 Dec 2022

Bibliographical note

Funding Information:
The first author would like to acknowledge that this investigation was financed in part by the Brazilian agencies: Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES) - Finance code 001 - Grant PROEX 803/2018 and CAPES-PRINT - Grant 88887.569759/2020-00, and Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPQ) - Grant 400933/2016-0 . The other two authors would like to acknowledge the support of the Engineering and Physical Sciences Research Council (EPSRC) via grant number EP/R006768/1 .

Publisher Copyright:
© 2022 Elsevier Ltd.

Research Groups and Themes

  • Engineering Mathematics Research Group

Keywords

  • ABC
  • Decision under uncertainty
  • Model selection
  • Nonlinear dynamics
  • Parameter identification
  • Reinforcement learning

Fingerprint

Dive into the research topics of 'Reinforcement learning and approximate Bayesian computation for model selection and parameter calibration applied to a nonlinear dynamical system'. Together they form a unique fingerprint.

Cite this