Reinforcement learning and approximate Bayesian computation (RL-ABC) for model selection and parameter calibration of time-varying systems

Thiago G Ritto*, Sandor Beregi*, David A W Barton

*Corresponding author for this work

Research output: Contribution to journalArticle (Academic Journal)peer-review

1 Citation (Scopus)
25 Downloads (Pure)

Abstract

This paper extends the recently developed methodology for model selection and parameter identification called RL-ABC (Ritto et al., 2022) (reinforced learning and approximate Bayesian computation) to time-varying systems. To tackle slowly-varying systems and detect abrupt changes, new features are proposed. (1) The probability of sampling the worst model has now a lower bound; because it cannot disappear, once it might be useful in the future as the system evolves. (2) A memory term (sliding window) is introduced such that past data can be forgotten whilst updating the reward; which might be useful depending on how fast the system changes. (3) The algorithm detects a change in the system by monitoring the models’ acceptance; a significant drop in acceptance indicates a change. If the system changes the algorithm is reset: new parameter ranges are computed and the rewards are restarted. To test the proposed strategy, new experimental data is obtained from a test rig with non-linear restoring force characteristics. The amplitude of the dynamical experiment is obtained with the control-based continuation strategy varying the excitation amplitude, and three Duffing-like models are used to represent the system. The results are consistent, and the strategy is able to detect changes and update parameter estimation and model predictions.
Original languageEnglish
Article number110458
JournalMechanical Systems and Signal Processing
Volume200
Issue number1
Early online date1 Jul 2023
DOIs
Publication statusPublished - 1 Oct 2023

Bibliographical note

Funding Information:
The first author would like to acknowledge that this investigation was financed in part by the Brazilian agencies: Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES) - Finance code 001 - Grant PROEX 803/2018 and CAPES-PRINT - Grant 88887.569759/2020-00 , and Fundação Carlos Chagas Filho de Amparo à Pesquisa do Estado do Rio de Janeiro (FAPERJ) - Grant E-26/201.183/2022 . The other two authors would like to acknowledge the support of the Engineering and Physical Sciences Research Council (EPSRC) via grant number EP/R006768/1 .

Publisher Copyright:
© 2023

Research Groups and Themes

  • Engineering Mathematics Research Group

Fingerprint

Dive into the research topics of 'Reinforcement learning and approximate Bayesian computation (RL-ABC) for model selection and parameter calibration of time-varying systems'. Together they form a unique fingerprint.

Cite this