Abstract
In light of the inherently complex and dynamic nature of real-world environments, incorporating risk measures is crucial for the robustness evaluation of deep learning models. In this work, we propose a Risk-Averse Certification framework for Bayesian neural networks called RAC-BNN. Our method leverages sampling and optimisation to compute a probabilistically sound approximation of the output set of a BNN, represented using a set of template polytopes. To enhance risk-aware robustness evaluation, we integrate a coherent distortion risk measure—Conditional Value at Risk (CVaR)—into the certification framework, providing probabilistic guarantees based on empirical distributions obtained through sampling. We validate RAC-BNN on a range of regression and classification benchmarks and compare its performance with a state-of-the-art method. The results show that RAC-BNN effectively quantifies robustness under worst-performing risky scenarios, and achieves tighter certified bounds and higher efficiency in complex tasks.
| Original language | English |
|---|---|
| Publication status | Accepted/In press - 8 Oct 2025 |
| Event | The 11th International Symposium on Dependable Software Engineering Theories, Tools and Applications - St Catherine's College , Oxford, United Kingdom Duration: 1 Dec 2025 → 3 Dec 2025 https://www.setta2025.uk/home |
Conference
| Conference | The 11th International Symposium on Dependable Software Engineering Theories, Tools and Applications |
|---|---|
| Abbreviated title | SETTA 2025 |
| Country/Territory | United Kingdom |
| City | Oxford |
| Period | 1/12/25 → 3/12/25 |
| Internet address |