Infinite time horizon safety of Bayesian neural networks

Bayesian neural networks (BNNs) place distributions over the weights of a neural network to model uncertainty in the data and the network’s prediction. We consider the problem of verifying safety when running a Bayesian neural network policy in a feedback loop with infinite time horizon systems. Com...

Full description

Saved in:
Bibliographic Details
Main Authors: LECHNER, Mathias, ZIKELIC, Dorde, CHATTERJEE, Krishnendu, HENZINGER, Thomas A.
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2021
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/9066
https://ink.library.smu.edu.sg/context/sis_research/article/10069/viewcontent/NeurIPS_2021_infinite_time_horizon_safety_of_bayesian_neural_networks_Paper.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-10069
record_format dspace
spelling sg-smu-ink.sis_research-100692024-08-01T15:27:14Z Infinite time horizon safety of Bayesian neural networks LECHNER, Mathias ZIKELIC, Dorde CHATTERJEE, Krishnendu HENZINGER, Thomas A. Bayesian neural networks (BNNs) place distributions over the weights of a neural network to model uncertainty in the data and the network’s prediction. We consider the problem of verifying safety when running a Bayesian neural network policy in a feedback loop with infinite time horizon systems. Compared to the existing sampling-based approaches, which are inapplicable to the infinite time horizon setting, we train a separate deterministic neural network that serves as an infinite time horizon safety certificate. In particular, we show that the certificate network guarantees the safety of the system over a subset of the BNN weight posterior’s support. Our method first computes a safe weight set and then alters the BNN’s weight posterior to reject samples outside this set. Moreover, we show how to extend our approach to a safe-exploration reinforcement learning setting, in order to avoid unsafe trajectories during the training of the policy. We evaluate our approach on a series of reinforcement learning benchmarks, including non-Lyapunovian safety specifications. 2021-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9066 info:doi/10.5555/3540261.3541039 https://ink.library.smu.edu.sg/context/sis_research/article/10069/viewcontent/NeurIPS_2021_infinite_time_horizon_safety_of_bayesian_neural_networks_Paper.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University OS and Networks
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic OS and Networks
spellingShingle OS and Networks
LECHNER, Mathias
ZIKELIC, Dorde
CHATTERJEE, Krishnendu
HENZINGER, Thomas A.
Infinite time horizon safety of Bayesian neural networks
description Bayesian neural networks (BNNs) place distributions over the weights of a neural network to model uncertainty in the data and the network’s prediction. We consider the problem of verifying safety when running a Bayesian neural network policy in a feedback loop with infinite time horizon systems. Compared to the existing sampling-based approaches, which are inapplicable to the infinite time horizon setting, we train a separate deterministic neural network that serves as an infinite time horizon safety certificate. In particular, we show that the certificate network guarantees the safety of the system over a subset of the BNN weight posterior’s support. Our method first computes a safe weight set and then alters the BNN’s weight posterior to reject samples outside this set. Moreover, we show how to extend our approach to a safe-exploration reinforcement learning setting, in order to avoid unsafe trajectories during the training of the policy. We evaluate our approach on a series of reinforcement learning benchmarks, including non-Lyapunovian safety specifications.
format text
author LECHNER, Mathias
ZIKELIC, Dorde
CHATTERJEE, Krishnendu
HENZINGER, Thomas A.
author_facet LECHNER, Mathias
ZIKELIC, Dorde
CHATTERJEE, Krishnendu
HENZINGER, Thomas A.
author_sort LECHNER, Mathias
title Infinite time horizon safety of Bayesian neural networks
title_short Infinite time horizon safety of Bayesian neural networks
title_full Infinite time horizon safety of Bayesian neural networks
title_fullStr Infinite time horizon safety of Bayesian neural networks
title_full_unstemmed Infinite time horizon safety of Bayesian neural networks
title_sort infinite time horizon safety of bayesian neural networks
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/9066
https://ink.library.smu.edu.sg/context/sis_research/article/10069/viewcontent/NeurIPS_2021_infinite_time_horizon_safety_of_bayesian_neural_networks_Paper.pdf
_version_ 1814047722096820224