Global PAC bounds for learning discrete time Markov chains

Learning models from observations of a system is a powerful tool with many applications. In this paper, we consider learning Discrete Time Markov Chains (DTMC), with different methods such as frequency estimation or Laplace smoothing. While models learnt with such methods converge asymptotically tow...

Full description

Saved in:
Bibliographic Details
Main Authors: BAZILLE, Hugo, GENEST, Blaise, JEGOUREL, Cyrille, SUN, Jun
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2020
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/6071
https://ink.library.smu.edu.sg/context/sis_research/article/7074/viewcontent/501999_1_En_Print.indd.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-7074
record_format dspace
spelling sg-smu-ink.sis_research-70742021-09-29T13:07:50Z Global PAC bounds for learning discrete time Markov chains BAZILLE, Hugo GENEST, Blaise JEGOUREL, Cyrille SUN, Jun Learning models from observations of a system is a powerful tool with many applications. In this paper, we consider learning Discrete Time Markov Chains (DTMC), with different methods such as frequency estimation or Laplace smoothing. While models learnt with such methods converge asymptotically towards the exact system, a more practical question in the realm of trusted machine learning is how accurate a model learnt with a limited time budget is. Existing approaches provide bounds on how close the model is to the original system, in terms of bounds on local (transition) probabilities, which has unclear implication on the global behavior. In this work, we provide global bounds on the error made by such a learning process, in terms of global behaviors formalized using temporal logic. More precisely, we propose a learning process ensuring a bound on the error in the probabilities of these properties. While such learning process cannot exist for the full LTL logic, we provide one ensuring a bound that is uniform over all the formulas of CTL. Further, given one timeto-failure property, we provide an improved learning algorithm. Interestingly, frequency estimation is sufficient for the latter, while Laplace smoothing is needed to ensure non-trivial uniform bounds for the full CTL logic. 2020-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6071 info:doi/10.1007/978-3-030-53291-8_17 https://ink.library.smu.edu.sg/context/sis_research/article/7074/viewcontent/501999_1_En_Print.indd.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Software Engineering
spellingShingle Software Engineering
BAZILLE, Hugo
GENEST, Blaise
JEGOUREL, Cyrille
SUN, Jun
Global PAC bounds for learning discrete time Markov chains
description Learning models from observations of a system is a powerful tool with many applications. In this paper, we consider learning Discrete Time Markov Chains (DTMC), with different methods such as frequency estimation or Laplace smoothing. While models learnt with such methods converge asymptotically towards the exact system, a more practical question in the realm of trusted machine learning is how accurate a model learnt with a limited time budget is. Existing approaches provide bounds on how close the model is to the original system, in terms of bounds on local (transition) probabilities, which has unclear implication on the global behavior. In this work, we provide global bounds on the error made by such a learning process, in terms of global behaviors formalized using temporal logic. More precisely, we propose a learning process ensuring a bound on the error in the probabilities of these properties. While such learning process cannot exist for the full LTL logic, we provide one ensuring a bound that is uniform over all the formulas of CTL. Further, given one timeto-failure property, we provide an improved learning algorithm. Interestingly, frequency estimation is sufficient for the latter, while Laplace smoothing is needed to ensure non-trivial uniform bounds for the full CTL logic.
format text
author BAZILLE, Hugo
GENEST, Blaise
JEGOUREL, Cyrille
SUN, Jun
author_facet BAZILLE, Hugo
GENEST, Blaise
JEGOUREL, Cyrille
SUN, Jun
author_sort BAZILLE, Hugo
title Global PAC bounds for learning discrete time Markov chains
title_short Global PAC bounds for learning discrete time Markov chains
title_full Global PAC bounds for learning discrete time Markov chains
title_fullStr Global PAC bounds for learning discrete time Markov chains
title_full_unstemmed Global PAC bounds for learning discrete time Markov chains
title_sort global pac bounds for learning discrete time markov chains
publisher Institutional Knowledge at Singapore Management University
publishDate 2020
url https://ink.library.smu.edu.sg/sis_research/6071
https://ink.library.smu.edu.sg/context/sis_research/article/7074/viewcontent/501999_1_En_Print.indd.pdf
_version_ 1770575809623359488