Learning-enabled decision-making for autonomous driving: framework and methodology

The growing adoption of autonomous vehicles (AVs) holds the promise of transforming transportation systems, enhancing traffic safety, and supporting environmental sustainability. Despite significant advancements in advanced perception systems for AVs, the effectiveness of these vehicles in real-worl...

Full description

Saved in:
Bibliographic Details
Main Author: Huang, Zhiyu
Other Authors: Lyu Chen
Format: Thesis-Doctor of Philosophy
Language:English
Published: Nanyang Technological University 2023
Subjects:
Online Access:https://hdl.handle.net/10356/172842
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-172842
record_format dspace
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Mechanical engineering
spellingShingle Engineering::Mechanical engineering
Huang, Zhiyu
Learning-enabled decision-making for autonomous driving: framework and methodology
description The growing adoption of autonomous vehicles (AVs) holds the promise of transforming transportation systems, enhancing traffic safety, and supporting environmental sustainability. Despite significant advancements in advanced perception systems for AVs, the effectiveness of these vehicles in real-world scenarios depends critically on advanced decision-making systems. Therefore, a critical challenge lies in developing safe, robust, and human-like decision-making systems for achieving higher levels of vehicle autonomy. Harnessing the power of artificial intelligence (AI) and machine learning (ML) is pivotal in developing such advanced decision-making systems. These techniques offer promising solutions for tackling key challenges in autonomous driving systems, including scalability, adaptability, and alignment with human driver behaviors. This thesis presents a comprehensive framework and a series of learning-based methodologies for decision-making in AVs, with the objective of improving the scalability, adaptability, and alignment of their decision-making systems. The learning-enabled decision-making framework consists of three essential modules: world model, actor, and cost. These modules work collaboratively to provide a unified solution for autonomous machine intelligence and can be learned individually or jointly from human driving data. The methodologies proposed in this thesis focus on three key areas: (1) creating ML-driven actors and cost functions that align with human values, (2) developing ML-based world models capable of accurately forecasting complex human interactions, and (3) designing ML-based decision-making architectures that seamlessly integrate these components. Initially, we explore each module individually. For the actor module, we propose an approach based on reinforcement learning (RL) that incorporates human prior knowledge into the training process. By combining imitation learning (IL) from human expert demonstrations with RL, a novel learning framework is developed to align agent behavior with human expectations. The results show improved sample efficiency, enabling the agent to navigate complex urban environments safely, while also generating distinct driving styles that closely resemble human driving behaviors. In the cost module, we focus on learning personalized cost functions underlying human driving behaviors using inverse reinforcement learning (IRL). We propose a maximum entropy IRL algorithm that infers the reward function based on a structural assumption about discrete latent intentions guiding continuous low-level control actions. The personalized cost learning method outperforms general cost modeling methods, leading to a more human-like driving experience. Regarding the world model module, the goal is to accurately represent the complexity and dynamics of human behaviors in real-world driving environments. We introduce a joint multi-agent prediction model based on Transformers, leveraging hierarchical game theory to model interactions among agents in driving scenarios. The model employs a novel Transformer decoder structure and a learning process that regulates agent behavior to respond to other agents' behaviors. Extensive experiments validate the model's state-of-the-art prediction accuracy and its capability to jointly reason about ego agent motion plans and other agents' behaviors. Following individual assessments, unified predictive decision-making architectures are presented, harmonizing the three modules. A comprehensive behavior planning framework that integrates all three core modules is proposed. It generates diverse trajectory proposals, forecasts multi-modal futures for other agents, and evaluates candidate plans using a learned cost function. Importantly, this framework includes a conditional motion predictor (CMP) that forecasts other agents' responses to the potential actions of the AV. The CMP improves prediction accuracy and facilitates the selection of human-like behaviors. Moreover, we propose a differentiable and integrated prediction-planning framework that enables end-to-end training of the entire decision-making system. The framework incorporates a differentiable nonlinear optimizer as the motion planner, allowing the gradients to flow between the three core modules and directly optimizing the planning performance. The experimental results demonstrate superior performance in both open-loop and closed-loop testing, effectively handling complex urban driving scenarios. This thesis not only makes significant contributions to the advancement of autonomous driving systems but also lays the foundation for future research in learning-based decision-making for a wide range of intelligent systems. It paves the way for safe and human-centered autonomy, while driving the evolution of intelligent transportation systems.
author2 Lyu Chen
author_facet Lyu Chen
Huang, Zhiyu
format Thesis-Doctor of Philosophy
author Huang, Zhiyu
author_sort Huang, Zhiyu
title Learning-enabled decision-making for autonomous driving: framework and methodology
title_short Learning-enabled decision-making for autonomous driving: framework and methodology
title_full Learning-enabled decision-making for autonomous driving: framework and methodology
title_fullStr Learning-enabled decision-making for autonomous driving: framework and methodology
title_full_unstemmed Learning-enabled decision-making for autonomous driving: framework and methodology
title_sort learning-enabled decision-making for autonomous driving: framework and methodology
publisher Nanyang Technological University
publishDate 2023
url https://hdl.handle.net/10356/172842
_version_ 1787590741672853504
spelling sg-ntu-dr.10356-1728422024-01-04T06:32:51Z Learning-enabled decision-making for autonomous driving: framework and methodology Huang, Zhiyu Lyu Chen School of Mechanical and Aerospace Engineering lyuchen@ntu.edu.sg Engineering::Mechanical engineering The growing adoption of autonomous vehicles (AVs) holds the promise of transforming transportation systems, enhancing traffic safety, and supporting environmental sustainability. Despite significant advancements in advanced perception systems for AVs, the effectiveness of these vehicles in real-world scenarios depends critically on advanced decision-making systems. Therefore, a critical challenge lies in developing safe, robust, and human-like decision-making systems for achieving higher levels of vehicle autonomy. Harnessing the power of artificial intelligence (AI) and machine learning (ML) is pivotal in developing such advanced decision-making systems. These techniques offer promising solutions for tackling key challenges in autonomous driving systems, including scalability, adaptability, and alignment with human driver behaviors. This thesis presents a comprehensive framework and a series of learning-based methodologies for decision-making in AVs, with the objective of improving the scalability, adaptability, and alignment of their decision-making systems. The learning-enabled decision-making framework consists of three essential modules: world model, actor, and cost. These modules work collaboratively to provide a unified solution for autonomous machine intelligence and can be learned individually or jointly from human driving data. The methodologies proposed in this thesis focus on three key areas: (1) creating ML-driven actors and cost functions that align with human values, (2) developing ML-based world models capable of accurately forecasting complex human interactions, and (3) designing ML-based decision-making architectures that seamlessly integrate these components. Initially, we explore each module individually. For the actor module, we propose an approach based on reinforcement learning (RL) that incorporates human prior knowledge into the training process. By combining imitation learning (IL) from human expert demonstrations with RL, a novel learning framework is developed to align agent behavior with human expectations. The results show improved sample efficiency, enabling the agent to navigate complex urban environments safely, while also generating distinct driving styles that closely resemble human driving behaviors. In the cost module, we focus on learning personalized cost functions underlying human driving behaviors using inverse reinforcement learning (IRL). We propose a maximum entropy IRL algorithm that infers the reward function based on a structural assumption about discrete latent intentions guiding continuous low-level control actions. The personalized cost learning method outperforms general cost modeling methods, leading to a more human-like driving experience. Regarding the world model module, the goal is to accurately represent the complexity and dynamics of human behaviors in real-world driving environments. We introduce a joint multi-agent prediction model based on Transformers, leveraging hierarchical game theory to model interactions among agents in driving scenarios. The model employs a novel Transformer decoder structure and a learning process that regulates agent behavior to respond to other agents' behaviors. Extensive experiments validate the model's state-of-the-art prediction accuracy and its capability to jointly reason about ego agent motion plans and other agents' behaviors. Following individual assessments, unified predictive decision-making architectures are presented, harmonizing the three modules. A comprehensive behavior planning framework that integrates all three core modules is proposed. It generates diverse trajectory proposals, forecasts multi-modal futures for other agents, and evaluates candidate plans using a learned cost function. Importantly, this framework includes a conditional motion predictor (CMP) that forecasts other agents' responses to the potential actions of the AV. The CMP improves prediction accuracy and facilitates the selection of human-like behaviors. Moreover, we propose a differentiable and integrated prediction-planning framework that enables end-to-end training of the entire decision-making system. The framework incorporates a differentiable nonlinear optimizer as the motion planner, allowing the gradients to flow between the three core modules and directly optimizing the planning performance. The experimental results demonstrate superior performance in both open-loop and closed-loop testing, effectively handling complex urban driving scenarios. This thesis not only makes significant contributions to the advancement of autonomous driving systems but also lays the foundation for future research in learning-based decision-making for a wide range of intelligent systems. It paves the way for safe and human-centered autonomy, while driving the evolution of intelligent transportation systems. Doctor of Philosophy 2023-12-26T06:43:43Z 2023-12-26T06:43:43Z 2023 Thesis-Doctor of Philosophy Huang, Z. (2023). Learning-enabled decision-making for autonomous driving: framework and methodology. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/172842 https://hdl.handle.net/10356/172842 10.32657/10356/172842 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University