Causal view of generalization

Causal reasoning, an essential cognitive ability in human intelligence, allows us to generalize past learning to solve present problems. Unfortunately, while machine learning prospers over the past decade by training powerful deep neural networks (DNN) on massive data, it still lacks the generalizat...

Full description

Saved in:
Bibliographic Details
Main Author: Yue, Zhongqi
Other Authors: Hanwang Zhang
Format: Thesis-Doctor of Philosophy
Language:English
Published: Nanyang Technological University 2023
Subjects:
Online Access:https://hdl.handle.net/10356/172269
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:Causal reasoning, an essential cognitive ability in human intelligence, allows us to generalize past learning to solve present problems. Unfortunately, while machine learning prospers over the past decade by training powerful deep neural networks (DNN) on massive data, it still lacks the generalization ability like us humans. Inspired by the important role of causality in human generalization, we take a causal view of machine generalization. We reveal that the spurious correlation in the training data is a confounder that prevents generalization, which can only be eliminated by causal intervention. In this thesis, we study three categories of causal intervention and contribute practical implementations to improve generalization: 1) backdoor adjustment, 2) invariant learning, and 3) learning disentangled representation. The proposed practical implementations are extensively evaluated by standard benchmarks and demonstrate state-of-the-art generalization performance in Few-Shot Learning, Unsupervised Domain Adaptation, Semi-Supervised Learning, Zero-Shot Learning, Open-Set Recognition, and Unsupervised Representation Learning.