Lean evolutionary machine learning by multitasking simpler and hard tasks

Many decisions in a machine learning (ML) pipeline involve non-differentiable and discontinuous objectives and search spaces. Examples include feature selection, model selection, and reinforcement learning where candidate solutions must be evaluated via a learning subsystem or through interactions w...

Full description

Saved in:
Bibliographic Details
Main Author: Zhang, Nick Shihui
Other Authors: Ong Yew Soon
Format: Thesis-Doctor of Philosophy
Language:English
Published: Nanyang Technological University 2023
Subjects:
Online Access:https://hdl.handle.net/10356/164300
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:Many decisions in a machine learning (ML) pipeline involve non-differentiable and discontinuous objectives and search spaces. Examples include feature selection, model selection, and reinforcement learning where candidate solutions must be evaluated via a learning subsystem or through interactions with complex environments. Evolutionary algorithms (EAs), a simple and largely problem-independent approach, are prominent gradient-free methods to handle such tasks. However, the sample complexity of iteratively evaluating populations of candidate solutions poses a steep challenge in terms of computational cost, especially when the learning subsystem is computationally expensive. Given the increasing volume of data in recent years, the resource needed to train machine learning models is rising steadily. As such, the tractability issue marks a barrier to the widespread adoption of evolutionary computation (EC) for ML, a subject that has sparked significant research interest in recent years. To this end, an emerging computational paradigm known as evolutionary multi-tasking (EMT) is making inroads towards resolving the computational tractability issue through the simultaneous handling of multiple optimization tasks in a single evolutionary run. The exploitation of shared information across tasks often leads to accelerated global convergence since the performant solutions found by similar tasks could serve as informative priors that rapidly and favourably bias the main task’s search. As opposed to prior works that often fall back on parallel computing hardware to resolve this big data problem of EAs, we aim for a software-centric solution. In particular, we first propose a unified probabilistic formulation for multitasking in evolutionary machine learning (EML) and then design algorithmic realizations in different contexts to achieve accelerated convergence characteristics. In this thesis, the basic premise is to achieve lean EML by concurrently evolving and transferring useful information from simpler to hard tasks. The EMT framework leverages the general idea of using a number of artificially generated (computationally cheaper) auxiliary tasks in a single multitask setting, with the objective of boosting evolutionary search on the main (target) task of interest. The unified probabilistic formulation for EMT serves to enable effective exploitation of transferable information while simultaneously guarding against threats of negative transfer. In addition, a novel strategy is used to allocate computational resources (i.e., size of the evolving population) to each of the tasks in a principled manner, minimizing the wastage of computational resources on tasks that are not useful to the target. Algorithmic realizations of the general idea are designed and implemented for two different domains of ML, namely, evolutionary feature selection and evolutionary reinforcement learning. Extensive experimental studies have been conducted to verify the efficacy of our proposed framework and algorithms. The results demonstrate that the proposed methods can achieve a competitive level of performance as existing algorithms while performing substantially fewer sample evaluations, thereby approaching the goal of lean EML.