TOWARDS EFFICIENT NEURAL NETWORK TRAINING: DATA AUGMENTATION AND DISTILLATION
The Hong Kong University of Science and Technology (Guangzhou)
Data Science and Analytics Thrust
PhD Thesis Proposal Examination
By Mr. Jiahang JIANG
Abstract
Deep learning has achieved remarkable success in the past decade and has become the primary choice in various research fields. In addition to optimizing the structure of deep learning models, researchers are increasingly paying attention to directly manipulating the original dataset, and many methods have been proposed for the dataset to enhance prediction accuracy, training efficiency, and generalization. In this paper, we specifically investigate the influence of data augmentation and distillation techniques, with the goal of improving both theoretical understanding and experimental performance.
TPE Committee
Chairperson: Prof. Fugee TSUNG
Prime Supervisor: Prof Wenjia WANG
Co-Supervisor: Prof Jia LI
Examiner: Prof Xinlei HE
Date
13 June 2024
Time
14:50:00 - 16:05:00
Location
E1-149