DSA & AIoT Joint Seminar

Knowledge Distillation: Towards Efficient and Compact Neural Networks

Deep neural networks usually have a large number of parameters, which makes them powerful in representation learning, but also limits their deployment in real-world application. Knowledge distillation, which aims to transfer the knowledge from an over-parameterized teacher model to an efficient student model, has become one of the most popular method for neural network compression and acceleration. In this report, I will introduce my works on knowledge distillation in the last five years, which mainly focus on two aspects: the fundamental problems in knowledge distillation, and how to apply knowledge distillation to more challenging tasks. Besides, I will introduce the challenge and opportunities of AI model compression in the decade of large models.

Linfeng ZHANG

Tsinghua University

Linfeng Zhang is a Ph.D. candidate in Tsinghua University. Before that, he obtained his bachelor degree in Northeastern University. Linfeng has been awarded with Microsoft Fellowship in 2020 to encourage his work in knowledge distillation and AI model compression. He has published 13 papers in top-tier conference and journals as the first author. Besides, his research results in AI model compression have been used in many companies such as Huawei, Kwai, DIDI, Polar Bear Technology, Intel and so on.

Date

23 January 2024

Time

09:30:00 - 10:30:00

Location

E3-2F-202, HKUST(GZ)

Join Link

Zoom Meeting ID:
842 2904 7286


Passcode: iott

Event Organizer

Data Science and Analytics Thrust
Internet of Things Thrust

Email

dsat@hkust-gz.edu.cn
iott@hkust-gz.edu.cn