PhD Qualifying-Exam

Algorithmic Perspectives on Efficient Fine-tuning of Large Language Models: A Survey

The Hong Kong University of Science and Technology (Guangzhou)

Data Science and Analytics Thrust

PhD Qualifying Examination

By Miss CHEN, Minping

Abstract

In recent years, the evolution of artificial intelligence has been marked by the rise of universal foundation models, with large language models (LLMs) leading the way. These models have demonstrated remarkable performance across diverse natural language tasks, including document understanding, complex reasoning, and code generation, et al. Despite their strong generalization abilities enabled by zero-shot learning, fine-tuning remains crucial for adapting LLMs to complex, domain-specific tasks. However, as model sizes grow, the resource demands of fine-tuning present significant challenges. This survey provides a comprehensive review of recent advances in efficient LLM fine-tuning, categorizing existing approaches into parameter efficient fine-tuning (PEFT) and memory-efficient fine-tuning. PEFT techniques aim to update a small subset of model parameters, thereby reducing memory overhead for gradient storage, while memory-efficient methods optimize memory usage across model weights, activations, and gradients. In addition, we introduce two novel methods: the Low-order Hybrid Optimizer (LoHO) and Low-Rank based Efficient Mask Learning (LoReML). LoHO mitigates the accuracy degradation and slow convergence typically associated with zeroth-order optimization, while LoReML eliminates the need for manual parameter selection and reduces the high cost of mask learning in existing masking-based methods. The experiments validate the effectiveness of our proposed approaches. Finally, we conclude with a discussion of future directions in the pursuit of efficient fine-tuning for LLMs.

PQE Committee

Chair of Committee: Prof. LUO Qiong

Prime Supervisor: Prof. WEN Zeyi

Co-Supervisor: Prof. TANG Guoming

Examiner: Prof. DING Ningning

Date

09 June 2025

Time

15:00:00 - 00:00:00