Unlocking LLM Potential: Investigating Graph Problems and Mathematical Reasoning
The Hong Kong University of Science and Technology (Guangzhou)
Data Science and Analytics Thrust
PhD Thesis Examination
By Mr. Nuo CHEN
ABSTRACT
Large language models (LLMs) excel at pattern matching but still struggle with rigorous reasoning. To systematically unlock their latent potential, we address key challenges in mathematical and graph-based reasoning, including: enabling large language models (LLMs) to perform self-reflection and calibration for better problem-solving, overcoming data scarcity, leveraging reinforcement learning to enhance reasoning capabilities, and bridging graph problem reasoning with general reasoning to improve overall LLM performance.
First, we propose IMR-TIP (Improving Math Reasoning with Tool-augmented Interleaf Prompting), an interactive reasoning framework that empowers LLMs to self-reflect and invoke external tools for solving mathematical problems. We further introduce MathOctopus, a multilingual mathematical reasoning model capable of addressing problems in over ten languages. To enhance generalization, we design a controllable math data generation method, significantly improving LLMs’ performance in mathematical tasks.
Having stabilised mathematical reasoning, we turn to structure as a catalyst for general reasoning. Graphs offer a universal abstraction, so we build GraphWiz, the first LLM fine-tuned specifically on graph computational problems. Its success motivates a larger vision: if graphs embody relations underlying many tasks, then pre-training on rich graph reasoning should lift performance everywhere. We curate GraphPile, a large, diverse corpus of graph-centric problems and tutorials, and pre-train GraphMind. Empirically, GraphMind transfers its relational bias to out-of-domain benchmarks, outperforming vanilla LLMs across arithmetic, commonsense, and program synthesis. Finally, we close the loop at the reward level.We propose Graph-PRM, a process-level reward model that grades each reasoning step by graph-problem correctness. Reinforcement learning with GraphPRM further sharpens both mathematical and graph reasoning, confirming our hypothesis that structured rewards amplify structured pre-training.
TEC
Chairperson: Prof Xin WANG
Prime Supervisor: Prof Jia LI
Co-Supervisor: Prof Yangqiu SONG
Examiners:
Prof Xiaowen CHU
Prof Wenjia WANG
Prof Enyan DAI
Prof Xu CHEN
Date
05 June 2025
Time
14:00:00 - 16:00:00
Location
E1-202, HKUST(GZ)
Join Link
Zoom Meeting ID: 968 3773 6868
Passcode: dsa2025
Event Organizer
Data Science and Analytics Thrust
dsarpg@hkust-gz.edu.cn