Nonparametric Methods for Stochastic Systems: Model Calibration and Online Decision-Making
The Hong Kong University of Science and Technology (Guangzhou)
数据科学与分析学域
PhD Thesis Examination
By Ms. Qingwen ZHANG
摘要
Modern stochastic systems, from computer simulations to real-time decision platforms, increasingly rely on flexible statistical methods that can adapt to complex, high-dimensional data without restrictive parametric assumptions. Nonparametric approaches have emerged as powerful tools for such challenges, offering robustness and scalability. This dissertation focuses on two problems in this domain: (1) improving model calibration for imperfect computer simulations, and (2) developing robust and efficient methods for online decision-making with high-dimensional covariates and fairness constraints.
For model calibration of imperfect computer simulations, we propose Sobolev calibration, a flexible nonparametric framework that minimizes discrepancies measured by reproducing kernel Hilbert space (RKHS) norm. Our approach addresses the potential overfitting problem in existing methods while establishing a novel theoretical connection between the classic L2 calibration and Kennedy-O’Hagan’s calibration. The proposed method achieves fast convergence rate, asymptotic normality, and semiparametric efficiency, and is supported by empirical validation.
In online decision-making, we tackle two challenges in contextual bandits. First, we propose an algorithm for non-parametric and high-dimensional settings using sparse additive models, which achieves sublinear regret with only polylogarithmic dependence on covariate dimensionality. This is the first such result for nonparametric contextual bandits. We also establish a lower bound, with the gap to the upper bound vanishing as smoothness increases. Second, we introduce a uniform δ-fairness constraint and design contextual bandit algorithms for linear and smooth reward functions to ensure equitable treatment among competing arms while maintaining decision efficiency. Our fair algorithms achieve near-optimal regret bounds in both parametric and nonparametric settings, representing the first systematic treatment of fairness in adversarial environments. Crucially, we expose the fragility of standard fair algorithms, proving that even Oe(1) reward corruption can induce persistent unfairness and Ω(T) regret. To address this, we develop robust variants that preserve fairness guarantees under attacks, with tight minimax regret bounds that adapt to corruption budgets.
The thesis makes three primary contributions: (1) a theoretically-grounded calibration framework that improves existing approaches; (2) novel bandit algorithms that overcome dimensionality challenges under nonparametric modeling; and (3) the first attack-resistant fairness guarantees for contextual bandits in both parametric and nonparametric regimes. These advancements are validated through rigorous theoretical analysis and extensive numerical experiments.
TEC
Chairperson: Prof Gareth John TYSON
Prime Supervisor: Prof Wenjia WANG
Co-Supervisor: Prof Yuan YAO
Examiners:
Prof Lei LI
Prof Zeyi WEN
Prof Xiaowei ZHANG
Prof Jun LUO
日期
15 August 2025
时间
15:00:00 - 17:00:00
地点
E3-201, HKUST(GZ)
Join Link
Zoom Meeting ID: 953 4380 9695
Passcode: dsa2025
主办方
Data Science and Analytics
联系邮箱
dsarpg@hkust-gz.edu.cn