Skip to main navigation Skip to search Skip to main content

The impacts of over-parameterization under distribution shift

  • Yifan HAO

Student thesis: Master's thesis

Abstract

In recent years, machine learning models have achieved success based on the independently and identically distributed (IID) assumption. However, this assumption can be easily violated in real-world applications, leading to both adversarial attack and the Out-of-Distribution (OOD) problem. Understanding how modern over-parameterized models behave under non-trivial distributional shifts is essential, as current theoretical understanding is insufficient. Here we investigate the performances of overparameterized model in terms of both adversarial robustness and OOD generalization under the commonly-adopted "benign overfitting" conditions [1]. Our results contains two parts: (i) for adversarial robustness, we prove that even if the ground truth itself is robust to adversarial examples, and the benignly overfitted model is benign in terms of the "standard" out-of-sample risk objective, this benign overfitting process can be harmful when out-of-sample data are subject to adversarial manipulation; (ii) for OOD generalization, we focus on a scenario where the benign overfitting estimators demonstrate a constant excess OOD loss, despite achieving zero excess in-distribution (ID) loss, and prove that further increasing the model’s parameterization can significantly reduce the OOD loss; (iii) we further show that model ensembles can also enhance the OOD testing loss, achieving a similar effect to increasing the model capacity. These results offer theoretical insights into the intriguing empirical phenomenon in "benign overfitting" regime.
Date of Award2024
Original languageEnglish
Awarding Institution
  • The Hong Kong University of Science and Technology
SupervisorKani CHEN (Supervisor)

Cite this

'