ScaleBiO: Scalable Bilevel Optimization for LLM Data Reweighting

Rui PAN, Dylan Zhang, Hanning Zhang, Xingyuan Pan, Minrui XU, Jipeng ZHANG, Renjie PI, Xiaoyu Wang, Tong ZHANG

Research output: Contribution to conferenceConference Paperpeer-review

Abstract

Bilevel optimization has shown its utility across various machine learning settings, yet most algorithms in practice require second-order information, making it challenging to scale them up. Only recently, a paradigm of first-order algorithms has emerged in the theoretical literature, capable of effectively addressing bilevel optimization problems. Nevertheless, the practical efficiency of this paradigm remains unverified, particularly in the context of large language models (LLMs). This paper introduces the first scalable instantiation of this paradigm called ScaleBiO, focusing on bilevel optimization for large-scale LLM data reweighting. By combining with a recently proposed memory-efficient training technique called LISA, our novel algorithm allows the paradigm to scale to 30B-sized LLMs on H100 GPUs, marking the first successful application of bilevel optimization under practical scenarios for large-sized LLMs. Empirically, extensive experiments on data reweighting verify the effectiveness of ScaleBiO for different-scaled models, including Llama-3-8B, Gemma-2-9B, Qwen-2-7B, and Qwen-2.5-32B, where bilevel optimization succeeds in instruction-following and math reasoning tasks, outperforming several popular baselines, including uniform sampling, influence-aware data filtering, and reference-model-based sampling methods. Theoretically, ScaleBiO ensures the optimality of the learned data weights, along with a convergence guarantee matching the conventional first-order bilevel optimization paradigm on smooth and strongly convex objectives.
Original languageEnglish
Publication statusPublished - Jul 2025
EventThe 63rd Annual Meeting of the Association for Computational Linguistics - Vienna, Austria
Duration: 27 Jul 20251 Aug 2025

Conference

ConferenceThe 63rd Annual Meeting of the Association for Computational Linguistics
Country/TerritoryAustria
CityVienna
Period27/07/251/08/25

Fingerprint

Dive into the research topics of 'ScaleBiO: Scalable Bilevel Optimization for LLM Data Reweighting'. Together they form a unique fingerprint.

Cite this