Efficient algorithms for distributionally robust optimization and its applications
Abstract
Details
- Title: Subtitle
- Efficient algorithms for distributionally robust optimization and its applications
- Creators
- Qi Qi
- Contributors
- Tianbao Yang (Advisor)Bijaya Adhikari (Committee Member)Peng P Jiang (Committee Member)Octav O Chipara (Committee Member)Qihang Q Lin (Committee Member)
- Resource Type
- Dissertation
- Degree Awarded
- Doctor of Philosophy (PhD), University of Iowa
- Degree in
- Computer Science
- Date degree season
- Autumn 2023
- Publisher
- University of Iowa
- DOI
- 10.25820/etd.006942
- Number of pages
- xi, 218 pages
- Copyright
- Copyright 2023 Qi Qi
- Language
- English
- Date submitted
- 12/03/2023
- Description illustrations
- Illustrations, tables, graphs, charts
- Description bibliographic
- Includes bibliographical references (pages 194-218).
- Public Abstract (ETD)
Distributionally Robust Optimization (DRO) is a cutting-edge approach in machine learning, enhancing model robustness by accounting for data uncertainty without relying on known probability distributions. Its effectiveness in tackling challenges like data imbalance, label noise, and fairness has attracted significant attention. However, existing algorithms for DRO optimization have limitations that hinder their scalability and efficiency.
This thesis focuses on developing efficient stochastic optimization algorithms for DRO problems and their real-world applications. From a theoretical standpoint, we introduce:
- RECOVER: an online variance reduction stochastic algorithm designed to address KL-divergence regularized DRO objectives, achieving optimal sample complexity for non-convex objectives.
- ABSGD: a moving average stochastic algorithm tailored for efficiently solving largescale data imbalanced and label noisy classification tasks, showcasing success in a prominent competition without incurring extra labeling costs.
- SCDRO: a stochastic compositional algorithm for solving constrained DRO objectives, demonstrating near-optimal complexity bounds for non-convex losses and optimal complexity for convex losses.
In terms of applications, beyond comprehensive studies on data imbalanced problems in long-tail image classification, we present a simple and effective DRO framework for addressing pair imbalance in deep metric learning. Furthermore, as an extension of our work on data imbalance, we propose a technique for directly optimizing the Area Under Precision-Recall Curves (AUPRC) metric for large-scale imbalance medical image problems.
- Academic Unit
- Computer Science
- Record Identifier
- 9984546943702771