R.I.P.
π»
Ghosted
Robust Generalization with Adaptive Optimal Transport Priors for Decision-Focused Learning
February 01, 2026 Β· Grace Period Β· π The 29th International Conference on Artificial Intelligence and Statistics (AISTATS), 2026
Authors
Haixiang Sun, Andrew L. Liu
arXiv ID
2602.01427
Category
stat.ML: Machine Learning (Stat)
Cross-listed
cs.LG,
stat.AP
Citations
0
Venue
The 29th International Conference on Artificial Intelligence and Statistics (AISTATS), 2026
Abstract
Few-shot learning requires models to generalize under limited supervision while remaining robust to distribution shifts. Existing Sinkhorn Distributionally Robust Optimization (DRO) methods provide theoretical guarantees but rely on a fixed reference distribution, which limits their adaptability. We propose a Prototype-Guided Distributionally Robust Optimization (PG-DRO) framework that learns class-adaptive priors from abundant base data via hierarchical optimal transport and embeds them into the Sinkhorn DRO formulation. This design enables few-shot information to be organically integrated into producing class-specific robust decisions that are both theoretically grounded and efficient, and further aligns the uncertainty set with transferable structural knowledge. Experiments show that PG-DRO achieves stronger robust generalization in few-shot scenarios, outperforming both standard learners and DRO baselines.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β Machine Learning (Stat)
R.I.P.
π»
Ghosted
Distilling the Knowledge in a Neural Network
R.I.P.
π»
Ghosted
Layer Normalization
R.I.P.
π»
Ghosted
Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning
R.I.P.
π»
Ghosted
Domain-Adversarial Training of Neural Networks
R.I.P.
π»
Ghosted