TY - GEN
T1 - A Theory of Learning with Competing Objectives and User Feedback
AU - Awasthi, Pranjal
AU - Cortes, Corinna
AU - Mansour, Yishay
AU - Mohri, Mehryar
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.
PY - 2024
Y1 - 2024
N2 - Large-scale deployed learning systems are often evaluated along multiple objectives or criteria. But, how can we learn or optimize such complex systems, with potentially conflicting or even incompatible objectives? How can we improve the system when user feedback becomes available, feedback possibly alerting to issues not previously optimized for by the system? We present a new theoretical model for learning and optimizing such complex systems. Rather than committing to a static or pre-defined tradeoff for the multiple objectives, our model is guided by the feedback received, which is used to update its internal state. Our model supports multiple objectives that can be of very general form and takes into account their potential incompatibilities. We consider both a stochastic and an adversarial setting. In the stochastic setting, we show that our framework can be naturally cast as a Markov Decision Process with stochastic losses, for which we give efficient vanishing regret algorithmic solutions. In the adversarial setting, we design efficient algorithms with competitive ratio guarantees. We also report the results of experiments with our stochastic algorithms validating their effectiveness.
AB - Large-scale deployed learning systems are often evaluated along multiple objectives or criteria. But, how can we learn or optimize such complex systems, with potentially conflicting or even incompatible objectives? How can we improve the system when user feedback becomes available, feedback possibly alerting to issues not previously optimized for by the system? We present a new theoretical model for learning and optimizing such complex systems. Rather than committing to a static or pre-defined tradeoff for the multiple objectives, our model is guided by the feedback received, which is used to update its internal state. Our model supports multiple objectives that can be of very general form and takes into account their potential incompatibilities. We consider both a stochastic and an adversarial setting. In the stochastic setting, we show that our framework can be naturally cast as a Markov Decision Process with stochastic losses, for which we give efficient vanishing regret algorithmic solutions. In the adversarial setting, we design efficient algorithms with competitive ratio guarantees. We also report the results of experiments with our stochastic algorithms validating their effectiveness.
KW - ML fairness
KW - Multiobjective optimization
KW - Online learning
UR - http://www.scopus.com/inward/record.url?scp=85200686215&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85200686215&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-63735-3_2
DO - 10.1007/978-3-031-63735-3_2
M3 - Conference contribution
AN - SCOPUS:85200686215
SN - 9783031637346
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 10
EP - 49
BT - Artificial Intelligence and Image Analysis - 18th International Symposium on Artificial Intelligence and Mathematics, ISAIM 2024, and 22nd International Workshop on Combinatorial Image Analysis, IWCIA 2024, Revised Selected Papers
A2 - Barneva, Reneta P.
A2 - Brimkov, Valentin E.
A2 - Brimkov, Valentin E.
A2 - Gentile, Claudio
A2 - Pacchiano, Aldo
PB - Springer Science and Business Media Deutschland GmbH
T2 - 18th International Symposium on Artificial Intelligence and Mathematics, ISAIM 2024, and 22nd International Workshop on Combinatorial Image Analysis, IWCIA 2024
Y2 - 8 January 2024 through 10 January 2024
ER -