TY - JOUR
T1 - The NeurIPS 2022 Neural MMO Challenge
T2 - 36th Annual Conference on Neural Information Processing Systems - Competition Track, NeurIPS 2022
AU - Liu, Enhong
AU - Suarez, Joseph
AU - You, Chenhui
AU - Wu, Bo
AU - Chen, Bingcheng
AU - Hu, Jun
AU - Chen, Jiaxin
AU - Zhu, Xiaolong
AU - Zhu, Clare
AU - Togelius, Julian
AU - Mohanty, Sharada
AU - Hong, Weijun
AU - Du, Rui
AU - Zhang, Yibing
AU - Wang, Qinwen
AU - Li, Xinhang
AU - Yuan, Zheng
AU - Li, Xiang
AU - Huang, Yuejia
AU - Zhang, Kun
AU - Yang, Hanhui
AU - Tang, Shiqi
AU - Isola, Phillip
N1 - Publisher Copyright:
© 2023 E. Liu et al.
PY - 2023
Y1 - 2023
N2 - In this paper, we present the results of the NeurIPS-2022 Neural MMO Challenge, which attracted 500 participants and received over 1,600 submissions. Like the previous IJCAI-2022 Neural MMO Challenge, it involved agents from 16 populations surviving in procedurally generated worlds by collecting resources and defeating opponents. This year’s competition runs on the latest v1.6 Neural MMO, which introduces new equipment, combat, trading, and a better scoring system. These elements combine to pose additional robustness and generalization challenges not present in previous competitions. This paper summarizes the design and results of the challenge, explores the potential of this environment as a benchmark for learning methods, and presents some practical reinforcement learning training approaches for complex tasks with sparse rewards. Additionally, we have open-sourced1 our baselines, including environment wrappers, benchmarks, and visualization tools for future research.
AB - In this paper, we present the results of the NeurIPS-2022 Neural MMO Challenge, which attracted 500 participants and received over 1,600 submissions. Like the previous IJCAI-2022 Neural MMO Challenge, it involved agents from 16 populations surviving in procedurally generated worlds by collecting resources and defeating opponents. This year’s competition runs on the latest v1.6 Neural MMO, which introduces new equipment, combat, trading, and a better scoring system. These elements combine to pose additional robustness and generalization challenges not present in previous competitions. This paper summarizes the design and results of the challenge, explores the potential of this environment as a benchmark for learning methods, and presents some practical reinforcement learning training approaches for complex tasks with sparse rewards. Additionally, we have open-sourced1 our baselines, including environment wrappers, benchmarks, and visualization tools for future research.
UR - http://www.scopus.com/inward/record.url?scp=85179138583&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85179138583&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85179138583
SN - 2640-3498
VL - 220
SP - 18
EP - 34
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
Y2 - 28 November 2022 through 9 December 2022
ER -