TY - GEN
T1 - Intercity Simulation of Human Mobility at Rare Events via Reinforcement Learning
AU - Pang, Yanbo
AU - Tsubouchi, Kota
AU - Yabe, Takahiro
AU - Sekimoto, Yoshihide
N1 - Publisher Copyright:
© 2020 ACM.
PY - 2020/11/3
Y1 - 2020/11/3
N2 - Agent-based simulations, combined with large scale mobility data, have been an effective method for understanding urban scale human dynamics. However, collecting such large scale human mobility datasets are especially difficult during rare events (e.g., natural disasters), reducing the performance of agent-based simulations. To tackle this problem, we develop an agent-based model that can simulate urban dynamics during rare events by learning from other cities using inverse reinforcement learning. More specifically, in our framework, agents imitate real human-beings' travel behavior from areas where rare events have occurred in the past (source area) and produce synthetic people movement in different cities where such rare events have never occurred (target area). Our framework contains three main stages: 1) recovering the reward function, where the people's travel patterns and preferences are learned from the source areas; 2) transferring the model of the source area to the target areas; 3) simulating the people movement based on learned model in the target area. We apply our approach in various cities for both normal and rare situations using real-world GPS data collected from more than 1 million people in Japan, and show higher simulation performance than previous models.
AB - Agent-based simulations, combined with large scale mobility data, have been an effective method for understanding urban scale human dynamics. However, collecting such large scale human mobility datasets are especially difficult during rare events (e.g., natural disasters), reducing the performance of agent-based simulations. To tackle this problem, we develop an agent-based model that can simulate urban dynamics during rare events by learning from other cities using inverse reinforcement learning. More specifically, in our framework, agents imitate real human-beings' travel behavior from areas where rare events have occurred in the past (source area) and produce synthetic people movement in different cities where such rare events have never occurred (target area). Our framework contains three main stages: 1) recovering the reward function, where the people's travel patterns and preferences are learned from the source areas; 2) transferring the model of the source area to the target areas; 3) simulating the people movement based on learned model in the target area. We apply our approach in various cities for both normal and rare situations using real-world GPS data collected from more than 1 million people in Japan, and show higher simulation performance than previous models.
KW - Human Mobility
KW - People Flow Simulation
KW - Reinforcement Learning
KW - Urban Computing
UR - http://www.scopus.com/inward/record.url?scp=85097275220&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85097275220&partnerID=8YFLogxK
U2 - 10.1145/3397536.3422244
DO - 10.1145/3397536.3422244
M3 - Conference contribution
AN - SCOPUS:85097275220
T3 - GIS: Proceedings of the ACM International Symposium on Advances in Geographic Information Systems
SP - 293
EP - 302
BT - Proceedings of the 28th International Conference on Advances in Geographic Information Systems, SIGSPATIAL GIS 2020
A2 - Lu, Chang-Tien
A2 - Wang, Fusheng
A2 - Trajcevski, Goce
A2 - Huang, Yan
A2 - Newsam, Shawn
A2 - Xiong, Li
PB - Association for Computing Machinery
T2 - 28th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, SIGSPATIAL GIS 2020
Y2 - 3 November 2020 through 6 November 2020
ER -