Abstract
Reinforcement learning is a machine learning framework that an agent repeatedly observes its reward from environment as a result of action to choose with high reward expectations. On the other hand, mobility models that satisfy statistical properties of human mobility patterns have been proposed. However, there is little study to consider a reasonable method of generating a mobility model based on reinforcement learning. In this paper, we proposed a method for generating a mobility model using reinforcement learning to earn rewards most efficiently. A dataset containing mobility traces of taxi cabs in San Francisco was used to propose the method where each taxi agent learns its actions selected by giving rewards, such as passenger fare, gasoline cost, and the distribution of the positions of past passengers. The \(\epsilon \)-greedy method was used to consider the method to perform reinforcement learning. The degree of learning was compared by changing the exploration parameter \(\epsilon \). As a result, it was found that the cumulative reward was the highest when \(\epsilon =0\), which is different from usual results in \(\epsilon \)-greedy method. This result come from an implicit exploration of taxi’s mobility patterns where taxi agents explored the places where many passengers can be found during the transfer of passengers.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Song, C., Qu, Z., Blumm, N., Barabási, A.-L.: Limits of predictability in human mobility. Science 327(5968), 1018–1021 (2010)
Pappalardo, L., Simini, F., Rinzivillo, S., Pedreschi, D., Giannotti, F., Barabási, A.-L.: Returners and explorers dichotomy in human mobility. Nat. Commun. 6, 8166 (2015)
González, M.C., Hidalgo, C.A., Barabási, A.-L.: Understanding individual human mobility patterns. Nature 453(7196), 779–782 (2008)
Rhee, I., Shin, M., Hong, S., Lee, K., Kim, S.J., Chong, S.: On the lévy-walk nature of human mobility. IEEE/ACM Trans. Netw. 19(3), 630–643 (2011)
Zaburdaev, V., Denisov, S., Klafter, J.: Lévy walks. Rev. Mod. Phys. 87(483), 483–530 (2015)
Song, C., Koren, T., Wang, P., Barabási, A.-L.: Modeling the scaling properties of human mobility. Nat. Phys. 6(10), 818–823 (2010)
Evans, M.R., Majumdar, S.N.: Diffusion with stochastic resetting. Phys. Rev. Lett. 106, 160601 (2011)
Fujihara, A., Miwa, H.: Homesick lévy walk and optimal forwarding criterion of utility-based routing under sequential encounters. Internet Things Inter-cooperative Computational Technologies for Collective Intelligence, vol. 460, pp. 207–231. Springer (2013)
Fujihara, A., Miwa, H.: Homesick lévy walk: a mobility model having Ichi-Go Ichi-e and scale-free properties of human encounters. In: 2014 IEEE 38th Annual International Computers, Software and Applications Conference, pp. 576–583. Springer (2014)
Fujihara, A.: Analyzing scale-free property on human serendipitous encounters using mobile phone data. In: MoMM 2015: Proceedings of the 13th International Conference on Advances in Mobile Computing and Multimedia, pp. 122–125. ACM (2015)
Sudo, A., et al.: Particle filter for real-time human mobility prediction following unprecedented disaster. In: Proceedings of the 24th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, vol. 5, pp. 1–10 (2016)
Pappalardo, L., Simini, F.: Data-driven generation of spatio-temporal routines in human mobility. Data Min Knowl. Disc. 32, 787–829 (2018). https://github.com/jonpappalord/DITRAS
Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. The MIT Press, Cambridge (2018)
OpenStreetMap. https://www.openstreetmap.org/ Geofabrik Download Server. https://download.geofabrik.de/north-america/us/california/norcal.html
SUMO: Simulation of Urban MObility. http://sumo.sourceforge.net/
Piorkowski, M., Sarafijanovic-Djukic, N., Grossglauser, M.: CRAWDAD dataset epfl/mobility (v.2009-02-24), traceset: cab (2009). https://crawdad.org/epfl/mobility/20090224/cab
Yellow Cab SF Cab Fares. https://yellowcabsf.com/service/cab-fares/
Bando, M., Hasebe, K., Nakayama, A., Shibata, A., Sugiyama, Y.: Structure stability of congestion in traffic dynamics. Jpn. J. Ind. Appl. Math. 11, 203 (1994). http://traffic.phys.cs.is.nagoya-u.ac.jp/mstf/sample/ov.html
Fujihara, A., Miwa, H.: Real-time disaster evacuation guidance using opportunistic communication. In: IEEE/IPSJ-SAINT 2012, pp. 326–331 (2012)
Fujihara, A., Miwa, H.: Effect of traffic volume in real-time disaster evacuation guidance using opportunistic communications. In: IEEE-INCoS-2012, pp. 457–462 (2012)
Fujihara, A., Miwa, H.: On the use of congestion information for rerouting in the disaster evacuation guidance using opportunistic communication. In: ADMNET 2013, pp. 563–568 (2013)
Fujihara, A., Miwa, H.: Disaster evacuation guidance using opportunistic communication: the potential for opportunity-based service. In: Big Data and Internet of Things: A Roadmap for Smart Environments, Studies in Computational Intelligence, vol. 546, pp. 425–446 (2014)
Fujihara, A., Miwa, H.: Necessary condition for self-organized follow-me evacuation guidance using opportunistic networking. In: INCoS 2014, pp. 213–220 (2014)
Acknowledgements
This work was partially supported by the Japan Society for the Promotion of Science (JSPS) through KAKENHI (Grants-in-Aid for Scientific Research) Grant Numbers 17K00141, 17H01742, and 20K11797.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Iwai, Y., Fujihara, A. (2021). Considering a Method for Generating Human Mobility Model by Reinforcement Learning. In: Barolli, L., Li, K., Miwa, H. (eds) Advances in Intelligent Networking and Collaborative Systems. INCoS 2020. Advances in Intelligent Systems and Computing, vol 1263. Springer, Cham. https://doi.org/10.1007/978-3-030-57796-4_12
Download citation
DOI: https://doi.org/10.1007/978-3-030-57796-4_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-57795-7
Online ISBN: 978-3-030-57796-4
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)