Abstract
Human Activity Recognition (HAR) can be defined as the automatic prediction of the regular human activities performed in our day-to-day life, such as walking, running, cooking, performing office work, etc. It is truly beneficial in the field of medical care services, for example, personal health care assistants, old-age care services, maintaining patient records for future help, etc. Input data to a HAR system can be (a) videos or still images capturing human activities, or (b) time-series data of human body movements while performing the activities taken from sensors in the smart devices like accelerometer, gyroscope, etc. In this work, we mainly focus on the second category of the input data. Here, we propose an ensemble of three classification models, namely CNN-Net, Encoded-Net, and CNN-LSTM, which is named as EnsemConvNet. Each of these classification models is built upon simple 1D Convolutional Neural Network (CNN) but differs in terms of the number of dense layers, kernel size used along with other key differences in the architecture. Each model accepts the time series data as a 2D matrix by taking a window of data at a time in order to infer information, which ultimately predicts the type of human activity. Classification outcome of the EnsemConvNet model is decided using various classifier combination methods that include majority voting, sum rule, product rule, and a score fusion approach called adaptive weighted approach. Three benchmark datasets, namely WISDM activity prediction, UniMiB SHAR, MobiAct, are used for evaluating our proposed model. We have compared our EnsemConvNet model with some existing deep learning models such as Multi Headed CNN, hybrid of CNN, and Long Short Term Memory (LSTM) models. The results obtained here establish the supremacy of the EnsemConvNet model over the other mentioned models.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Aggarwal JK, Xia L (2014) Letters, and undefined, “Human activity recognition from 3d data: A review,” Elsevier. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0167865514001299. Accessed 01 Jul 2020
Akhavian R, Behzadan AH (2016) Smartphone-based construction workers’ activity recognition and classification. Autom Constr 71(Part 2):198–209. https://doi.org/10.1016/j.autcon.2016.08.015
Anguita D, Ghio A, Oneto L, Parra X, Reyes-Ortiz JL (2013) A Public Domain Dataset for Human Activity Recognition Using Smartphones
Barua S, Islam M, Murase K (2011) On Neural, and undefined, “A novel synthetic minority oversampling technique for imbalanced data set learning,” Springer, Accessed 01 Jul 2020. [Online]. Available: https://doi.org/10.1007/978-3-642-24958-7_85.
Bhattacharya S, Shaw V, Singh PK, Sarkar R, Bhattacharjee D (2020) SV-NET: A Deep Learning Approach To Video Based Human Activity Recognition. Proceedings of the Eleventh International Conference on Soft Computing and Pattern Recognition, SoCPaR 2.”
Breiman L (2001) Random forests. Mach Learn 45(1):5–32. https://doi.org/10.1023/A:1010933404324
Bulling A (2014) “33 A Tutorial on Human Activity Recognition Using Body-Worn Inertial Sensors,” dl.acm.org 46(3). https://doi.org/10.1145/2499621.
Casilari E, Santoyo-Ramón J-A, Cano-García J-M (2017) “Analysis of Public Datasets for Wearable Fall Detection Systems,” mdpi.com. https://doi.org/10.3390/s17071513.
Cook DJ, Krishnan NC (2020) “Activity Learning: Discovering, Recognizing, and Predicting Human Behavior. Google Books.” https://books.google.co.in/books?hl=en&lr=&id=TMZ9BgAAQBAJ&oi=fnd&pg=PR11&dq=1.%09Cook,+D.J.+and+Krishnan,+N.C.,+2015.+Activity+learning:+discovering,+recognizing,+and+predicting+human+behavior+from+sensor+data.+John+Wiley+%26+Sons&ots=9qtJ5vnknl&sig=q9o1. Accessed 01 Jul 2020
Dietterich TG (2000) Ensemble methods in machine learning In: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 1857 LNCS, pp. 1–15. https://doi.org/10.1007/3-540-45014-9_1.
Ehatisham-ul-Haq M et al (2017) Authentication of smartphone users based on activity recognition and mobile sensing. Sensors 17(9):2043. https://doi.org/10.3390/s17092043
Frigo M, Johnson SG (2020) “FFTW: An adaptive software architecture for the FFT.” [Online]. Available: https://ieeexplore.ieee.org/abstract/document/681704/. Accessed 01 Jul 2020
Ghosh K, Ghosh S, Sen S, Sarkar R, Maulik U (2020) “A two-stage approach towards protein secondary structure classification,” Springer, Accessed 01 Jul 2020. [Online]. Available: https://doi.org/10.1007/s11517-020-02194-w.pdf.
Kwapisz JR, Weiss GM, Moore SA (2011) Activity recognition using cell phone accelerometers. ACM SIGKDD Explor Newsl 12(2):74–82. https://doi.org/10.1145/1964897.1964918
Maekawa T, Nakai D, Ohara K, Namioka Y (2016) Toward practical factory activity recognition: Unsupervised understanding of repetitive assembly work in a factory. In: UbiComp 2016 - Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, pp. 1088–1099. https://doi.org/10.1145/2971648.2971721.
Mi C, Xu R, Lin C-T (2019) “Real-time Recognition of Smartphone User Behavior Based on Prophet Algorithms,” no. 1, pp. 4–6. [Online]. Available: http://arxiv.org/abs/1909.08997. Accessed 01 Jul 2020
Micucci DI, Mobilio MI, Napoletano P (2017) UniMiB SHAR: a dataset for human activity recognition using acceleration data from smartphones. Appl Sci 7:1101. https://doi.org/10.3390/app7101101
Mubashir M, Shao L, Seed L (2013) A survey on fall detection: principles and approaches. Neurocomputing 100:144–152. https://doi.org/10.1016/j.neucom.2011.09.037
Panhwar M, Shah S, Syed M (2017) Journal of, and undefined 2017, “Smart phone based fall detection using auto regression modeling in a non-restrictive setting,” safetylit.org. [Online]. Available: https://www.safetylit.org/citations/index.php?fuseaction=citations.viewdetails&citationIds[]=citjournalarticle_553203_7. Accessed 01 Jul 2020
Quispe KM, Lima WS, DM Batista, Souto E (2018) MBOSS: A symbolic representation of human activity recognition using mobile sensors. Sensors 18(12): 4354. https://doi.org/10.3390/s18124354.
Rish I (2020) An empirical study of the naive Bayes classifier. [Online]. Available: https://www.cc.gatech.edu/~isbell/reading/papers/Rish.pdf. Accessed 01 Jul 2020
Rokach L, Maimon O (2005) Top-down induction of decision trees classifiers - a survey. IEEE Trans Syst Man Cybern Part C Appl Rev 35(4):476–487. https://doi.org/10.1109/TSMCC.2004.843247
Ronneberger O, Fischer P, Brox T (2020) U-Net: Convolutional Networks for Biomedical Image Segmentation. [Online]. Available: http://lmb.informatik.uni-freiburg.de/. Accessed 01 Jul 2020
Rosenblatt F 1958 Review and undefined 1958, “The perceptron: a probabilistic model for information storage and organization in the brain”. [Online]. Available: https://psycnet.apa.org/journals/rev/65/6/386/. Accessed 01 Jul 2020
Ruck DW, Rogers SK, Kabrisky M (1990) Feature Selection Using a Multilayer Perceptron. [Online]. Available: https://www.researchgate.net/profile/Teresa_Hawkes/project/Neurophysiology-based-computer-program-design-for-cyber-security/attachment/58f16b381042bf333c68ad87/AS:483205733195776@1492216632627/download/Ruck%2C+Rogers%2C+Kabrisky.pdf?context=ProjectUpdatesLog. Accessed 01 Jul 2020
Sadhukhan S, Mallick S, Singh PK, Sarkar R, Bhattacharjee D (2020) A comparative study of different feature descriptors for video-based human action recognition. Springer, Singapore, pp 35–52
Saunders C, Stitson MO, Weston J, Bottou L, Schölkopf B, Smola A (1998) Support Vector Machine Reference Manual. [Online]. Available: https://eprints.soton.ac.uk/258959/1/SVM_Reference.pdf. Accessed 01 Jul 2020
Sazonov E, Metcalfe K, Lopez-Meyer P, Tiffany S (2011) RF hand gesture sensor for monitoring of cigarette smoking. In: Proceedings of the International Conference on Sensing Technology, ICST, pp. 426–430. https://doi.org/10.1109/ICSensT.2011.6137014.
Shoaib M, Bosch S, Incel OD, Scholten H, Havinga PJM (2016) “Complex Human Activity Recognition Using Smartphone and Wrist-Worn Motion Sensors,” mdpi.com. https://doi.org/10.3390/s16040426.
Vavoulas G, Chatzaki C, Malliotakis T, Pediaditis M, Tsiknakis M 92016) The MobiAct Dataset: Recognition of Activities of Daily Living using Smartphones
Walse KH, Dharaskar RV (2016) Performance Evaluation of Classifiers on WISDM Dataset for Human Activity Recognition. Accessed 01 Jul 2020. [Online]. Available: https://doi.org/10.1145/2905055.2905232.
Wang J, Chen Y, Hao S, Peng X, Hu L (2019) Deep learning for sensor-based activity recognition: a survey. Pattern Recogn Lett 119:3–11. https://doi.org/10.1016/j.patrec.2018.02.010
Wiskott L, Sejnowski TJ (2002) Slow feature analysis: unsupervised learning of invariances. Neural Comput 14(4):715–770. https://doi.org/10.1162/089976602317318938
Xu Y, Lu Y (2015) Neurocomputing, and undefined 2015, “Adaptive weighted fusion: a novel fusion approach for image classification,” Elsevier. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0925231215007687. Accessed 01 Jul 2020
Xu J, He Z, Zhang Y (2019) CNN-LSTM combined network for iot enabled fall detection applications. J Phys 12044. https://doi.org/10.1088/1742-6596/1267/1/012044.
Zhang Y, Zhang Y, Zhang Z, Bao J, Song Y (2020) Human activity recognition based on time series analysis using U-Net. [Online]. Available: https://arxiv.org/abs/1809.08113. Accessed 01 Jul 2020
Zheng Z, Du J, Sun L, Huo M, Chen Y (2018) “TASG: An Augmented Classification Method for Impersonal HAR,” hindawi.com. https://doi.org/10.1155/2018/6751363.
Acknowledgements
Authors are grateful to the Center for Microprocessor Application for Training Education and Research (CMATER) of Computer Science & Engineering Department, Jadavpur University, for providing infrastructure facilities during progress of the work.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Mukherjee, D., Mondal, R., Singh, P.K. et al. EnsemConvNet: a deep learning approach for human activity recognition using smartphone sensors for healthcare applications. Multimed Tools Appl 79, 31663–31690 (2020). https://doi.org/10.1007/s11042-020-09537-7
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-020-09537-7