Abstract
Deep Learning in the field of Big Data has become essential for the analysis and perception of trends. Activation functions play a crucial role in the outcome of these deep learning frameworks. The existing activation functions are hugely focused on data translation from one neural layer to another. Although they have been proven useful and have given consistent results, they are static and mostly non-parametric. In this paper, we propose a new function for modified training of neural networks that is more flexible and adaptable to the data. The proposed catalysis function works over Rectified Linear Unit (ReLU), sigmoid, tanh and all other activation functions to provide adaptive feed-forward training. The function uses vector components of the activation function to provide variational flow of input. The performance of this algorithm is tested on Modified National Institute of Standards and Technology (MNIST) and Canadian Institute for Advanced Research (CIFAR10) datasets against the conventional activation functions. Visual Geometry Group (VGG) blocks and Residual Neural Network (ResNet) architectures are used for experimentation. The proposed function has shown significant improvements in comparison to the traditional functions with a 75 ± 2.5% acuuracy across activation functions. The adaptive nature of training has drastically decreased the probability of under-fitting. The parameterization has helped increase the data learning capacity of models. On performing sensitivity analysis, the catalysis activation show slight or no changes on varying initialization parameters.



















Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Reddy GT, Kumar Reddy MP, Lakshmanna K, Kaluri R, Rajput DS, Srivastava G, Baker T (2020) Analysis of dimensionality reduction techniques on big data. IEEE Access 8:54776–54788
Agrawal S, Sarkar S, Srivastava G, Maddikunta PKR, Gadekallu TR (2021) Genetically optimized prediction of remaining useful life. Sustainable Computing: Informatics and Systems 31:100565
Chen Y, Dai X, Liu M, Chen D, Yuan L, Liu Z (2020) Dynamic ReLU. arXiv preprint arXiv:2003.10027
Si J, Harris SL, Yfantis E (2018) A dynamic ReLU on neural network. In: 2018 IEEE 13th Dallas Circuits and Systems Conference (DCAS), IEEE, pp 1–6
Ahn H, Chung B, Yim C (2019) Super-resolution convolutional neural networks using modified and bilateral ReLU. In: 2019 International Conference on Electronics, Information, and Communication (ICEIC), IEEE, pp 1–4
Hu X, Niu P, Wang J, Zhang X (2019) A dynamic rectified linear activation units. IEEE Access 7:180409–16
Kim J, Kim S, Lee M (2015) Convolutional neural network with biologically inspired on/off relu. In: International Conference on Neural Information Processing, Springer, Cham, pp 316–323
Chung H, Lee SJ, Park JG (2016) Deep neural network using trainable activation functions. In: 2016 International Joint Conference on Neural Networks (IJCNN), IEEE, pp 348–352
Nwankpa C, Ijomah W, Gachagan A, Marshall S (2018) Activation functions: Comparison of trends in practice and research for deep learning. arXiv preprint arXiv:1811.03378
Pedamonti D (2018) Comparison of non-linear activation functions for deep neural networks on MNIST classification task. arXiv preprint arXiv:1804.02763
Yamashita R, Nishio M, Do RK, Togashi K (2018) Convolutional neural networks: an overview and application in radiology. Insights Into Imaging 9(4):611–29
Kumar C, Punitha R (2020) YOLOv3 and YOLOv4: Multiple object detection for surveillance applications. In: 2020 Third International Conference on Smart Systems and Inventive Technology (ICSSIT), IEEE, pp 1316–1321
Sharma N, Mandal R, Sharma R, Pal U, Blumenstein M (2018) Signature and logo detection using deep CNN for document image retrieval. In: 2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR), IEEE, pp 416–422
Xie Q, Luong MT, Hovy E, Le QV (2020) Self-training with noisy student improves imagenet classification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 10687–10698
Kolesnikov A, Beyer L, Zhai X, Puigcerver J, Yung J, Gelly S, Houlsby N (2019) Big transfer (BiT): General visual representation learning. arXiv preprint arXiv:1912.11370
Kirillov A, Wu Y, He K, Girshick R (2020) Pointrend: Image segmentation as rendering. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 9799–9808
Trottier L, Gigu P, Chaib-draa B (2017) Parametric exponential linear unit for deep convolutional neural networks. In: 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA), IEEE, pp 207–214
Svozil D, Kvasnicka V, Pospichal J (1997) Introduction to multi-layer feed-forward neural networks. Chemometrics and Intelligent Laboratory Systems 39(1):43–62
Narayan S (1997) The generalized sigmoid activation function: Competitive supervised learning. Inform Sci 99(1–2):69–82
Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767
Bochkovskiy A, Wang CY, Liao HY (2020) YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv preprint arXiv:2004.10934
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 770–778
Chen YT, Chen TS, Chen J (2018) A LeNet Based Convolution Neural Network for Image Steganalysis on Multiclass Classification. DEStech Transactions on Computer Science and Engineering, (ccme)
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556
Krizhevsky, A., Nair, V. and Hinton, G (2009) FAR-10 (canadian institute for advanced research). http://www.cs.toronto.edu/kriz/cifar.html, 5
LeCun, Y (998) The MNIST database of handwritten digits. http://yann.lecun.com/exdb/mnist/
Azmoodeh A, Dehghantanha A, Choo KK (2019) Big data and internet of things security and forensics: challenges and opportunities. In: Handbook of Big Data and IoT Security, Springer, Cham, pp 1–4
Zhang Q, Yang LT, Chen Z, Li P (2018) A survey on deep learning for big data. Inform Fusion 1(42):146–57
Banerjee C, Mukherjee T, Pasiliao E (2020) Feature representations using the reflected rectified linear unit (RReLU) activation. Big Data Mining and Analytics 3(2):102–120. https://doi.org/10.26599/BDMA.2019.9020024
Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, pp 249–256
He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE International Conference on Computer Vision, pp 1026–1034
Klambauer G, Unterthiner T, Mayr A, Hochreiter S (2017) Self-normalizing neural networks. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp 972–981
Biswas K, Banerjee S, Pandey AK (2021) TanhSoftÃ-Dynamic trainable activation functions for faster learning and better performance. IEEE Access 9:120613–120623. https://doi.org/10.1109/ACCESS.2021.3105355
Saha S, Mathur A, Pandey A, Arun Kumar H (2021) DiffAct: A unifying framework for activation functions. In: 2021 International Joint Conference on Neural Networks (IJCNN), pp 1–8, https://doi.org/10.1109/IJCNN52387.2021.9534391.
Pratama K, Kang DK (2021) Trainable activation function with differentiable negative side and adaptable rectified point. Appl Intell 51(3):1784–1801
Zhao H, Liu F, Li L, Luo C (2018) A novel softplus linear unit for deep convolutional neural networks. Appl Intell 48(7):1707–1720
Author information
Authors and Affiliations
Corresponding authors
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
Rights and permissions
About this article
Cite this article
Sarkar, S., Agrawal, S., Baker, T. et al. Catalysis of neural activation functions: Adaptive feed-forward training for big data applications. Appl Intell 52, 13364–13383 (2022). https://doi.org/10.1007/s10489-021-03082-y
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-021-03082-y