Vision-based human gesture detection is the task of forecasting a gesture, namely clapping or sign language gestures, or waving hello, utilizing various video frames. One of the attractive features of gesture detection is that it makes it possible for humans to interact with devices and computers without the necessity for an external input tool like a remote control or a mouse. Gesture detection from videos has various applications, like robot learning, control of consumer electronics computer games, and mechanical systems. This study leverages the Lion Swarm optimizer with a deep convolutional neural network (LSO-DCNN) for gesture recognition and classification. The purpose of the LSO-DCNN technique lies in the proper identification and categorization of various categories of gestures that exist in the input images. The presented LSO-DCNN model follows a three-step procedure. At the initial step, the 1D-convolutional neural network (1D-CNN) method derives a collection of feature vectors. In the second step, the LSO algorithm optimally chooses the hyperparameter values of the 1D-CNN model. At the final step, the extreme gradient boosting (XGBoost) classifier allocates proper classes, i.e., it recognizes the gestures efficaciously. To demonstrate the enhanced gesture classification results of the LSO-DCNN approach, a wide range of experimental results are investigated. The brief comparative study reported the improvements in the LSO-DCNN technique in the gesture recognition process.
Citation: Mashael Maashi, Mohammed Abdullah Al-Hagery, Mohammed Rizwanullah, Azza Elneil Osman. Deep convolutional neural network-based Leveraging Lion Swarm Optimizer for gesture recognition and classification[J]. AIMS Mathematics, 2024, 9(4): 9380-9393. doi: 10.3934/math.2024457
Vision-based human gesture detection is the task of forecasting a gesture, namely clapping or sign language gestures, or waving hello, utilizing various video frames. One of the attractive features of gesture detection is that it makes it possible for humans to interact with devices and computers without the necessity for an external input tool like a remote control or a mouse. Gesture detection from videos has various applications, like robot learning, control of consumer electronics computer games, and mechanical systems. This study leverages the Lion Swarm optimizer with a deep convolutional neural network (LSO-DCNN) for gesture recognition and classification. The purpose of the LSO-DCNN technique lies in the proper identification and categorization of various categories of gestures that exist in the input images. The presented LSO-DCNN model follows a three-step procedure. At the initial step, the 1D-convolutional neural network (1D-CNN) method derives a collection of feature vectors. In the second step, the LSO algorithm optimally chooses the hyperparameter values of the 1D-CNN model. At the final step, the extreme gradient boosting (XGBoost) classifier allocates proper classes, i.e., it recognizes the gestures efficaciously. To demonstrate the enhanced gesture classification results of the LSO-DCNN approach, a wide range of experimental results are investigated. The brief comparative study reported the improvements in the LSO-DCNN technique in the gesture recognition process.
[1] | C. Dewi, A. P. S Chen, H. J. Christanto, Deep learning for highly accurate hand recognition based on Yolov7 model, Big Data Cogn. Comput., 7 (2023), 53. https://doi.org/10.3390/bdcc7010053 doi: 10.3390/bdcc7010053 |
[2] | J. John, S. P. Deshpande, Hand gesture identification using deep learning and artificial neural networks: A review, Computational Intelligence for Engineering and Management Applications: Select Proceedings of CIEMA 2022, 2023,389–400. https://doi.org/10.1007/978-981-19-8493-8_30 doi: 10.1007/978-981-19-8493-8_30 |
[3] | R. Padmavathi, Expressive and Deployable Hand Gesture Recognition for Sign Way of Communication for Visually Impaired People, 2021. |
[4] | A. Agarwal, A. Das, Facial Gesture Recognition Based Real Time Gaming for Physically Impairment. In Artificial Intelligence: First International Symposium, ISAI 2022, 2023, Haldia, India, February 17–22, 2022, Revised Selected Papers (256–264). Cham: Springer Nature Switzerland. https://doi.org/10.1007/978-3-031-22485-0_23 |
[5] | V. Gorobets, C. Merkle, A. Kunz, Pointing, pairing and grouping gesture recognition in virtual reality, In Computers Helping People with Special Needs: 18th International Conference, ICCHP-AAATE 2022, Lecco, Italy, July 11–15, 2022, Proceedings, Part I (313–320), Cham: Springer International Publishing. https://doi.org/10.1007/978-3-031-08648-9_36 |
[6] | J. Gangrade, J. Bharti, Vision-based hand gesture recognition for Indian sign language using convolution neural network, IETE J. Res., 69 (2023), 723–732. https://doi.org/10.1007/978-3-031-08648-9_36 doi: 10.1007/978-3-031-08648-9_36 |
[7] | J. Li, C. Li, J. Han, Y. Shi, G. Bian, S. Zhou, Robust hand gesture recognition using HOG-9ULBP features and SVM model, Electronics, 11 (2022), 988. https://doi.org/10.1007/978-3-031-08648-9_36 doi: 10.1007/978-3-031-08648-9_36 |
[8] | D. Ryumin, D. Ivanko, E. Ryumina, Audio-visual speech and gesture recognition by sensors of mobile devices, Sensors, 23 (2023), 2284. https://doi.org/10.1007/978-3-031-08648-9_36 doi: 10.1007/978-3-031-08648-9_36 |
[9] | T. Sahana, S. Basu, M. Nasipuri, A. F. Mollah, MRCS: multi-radii circular signature based feature descriptor for hand gesture recognition, Multimed. Tools Appl., 81 (2022), 8539–8560. https://doi.org/10.1007/s11042-021-11743-w doi: 10.1007/s11042-021-11743-w |
[10] | S. Pandey, Automated Gesture Recognition and Speech Conversion Tool for Speech Impaired. In Proceedings of Third International Conference on Advances in Computer Engineering and Communication Systems: ICACECS 2022, (467–476), Singapore: Springer Nature Singapore, 2023. https://doi.org/10.1007/978-981-19-9228-5_39 |
[11] | Y. Sun, Y. Weng, B. Luo, G. Li, B. Tao, D. Jiang, et al., Gesture recognition algorithm based on multi‐scale feature fusion in RGB‐D images, IET Image Process., 17 (2023), 1280–1290. https://doi.org/10.1049/ipr2.12712 |
[12] | R. Barioul, O. Kanoun, k-Tournament grasshopper extreme learner for FMG-Based gesture recognition, Sensors, 23 (2023), 1096. https://doi.org/10.1049/ipr2.12712 doi: 10.1049/ipr2.12712 |
[13] | T. R. Gadekallu, M. Alazab, R. Kaluri, P. K. R Maddikunta, S. Bhattacharya, K. Lakshmanna, Hand gesture classification using a novel CNN-crow search algorithm, Complex Intell. Syst., 7 (2021), 1855–1868. https://doi.org/10.1049/ipr2.12712 doi: 10.1049/ipr2.12712 |
[14] | M. Yu, G. Li, D. Jiang, G. Jiang, F. Zeng, H. Zhao, et al., Application of PSO-RBF neural network in gesture recognition of continuous surface EMG signals, J. Intell. Fuzzy Syst., 38 (2020), 2469–2480. https://doi.org/10.3233/JIFS-179535 |
[15] | A. Sen, T. K. Mishra, R. Dash, A novel hand gesture detection and recognition system based on ensemble-based convolutional neural network, Multimed. Tools Appl., 81 (2022), 40043–40066. https://doi.org/10.3233/JIFS-179535 doi: 10.3233/JIFS-179535 |
[16] | Q. Gao, Z. Ju, Y. Chen, Q. Wang, C. Chi, An efficient RGB-D hand gesture detection framework for dexterous robot hand-arm teleoperation system, IEEE T. Hum-Mach Syst., 2022. https://doi.org/10.1109/THMS.2022.3206663 doi: 10.1109/THMS.2022.3206663 |
[17] | P. S. Neethu, R. Suguna, D. Sathish, An efficient method for human hand gesture detection and recognition using deep learning convolutional neural networks, Soft Comput., 24 (2020), 15239–15248. https://doi.org/10.1109/THMS.2022.3206663 doi: 10.1109/THMS.2022.3206663 |
[18] | X. Zhang, P. Han, L. Xu, F. Zhang, Y. Wang, L. Gao, Research on bearing fault diagnosis of wind turbine gearbox based on 1DCNN-PSO-SVM, IEEE Access, 8 (2020), 192248–192258. https://doi.org/10.1109/THMS.2022.3206663 doi: 10.1109/THMS.2022.3206663 |
[19] | J. Fu, J. Liu, D. Xie, Z. Sun, Application of fuzzy PID based on Stray Lion Swarm Optimization Algorithm in overhead crane system control, Mathematics, 11 (2023), 2170. https://doi.org/10.1109/THMS.2022.3206663 doi: 10.1109/THMS.2022.3206663 |
[20] | R. Jena, A. Shanableh, R. Al-Ruzouq, B. Pradhan, M. B. A. Gibril, M. A. Khalil, et al., Explainable Artificial Intelligence (XAI) model for earthquake spatial probability assessment in Arabian peninsula, Remote Sens., 15 (2023), 2248. https://doi.org/10.1109/THMS.2022.3206663 |
[21] | F. Wang, R. Hu, Y. Jin, Research on gesture image recognition method based on transfer learning, Procedia Comput. Sci., 187 (2021), 140–145. https://doi.org/10.1109/THMS.2022.3206663 doi: 10.1109/THMS.2022.3206663 |