LIVE EVENT DETECTION FOR PEOPLE’S SAFETY USING NLP AND DEEP LEARNING
Main Article Content
Abstract
In recent years, ensuring public safety during live events has become a critical challenge due to the increasing scale of gatherings and potential risks. This study proposes a novel approach to Live Event Detection for People’s Safety using audio data and the LightGBM classifier. The system leverages real-time audio streams to identify anomalies, such as loud disturbances, explosions, or unusual crowd behavior, which could indicate potential safety threats. Audio features are extracted using advanced signal processing techniques, including Mel-frequency cepstral coefficients (MFCCs), spectral contrast, and chroma features. These features are fed into a LightGBM classifier, which provides efficient and robust performance for real-time classification of event categories and potential risks. The proposed methodology is evaluated using diverse datasets comprising audio samples from live events, including concerts, sports, and emergency situations, to ensure a comprehensive understanding of normal and abnormal patterns. The LightGBM model demonstrates high accuracy, low latency, and scalability, making it suitable for deployment in real-time applications. Additionally, the system integrates a feedback loop for continuous model improvement based on new audio data. The results highlight the system's ability to enhance situational awareness and proactively alert authorities to potential risks, ensuring timely interventions. This approach demonstrates a significant step toward leveraging machine learning and audio analytics to improve public safety at live events.
Downloads
Metrics
Article Details
This work is licensed under a Creative Commons Attribution 4.0 International License.
You are free to:
- Share — copy and redistribute the material in any medium or format for any purpose, even commercially.
- Adapt — remix, transform, and build upon the material for any purpose, even commercially.
- The licensor cannot revoke these freedoms as long as you follow the license terms.
Under the following terms:
- Attribution — You must give appropriate credit , provide a link to the license, and indicate if changes were made . You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use.
- No additional restrictions — You may not apply legal terms or technological measures that legally restrict others from doing anything the license permits.
Notices:
You do not have to comply with the license for elements of the material in the public domain or where your use is permitted by an applicable exception or limitation .
No warranties are given. The license may not give you all of the permissions necessary for your intended use. For example, other rights such as publicity, privacy, or moral rights may limit how you use the material.
References
J.P. Bello, C. Mydlarz, J. Salamon, "Sound Analysis in Smart Cities," in: T. Virtanen, M.D. Plumbley, D. Ellis, Eds., Computational Analysis of Sound Scenes and Events, Springer International Publishing, Cham, Switzerland, 2018, pp. 373-397.
P. Zinemanas, M. Rocamora, M. Miron, F. Font, X. Serra, "An Interpretable Deep Learning Model for Automatic Sound Classification," Electronics, vol. 10, p. 850, 2021. doi: 10.3390/electronics10070850.
J.K. Das, A. Chakrabarty, M.J. Piran, "Environmental sound classification using convolution neural networks with different integrated loss functions," Expert Systems, vol. 39, e12804, 2021. doi: 10.1111/exsy.12804.
J.K. Das, A. Ghosh, A.K. Pal, S. Dutta, A. Chakrabarty, "Live event sound Classification Using Convolutional Neural Network and Long Short Term Memory Based on Multiple Features," in Proceedings of the 2020 Fourth International Conference on Intelligent Computing in Data Sciences (ICDS), Fez, Morocco, 21-23 October 2020, pp. 1-9. doi: 10.1109/ICDS50568.2020.9269108.
Z. Mushtaq, S.F. Su, "Efficient Classification of Environmental Sounds through Multiple Features Aggregation and Data Enhancement Techniques for Spectrogram Images," Symmetry, vol. 12, p. 1822, 2020. doi: 10.3390/sym12111822.
W. Mu, B. Yin, X. Huang, J. Xu, Z. Du, "Environmental sound classification using temporal-frequency attention based convolutional neural network," Scientific Reports, vol. 11, p. 21552, 2021. doi: 10.1038/s41598-021-00455-w.
T. Giannakopoulos, E. Spyrou, S.J. Perantonis, "Recognition of Live event sound Events Using Deep Context-Aware Feature Extractors and Handcrafted Features," in IFIP International Conference on Artificial Intelligence Applications and Innovations, J. MacIntyre, I. Maglogiannis, L. Iliadis, E. Pimenidis, Eds., Springer International Publishing, Cham, Switzerland, 2019, pp. 184-195.
J.S. Luz, M.C. Oliveira, F.H. Araújo, D.M. Magalhães, "Ensemble of handcrafted and deep features for live event sound classification," Applied Acoustics, vol. 175, p. 107819, 2021. doi: 10.1016/j.apacoust.2021.107819.
Y. Gong, Y. Chung, J.R. Glass, "AST: Audio Spectrogram Transformer," arXiv, arXiv:2104.01778, 2021.
İ. Türker, S. Aksu, "Connectogram—A graph-based time dependent representation for sounds," Applied Acoustics, vol. 191, p. 108660, 2022. doi: 10.1016/j.apacoust.2021.108660.
Q. Kong, Y. Xu, M. Plumbley, "Sound Event Detection of Weakly Labelled Data with CNN-Transformer and Automatic Threshold Optimization," IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 28, pp. 2450-2460, 2020. doi: 10.1109/TASLP.2020.3019456.
P. Gimeno, I. Viñals, A. Ortega, A. Miguel, E. Lleida, "Multiclass audio segmentation based on recurrent neural networks for broadcast domain data," EURASIP Journal on Audio, Speech, and Music Processing, vol. 2020, p. 5, 2020. doi: 10.1186/s13636-020-00178-2.
Z. Zhang, S. Xu, S. Zhang, T. Qiao, S. Cao, "Learning Attentive Representations for Environmental Sound Classification," IEEE Access, vol. 7, pp. 130327-130339, 2019. doi: 10.1109/ACCESS.2019.2940272.
Z. Zhang, S. Xu, S. Zhang, T. Qiao, S. Cao, "Attention based convolutional recurrent neural network for environmental sound classification," Neurocomputing, vol. 453, pp. 896-903, 2020. doi: 10.1016/j.neucom.2020.05.125.
T. Qiao, S. Zhang, S. Cao, S. Xu, "High Accurate Environmental Sound Classification: Sub-Spectrogram Segmentation versus Temporal-Frequency Attention Mechanism," Sensors, vol. 21, p. 5500, 2021. doi: 10.3390/s21165500.