LIVE EVENT DETECTION FOR PEOPLE’S SAFETY USING NLP AND DEEP LEARNING

Main Article Content

C. RASHMI
B. SOUMYA
A. HARIKA
D. HARATHI

Abstract

In recent years, ensuring public safety during live events has become a critical challenge due to the increasing scale of gatherings and potential risks. This study proposes a novel approach to Live Event Detection for People’s Safety using audio data and the LightGBM classifier. The system leverages real-time audio streams to identify anomalies, such as loud disturbances, explosions, or unusual crowd behavior, which could indicate potential safety threats. Audio features are extracted using advanced signal processing techniques, including Mel-frequency cepstral coefficients (MFCCs), spectral contrast, and chroma features. These features are fed into a LightGBM classifier, which provides efficient and robust performance for real-time classification of event categories and potential risks. The proposed methodology is evaluated using diverse datasets comprising audio samples from live events, including concerts, sports, and emergency situations, to ensure a comprehensive understanding of normal and abnormal patterns. The LightGBM model demonstrates high accuracy, low latency, and scalability, making it suitable for deployment in real-time applications. Additionally, the system integrates a feedback loop for continuous model improvement based on new audio data. The results highlight the system's ability to enhance situational awareness and proactively alert authorities to potential risks, ensuring timely interventions. This approach demonstrates a significant step toward leveraging machine learning and audio analytics to improve public safety at live events.

Downloads

Download data is not yet available.

Metrics

Metrics Loading ...

Article Details

How to Cite
RASHMI, C., SOUMYA, B. ., HARIKA, A. ., & HARATHI, D. . (2024). LIVE EVENT DETECTION FOR PEOPLE’S SAFETY USING NLP AND DEEP LEARNING. Turkish Journal of Computer and Mathematics Education (TURCOMAT), 15(3), 434–441. https://doi.org/10.61841/turcomat.v15i3.14956
Section
Original Article

References

J.P. Bello, C. Mydlarz, J. Salamon, "Sound Analysis in Smart Cities," in: T. Virtanen, M.D. Plumbley, D. Ellis, Eds., Computational Analysis of Sound Scenes and Events, Springer International Publishing, Cham, Switzerland, 2018, pp. 373-397.

P. Zinemanas, M. Rocamora, M. Miron, F. Font, X. Serra, "An Interpretable Deep Learning Model for Automatic Sound Classification," Electronics, vol. 10, p. 850, 2021. doi: 10.3390/electronics10070850.

J.K. Das, A. Chakrabarty, M.J. Piran, "Environmental sound classification using convolution neural networks with different integrated loss functions," Expert Systems, vol. 39, e12804, 2021. doi: 10.1111/exsy.12804.

J.K. Das, A. Ghosh, A.K. Pal, S. Dutta, A. Chakrabarty, "Live event sound Classification Using Convolutional Neural Network and Long Short Term Memory Based on Multiple Features," in Proceedings of the 2020 Fourth International Conference on Intelligent Computing in Data Sciences (ICDS), Fez, Morocco, 21-23 October 2020, pp. 1-9. doi: 10.1109/ICDS50568.2020.9269108.

Z. Mushtaq, S.F. Su, "Efficient Classification of Environmental Sounds through Multiple Features Aggregation and Data Enhancement Techniques for Spectrogram Images," Symmetry, vol. 12, p. 1822, 2020. doi: 10.3390/sym12111822.

W. Mu, B. Yin, X. Huang, J. Xu, Z. Du, "Environmental sound classification using temporal-frequency attention based convolutional neural network," Scientific Reports, vol. 11, p. 21552, 2021. doi: 10.1038/s41598-021-00455-w.

T. Giannakopoulos, E. Spyrou, S.J. Perantonis, "Recognition of Live event sound Events Using Deep Context-Aware Feature Extractors and Handcrafted Features," in IFIP International Conference on Artificial Intelligence Applications and Innovations, J. MacIntyre, I. Maglogiannis, L. Iliadis, E. Pimenidis, Eds., Springer International Publishing, Cham, Switzerland, 2019, pp. 184-195.

J.S. Luz, M.C. Oliveira, F.H. Araújo, D.M. Magalhães, "Ensemble of handcrafted and deep features for live event sound classification," Applied Acoustics, vol. 175, p. 107819, 2021. doi: 10.1016/j.apacoust.2021.107819.

Y. Gong, Y. Chung, J.R. Glass, "AST: Audio Spectrogram Transformer," arXiv, arXiv:2104.01778, 2021.

İ. Türker, S. Aksu, "Connectogram—A graph-based time dependent representation for sounds," Applied Acoustics, vol. 191, p. 108660, 2022. doi: 10.1016/j.apacoust.2021.108660.

Q. Kong, Y. Xu, M. Plumbley, "Sound Event Detection of Weakly Labelled Data with CNN-Transformer and Automatic Threshold Optimization," IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 28, pp. 2450-2460, 2020. doi: 10.1109/TASLP.2020.3019456.

P. Gimeno, I. Viñals, A. Ortega, A. Miguel, E. Lleida, "Multiclass audio segmentation based on recurrent neural networks for broadcast domain data," EURASIP Journal on Audio, Speech, and Music Processing, vol. 2020, p. 5, 2020. doi: 10.1186/s13636-020-00178-2.

Z. Zhang, S. Xu, S. Zhang, T. Qiao, S. Cao, "Learning Attentive Representations for Environmental Sound Classification," IEEE Access, vol. 7, pp. 130327-130339, 2019. doi: 10.1109/ACCESS.2019.2940272.

Z. Zhang, S. Xu, S. Zhang, T. Qiao, S. Cao, "Attention based convolutional recurrent neural network for environmental sound classification," Neurocomputing, vol. 453, pp. 896-903, 2020. doi: 10.1016/j.neucom.2020.05.125.

T. Qiao, S. Zhang, S. Cao, S. Xu, "High Accurate Environmental Sound Classification: Sub-Spectrogram Segmentation versus Temporal-Frequency Attention Mechanism," Sensors, vol. 21, p. 5500, 2021. doi: 10.3390/s21165500.