Deep Learning Multi-Target Underwater Acoustic Intelligence System
Saio Alusine Marrah,
Abu Bakarr Koroma,
Sayo Nakeleh Turay,
Gibrilla Deen Kamara,
Mabinty Marrah,
Mohamed Thoronka and
Paul Conteh
Additional contact information
Saio Alusine Marrah: School of Software Engineering, University of Electronic Science and Technology of China (UESTC), Chengdu, China.
Abu Bakarr Koroma: University of Electronic Science and Technology of China (CWAS of UESTC), School of Management and Economics (SME), (UESTC), Chengdu, China
Sayo Nakeleh Turay: Institute of Public Administration and Management, University of Sierra Leone, Freetown, Sierra Leone.
Gibrilla Deen Kamara: School of Information and Communication Engineering, University of Electronic Science and Technology of China (UESTC), Chengdu, China.
Mabinty Marrah: Limkokwing University of Creative Technology, Freetown, Sierra Leone.
Mohamed Thoronka: College of Software Engineering, Nankai University, Tianjin, China.
Paul Conteh: College of Software Engineering, Nankai University, Tianjin, China.
Journal of Scientific Reports, 2026, vol. 12, issue 1, 20-39
Abstract:
Increased complexity of underwater acoustic space poses great difficulty in the accurate and real-time identification of the target, especially in low signal-to-noise ratio (SNR) environments and in mixed-source sound waves. This research combines a Deep Learning Multi-Target Underwater Acoustic Intelligence System that incorporates a Convolutional Neural Network (CNN) with a Transformer-based attention encoder to obtain robust and explainable multi-classification of underwater acoustic signals. The model uses a trained resnet-50 backbone that computes spectral features (localized) in the form of Short-Time Fourier transform (STFT) spectrograms and then uses a multi-head self-attention mechanism to note long-term temporal features. A parallel attention fusion layer is used to facilitate both spatial and temporal representations, and the Focal Loss reconstruction makes weak or minority classes, including low-energy biological calls, more sensitive. The data set, containing real and simulated underwater records in five categories, namely ships, submarines, marine mammals, ambient noise and environmental interference was supplemented to display a variety of SNR conditions between 0 and 25 dB. Experimentally, the hybrid model has been found to be more precise and energy-efficient than CNN-only, LSTM, and Transformer-only baselines because it has 98.1 percent classification using a fixed modulus and above 90 percent classification in 0 dB SNR. Moreover, the model exhibits an average throughput rate of 370 frames per second with a real-time inference rate of 3.4 ms per frame, and hence the model is suitable in autonomous underwater vehicles (AUVs) and marine surveillance systems. Grad-CAM images verify that the attention module is concentrated on acoustical significance spectral areas, which proves the interpretability and cognitive openness of the model. On the whole, the hybrid framework represents a considerable breakthrough in the underwater acoustic intelligence sphere as it combines strengths, precision, and comprehensibility, and preconditions the emergence of intelligent sensing and real-time maritime observation platforms of the next generation.
Keywords: Underwater acoustic intelligence; Deep learning; CNN–Transformer hybrid; Multi-target classification; Signal-to-noise robustness (search for similar items in EconPapers)
Date: 2026
References: Add references at CitEc
Citations:
Downloads: (external link)
https://ijsab.com/wp-content/uploads/1152.pdf (application/pdf)
https://www.ijsab.com/jsr-volume-12-issue-1/8465 (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:aif:report:v:11:y:2026:i:1:p:20-39
DOI: 10.58970/JSR.1152
Access Statistics for this article
Journal of Scientific Reports is currently edited by Dr. Md Shamim Hossain
More articles in Journal of Scientific Reports from IJSAB International
Bibliographic data for series maintained by Farjana Rahman ().