A Dual-Attentive Multimodal Fusion Method for Fault Diagnosis Under Varying Working Conditions
Yan Chu (),
Leqi Zhu and
Mingfeng Lu
Additional contact information
Yan Chu: School of Finance, Shanghai Lixin University of Accounting and Finance, Shanghai 201620, China
Leqi Zhu: Research Institute for Data Management and Innovation, Nanjing University, Suzhou 215163, China
Mingfeng Lu: School of Management, Hefei University of Technology, Hefei 230002, China
Mathematics, 2025, vol. 13, issue 11, 1-29
Abstract:
Deep learning-based fault diagnosis methods have gained extensive attention in recent years due to their outstanding performance. The model input can take the form of multiple domains, such as the time domain, frequency domain, and time–frequency domain, with commonalities and differences between them. Fusing multimodal features is crucial for enhancing diagnostic effectiveness. In addition, original signals typically exhibit nonstationary characteristics influenced by varying working conditions. In this paper, a dual-attentive multimodal fusion method combining a multiscale dilated CNN (DAMFM-MD) is proposed for rotating machinery fault diagnosis. Firstly, multimodal data are constructed by combining original signals, FFT-based frequency spectra, and STFT-based time–frequency images. Secondly, a three-branch multiscale CNN is developed for discriminative feature learning to consider nonstationary factors. Finally, a two-stage sequential fusion is designed to achieve multimodal complementary fusion considering the features with commonality and differentiation. The performance of the proposed method was experimentally verified through a series of industrial case analyses. The proposed DAMFM-MD method achieves the best F-score of 99.95%, an accuracy of 99.96%, and a recall of 99.95% across four sub-datasets, with an average fault diagnosis response time per sample of 1.095 milliseconds, outperforming state-of-the-art methods.
Keywords: fault diagnosis; deep learning; convolutional neural network; dilation convolution; attention mechanism; multimodal feature fusion (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2025
References: View references in EconPapers View complete reference list from CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2227-7390/13/11/1868/pdf (application/pdf)
https://www.mdpi.com/2227-7390/13/11/1868/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:13:y:2025:i:11:p:1868-:d:1670956
Access Statistics for this article
Mathematics is currently edited by Ms. Emma He
More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().