基于Transformer的音频输入机器故障检测 / Transformer Based Machine Fault Detection From Audio Input
1️⃣ 一句话总结
这篇论文提出使用Transformer模型来分析机器音频,以检测故障,并证明它在分析声音频谱图方面比传统的卷积神经网络(CNN)更具潜力,尤其是在数据充足的情况下。
In recent years, Sound AI is being increasingly used to predict machine failures. By attaching a microphone to the machine of interest, one can get real time data on machine behavior from the field. Traditionally, Convolutional Neural Net (CNN) architectures have been used to analyze spectrogram images generated from the sounds captured and predict if the machine is functioning as expected. CNN architectures seem to work well empirically even though they have biases like locality and parameter-sharing which may not be completely relevant for spectrogram analysis. With the successful application of transformer-based models in the field of image processing starting with Vision Transformer (ViT) in 2020, there has been significant interest in leveraging these in the field of Sound AI. Since transformer-based architectures have significantly lower inductive biases, they are expected to perform better than CNNs at spectrogram analysis given enough data. This paper demonstrates the effectiveness of transformer-driven architectures in analyzing Sound data and compares the embeddings they generate with CNNs on the specific task of machine fault detection.
基于Transformer的音频输入机器故障检测 / Transformer Based Machine Fault Detection From Audio Input
这篇论文提出使用Transformer模型来分析机器音频,以检测故障,并证明它在分析声音频谱图方面比传统的卷积神经网络(CNN)更具潜力,尤其是在数据充足的情况下。
源自 arXiv: 2604.12733