Research Article

Multiscale Time-Frequency Sparse Transformer Based on Partly Interpretable Method for Bearing Fault Diagnosis

Figure 2

Self-attention mechanism. (a) The scaled dot-product self-attention used in MSA, (b) the sparse scaled dot-product self-attention used in MTFST, and (c) the examples of different attention mechanisms.
(a)
(b)
(c)