Abstract
Deep learning frameworks(such as deep convolutional networks) require data to have a regular shape. However, discrete features extracted from heterogeneous data cannot be collected in a regular shape to convolute. In this article, a Two-Dimensional Discrete Feature Based Spatial Attention CapsNet(TDACAPS) is proposed to convert one-dimensional discrete features into two-dimensional structured data through Cartesian Product for surface electromyogram(sEMG) signal recognition. sEMG signal varies from person to person is the main signal source of prosthetic control. Our model transforms multi-angle discrete features into structured data to find the inherent law of sEMG signal. Due to uneven information distribution of structured data, this model combines capsule network with attention mechanism to place emphasis on abundant information regions and reduce ancillary information loss. Extensive experiments show our model yields an improvement for sEMG signal recognition of almost 3% than capsule network and other neural networks under different conditions. Our attention mechanism that employs overlapping pooling to search feature map weight is preferable to the squeeze-and-excitation module, convolutional block attention module and others. Moreover, we validate that our model has great expansibility on Wine Quality Dataset and Breast Cancer Wisconsin.
| Original language | English |
|---|---|
| Pages (from-to) | 3503-3520 |
| Number of pages | 18 |
| Journal | Applied Intelligence |
| Volume | 50 |
| Issue number | 10 |
| DOIs | |
| State | Published - 1 Oct 2020 |
| Externally published | Yes |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 3 Good Health and Well-being
Keywords
- Capsule network
- Spatial attention
- Two-Dimensional discrete feature
- sEMG signal
Fingerprint
Dive into the research topics of 'Two-dimensional discrete feature based spatial attention CapsNet For sEMG signal recognition'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver