Abstract
Pooling operations, essential for neural networks (NNs), reduce feature
map dimensions while preserving key features and enhancing spatial
invariance. Traditional pooling methods often miss the feature maps’
alternating current (AC) components. This study introduces a novel
global pooling technique utilizing spectral self-attention, leveraging
the discrete cosine transform (DCT) for spectral analysis and a
self-attention mechanism for assessing frequency component significance.
This approach allows for efficient feature synthesis through weighted
averaging, significantly boosting TOP-1 accuracy with minimal parameter
increase, outperforming existing models.