Pooling operations, essential for neural networks (NNs), reduce feature map dimensions while preserving key features and enhancing spatial invariance. Traditional pooling methods often miss the feature maps’ alternating current (AC) components. This study introduces a novel global pooling technique utilizing spectral self-attention, leveraging the discrete cosine transform (DCT) for spectral analysis and a self-attention mechanism for assessing frequency component significance. This approach allows for efficient feature synthesis through weighted averaging, significantly boosting TOP-1 accuracy with minimal parameter increase, outperforming existing models.