Sorting Convolution Operation for Achieving Rotational Invariance
Mo, Hanlin; Zhao, Guoying (2024-03-26)
Mo, Hanlin
Zhao, Guoying
IEEE
26.03.2024
H. Mo and G. Zhao, "Sorting Convolution Operation for Achieving Rotational Invariance," in IEEE Signal Processing Letters, vol. 31, pp. 1199-1203, 2024, doi: 10.1109/LSP.2024.3381909
https://creativecommons.org/licenses/by/4.0/
© 2024 The Authors. This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
https://creativecommons.org/licenses/by/4.0/
© 2024 The Authors. This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
https://creativecommons.org/licenses/by/4.0/
Julkaisun pysyvä osoite on
https://urn.fi/URN:NBN:fi:oulu-202405143445
https://urn.fi/URN:NBN:fi:oulu-202405143445
Tiivistelmä
Abstract
The topic of achieving rotational invariance in convolutional neural networks (CNNs) has gained considerable attention recently, as this invariance is crucial for many computer vision tasks. In this letter, we propose a sorting convolution operation ( SConv ), which achieves invariance to arbitrary rotations without additional learnable parameters or data augmentation. It can directly replace conventional convolution operations in a classic CNN model to achieve the model's rotational invariance. Based on MNIST-rot dataset, we first analyze the impact of convolution kernel size, sampling grid and sorting method on SConv ’s rotational invariance, and compare our method with previous rotation-invariant CNN models. Then, we combine SConv with VGG, ResNet and DenseNet, and conduct classification experiments on texture and remote sensing image datasets. The results show that SConv significantly improves the performance of these models, especially when training data is limited.
The topic of achieving rotational invariance in convolutional neural networks (CNNs) has gained considerable attention recently, as this invariance is crucial for many computer vision tasks. In this letter, we propose a sorting convolution operation ( SConv ), which achieves invariance to arbitrary rotations without additional learnable parameters or data augmentation. It can directly replace conventional convolution operations in a classic CNN model to achieve the model's rotational invariance. Based on MNIST-rot dataset, we first analyze the impact of convolution kernel size, sampling grid and sorting method on SConv ’s rotational invariance, and compare our method with previous rotation-invariant CNN models. Then, we combine SConv with VGG, ResNet and DenseNet, and conduct classification experiments on texture and remote sensing image datasets. The results show that SConv significantly improves the performance of these models, especially when training data is limited.
Kokoelmat
- Avoin saatavuus [38618]