Deep Separable Hypercomplex Networks
DOI:
https://doi.org/10.32473/flairs.36.133540Palabras clave:
Hypercomplex networks, Separable Convolution, Separable hypercomplex networks, Quaternion CNNs, Vectormap CNNsResumen
Deep hypercomplex-inspired convolutional neural networks (CNNs) have recently enhanced feature extraction for image classification by allowing weight sharing across input channels. This makes it possible to improve the representation acquisition abilities of the networks. Hypercomplex-inspired networks, however, still incur higher computational costs than standard CNNs.
This paper reduces this cost by decomposing a quaternion 2D convolutional module into two consecutive separable vectormap modules.
In addition, we use 4 and 5D parameterized hypercomplex multiplication-based fully connected layers. Incorporating both yields our proposed hypercomplex CNN, a novel architecture that can be assembled to construct deep separable hypercomplex networks (SHNNs) for image classification.
We conduct experiments on CIFAR, SVHN, and Tiny ImageNet datasets and achieve better performance using fewer trainable parameters and FLOPS. Our proposed model achieves almost 2% higher performance for CIFAR and SVHN datasets and more than 3% for the ImageNet-Tiny dataset and takes 84%, 35%, and 51% fewer parameters than the ResNets, quaternion, and vectormap networks, respectively.
Also, we achieve state-of-the-art performance on CIFAR benchmarks in hypercomplex space.
Descargas
Publicado
Cómo citar
Número
Sección
Licencia
Derechos de autor 2023 Nazmul Shahadat, Anthony S. Maida
Esta obra está bajo una licencia internacional Creative Commons Atribución-NoComercial 4.0.