Adaptive convolution kernel for artificial neural networks

F. Boray Tek*, İlker Çam, Deniz Karlı

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

11 Citations (Scopus)

Abstract

Many deep neural networks are built by using stacked convolutional layers of fixed and single size (often 3 × 3) kernels. This paper describes a method for learning the size of convolutional kernels to provide varying size kernels in a single layer. The method utilizes a differentiable, and therefore backpropagation-trainable Gaussian envelope which can grow or shrink in a base grid. Our experiments compared the proposed adaptive layers to ordinary convolution layers in a simple two-layer network, a deeper residual network, and a U-Net architecture. The results in the popular image classification datasets such as MNIST, MNIST-CLUTTERED, CIFAR-10, Fashion, and “Faces in the Wild” showed that the adaptive kernels can provide statistically significant improvements on ordinary convolution kernels. A segmentation experiment in the Oxford-Pets dataset demonstrated that replacing ordinary convolution layers in a U-shaped network with 7 × 7 adaptive layers can improve its learning performance and ability to generalize.

Original languageEnglish
Article number103015
JournalJournal of Visual Communication and Image Representation
Volume75
DOIs
Publication statusPublished - Feb 2021
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2021 Elsevier Inc.

Keywords

  • Adaptive convolution
  • Image classification
  • Multi-scale convolution
  • Residual networks

Fingerprint

Dive into the research topics of 'Adaptive convolution kernel for artificial neural networks'. Together they form a unique fingerprint.

Cite this