
Dynamic Patch Convolution (DPConv)
Author(s) -
Shuchao Deng
Publication year - 2021
Publication title -
journal of physics. conference series
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.21
H-Index - 85
eISSN - 1742-6596
pISSN - 1742-6588
DOI - 10.1088/1742-6596/1865/4/042079
Subject(s) - convolution (computer science) , kernel (algebra) , convolutional neural network , block (permutation group theory) , computer science , representation (politics) , dimension (graph theory) , algorithm , block size , feature (linguistics) , pattern recognition (psychology) , artificial intelligence , mathematics , artificial neural network , key (lock) , discrete mathematics , geometry , linguistics , philosophy , computer security , politics , political science , pure mathematics , law
Lightweight Convolutional Neural Networks (CNNs) due to the small amount of calculation and performance degradation, budget constraints depth (number of convolution layers), and width of CNN (number of channels), resulting in limited representation capabilities. To solve this problem, this paper proposes a new dynamic block convolution design to increase the depth or width of the model complexity without increasing the network. Instead of using a single convolution kernel in each layer, dynamic block convolution dynamically aggregates multiple parallel convolution kernels based on input-related attention. Assembling multiple kernels is not only computationally efficient, but also due to the small size of the kernel, but also has more representation capabilities. These kernels are aggregated in a non-linear manner through attention. The two methods of spatial dimension block and multi-head are used in the traditional CNN attention design. The image is divided into small images and then multiple attention mechanism weights are used for each small image. Compared with the traditional CNN attention, the experimental effect is better.