z-logo
open-access-imgOpen Access
Hierarchical bilinear convolutional neural network for image classification
Author(s) -
Zhang Xiang,
Tang Lei,
Luo Hangzai,
Zhong Sheng,
Guan Ziyu,
Chen Long,
Zhao Chao,
Peng Jinye,
Fan Jianping
Publication year - 2021
Publication title -
iet computer vision
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.38
H-Index - 37
eISSN - 1751-9640
pISSN - 1751-9632
DOI - 10.1049/cvi2.12023
Subject(s) - computer science , convolutional neural network , artificial intelligence , discriminative model , pattern recognition (psychology) , feature (linguistics) , bilinear interpolation , context (archaeology) , contextual image classification , feature vector , machine learning , image (mathematics) , computer vision , paleontology , philosophy , linguistics , biology
Image classification is one of the mainstream tasks of computer vision. However, the most existing methods use labels of the same granularity level for training. This leads to ignoring the hierarchy that may help to differentiate different visual objects better. Embedding hierarchical information into the convolutional neural networks (CNNs) can effectively regulate the semantic space and thus reduce the ambiguity of prediction. To this end, a multi‐task learning framework, named as Hierarchical Bilinear Convolutional Neural Network (HB‐CNN), is developed by seamlessly integrating CNNs with multi‐task learning over the hierarchical visual concept structures. Specifically, the labels with a tree structure are used as the supervision to hierarchically train multiple branch networks. In this way, the model can not only learn additional information (e.g. context information) as the coarse‐level category features, but also focus the learned fine‐level category features on the object properties. To smoothly pass hierarchical conceptual information and encourage feature reuse, a connectivity pattern is proposed to connect features at different levels. Furthermore, a bilinear module is embedded to generalise various orderless texture feature descriptors so that our model can capture more discriminative features. The proposed method is extensively evaluated on the CIFAR‐10, CIFAR‐100, and ‘Orchid’ Plant image sets. The experimental results show the effectiveness and superiority of our method.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here