scholarly journals Local Importance Representation Convolutional Neural Network for Fine-Grained Image Classification

Symmetry ◽  
2018 ◽  
Vol 10 (10) ◽  
pp. 479 ◽  
Author(s):  
Yadong Yang ◽  
Xiaofeng Wang ◽  
Hengzheng Zhang

Compared with ordinary image classification tasks, fine-grained image classification is closer to real-life scenes. Its key point is how to find the local areas with sufficient discrimination and perform effective feature learning. Based on a bilinear convolutional neural network (B-CNN), this paper designs a local importance representation convolutional neural network (LIR-CNN) model, which can be divided into three parts. Firstly, the super-pixel segmentation convolution method is used for the input layer of the model. It allows the model to receive images of different sizes and fully considers the complex geometric deformation of the images. Then, we replaced the standard convolution of B-CNN with the proposed local importance representation convolution. It can score each local area of the image using learning to distinguish their importance. Finally, channelwise convolution is proposed and it plays an important role in balancing lightweight network and classification accuracy. Experimental results on the benchmark datasets (e.g., CUB-200-2011, FGVC-Aircraft, and Stanford Cars) showed that the LIR-CNN model had good performance in fine-grained image classification tasks.

2021 ◽  
Vol 2021 ◽  
pp. 1-6
Author(s):  
Qiang Cai ◽  
Fenghai Li ◽  
Yifan Chen ◽  
Haisheng Li ◽  
Jian Cao ◽  
...  

Along with the strong representation of the convolutional neural network (CNN), image classification tasks have achieved considerable progress. However, majority of works focus on designing complicated and redundant architectures for extracting informative features to improve classification performance. In this study, we concentrate on rectifying the incomplete outputs of CNN. To be concrete, we propose an innovative image classification method based on Label Rectification Learning (LRL) through kernel extreme learning machine (KELM). It mainly consists of two steps: (1) preclassification, extracting incomplete labels through a pretrained CNN, and (2) label rectification, rectifying the generated incomplete labels by the KELM to obtain the rectified labels. Experiments conducted on publicly available datasets demonstrate the effectiveness of our method. Notably, our method is extensible which can be easily integrated with off-the-shelf networks for improving performance.


2021 ◽  
Vol 2021 ◽  
pp. 1-8
Author(s):  
Lin Liu

HTP test in psychometrics is a widely studied and applied psychological assessment technique. HTP test is a kind of projection test, which refers to the free expression of painting itself and its creativity. Therefore, the form of group psychological counselling is widely used in mental health education. Compared with traditional neural networks, deep learning networks have deeper and more network layers and can learn more complex processing functions. In this stage, image recognition technology can be used as an assistant of human vision. People can quickly get the information in the picture through retrieval. For example, you can take a picture of an object that is difficult to describe and quickly search the content related to it. Convolutional neural network, which is widely used in the image classification task of computer vision, can automatically complete feature learning on the data without manual feature extraction. Compared with the traditional test, the test can reflect the painting characteristics of different groups. After quantitative scoring, it has good reliability and validity. It has high application value in psychological evaluation, especially in the diagnosis of mental diseases. This paper focuses on the subjectivity of HTP evaluation. Convolutional neural network is a mature technology in deep learning. The traditional HTP assessment process relies on the experience of researchers to extract painting features and classification.


2020 ◽  
Vol 12 (12) ◽  
pp. 2033 ◽  
Author(s):  
Xiaofei Yang ◽  
Xiaofeng Zhang ◽  
Yunming Ye ◽  
Raymond Y. K. Lau ◽  
Shijian Lu ◽  
...  

Accurate hyperspectral image classification has been an important yet challenging task for years. With the recent success of deep learning in various tasks, 2-dimensional (2D)/3-dimensional (3D) convolutional neural networks (CNNs) have been exploited to capture spectral or spatial information in hyperspectral images. On the other hand, few approaches make use of both spectral and spatial information simultaneously, which is critical to accurate hyperspectral image classification. This paper presents a novel Synergistic Convolutional Neural Network (SyCNN) for accurate hyperspectral image classification. The SyCNN consists of a hybrid module that combines 2D and 3D CNNs in feature learning and a data interaction module that fuses spectral and spatial hyperspectral information. Additionally, it introduces a 3D attention mechanism before the fully-connected layer which helps filter out interfering features and information effectively. Extensive experiments over three public benchmarking datasets show that our proposed SyCNNs clearly outperform state-of-the-art techniques that use 2D/3D CNNs.


2018 ◽  
Vol 8 (12) ◽  
pp. 2529 ◽  
Author(s):  
Xiaoqing Wang ◽  
Xiangjun Wang

When large-scale annotated data are not available for certain image classification tasks, training a deep convolutional neural network model becomes challenging. Some recent domain adaptation methods try to solve this problem using generative adversarial networks and have achieved promising results. However, these methods are based on a shared latent space assumption and they do not consider the situation when shared high level representations in different domains do not exist or are not ideal as they assumed. To overcome this limitation, we propose a neural network structure called coupled generative adversarial autoencoders (CGAA) that allows a pair of generators to learn the high-level differences between two domains by sharing only part of the high-level layers. Additionally, by introducing a class consistent loss calculated by a stand-alone classifier into the generator optimization, our model is able to generate class invariant style-transferred images suitable for classification tasks in domain adaptation. We apply CGAA to several domain transferred image classification scenarios including several benchmark datasets. Experiment results have shown that our method can achieve state-of-the-art classification results.


2019 ◽  
Vol 6 (1) ◽  
Author(s):  
Samir S. Yadav ◽  
Shivajirao M. Jadhav

AbstractMedical image classification plays an essential role in clinical treatment and teaching tasks. However, the traditional method has reached its ceiling on performance. Moreover, by using them, much time and effort need to be spent on extracting and selecting classification features. The deep neural network is an emerging machine learning method that has proven its potential for different classification tasks. Notably, the convolutional neural network dominates with the best results on varying image classification tasks. However, medical image datasets are hard to collect because it needs a lot of professional expertise to label them. Therefore, this paper researches how to apply the convolutional neural network (CNN) based algorithm on a chest X-ray dataset to classify pneumonia. Three techniques are evaluated through experiments. These are linear support vector machine classifier with local rotation and orientation free features, transfer learning on two convolutional neural network models: Visual Geometry Group i.e., VGG16 and InceptionV3, and a capsule network training from scratch. Data augmentation is a data preprocessing method applied to all three methods. The results of the experiments show that data augmentation generally is an effective way for all three algorithms to improve performance. Also, Transfer learning is a more useful classification method on a small dataset compared to a support vector machine with oriented fast and rotated binary (ORB) robust independent elementary features and capsule network. In transfer learning, retraining specific features on a new target dataset is essential to improve performance. And, the second important factor is a proper network complexity that matches the scale of the dataset.


Sign in / Sign up

Export Citation Format

Share Document