scholarly journals A Novel Bilinear Feature and Multi-Layer Fused Convolutional Neural Network for Tactile Shape Recognition

Sensors ◽  
2020 ◽  
Vol 20 (20) ◽  
pp. 5822
Author(s):  
Jie Chu ◽  
Jueping Cai ◽  
He Song ◽  
Yuxin Zhang ◽  
Linyu Wei

Convolutional neural networks (CNNs) can automatically learn features from pressure information, and some studies have applied CNNs for tactile shape recognition. However, the limited density of the sensor and its flexibility requirement lead the obtained tactile images to have a low-resolution and blurred. To address this issue, we propose a bilinear feature and multi-layer fused convolutional neural network (BMF-CNN). The bilinear calculation of the feature improves the feature extraction capability of the network. Meanwhile, the multi-layer fusion strategy exploits the complementarity of different layers to enhance the feature utilization efficiency. To validate the proposed method, a 26 class letter-shape tactile image dataset with complex edges was constructed. The BMF-CNN model achieved a 98.64% average accuracy of tactile shape. The results show that BMF-CNN can deal with tactile shapes more effectively than traditional CNN and artificial feature methods.

2021 ◽  
Vol 11 (9) ◽  
pp. 4292
Author(s):  
Mónica Y. Moreno-Revelo ◽  
Lorena Guachi-Guachi ◽  
Juan Bernardo Gómez-Mendoza ◽  
Javier Revelo-Fuelagán ◽  
Diego H. Peluffo-Ordóñez

Automatic crop identification and monitoring is a key element in enhancing food production processes as well as diminishing the related environmental impact. Although several efficient deep learning techniques have emerged in the field of multispectral imagery analysis, the crop classification problem still needs more accurate solutions. This work introduces a competitive methodology for crop classification from multispectral satellite imagery mainly using an enhanced 2D convolutional neural network (2D-CNN) designed at a smaller-scale architecture, as well as a novel post-processing step. The proposed methodology contains four steps: image stacking, patch extraction, classification model design (based on a 2D-CNN architecture), and post-processing. First, the images are stacked to increase the number of features. Second, the input images are split into patches and fed into the 2D-CNN model. Then, the 2D-CNN model is constructed within a small-scale framework, and properly trained to recognize 10 different types of crops. Finally, a post-processing step is performed in order to reduce the classification error caused by lower-spatial-resolution images. Experiments were carried over the so-named Campo Verde database, which consists of a set of satellite images captured by Landsat and Sentinel satellites from the municipality of Campo Verde, Brazil. In contrast to the maximum accuracy values reached by remarkable works reported in the literature (amounting to an overall accuracy of about 81%, a f1 score of 75.89%, and average accuracy of 73.35%), the proposed methodology achieves a competitive overall accuracy of 81.20%, a f1 score of 75.89%, and an average accuracy of 88.72% when classifying 10 different crops, while ensuring an adequate trade-off between the number of multiply-accumulate operations (MACs) and accuracy. Furthermore, given its ability to effectively classify patches from two image sequences, this methodology may result appealing for other real-world applications, such as the classification of urban materials.


2021 ◽  
Vol 55 (4) ◽  
pp. 88-98
Author(s):  
Maria Inês Pereira ◽  
Pedro Nuno Leite ◽  
Andry Maykol Pinto

Abstract The maritime industry has been following the paradigm shift toward the automation of typically intelligent procedures, with research regarding autonomous surface vehicles (ASVs) having seen an upward trend in recent years. However, this type of vehicle cannot be employed on a full scale until a few challenges are solved. For example, the docking process of an ASV is still a demanding task that currently requires human intervention. This research work proposes a volumetric convolutional neural network (vCNN) for the detection of docking structures from 3-D data, developed according to a balance between precision and speed. Another contribution of this article is a set of synthetically generated data regarding the context of docking structures. The dataset is composed of LiDAR point clouds, stereo images, GPS, and Inertial Measurement Unit (IMU) information. Several robustness tests carried out with different levels of Gaussian noise demonstrated an average accuracy of 93.34% and a deviation of 5.46% for the worst case. Furthermore, the system was fine-tuned and evaluated in a real commercial harbor, achieving an accuracy of over 96%. The developed classifier is able to detect different types of structures and works faster than other state-of-the-art methods that establish their performance in real environments.


2017 ◽  
Vol 10 (27) ◽  
pp. 1329-1342 ◽  
Author(s):  
Javier O. Pinzon Arenas ◽  
Robinson Jimenez Moreno ◽  
Paula C. Useche Murillo

This paper presents the implementation of a Region-based Convolutional Neural Network focused on the recognition and localization of hand gestures, in this case 2 types of gestures: open and closed hand, in order to achieve the recognition of such gestures in dynamic backgrounds. The neural network is trained and validated, achieving a 99.4% validation accuracy in gesture recognition and a 25% average accuracy in RoI localization, which is then tested in real time, where its operation is verified through times taken for recognition, execution behavior through trained and untrained gestures, and complex backgrounds.


2020 ◽  
Vol 2020 ◽  
pp. 1-7
Author(s):  
Ahmed Jawad A. AlBdairi ◽  
Zhu Xiao ◽  
Mohammed Alghaili

The interest in face recognition studies has grown rapidly in the last decade. One of the most important problems in face recognition is the identification of ethnics of people. In this study, a new deep learning convolutional neural network is designed to create a new model that can recognize the ethnics of people through their facial features. The new dataset for ethnics of people consists of 3141 images collected from three different nationalities. To the best of our knowledge, this is the first image dataset collected for the ethnics of people and that dataset will be available for the research community. The new model was compared with two state-of-the-art models, VGG and Inception V3, and the validation accuracy was calculated for each convolutional neural network. The generated models have been tested through several images of people, and the results show that the best performance was achieved by our model with a verification accuracy of 96.9%.


2020 ◽  
Vol 37 (9) ◽  
pp. 1661-1668
Author(s):  
Min Wang ◽  
Shudao Zhou ◽  
Zhong Yang ◽  
Zhanhua Liu

AbstractConventional classification methods are based on artificial experience to extract features, and each link is independent, which is a kind of “shallow learning.” As a result, the scope of the cloud category applied by this method is limited. In this paper, we propose a new convolutional neural network (CNN) with deep learning ability, called CloudA, for the ground-based cloud image recognition method. We use the Singapore Whole-Sky Imaging Categories (SWIMCAT) sample library and total-sky sample library to train and test CloudA. In particular, we visualize the cloud features captured by CloudA using the TensorBoard visualization method, and these features can help us to understand the process of ground-based cloud classification. We compare this method with other commonly used methods to explore the feasibility of using CloudA to classify ground-based cloud images, and the evaluation of a large number of experiments show that the average accuracy of this method is nearly 98.63% for ground-based cloud classification.


2019 ◽  
Author(s):  
Marcelo Vilela Vizoni ◽  
Aparecido Nilceu Marana

This paper presents a new method for person authentication that relies on the fusion of two biometric authentication methods based, respectively, on ocular deep features and facial deep features. In our work, the deep features are extracted from the regions of interest by using a very deep CNN (Convolutional Neural Network). Another interesting aspect of our work is that, instead of using directly the deep features as input for the authentication methods, we use the difference between the probe and gallery deep features. So, our method adopts a pairwise strategy. Support Vector Machine classifiers are trained separately for each approach. The fusion of the ocular and the facial based methods are carried out in the score level. The proposed method was assessed with a facial database taken under uncontrolled environment and reached good results. Besides, the fusion strategy proposed in this work showed better results than the results obtained by each individual method.


2020 ◽  
Vol 5 (2) ◽  
pp. 83-88
Author(s):  
Hedi Pandowo

Deep Learning is part of the scientific field of Machine Learning and Machine Learning is part of Artificial Intelligence science. Deep Learning has extraordinary capabilities by using a hardware Graphical Processing Unit (GPU) so that the artificial requirement network can run faster than using a Personal Computer Unit (CPU). Especially in terms of object classification in images using existing methods in the Convolutional Neural Network (CNN). The method used in this research is Preprocessing and Processing of Input Data, Training Process in which CNN is trained to obtain high accuracy from the classification carried out and the Testing Process which is a classification process using weights and bias from the results of the training process. This type of research is a pre experimental design (pre experimental design). The results of the object image classification test with different levels of confusion in the Concrete database with the Mix Design K-125, K-150, K-250 and K-300 produce an average accuracy value. This is also relevant to measuring the failure rate of concrete or slump


2020 ◽  
Author(s):  
Yang Xu ◽  
Ting Ting Qiu

With the improvement of people's living standards, the demand for health monitoring and exercise detection is increasing. It is of great significance to study human activity recognition methods that are different from traditional feature extraction methods. This article uses convolutional neural network algorithms in deep learning to automatically extract features of activities related to human life. It uses a stochastic gradient descent algorithm to optimize the parameters of the convolutional neural network. The trained network model is compressed on STM32CubeMX-AI. Finally, this article introduces the use of neural networks on embedded devices to recognize six human activities of daily life, such as sitting, standing, walking, jogging, upstairs and downstairs. The acceleration sensor related to human activity information is used to obtain the relevant characteristics of the activity, thereby solving the human activity recognition (HAR) problem. The network structure of the constructed CNN model is shown in Figure 1, including an input layer, two convolutional layers and two pooling layers. After comparing the average accuracy of each set of experiments and the test set of the best model obtained from it, the best model is then selected.


2019 ◽  
Author(s):  
Krishanveer Gangwar ◽  
Vimal Kumar ◽  
Ajay Kr. Singh ◽  
Vijay Kr. Sharma

Sign in / Sign up

Export Citation Format

Share Document