scholarly journals PVCLN: Point-View Complementary Learning Network for 3D Shape Recognition

IEEE Access ◽  
2021 ◽  
Vol 9 ◽  
pp. 3451-3460
Author(s):  
Shanlin Sun ◽  
Yun Li ◽  
Minjie Ren ◽  
Guo Li ◽  
Xing Yao
Electronics ◽  
2020 ◽  
Vol 9 (4) ◽  
pp. 649
Author(s):  
Long Hoang ◽  
Suk-Hwan Lee ◽  
Ki-Ryong Kwon

3D shape recognition becomes necessary due to the popularity of 3D data resources. This paper aims to introduce the new method, hybrid deep learning network convolution neural network–support vector machine (CNN–SVM), for 3D recognition. The vertices of the 3D mesh are interpolated to be converted into Point Clouds; those Point Clouds are rotated for 3D data augmentation. We obtain and store the 2D projection of this 3D augmentation data in a 32 × 32 × 12 matrix, the input data of CNN–SVM. An eight-layer CNN is used as the algorithm for feature extraction, then SVM is applied for classifying feature extraction. Two big datasets, ModelNet40 and ModelNet10, of the 3D model are used for model validation. Based on our numerical experimental results, CNN–SVM is more accurate and efficient than other methods. The proposed method is 13.48% more accurate than the PointNet method in ModelNet10 and 8.5% more precise than 3D ShapeNets for ModelNet40. The proposed method works with both the 3D model in the augmented/virtual reality system and in the 3D Point Clouds, an output of the LIDAR sensor in autonomously driving cars.


Author(s):  
Jie Nie ◽  
Zhi-Qiang Wei ◽  
Weizhi Nie ◽  
An-An Liu

Three-dimensional (3D) shape recognition is a popular topic and has potential application value in the field of computer vision. With the recent proliferation of deep learning, various deep learning models have achieved state-of-the-art performance. Among them, multiview-based 3D shape representation has received increased attention in recent years, and related approaches have shown significant improvement in 3D shape recognition. However, these methods focus on feature learning based on the design of the network and ignore the correlation among views. In this article, we propose a novel progressive feature guide learning network (PGNet) that focuses on the correlation among multiple views and integrates multiple modalities for 3D shape recognition. In particular, we propose two information fusion schemes from visual and feature aspects. The visual fusion scheme focuses on the view level and employs the soft-attention model to define the weights of views for visual information fusion. The feature fusion scheme focuses on the feature dimension information and employs the quantified feature as the mask to further optimize the feature. These two schemes jointly construct a PGNet for 3D shape representation. The classic ModelNet40 and ShapeNetCore55 datasets are applied to demonstrate the performance of our approach. The corresponding experiment also demonstrates the superiority of our approach.


Author(s):  
Huazhen Chu ◽  
Chao Le ◽  
Rongquan Wang ◽  
Xi Li ◽  
Huimin Ma
Keyword(s):  

2013 ◽  
Vol 106 (3) ◽  
pp. 332-341 ◽  
Author(s):  
Oliver J. Woodford ◽  
Minh-Tri Pham ◽  
Atsuto Maki ◽  
Frank Perbet ◽  
Björn Stenger

Sign in / Sign up

Export Citation Format

Share Document