scholarly journals Efficient ConvNet Feature Extraction with Multiple RoI Pooling for Landmark-Based Visual Localization of Autonomous Vehicles

2017 ◽  
Vol 2017 ◽  
pp. 1-14 ◽  
Author(s):  
Yi Hou ◽  
Hong Zhang ◽  
Shilin Zhou ◽  
Huanxin Zou

Efficient and robust visual localization is important for autonomous vehicles. By achieving impressive localization accuracy under conditions of significant changes, ConvNet landmark-based approach has attracted the attention of people in several research communities including autonomous vehicles. Such an approach relies heavily on the outstanding discrimination power of ConvNet features to match detected landmarks between images. However, a major challenge of this approach is how to extract discriminative ConvNet features efficiently. To address this challenging, inspired by the high efficiency of the region of interest (RoI) pooling layer, we propose a Multiple RoI (MRoI) pooling technique, an enhancement of RoI, and a simple yet efficient ConvNet feature extraction method. Our idea is to leverage MRoI pooling to exploit multilevel and multiresolution information from multiple convolutional layers and then fuse them to improve the discrimination capacity of the final ConvNet features. The main advantages of our method are (a) high computational efficiency for real-time applications; (b) GPU memory efficiency for mobile applications; and (c) use of pretrained model without fine-tuning or retraining for easy implementation. Experimental results on four datasets have demonstrated not only the above advantages but also the high discriminating power of the extracted ConvNet features with state-of-the-art localization accuracy.

2014 ◽  
Vol 1008-1009 ◽  
pp. 1509-1512
Author(s):  
Qing E Wu ◽  
Hong Wang ◽  
Li Fen Ding

To carry out an effective classification and recognition for target, this paper studied the target owned characteristics, discussed a decryption algorithm, gave a feature extraction method based on the decryption process, and extracted the feature of palmprint in region of interest. Moreover, this paper used the wavelet transform to extract the energy feature of target, gave an approach on matching and recognition to improve the correctness and efficiency of existing recognition approaches, and compared it with existing approaches of palmprint recognition by experiments. The experiment results show that the correct recognition rate of the approach in this paper is improved averagely by 2.34% than that of the existing recognition approaches.


2017 ◽  
Vol 2017 ◽  
pp. 1-15 ◽  
Author(s):  
Min Dong ◽  
Zhe Wang ◽  
Chenghui Dong ◽  
Xiaomin Mu ◽  
Yide Ma

Goal. Breast cancer is becoming one of the most common cancers among women. Early detection can help increase the survival rates. Feature extraction directly affects diagnosis result. In this work, a novel feature extraction method based on Dual Contourlet Transform (Dual-CT) is presented, and improved K nearest neighbor (KNN) is employed to improve the classification performance. Method. This presented method includes three main sections: firstly, the Region of Interest (ROI) is cropped manually according to gold standard from Mammographic Image Analysis Society (MIAS) database; secondly, the ROIs are decomposed into different resolution levels using Dual-CT, contourlet, and wavelet; a set of texture features are extracted. Then improved KNN and traditional KNN are implemented for classification. Experiments are performed on 324 ROIs which include 206 normal cases and 118 abnormal cases; the abnormal cases are composed of 66 benign cases and 52 malignant cases. Results. Experimental results prove the validity and superiority of Dual-CT-based feature and improved KNN. In particular, 94.14% and 95.76% classification accuracy is achieved based on Dual-CT domain. Moreover, the proposed method is comparable with state-of-the-art methods in terms of accuracy. Contribution. Dual-CT-based feature is used for analyzing mammogram and can help improve breast cancer diagnosis accuracy.


2019 ◽  
Author(s):  
William Barcellos ◽  
Nicolas Hiroaki Shitara ◽  
Carolina Toledo Ferraz ◽  
Raissa Tavares Vieira Queiroga ◽  
Jose Hiroki Saito ◽  
...  

The aim of this paper is to evaluate the performance of Transfer Learning techniques applied in Convolucional Neural Networks for biometric periocular classification. Two aspects of Transfer Learning were evaluated: the technique known as Fine Tuning and the technique known as Feature Extraction. Two CNN architectures were evaluated, the AlexNet and the VGG-16, and two image databases were used. These two databases have different characteristics regarding the method of acquisition, the amount of classes, the class balancing, and the number of elements in each class. Three experiments were conducted to evaluate the performance of the CNNs. In the first experiment we measured the Feature Extraction accuracy, and in the second one we evaluated the Fine Tuning performance. In the third experiment, we used the AlexNet for Fine Tuning in one database, and then, the FC7 layer of this trained CNN was used for Feature Extraction in the other database. We concluded that the data quality (the presence or not of class samples in the training set), class imbalance (different number of elements in each class) and the selection method of the training and testing, directly influence the CNN accuracy. The Feature Extraction method, by being more simple and does not require network training, has lower accuracy than Fine Tuning. Furthermore, Fine Tuning a CNN with periocular's images from one database, doesn't increase the accuracy of this CNN in Feature Extraction mode for another periocular's database. The accuracy is quite similar to that obtained by the original pre-trained network


Sensors ◽  
2021 ◽  
Vol 21 (16) ◽  
pp. 5283
Author(s):  
Tahira Nazir ◽  
Marriam Nawaz ◽  
Junaid Rashid ◽  
Rabbia Mahum ◽  
Momina Masood ◽  
...  

Diabetic retinopathy (DR) is an eye disease that alters the blood vessels of a person suffering from diabetes. Diabetic macular edema (DME) occurs when DR affects the macula, which causes fluid accumulation in the macula. Efficient screening systems require experts to manually analyze images to recognize diseases. However, due to the challenging nature of the screening method and lack of trained human resources, devising effective screening-oriented treatment is an expensive task. Automated systems are trying to cope with these challenges; however, these methods do not generalize well to multiple diseases and real-world scenarios. To solve the aforementioned issues, we propose a new method comprising two main steps. The first involves dataset preparation and feature extraction and the other relates to improving a custom deep learning based CenterNet model trained for eye disease classification. Initially, we generate annotations for suspected samples to locate the precise region of interest, while the other part of the proposed solution trains the Center Net model over annotated images. Specifically, we use DenseNet-100 as a feature extraction method on which the one-stage detector, CenterNet, is employed to localize and classify the disease lesions. We evaluated our method over challenging datasets, namely, APTOS-2019 and IDRiD, and attained average accuracy of 97.93% and 98.10%, respectively. We also performed cross-dataset validation with benchmark EYEPACS and Diaretdb1 datasets. Both qualitative and quantitative results demonstrate that our proposed approach outperforms state-of-the-art methods due to more effective localization power of CenterNet, as it can easily recognize small lesions and deal with over-fitted training data. Our proposed framework is proficient in correctly locating and classifying disease lesions. In comparison to existing DR and DME classification approaches, our method can extract representative key points from low-intensity and noisy images and accurately classify them. Hence our approach can play an important role in automated detection and recognition of DR and DME lesions.


2022 ◽  
Vol 22 (1) ◽  
Author(s):  
Zhao Shuai ◽  
Diao Xiaolin ◽  
Yuan Jing ◽  
Huo Yanni ◽  
Cui Meng ◽  
...  

Abstract Background Automated ICD coding on medical texts via machine learning has been a hot topic. Related studies from medical field heavily relies on conventional bag-of-words (BoW) as the feature extraction method, and do not commonly use more complicated methods, such as word2vec (W2V) and large pretrained models like BERT. This study aimed at uncovering the most effective feature extraction methods for coding models by comparing BoW, W2V and BERT variants. Methods We experimented with a Chinese dataset from Fuwai Hospital, which contains 6947 records and 1532 unique ICD codes, and a public Spanish dataset, which contains 1000 records and 2557 unique ICD codes. We designed coding tasks with different code frequency thresholds (denoted as $$f_s$$ f s ), with a lower threshold indicating a more complex task. Using traditional classifiers, we compared BoW, W2V and BERT variants on accomplishing these coding tasks. Results When $$f_s$$ f s was equal to or greater than 140 for Fuwai dataset, and 60 for the Spanish dataset, the BERT variants with the whole network fine-tuned was the best method, leading to a Micro-F1 of 93.9% for Fuwai data when $$f_s=200$$ f s = 200 , and a Micro-F1 of 85.41% for the Spanish dataset when $$f_s=180$$ f s = 180 . When $$f_s$$ f s fell below 140 for Fuwai dataset, and 60 for the Spanish dataset, BoW turned out to be the best, leading to a Micro-F1 of 83% for Fuwai dataset when $$f_s=20$$ f s = 20 , and a Micro-F1 of 39.1% for the Spanish dataset when $$f_s=20$$ f s = 20 . Our experiments also showed that both the BERT variants and BoW possessed good interpretability, which is important for medical applications of coding models. Conclusions This study shed light on building promising machine learning models for automated ICD coding by revealing the most effective feature extraction methods. Concretely, our results indicated that fine-tuning the whole network of the BERT variants was the optimal method for tasks covering only frequent codes, especially codes that represented unspecified diseases, while BoW was the best for tasks involving both frequent and infrequent codes. The frequency threshold where the best-performing method varied differed between different datasets due to factors like language and codeset.


Sign in / Sign up

Export Citation Format

Share Document