scholarly journals CapPlant: a capsule network based framework for plant disease classification

2021 ◽  
Vol 7 ◽  
pp. e752
Author(s):  
Omar Bin Samin ◽  
Maryam Omar ◽  
Musadaq Mansoor

Accurate disease classification in plants is important for a profound understanding of their growth and health. Recognizing diseases in plants from images is one of the critical and challenging problem in agriculture. In this research, a deep learning architecture model (CapPlant) is proposed that utilizes plant images to predict whether it is healthy or contain some disease. The prediction process does not require handcrafted features; rather, the representations are automatically extracted from input data sequence by architecture. Several convolutional layers are applied to extract and classify features accordingly. The last convolutional layer in CapPlant is replaced by state-of-the-art capsule layer to incorporate orientational and relative spatial relationship between different entities of a plant in an image to predict diseases more precisely. The proposed architecture is tested on the PlantVillage dataset, which contains more than 50,000 images of infected and healthy plants. Significant improvements in terms of prediction accuracy has been observed using the CapPlant model when compared with other plant disease classification models. The experimental results on the developed model have achieved an overall test accuracy of 93.01%, with F1 score of 93.07%.

2021 ◽  
Vol 40 (1) ◽  
pp. 1025-1036
Author(s):  
Patrick Kwabena Mensah ◽  
Benjamin Asubam Weyori ◽  
Mighty Abra Ayidzoe

Capsule Networks (CapsNets) excel on simple image recognition problems. However, they fail to perform on complex images with high similarity and background objects. This paper proposes Local Binary Pattern (LBP) k-means routing and evaluates its performance on three publicly available plant disease datasets containing images with high similarity and background objects. The proposed routing algorithm adopts the squared Euclidean distance, sigmoid function, and a ‘simple-squash’ in place of dot product, SoftMax normalizer, and the squashing function found respectively in the dynamic routing algorithm. Extensive experiments conducted on the three datasets showed that the proposed model achieves consistent improvement in test accuracy across the three datasets as well as allowing an increase in the number of routing iterations with no performance degradation. The proposed model outperformed a baseline CapsNet by 8.37% on the tomato dataset with an overall test accuracy of 98.80%, comparable to state-of-the-art models on the same datasets.


Plants ◽  
2020 ◽  
Vol 9 (10) ◽  
pp. 1319
Author(s):  
Muhammad Hammad Saleem ◽  
Johan Potgieter ◽  
Khalid Mahmood Arif

Recently, plant disease classification has been done by various state-of-the-art deep learning (DL) architectures on the publicly available/author generated datasets. This research proposed the deep learning-based comparative evaluation for the classification of plant disease in two steps. Firstly, the best convolutional neural network (CNN) was obtained by conducting a comparative analysis among well-known CNN architectures along with modified and cascaded/hybrid versions of some of the DL models proposed in the recent researches. Secondly, the performance of the best-obtained model was attempted to improve by training through various deep learning optimizers. The comparison between various CNNs was based on performance metrics such as validation accuracy/loss, F1-score, and the required number of epochs. All the selected DL architectures were trained in the PlantVillage dataset which contains 26 different diseases belonging to 14 respective plant species. Keras with TensorFlow backend was used to train deep learning architectures. It is concluded that the Xception architecture trained with the Adam optimizer attained the highest validation accuracy and F1-score of 99.81% and 0.9978 respectively which is comparatively better than the previous approaches and it proves the novelty of the work. Therefore, the method proposed in this research can be applied to other agricultural applications for transparent detection and classification purposes.


2021 ◽  
pp. 1-13 ◽  
Author(s):  
Bhabendu Kumar Mohanta ◽  
Debasish Jena ◽  
Niva Mohapatra ◽  
Somula Ramasubbareddy ◽  
Bharat S. Rawal

Smart city has come a long way since the development of emerging technology like Information and communications technology (ICT), Internet of Things (IoT), Machine Learning (ML), Block chain and Artificial Intelligence. The Intelligent Transportation System (ITS) is an important application in a rapidly growing smart city. Prediction of the automotive accident severity plays a very crucial role in the smart transportation system. The main motive behind this research is to determine the specific features which could affect vehicle accident severity. In this paper, some of the classification models, specifically Logistic Regression, Artificial Neural network, Decision Tree, K-Nearest Neighbors, and Random Forest have been implemented for predicting the accident severity. All the models have been verified, and the experimental results prove that these classification models have attained considerable accuracy. The paper also explained a secure communication architecture model for secure information exchange among all the components associated with the ITS. Finally paper implemented web base Message alert system which will be used for alert the users through smart IoT devices.


2021 ◽  
Vol 11 (6) ◽  
pp. 2511
Author(s):  
Julian Hatwell ◽  
Mohamed Medhat Gaber ◽  
R. Muhammad Atif Azad

This research presents Gradient Boosted Tree High Importance Path Snippets (gbt-HIPS), a novel, heuristic method for explaining gradient boosted tree (GBT) classification models by extracting a single classification rule (CR) from the ensemble of decision trees that make up the GBT model. This CR contains the most statistically important boundary values of the input space as antecedent terms. The CR represents a hyper-rectangle of the input space inside which the GBT model is, very reliably, classifying all instances with the same class label as the explanandum instance. In a benchmark test using nine data sets and five competing state-of-the-art methods, gbt-HIPS offered the best trade-off between coverage (0.16–0.75) and precision (0.85–0.98). Unlike competing methods, gbt-HIPS is also demonstrably guarded against under- and over-fitting. A further distinguishing feature of our method is that, unlike much prior work, our explanations also provide counterfactual detail in accordance with widely accepted recommendations for what makes a good explanation.


Author(s):  
My Kieu ◽  
Andrew D. Bagdanov ◽  
Marco Bertini

Pedestrian detection is a canonical problem for safety and security applications, and it remains a challenging problem due to the highly variable lighting conditions in which pedestrians must be detected. This article investigates several domain adaptation approaches to adapt RGB-trained detectors to the thermal domain. Building on our earlier work on domain adaptation for privacy-preserving pedestrian detection, we conducted an extensive experimental evaluation comparing top-down and bottom-up domain adaptation and also propose two new bottom-up domain adaptation strategies. For top-down domain adaptation, we leverage a detector pre-trained on RGB imagery and efficiently adapt it to perform pedestrian detection in the thermal domain. Our bottom-up domain adaptation approaches include two steps: first, training an adapter segment corresponding to initial layers of the RGB-trained detector adapts to the new input distribution; then, we reconnect the adapter segment to the original RGB-trained detector for final adaptation with a top-down loss. To the best of our knowledge, our bottom-up domain adaptation approaches outperform the best-performing single-modality pedestrian detection results on KAIST and outperform the state of the art on FLIR.


2021 ◽  
Vol 15 (6) ◽  
pp. 1-21
Author(s):  
Huandong Wang ◽  
Yong Li ◽  
Mu Du ◽  
Zhenhui Li ◽  
Depeng Jin

Both app developers and service providers have strong motivations to understand when and where certain apps are used by users. However, it has been a challenging problem due to the highly skewed and noisy app usage data. Moreover, apps are regarded as independent items in existing studies, which fail to capture the hidden semantics in app usage traces. In this article, we propose App2Vec, a powerful representation learning model to learn the semantic embedding of apps with the consideration of spatio-temporal context. Based on the obtained semantic embeddings, we develop a probabilistic model based on the Bayesian mixture model and Dirichlet process to capture when , where , and what semantics of apps are used to predict the future usage. We evaluate our model using two different app usage datasets, which involve over 1.7 million users and 2,000+ apps. Evaluation results show that our proposed App2Vec algorithm outperforms the state-of-the-art algorithms in app usage prediction with a performance gap of over 17.0%.


2021 ◽  
Vol 11 (9) ◽  
pp. 4248
Author(s):  
Hong Hai Hoang ◽  
Bao Long Tran

With the rapid development of cameras and deep learning technologies, computer vision tasks such as object detection, object segmentation and object tracking are being widely applied in many fields of life. For robot grasping tasks, object segmentation aims to classify and localize objects, which helps robots to be able to pick objects accurately. The state-of-the-art instance segmentation network framework, Mask Region-Convolution Neural Network (Mask R-CNN), does not always perform an excellent accurate segmentation at the edge or border of objects. The approach using 3D camera, however, is able to extract the entire (foreground) objects easily but can be difficult or require a large amount of computation effort to classify it. We propose a novel approach, in which we combine Mask R-CNN with 3D algorithms by adding a 3D process branch for instance segmentation. Both outcomes of two branches are contemporaneously used to classify the pixels at the edge objects by dealing with the spatial relationship between edge region and mask region. We analyze the effectiveness of the method by testing with harsh cases of object positions, for example, objects are closed, overlapped or obscured by each other to focus on edge and border segmentation. Our proposed method is about 4 to 7% higher and more stable in IoU (intersection of union). This leads to a reach of 46% of mAP (mean Average Precision), which is a higher accuracy than its counterpart. The feasibility experiment shows that our method could be a remarkable promoting for the research of the grasping robot.


1977 ◽  
Vol 86 (6_suppl) ◽  
pp. 1-20 ◽  
Author(s):  
Juergen Tonndorf

Work on cochlear prostheses for the auditory rehabilitation of the profoundly deaf represents a challenging problem. Some early, but perhaps premature, surgical attempts have helped to bring the entire issue into focus. Systemic studies are now under way in many different places. Although the purely engineering problems as well as the surgical ones appear solvable at this time, the remaining unsolved problems lie in two areas: 1) the bioengineering interfacing, i.e., the search for methods needed to connect an engineering (electronic) device to the neural auditory system in an efficient manner; and 2) clinical tests for the assessment of the functional state of the cochlear nerve.


Sign in / Sign up

Export Citation Format

Share Document