scholarly journals A Deep Learning Approach for Managing Medical Consumable Materials in Intensive Care Units via Convolutional Neural Networks: Technical Proof-of-Concept Study

10.2196/14806 ◽  
2019 ◽  
Vol 7 (4) ◽  
pp. e14806 ◽  
Author(s):  
Arne Peine ◽  
Ahmed Hallawa ◽  
Oliver Schöffski ◽  
Guido Dartmann ◽  
Lejla Begic Fazlic ◽  
...  

Background High numbers of consumable medical materials (eg, sterile needles and swabs) are used during the daily routine of intensive care units (ICUs) worldwide. Although medical consumables largely contribute to total ICU hospital expenditure, many hospitals do not track the individual use of materials. Current tracking solutions meeting the specific requirements of the medical environment, like barcodes or radio frequency identification, require specialized material preparation and high infrastructure investment. This impedes the accurate prediction of consumption, leads to high storage maintenance costs caused by large inventories, and hinders scientific work due to inaccurate documentation. Thus, new cost-effective and contactless methods for object detection are urgently needed. Objective The goal of this work was to develop and evaluate a contactless visual recognition system for tracking medical consumable materials in ICUs using a deep learning approach on a distributed client-server architecture. Methods We developed Consumabot, a novel client-server optical recognition system for medical consumables, based on the convolutional neural network model MobileNet implemented in Tensorflow. The software was designed to run on single-board computer platforms as a detection unit. The system was trained to recognize 20 different materials in the ICU, while 100 sample images of each consumable material were provided. We assessed the top-1 recognition rates in the context of different real-world ICU settings: materials presented to the system without visual obstruction, 50% covered materials, and scenarios of multiple items. We further performed an analysis of variance with repeated measures to quantify the effect of adverse real-world circumstances. Results Consumabot reached a >99% reliability of recognition after about 60 steps of training and 150 steps of validation. A desirable low cross entropy of <0.03 was reached for the training set after about 100 iteration steps and after 170 steps for the validation set. The system showed a high top-1 mean recognition accuracy in a real-world scenario of 0.85 (SD 0.11) for objects presented to the system without visual obstruction. Recognition accuracy was lower, but still acceptable, in scenarios where the objects were 50% covered (P<.001; mean recognition accuracy 0.71; SD 0.13) or multiple objects of the target group were present (P=.01; mean recognition accuracy 0.78; SD 0.11), compared to a nonobstructed view. The approach met the criteria of absence of explicit labeling (eg, barcodes, radio frequency labeling) while maintaining a high standard for quality and hygiene with minimal consumption of resources (eg, cost, time, training, and computational power). Conclusions Using a convolutional neural network architecture, Consumabot consistently achieved good results in the classification of consumables and thus is a feasible way to recognize and register medical consumables directly to a hospital’s electronic health record. The system shows limitations when the materials are partially covered, therefore identifying characteristics of the consumables are not presented to the system. Further development of the assessment in different medical circumstances is needed.

2021 ◽  
Author(s):  
Wael Alnahari

Abstract In this paper, I proposed an iris recognition system by using deep learning via neural networks (CNN). Although CNN is used for machine learning, the recognition is achieved by building a non-trained CNN network with multiple layers. The main objective of the code the test pictures’ category (aka person name) with a high accuracy rate after having extracted enough features from training pictures of the same category which are obtained from a that I added to the code. I used IITD iris which included 10 iris pictures for 223 people.


Author(s):  
Dr. I. Jeena Jacob

The biometric recognition plays a significant and a unique part in the applications that are based on the personal identification. This is because of the stability, irreplaceability and the uniqueness that is found in the biometric traits of the humans. Currently the deep learning techniques that are capable of strongly generalizing and automatically learning, with the enhanced accuracy is utilized for the biometric recognition to develop an efficient biometric system. But the poor noise removal abilities and the accuracy degradation caused due to the very small disturbances has made the conventional means of the deep learning that utilizes the convolutional neural network incompatible for the biometric recognition. So the capsule neural network replaces the CNN due to its high accuracy in the recognition and the classification, due to its learning capacities and the ability to be trained with the limited number of samples compared to the CNN (convolutional neural network). The frame work put forward in the paper utilizes the capsule network with the fuzzified image enhancement for the retina based biometric recognition as it is a highly secure and reliable basis of person identification as it is layered behind the eye and cannot be counterfeited. The method was tested with the dataset of face 95 database and the CASIA-Iris-Thousand, and was found to be 99% accurate with the error rate convergence of 0.3% to .5%


2019 ◽  
Vol 34 (11) ◽  
pp. 4924-4931 ◽  
Author(s):  
Daichi Kitaguchi ◽  
Nobuyoshi Takeshita ◽  
Hiroki Matsuzaki ◽  
Hiroaki Takano ◽  
Yohei Owada ◽  
...  

2018 ◽  
Vol 132 ◽  
pp. 679-688 ◽  
Author(s):  
Sakshi Indolia ◽  
Anil Kumar Goswami ◽  
S.P. Mishra ◽  
Pooja Asopa

2021 ◽  
Vol 11 (21) ◽  
pp. 10301
Author(s):  
Muhammad Shoaib Farooq ◽  
Attique Ur Rehman ◽  
Muhammad Idrees ◽  
Muhammad Ahsan Raza ◽  
Jehad Ali ◽  
...  

COVID-19 has been difficult to diagnose and treat at an early stage all over the world. The numbers of patients showing symptoms for COVID-19 have caused medical facilities at hospitals to become unavailable or overcrowded, which is a major challenge. Studies have recently allowed us to determine that COVID-19 can be diagnosed with the aid of chest X-ray images. To combat the COVID-19 outbreak, developing a deep learning (DL) based model for automated COVID-19 diagnosis on chest X-ray is beneficial. In this research, we have proposed a customized convolutional neural network (CNN) model to detect COVID-19 from chest X-ray images. The model is based on nine layers which uses a binary classification method to differentiate between COVID-19 and normal chest X-rays. It provides COVID-19 detection early so the patients can be admitted in a timely fashion. The proposed model was trained and tested on two publicly available datasets. Cross-dataset studies are used to assess the robustness in a real-world context. Six hundred X-ray images were used for training and two hundred X-rays were used for validation of the model. The X-ray images of the dataset were preprocessed to improve the results and visualized for better analysis. The developed algorithm reached 98% precision, recall and f1-score. The cross-dataset studies also demonstrate the resilience of deep learning algorithms in a real-world context with 98.5 percent accuracy. Furthermore, a comparison table was created which shows that our proposed model outperforms other relative models in terms of accuracy. The quick and high-performance of our proposed DL-based customized model identifies COVID-19 patients quickly, which is helpful in controlling the COVID-19 outbreak.


Author(s):  
Oyeniran Oluwashina Akinloye ◽  
Oyebode Ebenezer Olukunle

Numerous works have been proposed and implemented in computerization of various human languages, nevertheless, miniscule effort have also been made so as to put Yorùbá Handwritten Character on the map of Optical Character Recognition. This study presents a novel technique in the development of Yorùbá alphabets recognition system through the use of deep learning. The developed model was implemented on Matlab R2018a environment using the developed framework where 10,500 samples of dataset were for training and 2100 samples were used for testing. The training of the developed model was conducted using 30 Epoch, at 164 iteration per epoch while the total iteration is 4920 iterations. Also, the training period was estimated to 11296 minutes 41 seconds. The model yielded the network accuracy of 100% while the accuracy of the test set is 97.97%, with F1 score of 0.9800, Precision of 0.9803 and Recall value of 0.9797.


IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 171548-171558 ◽  
Author(s):  
Jiaying Wang ◽  
Yaxin Li ◽  
Jing Shan ◽  
Jinling Bao ◽  
Chuanyu Zong ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document