scholarly journals A Systematic Review of Federated Learning in the Healthcare Area: From the Perspective of Data Properties and Applications

2021 ◽  
Vol 11 (23) ◽  
pp. 11191
Author(s):  
Prayitno ◽  
Chi-Ren Shyu ◽  
Karisma Trinanda Putra ◽  
Hsing-Chung Chen ◽  
Yuan-Yu Tsai ◽  
...  

Recent advances in deep learning have shown many successful stories in smart healthcare applications with data-driven insight into improving clinical institutions’ quality of care. Excellent deep learning models are heavily data-driven. The more data trained, the more robust and more generalizable the performance of the deep learning model. However, pooling the medical data into centralized storage to train a robust deep learning model faces privacy, ownership, and strict regulation challenges. Federated learning resolves the previous challenges with a shared global deep learning model using a central aggregator server. At the same time, patient data remain with the local party, maintaining data anonymity and security. In this study, first, we provide a comprehensive, up-to-date review of research employing federated learning in healthcare applications. Second, we evaluate a set of recent challenges from a data-centric perspective in federated learning, such as data partitioning characteristics, data distributions, data protection mechanisms, and benchmark datasets. Finally, we point out several potential challenges and future research directions in healthcare applications.

2018 ◽  
Vol 19 (9) ◽  
pp. 2817 ◽  
Author(s):  
Haixia Long ◽  
Bo Liao ◽  
Xingyu Xu ◽  
Jialiang Yang

Protein hydroxylation is one type of post-translational modifications (PTMs) playing critical roles in human diseases. It is known that protein sequence contains many uncharacterized residues of proline and lysine. The question that needs to be answered is: which residue can be hydroxylated, and which one cannot. The answer will not only help understand the mechanism of hydroxylation but can also benefit the development of new drugs. In this paper, we proposed a novel approach for predicting hydroxylation using a hybrid deep learning model integrating the convolutional neural network (CNN) and long short-term memory network (LSTM). We employed a pseudo amino acid composition (PseAAC) method to construct valid benchmark datasets based on a sliding window strategy and used the position-specific scoring matrix (PSSM) to represent samples as inputs to the deep learning model. In addition, we compared our method with popular predictors including CNN, iHyd-PseAAC, and iHyd-PseCp. The results for 5-fold cross-validations all demonstrated that our method significantly outperforms the other methods in prediction accuracy.


2020 ◽  
Vol 36 (12) ◽  
pp. 3856-3862
Author(s):  
Di Jin ◽  
Peter Szolovits

Abstract Motivation In evidence-based medicine, defining a clinical question in terms of the specific patient problem aids the physicians to efficiently identify appropriate resources and search for the best available evidence for medical treatment. In order to formulate a well-defined, focused clinical question, a framework called PICO is widely used, which identifies the sentences in a given medical text that belong to the four components typically reported in clinical trials: Participants/Problem (P), Intervention (I), Comparison (C) and Outcome (O). In this work, we propose a novel deep learning model for recognizing PICO elements in biomedical abstracts. Based on the previous state-of-the-art bidirectional long-short-term memory (bi-LSTM) plus conditional random field architecture, we add another layer of bi-LSTM upon the sentence representation vectors so that the contextual information from surrounding sentences can be gathered to help infer the interpretation of the current one. In addition, we propose two methods to further generalize and improve the model: adversarial training and unsupervised pre-training over large corpora. Results We tested our proposed approach over two benchmark datasets. One is the PubMed-PICO dataset, where our best results outperform the previous best by 5.5%, 7.9% and 5.8% for P, I and O elements in terms of F1 score, respectively. And for the other dataset named NICTA-PIBOSO, the improvements for P/I/O elements are 3.9%, 15.6% and 1.3% in F1 score, respectively. Overall, our proposed deep learning model can obtain unprecedented PICO element detection accuracy while avoiding the need for any manual feature selection. Availability and implementation Code is available at https://github.com/jind11/Deep-PICO-Detection.


Energies ◽  
2020 ◽  
Vol 13 (22) ◽  
pp. 5987
Author(s):  
Jerol Soibam ◽  
Achref Rabhi ◽  
Ioanna Aslanidou ◽  
Konstantinos Kyprianidis ◽  
Rebei Bel Fdhila

Subcooled flow boiling occurs in many industrial applications where enormous heat transfer is needed. Boiling is a complex physical process that involves phase change, two-phase flow, and interactions between heated surfaces and fluids. In general, boiling heat transfer is usually predicted by empirical or semiempirical models, which are horizontal to uncertainty. In this work, a data-driven method based on artificial neural networks has been implemented to study the heat transfer behavior of a subcooled boiling model. The proposed method considers the near local flow behavior to predict wall temperature and void fraction of a subcooled minichannel. The input of the network consists of pressure gradients, momentum convection, energy convection, turbulent viscosity, liquid and gas velocities, and surface information. The outputs of the models are based on the quantities of interest in a boiling system wall temperature and void fraction. To train the network, high-fidelity simulations based on the Eulerian two-fluid approach are carried out for varying heat flux and inlet velocity in the minichannel. Two classes of the deep learning model have been investigated for this work. The first one focuses on predicting the deterministic value of the quantities of interest. The second one focuses on predicting the uncertainty present in the deep learning model while estimating the quantities of interest. Deep ensemble and Monte Carlo Dropout methods are close representatives of maximum likelihood and Bayesian inference approach respectively, and they are used to derive the uncertainty present in the model. The results of this study prove that the models used here are capable of predicting the quantities of interest accurately and are capable of estimating the uncertainty present. The models are capable of accurately reproducing the physics on unseen data and show the degree of uncertainty when there is a shift of physics in the boiling regime.


Energies ◽  
2021 ◽  
Vol 14 (12) ◽  
pp. 3629
Author(s):  
Dongkwon Han ◽  
Sunil Kwon

Reservoir modeling to predict shale reservoir productivity is considerably uncertain and time consuming. Since we need to simulate the physical phenomenon of multi-stage hydraulic fracturing. To overcome these limitations, this paper presents an alternative proxy model based on data-driven deep learning model. Furthermore, this study not only proposes the development process of a proxy model, but also verifies using field data for 1239 horizontal wells from the Montney shale formation in Alberta, Canada. A deep neural network (DNN) based on multi-layer perceptron was applied to predict the cumulative gas production as the dependent variable. The independent variable is largely divided into four types: well information, completion and hydraulic fracturing and production data. It was found that the prediction performance was better when using a principal component with a cumulative contribution of 85% using principal component analysis that extracts important information from multivariate data, and when predicting with a DNN model using 6 variables calculated through variable importance analysis. Hence, to develop a reliable deep learning model, sensitivity analysis of hyperparameters was performed to determine one-hot encoding, dropout, activation function, learning rate, hidden layer number and neuron number. As a result, the best prediction of the mean absolute percentage error of the cumulative gas production improved to at least 0.2% and up to 9.1%. The novel approach of this study can also be applied to other shale formations. Furthermore, a useful guide for economic analysis and future development plans of nearby reservoirs.


2020 ◽  
Vol 10 (20) ◽  
pp. 7299 ◽  
Author(s):  
Jinsung Kim ◽  
Jin-Kook Lee

This paper describes an approach for identifying and appending interior design style information stochastically with reference images and a deep-learning model. In the field of interior design, design style is a useful concept and has played an important role in helping people understand and communicate interior design. Previous studies have focused on how the interior design style categories can be defined. On the other hand, this paper focuses on how stochastically recognizing the design style of given interior design reference images using a deep learning-based data-driven approach. The proposed method can be summarized as follows: (1) data preparation based on a general design style definition, (2) implementing an interior design style recognition model using a pre-trained VGG16 model, (3) training and evaluating the trained model, and (4) demonstration of stochastic detection of interior design styles for reference images. The result shows that the trained model automatically recognizes the design styles of given interior images with probability values. The recognition results, model, and trained image set contribute to facilitating the management and utilization of an interior design references database.


2020 ◽  
Vol 10 (24) ◽  
pp. 8934
Author(s):  
Yan He ◽  
Bin Fu ◽  
Jian Yu ◽  
Renfa Li ◽  
Rucheng Jiang

Wireless and mobile health applications promote the development of smart healthcare. Effective diagnosis and feedbacks of remote health data pose significant challenges due to streaming data, high noise, network latency and user privacy. Therefore, we explore efficient edge and cloud design to maintain electrocardiogram classification performance while reducing the communication cost. These contributions include: (1) We introduce a hybrid smart medical architecture named edge convolutional neural networks (EdgeCNN) that balances the capability of edge and cloud computing to address the issue for agile learning of healthcare data from IoT devices. (2) We present an effective deep learning model for electrocardiogram (ECG) inference, which can be deployed to run on edge smart devices for low-latency diagnosis. (3) We design a data enhancement method for ECG based on deep convolutional generative adversarial network to expand ECG data volume. (4) We carried out experiments on two representative datasets to evaluate the effectiveness of the deep learning model of ECG classification based on EdgeCNN. EdgeCNN shows superior to traditional cloud medical systems in terms of network Input/Output (I/O) pressure, architecture cost and system high availability. The deep learning model not only ensures high diagnostic accuracy, but also has advantages in aspect of inference time, storage, running memory and power consumption.


2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Benzhen Guo ◽  
Yanli Ma ◽  
Jingjing Yang ◽  
Zhihui Wang

Introduction. Health monitoring and remote diagnosis can be realized through Smart Healthcare. In view of the existing problems such as simple measurement parameters of wearable devices, huge computing pressure of cloud servers, and lack of individualization of diagnosis, a novel Cloud-Internet of Things (C-IOT) framework for medical monitoring is put forward. Methods. Smart phones are adopted as gateway devices to achieve data standardization and preprocess to generate health gray-scale map uploaded to the cloud server. The cloud server realizes the business logic processing and uses the deep learning model to carry out the gray-scale map calculation of health parameters. A deep learning model based on the convolution neural network (CNN) is constructed, in which six volunteers are selected to participate in the experiment, and their health data are marked by private doctors to generate initial data set. Results. Experimental results show the feasibility of the proposed framework. The test data set is used to test the CNN model after training; the forecast accuracy is over 77.6%. Conclusion. The CNN model performs well in the recognition of health status. Collectively, this Smart Healthcare System is expected to assist doctors by improving the diagnosis of health status in clinical practice.


2021 ◽  
Vol 30 (2) ◽  
pp. 1-31
Author(s):  
Deqing Zou ◽  
Yawei Zhu ◽  
Shouhuai Xu ◽  
Zhen Li ◽  
Hai Jin ◽  
...  

Detecting software vulnerabilities is an important problem and a recent development in tackling the problem is the use of deep learning models to detect software vulnerabilities. While effective, it is hard to explain why a deep learning model predicts a piece of code as vulnerable or not because of the black-box nature of deep learning models. Indeed, the interpretability of deep learning models is a daunting open problem. In this article, we make a significant step toward tackling the interpretability of deep learning model in vulnerability detection. Specifically, we introduce a high-fidelity explanation framework, which aims to identify a small number of tokens that make significant contributions to a detector’s prediction with respect to an example. Systematic experiments show that the framework indeed has a higher fidelity than existing methods, especially when features are not independent of each other (which often occurs in the real world). In particular, the framework can produce some vulnerability rules that can be understood by domain experts for accepting a detector’s outputs (i.e., true positives) or rejecting a detector’s outputs (i.e., false-positives and false-negatives). We also discuss limitations of the present study, which indicate interesting open problems for future research.


2020 ◽  
Vol 6 (29) ◽  
pp. eaba1482
Author(s):  
Gang Zheng ◽  
Xiaofeng Li ◽  
Rong-Hua Zhang ◽  
Bin Liu

Forecasting fields of oceanic phenomena has long been dependent on physical equation–based numerical models. The challenge is that many natural processes need to be considered for understanding complicated phenomena. In contrast, rules of the processes are already embedded in the time-series observation itself. Thus, inspired by largely available satellite remote sensing data and the advance of deep learning technology, we developed a purely satellite data–driven deep learning model for forecasting the sea surface temperature evolution associated with a typical phenomenon: a tropical instability wave. During the testing period of 9 years (2010–2019), our model accurately and efficiently forecasts the sea surface temperature field. This study demonstrates the strong potential of the satellite data–driven deep learning model as an alternative to traditional numerical models for forecasting oceanic phenomena.


Sign in / Sign up

Export Citation Format

Share Document