scholarly journals Artificial Intelligence-Based Digital Image Steganalysis

2021 ◽  
Vol 2021 ◽  
pp. 1-9
Author(s):  
Ahmed I. Iskanderani ◽  
Ibrahim M. Mehedi ◽  
Abdulah Jeza Aljohani ◽  
Mohammad Shorfuzzaman ◽  
Farzana Akther ◽  
...  

Recently, deep learning-based models are being extensively utilized for steganalysis. However, deep learning models suffer from overfitting and hyperparameter tuning issues. Therefore, in this paper, an efficient θ -nondominated sorting genetic algorithm- ( θ NSGA-) III based densely connected convolutional neural network (DCNN) model is proposed for image steganalysis. θ NSGA-III is utilized to tune the initial parameters of DCNN model. It can control the accuracy and f-measure of the DCNN model by utilizing them as the multiobjective fitness function. Extensive experiments are drawn on STEGRT1 dataset. Comparison of the proposed model is also drawn with the competitive steganalysis model. Performance analyses reveal that the proposed model outperforms the existing steganalysis models in terms of various performance metrics.

2021 ◽  
Vol 2021 ◽  
pp. 1-13
Author(s):  
G. Kothai ◽  
E. Poovammal ◽  
Gaurav Dhiman ◽  
Kadiyala Ramana ◽  
Ashutosh Sharma ◽  
...  

The vehicular adhoc network (VANET) is an emerging research topic in the intelligent transportation system that furnishes essential information to the vehicles in the network. Nearly 150 thousand people are affected by the road accidents that must be minimized, and improving safety is required in VANET. The prediction of traffic congestions plays a momentous role in minimizing accidents in roads and improving traffic management for people. However, the dynamic behavior of the vehicles in the network degrades the rendition of deep learning models in predicting the traffic congestion on roads. To overcome the congestion problem, this paper proposes a new hybrid boosted long short-term memory ensemble (BLSTME) and convolutional neural network (CNN) model that ensemble the powerful features of CNN with BLSTME to negotiate the dynamic behavior of the vehicle and to predict the congestion in traffic effectively on roads. The CNN extracts the features from traffic images, and the proposed BLSTME trains and strengthens the weak classifiers for the prediction of congestion. The proposed model is developed using Tensor flow python libraries and are tested in real traffic scenario simulated using SUMO and OMNeT++. The extensive experimentations are carried out, and the model is measured with the performance metrics likely prediction accuracy, precision, and recall. Thus, the experimental result shows 98% of accuracy, 96% of precision, and 94% of recall. The results complies that the proposed model clobbers the other existing algorithms by furnishing 10% higher than deep learning models in terms of stability and performance.


Sensors ◽  
2021 ◽  
Vol 21 (4) ◽  
pp. 1064
Author(s):  
I Nyoman Kusuma Wardana ◽  
Julian W. Gardner ◽  
Suhaib A. Fahmy

Accurate air quality monitoring requires processing of multi-dimensional, multi-location sensor data, which has previously been considered in centralised machine learning models. These are often unsuitable for resource-constrained edge devices. In this article, we address this challenge by: (1) designing a novel hybrid deep learning model for hourly PM2.5 pollutant prediction; (2) optimising the obtained model for edge devices; and (3) examining model performance running on the edge devices in terms of both accuracy and latency. The hybrid deep learning model in this work comprises a 1D Convolutional Neural Network (CNN) and a Long Short-Term Memory (LSTM) to predict hourly PM2.5 concentration. The results show that our proposed model outperforms other deep learning models, evaluated by calculating RMSE and MAE errors. The proposed model was optimised for edge devices, the Raspberry Pi 3 Model B+ (RPi3B+) and Raspberry Pi 4 Model B (RPi4B). This optimised model reduced file size to a quarter of the original, with further size reduction achieved by implementing different post-training quantisation. In total, 8272 hourly samples were continuously fed to the edge device, with the RPi4B executing the model twice as fast as the RPi3B+ in all quantisation modes. Full-integer quantisation produced the lowest execution time, with latencies of 2.19 s and 4.73 s for RPi4B and RPi3B+, respectively.


Webology ◽  
2021 ◽  
Vol 18 (Special Issue 04) ◽  
pp. 944-962
Author(s):  
K. Niha ◽  
Dr.S. Amutha ◽  
Dr. Aisha Banu

Disease in plants are a great challenge in the advancement of agriculture which affects farmers yield and the plants. In this modern research deep learning models got a spot light by increasing plant detection accuracy and classification. The proposed CNN (Convolutional Neural Network) model detect seven plant diseases out of healthy leaf, where the dataset considered in this work contain 8685 leaf images from Plant Village Dataset. The proposed modals performance are evaluated with respect to the performance metrics (F1 score, Precision and Recall) and are compared with SVM and ANN. Where the proposed CNN model outperforms the rest with the accuracy of 96.2% and the F1 score greater than 95%. The feasibility of the proposed model in plant detection and classification may provide a solution to the problem faced by farmers.


2020 ◽  
Vol 13 (2) ◽  
pp. 215-227
Author(s):  
Dalila Bouras ◽  
Mohamed Amroune ◽  
Hakim Bendjenna ◽  
Nabiha Azizi

Background: Nowadays, with the appearance of web 2.0, more users express their opinions, judgments, and thoughts towards certain objects, services, organizations, and their attributes via social networking, forum entries, websites, and blogs and so on. In this way, the volume of raw content generated by these users will increase rapidly with enormous size, where people often find difficulties in identifying and summarizing fine-grained sentiments buried in the opinion-rich resources. The traditional opinion mining techniques, which focused on the overall sentiment of the review, fails to uncover the sentiments expressed on the aspects of the reviewed entity. For that, researchers in Aspect-based opinion mining community try to solve and handle this problem. Objective: Our proposed study aims to present, survey and compare in the first place the important recent Aspect-based opinion mining approaches relevant to important languages such English, Arabic and Chinese and commonly datasets used in literature so that future researchers could improve their results. The cited approaches used the last techniques in the area on Opinion mining field, relevant to the Deep Learning models. In the second place, we try to highlight and give special attention to the Arabic language by introducing a dashboard of deep learning methods dedicated to the Arabic language. Finally, we emphasize the research gaps and future challenges in both English and Arabic languages that provide some new potential research fields. Methods: We have carefully summarized 48 models according to their algorithm into three categories: supervised, semi-supervised and unsupervised. Due to a large number of approaches with diverse datasets and techniques, we propose some statistical graphics to compare different experimentation results namely precision, Recall, and F-measure. Also, the study has conferred a comparative analysis and a comprehensive discussion of different approaches and techniques dedicated to the aspect extraction sub-task using the new tendency that of deep learning on both Arabic, English and Chinese language. We have introduced some future challenges, research gaps, and new trends in the opinion mining task, which need more efforts and investigations to present new solutions that make the opinion mining field more pervasive and give more ideas about the different researches done in the field of OM. Methods: We have compared the different approaches and techniques dedicated to the extraction of aspects using the new tendency that of deep learning. Our contribution illustrates the add values given by deep learning models in the treatment of user reviews expressed in the Arabic language. At the same time, this work is mainly based on the use of the evaluation performance metrics (precision, recall, and f-measure).


Technologies ◽  
2021 ◽  
Vol 9 (1) ◽  
pp. 14
Author(s):  
James Dzisi Gadze ◽  
Akua Acheampomaa Bamfo-Asante ◽  
Justice Owusu Agyemang ◽  
Henry Nunoo-Mensah ◽  
Kwasi Adu-Boahen Opare

Software-Defined Networking (SDN) is a new paradigm that revolutionizes the idea of a software-driven network through the separation of control and data planes. It addresses the problems of traditional network architecture. Nevertheless, this brilliant architecture is exposed to several security threats, e.g., the distributed denial of service (DDoS) attack, which is hard to contain in such software-based networks. The concept of a centralized controller in SDN makes it a single point of attack as well as a single point of failure. In this paper, deep learning-based models, long-short term memory (LSTM) and convolutional neural network (CNN), are investigated. It illustrates their possibility and efficiency in being used in detecting and mitigating DDoS attack. The paper focuses on TCP, UDP, and ICMP flood attacks that target the controller. The performance of the models was evaluated based on the accuracy, recall, and true negative rate. We compared the performance of the deep learning models with classical machine learning models. We further provide details on the time taken to detect and mitigate the attack. Our results show that RNN LSTM is a viable deep learning algorithm that can be applied in the detection and mitigation of DDoS in the SDN controller. Our proposed model produced an accuracy of 89.63%, which outperformed linear-based models such as SVM (86.85%) and Naive Bayes (82.61%). Although KNN, which is a linear-based model, outperformed our proposed model (achieving an accuracy of 99.4%), our proposed model provides a good trade-off between precision and recall, which makes it suitable for DDoS classification. In addition, it was realized that the split ratio of the training and testing datasets can give different results in the performance of a deep learning algorithm used in a specific work. The model achieved the best performance when a split of 70/30 was used in comparison to 80/20 and 60/40 split ratios.


Author(s):  
S. Arokiaraj ◽  
Dr. N. Viswanathan

With the advent of Internet of things(IoT),HA (HA) recognition has contributed the more application in health care in terms of diagnosis and Clinical process. These devices must be aware of human movements to provide better aid in the clinical applications as well as user’s daily activity.Also , In addition to machine and deep learning algorithms, HA recognition systems has significantly improved in terms of high accurate recognition. However, the most of the existing models designed needs improvisation in terms of accuracy and computational overhead. In this research paper, we proposed a BAT optimized Long Short term Memory (BAT-LSTM) for an effective recognition of human activities using real time IoT systems. The data are collected by implanting the Internet of things) devices invasively. Then, proposed BAT-LSTM is deployed to extract the temporal features which are then used for classification to HA. Nearly 10,0000 dataset were collected and used for evaluating the proposed model. For the validation of proposed framework, accuracy, precision, recall, specificity and F1-score parameters are chosen and comparison is done with the other state-of-art deep learning models. The finding shows the proposed model outperforms the other learning models and finds its suitability for the HA recognition.


2021 ◽  
Author(s):  
Nithin G R ◽  
Nitish Kumar M ◽  
Venkateswaran Narasimhan ◽  
Rajanikanth Kakani ◽  
Ujjwal Gupta ◽  
...  

Pansharpening is the task of creating a High-Resolution Multi-Spectral Image (HRMS) by extracting and infusing pixel details from the High-Resolution Panchromatic Image into the Low-Resolution Multi-Spectral (LRMS). With the boom in the amount of satellite image data, researchers have replaced traditional approaches with deep learning models. However, existing deep learning models are not built to capture intricate pixel-level relationships. Motivated by the recent success of self-attention mechanisms in computer vision tasks, we propose Pansformers, a transformer-based self-attention architecture, that computes band-wise attention. A further improvement is proposed in the attention network by introducing a Multi-Patch Attention mechanism, which operates on non-overlapping, local patches of the image. Our model is successful in infusing relevant local details from the Panchromatic image while preserving the spectral integrity of the MS image. We show that our Pansformer model significantly improves the performance metrics and the output image quality on imagery from two satellite distributions IKONOS and LANDSAT-8.


Electronics ◽  
2020 ◽  
Vol 9 (3) ◽  
pp. 445 ◽  
Author(s):  
Laith Alzubaidi ◽  
Omran Al-Shamma ◽  
Mohammed A. Fadhel ◽  
Laith Farhan ◽  
Jinglan Zhang ◽  
...  

Breast cancer is a significant factor in female mortality. An early cancer diagnosis leads to a reduction in the breast cancer death rate. With the help of a computer-aided diagnosis system, the efficiency increased, and the cost was reduced for the cancer diagnosis. Traditional breast cancer classification techniques are based on handcrafted features techniques, and their performance relies upon the chosen features. They also are very sensitive to different sizes and complex shapes. However, histopathological breast cancer images are very complex in shape. Currently, deep learning models have become an alternative solution for diagnosis, and have overcome the drawbacks of classical classification techniques. Although deep learning has performed well in various tasks of computer vision and pattern recognition, it still has some challenges. One of the main challenges is the lack of training data. To address this challenge and optimize the performance, we have utilized a transfer learning technique which is where the deep learning models train on a task, and then fine-tune the models for another task. We have employed transfer learning in two ways: Training our proposed model first on the same domain dataset, then on the target dataset, and training our model on a different domain dataset, then on the target dataset. We have empirically proven that the same domain transfer learning optimized the performance. Our hybrid model of parallel convolutional layers and residual links is utilized to classify hematoxylin–eosin-stained breast biopsy images into four classes: invasive carcinoma, in-situ carcinoma, benign tumor and normal tissue. To reduce the effect of overfitting, we have augmented the images with different image processing techniques. The proposed model achieved state-of-the-art performance, and it outperformed the latest methods by achieving a patch-wise classification accuracy of 90.5%, and an image-wise classification accuracy of 97.4% on the validation set. Moreover, we have achieved an image-wise classification accuracy of 96.1% on the test set of the microscopy ICIAR-2018 dataset.


Author(s):  
Yasir Hussain ◽  
Zhiqiu Huang ◽  
Yu Zhou ◽  
Senzhang Wang

In recent years, deep learning models have shown great potential in source code modeling and analysis. Generally, deep learning-based approaches are problem-specific and data-hungry. A challenging issue of these approaches is that they require training from scratch for a different related problem. In this work, we propose a transfer learning-based approach that significantly improves the performance of deep learning-based source code models. In contrast to traditional learning paradigms, transfer learning can transfer the knowledge learned in solving one problem into another related problem. First, we present two recurrent neural network-based models RNN and GRU for the purpose of transfer learning in the domain of source code modeling. Next, via transfer learning, these pre-trained (RNN and GRU) models are used as feature extractors. Then, these extracted features are combined into attention learner for different downstream tasks. The attention learner leverages from the learned knowledge of pre-trained models and fine-tunes them for a specific downstream task. We evaluate the performance of the proposed approach with extensive experiments with the source code suggestion task. The results indicate that the proposed approach outperforms the state-of-the-art models in terms of accuracy, precision, recall and F-measure without training the models from scratch.


Author(s):  
Mohammad Shahab Uddin ◽  
Jiang Li

Deep learning models are data driven. For example, the most popular convolutional neural network (CNN) model used for image classification or object detection requires large labeled databases for training to achieve competitive performances. This requirement is not difficult to be satisfied in the visible domain since there are lots of labeled video and image databases available nowadays. However, given the less popularity of infrared (IR) camera, the availability of labeled infrared videos or image databases is limited. Therefore, training deep learning models in infrared domain is still challenging. In this chapter, we applied the pix2pix generative adversarial network (Pix2Pix GAN) and cycle-consistent GAN (Cycle GAN) models to convert visible videos to infrared videos. The Pix2Pix GAN model requires visible-infrared image pairs for training while the Cycle GAN relaxes this constraint and requires only unpaired images from both domains. We applied the two models to an open-source database where visible and infrared videos provided by the signal multimedia and telecommunications laboratory at the Federal University of Rio de Janeiro. We evaluated conversion results by performance metrics including Inception Score (IS), Frechet Inception Distance (FID) and Kernel Inception Distance (KID). Our experiments suggest that cycle-consistent GAN is more effective than pix2pix GAN for generating IR images from optical images.


Sign in / Sign up

Export Citation Format

Share Document