scholarly journals Marine Vision-Based Situational Awareness Using Discriminative Deep Learning: A Survey

2021 ◽  
Vol 9 (4) ◽  
pp. 397
Author(s):  
Dalei Qiao ◽  
Guangzhong Liu ◽  
Taizhi Lv ◽  
Wei Li ◽  
Juan Zhang

The primary task of marine surveillance is to construct a perfect marine situational awareness (MSA) system that serves to safeguard national maritime rights and interests and to maintain blue homeland security. Progress in maritime wireless communication, developments in artificial intelligence, and automation of marine turbines together imply that intelligent shipping is inevitable in future global shipping. Computer vision-based situational awareness provides visual semantic information to human beings that approximates eyesight, which makes it likely to be widely used in the field of intelligent marine transportation. We describe how we combined the visual perception tasks required for marine surveillance with those required for intelligent ship navigation to form a marine computer vision-based situational awareness complex and investigated the key technologies they have in common. Deep learning was a prerequisite activity. We summarize the progress made in four aspects of current research: full scene parsing of an image, target vessel re-identification, target vessel tracking, and multimodal data fusion with data from visual sensors. The paper gives a summary of research to date to provide background for this work and presents brief analyses of existing problems, outlines some state-of-the-art approaches, reviews available mainstream datasets, and indicates the likely direction of future research and development. As far as we know, this paper is the first review of research into the use of deep learning in situational awareness of the ocean surface. It provides a firm foundation for further investigation by researchers in related fields.

Data ◽  
2018 ◽  
Vol 3 (3) ◽  
pp. 28 ◽  
Author(s):  
Kasthurirangan Gopalakrishnan

Deep learning, more specifically deep convolutional neural networks, is fast becoming a popular choice for computer vision-based automated pavement distress detection. While pavement image analysis has been extensively researched over the past three decades or so, recent ground-breaking achievements of deep learning algorithms in the areas of machine translation, speech recognition, and computer vision has sparked interest in the application of deep learning to automated detection of distresses in pavement images. This paper provides a narrative review of recently published studies in this field, highlighting the current achievements and challenges. A comparison of the deep learning software frameworks, network architecture, hyper-parameters employed by each study, and crack detection performance is provided, which is expected to provide a good foundation for driving further research on this important topic in the context of smart pavement or asset management systems. The review concludes with potential avenues for future research; especially in the application of deep learning to not only detect, but also characterize the type, extent, and severity of distresses from 2D and 3D pavement images.


2020 ◽  
Vol 32 (5) ◽  
pp. 829-864 ◽  
Author(s):  
Jing Gao ◽  
Peng Li ◽  
Zhikui Chen ◽  
Jianing Zhang

With the wide deployments of heterogeneous networks, huge amounts of data with characteristics of high volume, high variety, high velocity, and high veracity are generated. These data, referred to multimodal big data, contain abundant intermodality and cross-modality information and pose vast challenges on traditional data fusion methods. In this review, we present some pioneering deep learning models to fuse these multimodal big data. With the increasing exploration of the multimodal big data, there are still some challenges to be addressed. Thus, this review presents a survey on deep learning for multimodal data fusion to provide readers, regardless of their original community, with the fundamentals of multimodal deep learning fusion method and to motivate new multimodal data fusion techniques of deep learning. Specifically, representative architectures that are widely used are summarized as fundamental to the understanding of multimodal deep learning. Then the current pioneering multimodal data fusion deep learning models are summarized. Finally, some challenges and future topics of multimodal data fusion deep learning models are described.


Entropy ◽  
2020 ◽  
Vol 22 (10) ◽  
pp. 1174
Author(s):  
Ashish Kumar Gupta ◽  
Ayan Seal ◽  
Mukesh Prasad ◽  
Pritee Khanna

Detection and localization of regions of images that attract immediate human visual attention is currently an intensive area of research in computer vision. The capability of automatic identification and segmentation of such salient image regions has immediate consequences for applications in the field of computer vision, computer graphics, and multimedia. A large number of salient object detection (SOD) methods have been devised to effectively mimic the capability of the human visual system to detect the salient regions in images. These methods can be broadly categorized into two categories based on their feature engineering mechanism: conventional or deep learning-based. In this survey, most of the influential advances in image-based SOD from both conventional as well as deep learning-based categories have been reviewed in detail. Relevant saliency modeling trends with key issues, core techniques, and the scope for future research work have been discussed in the context of difficulties often faced in salient object detection. Results are presented for various challenging cases for some large-scale public datasets. Different metrics considered for assessment of the performance of state-of-the-art salient object detection models are also covered. Some future directions for SOD are presented towards end.


Author(s):  
Aparna U ◽  
Athira B ◽  
Anuja M V ◽  
Aswathy Ramakrishnan ◽  
Divya R

Collapse of man-made structures, such as buildings and bridges earth quakes and fire accident, occur with varying frequency across the world. In such a scenario, the survived human beings are likely to get trapped in the cavities created by collapsed building material. During post disaster rescue operations, searchand-rescue crews have a very limited or no knowledge of the presence, location, and number of the trapped victims. Deep learning is a fast-growing domain of machine learning, mainly for solving problems in computer vision. One of the implementation of deep learning is detection of objects including humans, based on video stream. Thus, the presence of a human buried under earthquake rubble or hidden behind barriers can be identified using deep learning. This is done with the help of USB camera which can be inserted into the rubble. Spotter also gives an audio message about the location of the human presence and gives the area where the human is likely to be present. Human detection is done with the help of Computer Vision using OpenCV.


2021 ◽  
Vol 2079 (1) ◽  
pp. 012029
Author(s):  
Xueyuan Liu

Abstract The process of using CNN (Convolutional Neural Network) to blend the contents of a picture with different styles is called neural style transfer (NST). The purpose of this paper is to introduce current progress of NST, and introduce in detail the classification of the main NST algorithms based on deep learning, and make qualitative and quantitative comparisons of different algorithms, and then analyze the application prospects of image style migration in related fields, and finally summarize the existing problems and future research directions of NST.


Author(s):  
Anuraag Velamati Et.al

The world is quickly and continuously advancing towards better technological advancements that will make life quite easier for us, human beings [22]. Humans are looking for more interactive and advanced ways to improve their learning. One such dream is making a machine think like a computer, which lead to innovations like AI and deep learning [25]. The world is running at a higher pace in the domain of AI, deep learning, robotics and machine learning Using this knowledge and technology, we could develop anything right now [36]. As a part of sub-domain, the introduction of Convolution Neural Networks made deep learning extensively strong in the domain of image classification and detection [1]..The research that we have conducted is one of its kind. Our research used Convolution Neural Network, TensorFlow and Keras.


Author(s):  
Asako Kanezaki ◽  
Ryohei Kuga ◽  
Yusuke Sugano ◽  
Yasuyuki Matsushita

Sensors ◽  
2020 ◽  
Vol 20 (23) ◽  
pp. 6856
Author(s):  
Su Mu ◽  
Meng Cui ◽  
Xiaodi Huang

Multimodal learning analytics (MMLA), which has become increasingly popular, can help provide an accurate understanding of learning processes. However, it is still unclear how multimodal data is integrated into MMLA. By following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines, this paper systematically surveys 346 articles on MMLA published during the past three years. For this purpose, we first present a conceptual model for reviewing these articles from three dimensions: data types, learning indicators, and data fusion. Based on this model, we then answer the following questions: 1. What types of data and learning indicators are used in MMLA, together with their relationships; and 2. What are the classifications of the data fusion methods in MMLA. Finally, we point out the key stages in data fusion and the future research direction in MMLA. Our main findings from this review are (a) The data in MMLA are classified into digital data, physical data, physiological data, psychometric data, and environment data; (b) The learning indicators are behavior, cognition, emotion, collaboration, and engagement; (c) The relationships between multimodal data and learning indicators are one-to-one, one-to-any, and many-to-one. The complex relationships between multimodal data and learning indicators are the key for data fusion; (d) The main data fusion methods in MMLA are many-to-one, many-to-many and multiple validations among multimodal data; and (e) Multimodal data fusion can be characterized by the multimodality of data, multi-dimension of indicators, and diversity of methods.


Sensors ◽  
2020 ◽  
Vol 20 (15) ◽  
pp. 4220 ◽  
Author(s):  
Jamil Fayyad ◽  
Mohammad A. Jaradat ◽  
Dominique Gruyer ◽  
Homayoun Najjaran

Autonomous vehicles (AV) are expected to improve, reshape, and revolutionize the future of ground transportation. It is anticipated that ordinary vehicles will one day be replaced with smart vehicles that are able to make decisions and perform driving tasks on their own. In order to achieve this objective, self-driving vehicles are equipped with sensors that are used to sense and perceive both their surroundings and the faraway environment, using further advances in communication technologies, such as 5G. In the meantime, local perception, as with human beings, will continue to be an effective means for controlling the vehicle at short range. In the other hand, extended perception allows for anticipation of distant events and produces smarter behavior to guide the vehicle to its destination while respecting a set of criteria (safety, energy management, traffic optimization, comfort). In spite of the remarkable advancements of sensor technologies in terms of their effectiveness and applicability for AV systems in recent years, sensors can still fail because of noise, ambient conditions, or manufacturing defects, among other factors; hence, it is not advisable to rely on a single sensor for any of the autonomous driving tasks. The practical solution is to incorporate multiple competitive and complementary sensors that work synergistically to overcome their individual shortcomings. This article provides a comprehensive review of the state-of-the-art methods utilized to improve the performance of AV systems in short-range or local vehicle environments. Specifically, it focuses on recent studies that use deep learning sensor fusion algorithms for perception, localization, and mapping. The article concludes by highlighting some of the current trends and possible future research directions.


Sign in / Sign up

Export Citation Format

Share Document