Quantitative performance analysis of object detection algorithms on underwater video footage

Author(s):  
Isaak Kavasidis ◽  
Simone Palazzo
Author(s):  
Muhammad Ahmed ◽  
Khurram Azeem Hashmi ◽  
Alain Pagani ◽  
Marcus Liwicki ◽  
Didier Stricker ◽  
...  

Recent progress in deep learning has led to accurate and efficient generic object detection networks. Training of highly reliable models depends on large datasets with highly textured and rich images. However, in real-world scenarios, the performance of the generic object detection system decreases when (i) occlusions hide the objects, (ii) objects are present in low-light images, or (iii) they are merged with background information. In this paper, we refer to all these situations as challenging environments. With the recent rapid development in generic object detection algorithms, notable progress has been observed in the field of object detection in challenging environments. However, there is no consolidated reference to cover state-of-the-art in this domain. To the best of our knowledge, this paper presents the first comprehensive overview, covering recent approaches that have tackled the problem of object detection in challenging environments. Furthermore, we present the quantitative and qualitative performance analysis of these approaches and discuss the currently available challenging datasets. Moreover, this paper investigates the performance of current state-of-the-art generic object detection algorithms by benchmarking results on the three well-known challenging datasets. Finally, we highlight several current shortcomings and outline future directions.


Sensors ◽  
2021 ◽  
Vol 21 (15) ◽  
pp. 5116
Author(s):  
Muhammad Ahmed ◽  
Khurram Azeem Hashmi ◽  
Alain Pagani ◽  
Marcus Liwicki ◽  
Didier Stricker ◽  
...  

Recent progress in deep learning has led to accurate and efficient generic object detection networks. Training of highly reliable models depends on large datasets with highly textured and rich images. However, in real-world scenarios, the performance of the generic object detection system decreases when (i) occlusions hide the objects, (ii) objects are present in low-light images, or (iii) they are merged with background information. In this paper, we refer to all these situations as challenging environments. With the recent rapid development in generic object detection algorithms, notable progress has been observed in the field of deep learning-based object detection in challenging environments. However, there is no consolidated reference to cover the state of the art in this domain. To the best of our knowledge, this paper presents the first comprehensive overview, covering recent approaches that have tackled the problem of object detection in challenging environments. Furthermore, we present a quantitative and qualitative performance analysis of these approaches and discuss the currently available challenging datasets. Moreover, this paper investigates the performance of current state-of-the-art generic object detection algorithms by benchmarking results on the three well-known challenging datasets. Finally, we highlight several current shortcomings and outline future directions.


Author(s):  
Samuel Humphries ◽  
Trevor Parker ◽  
Bryan Jonas ◽  
Bryan Adams ◽  
Nicholas J Clark

Quick identification of building and roads is critical for execution of tactical US military operations in an urban environment. To this end, a gridded, referenced, satellite images of an objective, often referred to as a gridded reference graphic or GRG, has become a standard product developed during intelligence preparation of the environment. At present, operational units identify key infrastructure by hand through the work of individual intelligence officers. Recent advances in Convolutional Neural Networks, however, allows for this process to be streamlined through the use of object detection algorithms. In this paper, we describe an object detection algorithm designed to quickly identify and label both buildings and road intersections present in an image. Our work leverages both the U-Net architecture as well the SpaceNet data corpus to produce an algorithm that accurately identifies a large breadth of buildings and different types of roads. In addition to predicting buildings and roads, our model numerically labels each building by means of a contour finding algorithm. Most importantly, the dual U-Net model is capable of predicting buildings and roads on a diverse set of test images and using these predictions to produce clean GRGs.


2021 ◽  
Author(s):  
Alexis Koulidis ◽  
Mohamed Abdullatif ◽  
Ahmed Galal Abdel-Kader ◽  
Mohammed-ilies Ayachi ◽  
Shehab Ahmed ◽  
...  

Abstract Surface data measurement and analysis are an established mean of detecting drillstring low-frequency torsional vibration or stick-slip. The industry has also developed models that link surface torque and downhole drill bit rotational speed. Cameras provide an alternative noninvasive approach to existing wired/wireless sensors used to gather such surface data. The results of a preliminary field assessment of drilling dynamics utilizing camera-based drillstring monitoring are presented in this work. Detection and timing of events from the video are performed using computer vision techniques and object detection algorithms. A real-time interest point tracker utilizing homography estimation and sparse optical flow point tracking is deployed. We use a fully convolutional deep neural network trained to detect interest points and compute their accompanying descriptors. The detected points and descriptors are matched across video sequences and used for drillstring rotation detection and speed estimation. When the drillstring's vibration is invisible to the naked eye, the point tracking algorithm is preceded with a motion amplification function based on another deep convolutional neural network. We have clearly demonstrated the potential of camera-based noninvasive approaches to surface drillstring dynamics data acquisition and analysis. Through the application of real-time object detection algorithms on rig video feed, surface events were detected and timed. We were also able to estimate drillstring rotary speed and motion profile. Torsional drillstring modes can be identified and correlated with drilling parameters and bottomhole assembly design. A novel vibration array sensing approach based on a multi-point tracking algorithm is also proposed. A vibration threshold setting was utilized to enable an additional motion amplification function providing seamless assessment for multi-scale vibration measurement. Cameras were typically devices to acquire images/videos for offline automated assessment (recently) or online manual monitoring (mainly), this work has shown how fog/edge computing makes it possible for these cameras to be "conscious" and "intelligent," hence play a critical role in automation/digitalization of drilling rigs. We showcase their preliminary application as drilling dynamics and rig operations sensors in this work. Cameras are an ideal sensor for a drilling environment since they can be installed anywhere on a rig to perform large-scale live video analytics on drilling processes.


Sensors ◽  
2018 ◽  
Vol 18 (10) ◽  
pp. 3415 ◽  
Author(s):  
Jinpeng Zhang ◽  
Jinming Zhang ◽  
Shan Yu

In the image object detection task, a huge number of candidate boxes are generated to match with a relatively very small amount of ground-truth boxes, and through this method the learning samples can be created. But in fact the vast majority of the candidate boxes do not contain valid object instances and should be recognized and rejected during the training and evaluation of the network. This leads to extra high computation burden and a serious imbalance problem between object and none-object samples, thereby impeding the algorithm’s performance. Here we propose a new heuristic sampling method to generate candidate boxes for two-stage detection algorithms. It is generally applicable to the current two-stage detection algorithms to improve their detection performance. Experiments on COCO dataset showed that, relative to the baseline model, this new method could significantly increase the detection accuracy and efficiency.


2020 ◽  
Vol 219 (10) ◽  
Author(s):  
Dominic Waithe ◽  
Jill M. Brown ◽  
Katharina Reglinski ◽  
Isabel Diez-Sevilla ◽  
David Roberts ◽  
...  

Object detection networks are high-performance algorithms famously applied to the task of identifying and localizing objects in photography images. We demonstrate their application for the classification and localization of cells in fluorescence microscopy by benchmarking four leading object detection algorithms across multiple challenging 2D microscopy datasets. Furthermore we develop and demonstrate an algorithm that can localize and image cells in 3D, in close to real time, at the microscope using widely available and inexpensive hardware. Furthermore, we exploit the fast processing of these networks and develop a simple and effective augmented reality (AR) system for fluorescence microscopy systems using a display screen and back-projection onto the eyepiece. We show that it is possible to achieve very high classification accuracy using datasets with as few as 26 images present. Using our approach, it is possible for relatively nonskilled users to automate detection of cell classes with a variety of appearances and enable new avenues for automation of fluorescence microscopy acquisition pipelines.


Sign in / Sign up

Export Citation Format

Share Document