Vision based distance estimation from single RGB camera using field of view and magnification measurements –an AI based non triangulation technique for person distance estimation in surveillance areas

2020 ◽  
pp. 1-17
Author(s):  
P.J.A. Alphonse ◽  
K.V. Sriharsha

Depth data from conventional cameras in monitoring fields provides a thorough assessment of human behavior. In this context, the depth of each viewpoint must be calculated using binocular stereo, which requires two cameras to retrieve 3D data. In networked surveillance environments, this drives excess energy and also provides extra infrastructure. We launched a new computational photographic technique for depth estimation using a single camera based on the ideas of perspective projection and lens magnification property. The person to camera distance (or depth) is obtained from understanding the focal length, field of view and magnification characteristics. Prior to finding distance, initially real height is estimated using Human body anthropometrics. These metrics are given as inputs to the Gradient-Boosting machine learning algorithm for estimating Real Height. And then magnification and Field of View measurements are extracted for each sample. The depth (or distance) is predicted on the basis of the geometrical relationship between field of view, magnification and camera at object distance. Using physical distance and height measurements taken in real time as ground truth, experimental validation is performed and it is inferred that with in 3m–7 m range, both in indoor and outdoor environments, the camera to person distance (Preddist) anticipated from field of view and magnification is 91% correlated with actual depth at a confidence point of 95% with RMSE of 0.579.

2021 ◽  
Vol 3 (6) ◽  
Author(s):  
P. J. A. Alphonse ◽  
K. V. Sriharsha

AbstractIn recent years, with increase in concern about public safety and security, human movements or action sequences are highly valued when dealing with suspicious and criminal activities. In order to estimate the position and orientation related to human movements, depth information is needed. This is obtained by fusing data obtained from multiple cameras at different viewpoints. In practice, whenever occlusion occurs in a surveillance environment, there may be a pixel-to-pixel correspondence between two images captured from two cameras and, as a result, depth information may not be accurate. Moreover use of more than one camera exclusively adds burden to the surveillance infrastructure. In this study, we present a mathematical model for acquiring object depth information using single camera by capturing the in focused portion of an object from a single image. When camera is in-focus, with the reference to camera lens center, for a fixed focal length for each aperture setting, the object distance is varied. For each aperture reading, for the corresponding distance, the object distance (or depth) is estimated by relating the three parameters namely lens aperture radius, object distance and object size in image plane. The results show that the distance computed from the relationship approximates actual with a standard error estimate of 2.39 to 2.54, when tested on Nikon and Cannon versions with an accuracy of 98.1% at 95% confidence level.


Sensors ◽  
2020 ◽  
Vol 20 (14) ◽  
pp. 3860
Author(s):  
Namhoon Kim ◽  
Junsu Bae ◽  
Cheolhwan Kim ◽  
Soyeon Park ◽  
Hong-Gyoo Sohn

This paper proposes a technique to estimate the distance between an object and a rolling shutter camera using a single image. The implementation of this technique uses the principle of the rolling shutter effect (RSE), a distortion within the rolling-shutter-type camera. The proposed technique has a mathematical strength compared to other single photo-based distance estimation methods that do not consider the geometric arrangement. The relationship between the distance and RSE angle was derived using the camera parameters (focal length, shutter speed, image size, etc.). Mathematical equations were derived for three different scenarios. The mathematical model was verified through experiments using a Nikon D750 and Nikkor 50 mm lens mounted on a car with varying speeds, object distances, and camera parameters. The results show that the mathematical model provides an accurate distance estimation of an object. The distance estimation error using the RSE due to the change in speed remained stable at approximately 10 cm. However, when the distance between the object and camera was more than 10 m, the estimated distance was sensitive to the RSE and the error increased dramatically.


Author(s):  
Shubhada Mone ◽  
Nihar Salunke ◽  
Omkar Jadhav ◽  
Arjun Barge ◽  
Nikhil Magar

With the easy availability of technology, smartphones are playing an important role in every person’s life. Also, with the advancements in computer vision based research, Automatic Driving cars, Object Recognition, Depth Map Prediction, Object Distance Estimation, have reached commendable levels of intelligence and accuracy. Combining the research and technological advancements, we can be hopeful in creating a computer vision based mobile-application which will help guide visually disabled people in performing their day to day tasks with easily available mobile applications. With our study, the visually disabled can perform simple tasks like outdoor/indoor navigation without encountering obstacles, also they can avoid accidental collisions with objects in their surroundings. Currently, there are very few applications which provide the same assistance to the visually impaired. Using physical tools like sticks is a very common practice when it comes to avoiding obstacles in a visually disabled person’s path. Our study will be focused on object detection and depth estimation techniques- two of the most popular and advanced fields in Intelligent Computer vision studies. We have explored more on the traditional challenges and future hopes of incorporating these techniques on embedded devices.


Author(s):  
A. V. Crewe ◽  
J. Wall ◽  
L. M. Welter

A scanning microscope using a field emission source has been described elsewhere. This microscope has now been improved by replacing the single magnetic lens with a high quality lens of the type described by Ruska. This lens has a focal length of 1 mm and a spherical aberration coefficient of 0.5 mm. The final spot size, and therefore the microscope resolution, is limited by the aberration of this lens to about 6 Å.The lens has been constructed very carefully, maintaining a tolerance of + 1 μ on all critical surfaces. The gun is prealigned on the lens to form a compact unit. The only mechanical adjustments are those which control the specimen and the tip positions. The microscope can be used in two modes. With the lens off and the gun focused on the specimen, the resolution is 250 Å over an undistorted field of view of 2 mm. With the lens on,the resolution is 20 Å or better over a field of view of 40 microns. The magnification can be accurately varied by attenuating the raster current.


Author(s):  
Louis Lecrosnier ◽  
Redouane Khemmar ◽  
Nicolas Ragot ◽  
Benoit Decoux ◽  
Romain Rossi ◽  
...  

This paper deals with the development of an Advanced Driver Assistance System (ADAS) for a smart electric wheelchair in order to improve the autonomy of disabled people. Our use case, built from a formal clinical study, is based on the detection, depth estimation, localization and tracking of objects in wheelchair’s indoor environment, namely: door and door handles. The aim of this work is to provide a perception layer to the wheelchair, enabling this way the detection of these keypoints in its immediate surrounding, and constructing of a short lifespan semantic map. Firstly, we present an adaptation of the YOLOv3 object detection algorithm to our use case. Then, we present our depth estimation approach using an Intel RealSense camera. Finally, as a third and last step of our approach, we present our 3D object tracking approach based on the SORT algorithm. In order to validate all the developments, we have carried out different experiments in a controlled indoor environment. Detection, distance estimation and object tracking are experimented using our own dataset, which includes doors and door handles.


2021 ◽  
Vol 4 (1) ◽  
Author(s):  
Peter M. Maloca ◽  
Philipp L. Müller ◽  
Aaron Y. Lee ◽  
Adnan Tufail ◽  
Konstantinos Balaskas ◽  
...  

AbstractMachine learning has greatly facilitated the analysis of medical data, while the internal operations usually remain intransparent. To better comprehend these opaque procedures, a convolutional neural network for optical coherence tomography image segmentation was enhanced with a Traceable Relevance Explainability (T-REX) technique. The proposed application was based on three components: ground truth generation by multiple graders, calculation of Hamming distances among graders and the machine learning algorithm, as well as a smart data visualization (‘neural recording’). An overall average variability of 1.75% between the human graders and the algorithm was found, slightly minor to 2.02% among human graders. The ambiguity in ground truth had noteworthy impact on machine learning results, which could be visualized. The convolutional neural network balanced between graders and allowed for modifiable predictions dependent on the compartment. Using the proposed T-REX setup, machine learning processes could be rendered more transparent and understandable, possibly leading to optimized applications.


2021 ◽  
Vol 2021 (1) ◽  
Author(s):  
Xiang Li ◽  
Jianzheng Liu ◽  
Jessica Baron ◽  
Khoa Luu ◽  
Eric Patterson

AbstractRecent attention to facial alignment and landmark detection methods, particularly with application of deep convolutional neural networks, have yielded notable improvements. Neither these neural-network nor more traditional methods, though, have been tested directly regarding performance differences due to camera-lens focal length nor camera viewing angle of subjects systematically across the viewing hemisphere. This work uses photo-realistic, synthesized facial images with varying parameters and corresponding ground-truth landmarks to enable comparison of alignment and landmark detection techniques relative to general performance, performance across focal length, and performance across viewing angle. Recently published high-performing methods along with traditional techniques are compared in regards to these aspects.


2018 ◽  
Author(s):  
Jatin Kumar ◽  
Qianxiao Li ◽  
Karen Y.T. Tang ◽  
Tonio Buonassisi ◽  
Anibal L. Gonzalez-Oyarce ◽  
...  

<div><div><div><p>Inverse design is an outstanding challenge in disordered systems with multiple length scales such as polymers, particularly when designing polymers with desired phase behavior. We demonstrate high-accuracy tuning of poly(2-oxazoline) cloud point via machine learning. With a design space of four repeating units and a range of molecular masses, we achieve an accuracy of 4°C root mean squared error (RMSE) in a temperature range of 24– 90°C, employing gradient boosting with decision trees. The RMSE is >3x better than linear and polynomial regression. We perform inverse design via particle-swarm optimization, predicting and synthesizing 17 polymers with constrained design at 4 target cloud points from 37 to 80°C. Our approach challenges the status quo in polymer design with a machine learning algorithm, that is capable of fast and systematic discovery of new polymers.</p></div></div></div>


Sign in / Sign up

Export Citation Format

Share Document