scholarly journals The Real-Time Depth Estimation for an Occluded Person Based on a Single Image and OpenPose Method

Mathematics ◽  
2020 ◽  
Vol 8 (8) ◽  
pp. 1333 ◽  
Author(s):  
Yu-Shiuan Tsai ◽  
Li-Heng Hsu ◽  
Yi-Zeng Hsieh ◽  
Shih-Syun Lin

In recent years, the breakthrough of neural networks and the rise of deep learning have led to the advancement of machine vision, which has been commonly used in the practical application of image recognition. Automobiles, drones, portable devices, behavior recognition, indoor positioning and many other industries also rely on the integrated application, and require the support of deep learning and machine vision. As for these technologies, there is a high demand for the accuracy related to the recognition of portraits or objects. The recognition of human figures is also a research goal that has drawn great attention in various fields. However, the portrait will be affected by various factors such as height, weight, posture, angle and whether it is covered or not, which affects the accuracy of recognition. This paper applies the application of deep learning to portraits with different poses and angles, especially the actual distance of a single lens for the shadowed portrait (depth estimation), so that it can be used for automatic control of drones in the future. Traditional methods for calculating depth using images are mainly divided into three types: one—single-lens estimation, two—lens estimation, and three—optical band estimation. In view of the fact that both the second and third categories require relatively large and expensive equipment to effectively perform distance calculations, numerous methods for calculating distance using a single lens have recently been produced. However, whether it is the use of traditional “units of distance measurement calibration”, “defocus distance measurement”, or the “three-dimensional grid space messages distance measurement method”, all of these face corresponding difficulties and problems. Additionally, they have to deal with outside disturbances and process the shadowed image. Therefore, under the new research method, OpenPose, which is proposed by Carnegie Mellon University, this paper intends to propose a depth algorithm for a single-lens occluded portrait to estimate the actual portrait distance for different poses, angles of view and obscuration.

Polymers ◽  
2019 ◽  
Vol 11 (1) ◽  
pp. 176 ◽  
Author(s):  
Wei Dong ◽  
Xuan Wang ◽  
Bo Tian ◽  
Yuguang Liu ◽  
Zaixing Jiang ◽  
...  

Aromatic voltage stabilizers can improve the dielectric properties of cross-linked polyethylene (XLPE); however, their poor compatibility with XLPE hinders their practical application. Improving the compatibility of aromatic voltage stabilizers with XLPE has, therefore, become a new research goal. Herein 1-(4-vinyloxy)phenylethenone (VPE) was prepared and characterized. It can be grafted onto polyethylene molecules during the cross-linking processes to promote stability of the aromatic voltage stabilizers in XLPE. Fourier transform infrared spectroscopy confirmed that VPE was successfully grafted onto XLPE, and effectively inhibited thermal migration. Thermogravimetric analysis showed that the grafted VPE/XLPE composite exhibits a better thermal stability than a VPE/PE blend composite. Evaluation of the electrical properties showed that the breakdown strength and electrical tree initiation voltage of the VPE/XLPE composite were increased by 15.5% and 39.6%, respectively, when compared to those of bare XLPE. After thermal aging, the breakdown strength and electrical tree initiation voltage of the VPE/XLPE composite were increased by 9.4% and 25.8%, respectively, in comparison to those of bare XLPE, which indicates that the grafted voltage stabilizer can effectively inhibit its migration and enhance the stability of the composite material.


Sensors ◽  
2019 ◽  
Vol 19 (17) ◽  
pp. 3738 ◽  
Author(s):  
Abozar Nasirahmadi ◽  
Barbara Sturm ◽  
Sandra Edwards ◽  
Knut-Håkan Jeppsson ◽  
Anne-Charlotte Olsson ◽  
...  

Posture detection targeted towards providing assessments for the monitoring of health and welfare of pigs has been of great interest to researchers from different disciplines. Existing studies applying machine vision techniques are mostly based on methods using three-dimensional imaging systems, or two-dimensional systems with the limitation of monitoring under controlled conditions. Thus, the main goal of this study was to determine whether a two-dimensional imaging system, along with deep learning approaches, could be utilized to detect the standing and lying (belly and side) postures of pigs under commercial farm conditions. Three deep learning-based detector methods, including faster regions with convolutional neural network features (Faster R-CNN), single shot multibox detector (SSD) and region-based fully convolutional network (R-FCN), combined with Inception V2, Residual Network (ResNet) and Inception ResNet V2 feature extractions of RGB images were proposed. Data from different commercial farms were used for training and validation of the proposed models. The experimental results demonstrated that the R-FCN ResNet101 method was able to detect lying and standing postures with higher average precision (AP) of 0.93, 0.95 and 0.92 for standing, lying on side and lying on belly postures, respectively and mean average precision (mAP) of more than 0.93.


1991 ◽  
Vol 24 (6) ◽  
pp. 171-177 ◽  
Author(s):  
Zeng Fantang ◽  
Xu Zhencheng ◽  
Chen Xiancheng

A real-time mathematical model for three-dimensional tidal flow and water quality is presented in this paper. A control-volume-based difference method and a “power interpolation distribution” advocated by Patankar (1984) have been employed, and a concept of “separating the top-layer water” has been developed to solve the movable boundary problem. The model is unconditionally stable and convergent. Practical application of the model is illustrated by an example for the Pearl River Estuary.


Author(s):  
Dan Luo

Background: As known that the semi-supervised algorithm is a classical algorithm in semi-supervised learning algorithm. Methods: In the paper, it proposed improved cooperative semi-supervised learning algorithm, and the algorithm process is presented in detailed, and it is adopted to predict unlabeled electronic components image. Results: In the experiments of classification and recognition of electronic components, it show that through the method the accuracy the proposed algorithm in electron device image recognition can be significantly improved, the improved algorithm can be used in the actual recognition process . Conclusion: With the continuous development of science and technology, machine vision and deep learning will play a more important role in people's life in the future. The subject research based on the identification of the number of components is bound to develop towards the direction of high precision and multi-dimension, which will greatly improve the production efficiency of electronic components industry.


2019 ◽  
Vol 46 (7) ◽  
pp. 3180-3193 ◽  
Author(s):  
Ran Zhou ◽  
Aaron Fenster ◽  
Yujiao Xia ◽  
J. David Spence ◽  
Mingyue Ding

Author(s):  
Qiang Yu ◽  
Feiqiang Liu ◽  
Long Xiao ◽  
Zitao Liu ◽  
Xiaomin Yang

Deep-learning (DL)-based methods are of growing importance in the field of single image super-resolution (SISR). The practical application of these DL-based models is a remaining problem due to the requirement of heavy computation and huge storage resources. The powerful feature maps of hidden layers in convolutional neural networks (CNN) help the model learn useful information. However, there exists redundancy among feature maps, which can be further exploited. To address these issues, this paper proposes a lightweight efficient feature generating network (EFGN) for SISR by constructing the efficient feature generating block (EFGB). Specifically, the EFGB can conduct plain operations on the original features to produce more feature maps with parameters slightly increasing. With the help of these extra feature maps, the network can extract more useful information from low resolution (LR) images to reconstruct the desired high resolution (HR) images. Experiments conducted on the benchmark datasets demonstrate that the proposed EFGN can outperform other deep-learning based methods in most cases and possess relatively lower model complexity. Additionally, the running time measurement indicates the feasibility of real-time monitoring.


Sensors ◽  
2021 ◽  
Vol 21 (6) ◽  
pp. 1952
Author(s):  
May Phu Paing ◽  
Supan Tungjitkusolmun ◽  
Toan Huy Bui ◽  
Sarinporn Visitsattapongse ◽  
Chuchart Pintavirooj

Automated segmentation methods are critical for early detection, prompt actions, and immediate treatments in reducing disability and death risks of brain infarction. This paper aims to develop a fully automated method to segment the infarct lesions from T1-weighted brain scans. As a key novelty, the proposed method combines variational mode decomposition and deep learning-based segmentation to take advantages of both methods and provide better results. There are three main technical contributions in this paper. First, variational mode decomposition is applied as a pre-processing to discriminate the infarct lesions from unwanted non-infarct tissues. Second, overlapped patches strategy is proposed to reduce the workload of the deep-learning-based segmentation task. Finally, a three-dimensional U-Net model is developed to perform patch-wise segmentation of infarct lesions. A total of 239 brain scans from a public dataset is utilized to develop and evaluate the proposed method. Empirical results reveal that the proposed automated segmentation can provide promising performances with an average dice similarity coefficient (DSC) of 0.6684, intersection over union (IoU) of 0.5022, and average symmetric surface distance (ASSD) of 0.3932, respectively.


Sensors ◽  
2021 ◽  
Vol 21 (3) ◽  
pp. 884
Author(s):  
Chia-Ming Tsai ◽  
Yi-Horng Lai ◽  
Yung-Da Sun ◽  
Yu-Jen Chung ◽  
Jau-Woei Perng

Numerous sensors can obtain images or point cloud data on land, however, the rapid attenuation of electromagnetic signals and the lack of light in water have been observed to restrict sensing functions. This study expands the utilization of two- and three-dimensional detection technologies in underwater applications to detect abandoned tires. A three-dimensional acoustic sensor, the BV5000, is used in this study to collect underwater point cloud data. Some pre-processing steps are proposed to remove noise and the seabed from raw data. Point clouds are then processed to obtain two data types: a 2D image and a 3D point cloud. Deep learning methods with different dimensions are used to train the models. In the two-dimensional method, the point cloud is transferred into a bird’s eye view image. The Faster R-CNN and YOLOv3 network architectures are used to detect tires. Meanwhile, in the three-dimensional method, the point cloud associated with a tire is cut out from the raw data and is used as training data. The PointNet and PointConv network architectures are then used for tire classification. The results show that both approaches provide good accuracy.


Sign in / Sign up

Export Citation Format

Share Document