scholarly journals Towards Deep Object Detection Techniques for Phoneme Recognition

IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 54663-54680 ◽  
Author(s):  
Mohammed Algabri ◽  
Hassan Mathkour ◽  
Mohamed Abdelkader Bencherif ◽  
Mansour Alsulaiman ◽  
Mohamed Amine Mekhtiche
Sensors ◽  
2021 ◽  
Vol 21 (4) ◽  
pp. 1205
Author(s):  
Mohammed Algabri ◽  
Hassan Mathkour ◽  
Mansour M. Alsulaiman ◽  
Mohamed A. Bencherif

This study proposes using object detection techniques to recognize sequences of articulatory features (AFs) from speech utterances by treating AFs of phonemes as multi-label objects in speech spectrogram. The proposed system, called AFD-Obj, recognizes sequence of multi-label AFs in speech signal and localizes them. AFD-Obj consists of two main stages: firstly, we formulate the problem of AFs detection as an object detection problem and prepare the data to fulfill requirement of object detectors by generating a spectral three-channel image from the speech signal and creating the corresponding annotation for each utterance. Secondly, we use annotated images to train the proposed system to detect sequences of AFs and their boundaries. We test the system by feeding spectrogram images to the system, which will recognize and localize multi-label AFs. We investigated using these AFs to detect the utterance phonemes. YOLOv3-tiny detector is selected because of its real-time property and its support for multi-label detection. We test our AFD-Obj system on Arabic and English languages using KAPD and TIMIT corpora, respectively. Additionally, we propose using YOLOv3-tiny as an Arabic phoneme detection system (i.e., PD-Obj) to recognize and localize a sequence of Arabic phonemes from whole speech utterances. The proposed AFD-Obj and PD-Obj systems achieve excellent results for Arabic corpus and comparable to the state-of-the-art method for English corpus. Moreover, we showed that using only one-scale detection is suitable for AFs detection or phoneme recognition.


2021 ◽  
Vol 11 (13) ◽  
pp. 6016
Author(s):  
Jinsoo Kim ◽  
Jeongho Cho

For autonomous vehicles, it is critical to be aware of the driving environment to avoid collisions and drive safely. The recent evolution of convolutional neural networks has contributed significantly to accelerating the development of object detection techniques that enable autonomous vehicles to handle rapid changes in various driving environments. However, collisions in an autonomous driving environment can still occur due to undetected obstacles and various perception problems, particularly occlusion. Thus, we propose a robust object detection algorithm for environments in which objects are truncated or occluded by employing RGB image and light detection and ranging (LiDAR) bird’s eye view (BEV) representations. This structure combines independent detection results obtained in parallel through “you only look once” networks using an RGB image and a height map converted from the BEV representations of LiDAR’s point cloud data (PCD). The region proposal of an object is determined via non-maximum suppression, which suppresses the bounding boxes of adjacent regions. A performance evaluation of the proposed scheme was performed using the KITTI vision benchmark suite dataset. The results demonstrate the detection accuracy in the case of integration of PCD BEV representations is superior to when only an RGB camera is used. In addition, robustness is improved by significantly enhancing detection accuracy even when the target objects are partially occluded when viewed from the front, which demonstrates that the proposed algorithm outperforms the conventional RGB-based model.


Author(s):  
Yukta Lapsiya ◽  
Dhruvi Jain ◽  
Parshva Shah ◽  
Atul Kachare

2021 ◽  
Vol 23 (11) ◽  
pp. 159-165
Author(s):  
JAYANTH DWIJESH H P ◽  
◽  
SANDEEP S V ◽  
RASHMI S ◽  
◽  
...  

In today’s world, accurate and fast information is vital for safe aircraft landings. The purpose of an EMAS (Engineered Materials Arresting System) is to prevent an aeroplane from overrunning with no human injury and minimal damage to the aircraft. Although various algorithms for object detection analysis have been developed, only a few researchers have examined image analysis as a landing assist. Image intensity edges are employed in one system to detect the sides of a runway in an image sequence, allowing the runway’s 3-dimensional position and orientation to be approximated. A fuzzy network system is used to improve object detection and extraction from aerial images. In another system, multi-scale, multiplatform imagery is used to combine physiologically and geometrically inspired algorithms for recognizing objects from hyper spectral and/or multispectral (HS/MS) imagery. However, the similarity in the top view of runways, buildings, highways, and other objects is a disadvantage of these methods. We propose a new method for detecting and tracking the runway based on pattern matching and texture analysis of digital images captured by aircraft cameras. Edge detection techniques are used to recognize runways from aerial images. The edge detection algorithms employed in this paper are the Hough Transform, Canny Filter, and Sobel Filter algorithms, which result in efficient detection.


2021 ◽  
Author(s):  
S J Fiona G Sathiaraj ◽  
S J Evelyn G Sathiaraj ◽  
Laxmi Bewoor

2019 ◽  
Vol 2019 ◽  
pp. 1-15 ◽  
Author(s):  
Qimei Wang ◽  
Feng Qi ◽  
Minghe Sun ◽  
Jianhua Qu ◽  
Jie Xue

This study develops tomato disease detection methods based on deep convolutional neural networks and object detection models. Two different models, Faster R-CNN and Mask R-CNN, are used in these methods, where Faster R-CNN is used to identify the types of tomato diseases and Mask R-CNN is used to detect and segment the locations and shapes of the infected areas. To select the model that best fits the tomato disease detection task, four different deep convolutional neural networks are combined with the two object detection models. Data are collected from the Internet and the dataset is divided into a training set, a validation set, and a test set used in the experiments. The experimental results show that the proposed models can accurately and quickly identify the eleven tomato disease types and segment the locations and shapes of the infected areas.


Sign in / Sign up

Export Citation Format

Share Document