scholarly journals TUHAD: Taekwondo Unit Technique Human Action Dataset with Key Frame-Based CNN Action Recognition

Sensors ◽  
2020 ◽  
Vol 20 (17) ◽  
pp. 4871
Author(s):  
Jinkue Lee ◽  
Hoeryong Jung

In taekwondo, poomsae (i.e., form) competitions have no quantitative scoring standards, unlike gyeorugi (i.e., full-contact sparring) in the Olympics. Consequently, there are diverse fairness issues regarding poomsae evaluation, and the demand for quantitative evaluation tools is increasing. Action recognition is a promising approach, but the extreme and rapid actions of taekwondo complicate its application. This study established the Taekwondo Unit technique Human Action Dataset (TUHAD), which consists of multimodal image sequences of poomsae actions. TUHAD contains 1936 action samples of eight unit techniques performed by 10 experts and captured by two camera views. A key frame-based convolutional neural network architecture was developed for taekwondo action recognition, and its accuracy was validated for various input configurations. A correlation analysis of the input configuration and accuracy demonstrated that the proposed model achieved a recognition accuracy of up to 95.833% (lowest accuracy of 74.49%). This study contributes to the research and development of taekwondo action recognition.

2021 ◽  
Vol 11 (1) ◽  
pp. 28
Author(s):  
Ana Bárbara Cardoso ◽  
Bruno Martins ◽  
Jacinto Estima

This article describes a novel approach for toponym resolution with deep neural networks. The proposed approach does not involve matching references in the text against entries in a gazetteer, instead directly predicting geo-spatial coordinates. Multiple inputs are considered in the neural network architecture (e.g., the surrounding words are considered in combination with the toponym to disambiguate), using pre-trained contextual word embeddings (i.e., ELMo or BERT) as well as bi-directional Long Short-Term Memory units, which are both regularly used for modeling textual data. The intermediate representations are then used to predict a probability distribution over possible geo-spatial regions, and finally to predict the coordinates for the input toponym. The proposed model was tested on three datasets used on previous toponym resolution studies, specifically the (i) War of the Rebellion, (ii) Local–Global Lexicon, and (iii) SpatialML corpora. Moreover, we evaluated the effect of using (i) geophysical terrain properties as external information, including information on elevation or terrain development, among others, and (ii) additional data collected from Wikipedia articles, to further help with the training of the model. The obtained results show improvements using the proposed method, when compared to previous approaches, and specifically when BERT embeddings and additional data are involved.


2021 ◽  
Author(s):  
Hai-Hong Phan ◽  
Trung Tin Nguyen ◽  
Ngo Huu Phuc ◽  
Nguyen Huu Nhan ◽  
Do Minh Hieu ◽  
...  

Sensors ◽  
2020 ◽  
Vol 20 (7) ◽  
pp. 1825 ◽  
Author(s):  
Huy Hieu Pham ◽  
Houssam Salmane ◽  
Louahdi Khoudour ◽  
Alain Crouzil ◽  
Sergio A. Velastin ◽  
...  

We present a deep learning-based multitask framework for joint 3D human pose estimation and action recognition from RGB sensors using simple cameras. The approach proceeds along two stages. In the first, a real-time 2D pose detector is run to determine the precise pixel location of important keypoints of the human body. A two-stream deep neural network is then designed and trained to map detected 2D keypoints into 3D poses. In the second stage, the Efficient Neural Architecture Search (ENAS) algorithm is deployed to find an optimal network architecture that is used for modeling the spatio-temporal evolution of the estimated 3D poses via an image-based intermediate representation and performing action recognition. Experiments on Human3.6M, MSR Action3D and SBU Kinect Interaction datasets verify the effectiveness of the proposed method on the targeted tasks. Moreover, we show that the method requires a low computational budget for training and inference. In particular, the experimental results show that by using a monocular RGB sensor, we can develop a 3D pose estimation and human action recognition approach that reaches the performance of RGB-depth sensors. This opens up many opportunities for leveraging RGB cameras (which are much cheaper than depth cameras and extensively deployed in private and public places) to build intelligent recognition systems.


2015 ◽  
Vol 713-715 ◽  
pp. 2152-2155 ◽  
Author(s):  
Shao Ping Zhu

According to the problem that achieves robust human actions recognition from image sequences in computer vision, using the Iterative Querying Heuristic algorithm as a guide, a improved Multiple Instance Learning (MIL) method is proposed for human action recognition in video image sequences. Experiments show that the new method can quickly recognize human actions and achieve high recognition rates, and on the Weizmann database validate our analysis.


Author(s):  
Karthika Gidijala ◽  
◽  
Mansa Devi Pappu ◽  
Manasa Vavilapalli ◽  
Mahesh Kothuru ◽  
...  

Many different models of Convolution Neural Networks exist in the Deep Learning studies. The application and prudence of the algorithms is known only when they are implemented with strong datasets. The histopathological images of breast cancer are considered as to have much number of haphazard structures and textures. Dealing with such images is a challenging issue in deep learning. Working on wet labs and in coherence to the results many research have blogged with novel annotations in the research. In this paper, we are presenting a model that can work efficiently on the raw images with different resolutions and alleviating with the problems of the presence of the structures and textures. The proposed model achieves considerably good results useful for decision making in cancer diagnosis.


2021 ◽  
Vol 11 (18) ◽  
pp. 8628
Author(s):  
Kang-Moon Park ◽  
Donghoon Shin ◽  
Sung-Do Chi

This paper proposes a deep neural network structuring methodology through a genetic algorithm (GA) using chromosome non-disjunction. The proposed model includes methods for generating and tuning the neural network architecture without the aid of human experts. Since the original neural architecture search (henceforth, NAS) was announced, NAS techniques, such as NASBot, NASGBO and CoDeepNEAT, have been widely adopted in order to improve cost- and/or time-effectiveness for human experts. In these models, evolutionary algorithms (EAs) are employed to effectively enhance the accuracy of the neural network architecture. In particular, CoDeepNEAT uses a constructive GA starting from minimal architecture. This will only work quickly if the solution architecture is small. On the other hand, the proposed methodology utilizes chromosome non-disjunction as a new genetic operation. Our approach differs from previous methodologies in that it includes a destructive approach as well as a constructive approach, and is similar to pruning methodologies, which realizes tuning of the previous neural network architecture. A case study applied to the sentence word ordering problem and AlexNet for CIFAR-10 illustrates the applicability of the proposed methodology. We show from the simulation studies that the accuracy of the model was improved by 0.7% compared to the conventional model without human expert.


Author(s):  
Huifeng Guo ◽  
Ruiming TANG ◽  
Yunming Ye ◽  
Zhenguo Li ◽  
Xiuqiang He

Learning sophisticated feature interactions behind user behaviors is critical in maximizing CTR for recommender systems. Despite great progress, existing methods seem to have a strong bias towards low- or high-order interactions, or require expertise feature engineering. In this paper, we show that it is possible to derive an end-to-end learning model that emphasizes both low- and high-order feature interactions. The proposed model, DeepFM, combines the power of factorization machines for recommendation and deep learning for feature learning in a new neural network architecture. Compared to the latest Wide & Deep model from Google, DeepFM has a shared input to its "wide" and "deep" parts, with no need of feature engineering besides raw features. Comprehensive experiments are conducted to demonstrate the effectiveness and efficiency of DeepFM over the existing models for CTR prediction, on both benchmark data and commercial data.


Sign in / Sign up

Export Citation Format

Share Document