Entropy Based Fuzzy C Means Clustering and Key Frame Extraction for Sports Video Summarization

Author(s):  
Shanmukhappa Angadi ◽  
Vilas Naik
2021 ◽  
Vol 2021 ◽  
pp. 1-9
Author(s):  
Xiaoping Guo

Traditional text annotation-based video retrieval is done by manually labeling videos with text, which is inefficient and highly subjective and generally cannot accurately describe the meaning of videos. Traditional content-based video retrieval uses convolutional neural networks to extract the underlying feature information of images to build indexes and achieves similarity retrieval of video feature vectors according to certain similarity measure algorithms. In this paper, by studying the characteristics of sports videos, we propose the histogram difference method based on using transfer learning and the four-step method based on block matching for mutation detection and fading detection of video shots, respectively. By adaptive thresholding, regions with large frame difference changes are marked as candidate regions for shots, and then the shot boundaries are determined by mutation detection algorithm. Combined with the characteristics of sports video, this paper proposes a key frame extraction method based on clustering and optical flow analysis, and experimental comparison with the traditional clustering method. In addition, this paper proposes a key frame extraction algorithm based on clustering and optical flow analysis for key frame extraction of sports video. The algorithm effectively removes the redundant frames, and the extracted key frames are more representative. Through extensive experiments, the keyword fuzzy finding algorithm based on improved deep neural network and ontology semantic expansion proposed in this paper shows a more desirable retrieval performance, and it is feasible to use this method for video underlying feature extraction, annotation, and keyword finding, and one of the outstanding features of the algorithm is that it can quickly and effectively retrieve the desired video in a large number of Internet video resources, reducing the false detection rate and leakage rate while improving the fidelity, which basically meets people’s daily needs.


In videos, detecting text with multifaceted scenarios is perplexing. Texts in those videos have content full data facts that will be applied for various applications. Here, a system is proposed to enrich the text detection process from video. Here, a new method is implemented that detects Tamil text based on Gradient Vector Flow (GVF) and fuzzy c-means. First the video is split into number of frames. To circumvent temporal redundancy in each frame, a key frame is chosen and the frame where the text be located is identified to be the key frame. The dominant edge pixel is identified in that frame by the sobel edge map. Edge components are detected conforming towards the dominant pixel in sobel detector for constructing Text Candidates (TC).Clustering of a pixel is performed to detect text by using fuzzy c means clustering algorithm. Finally text is detected.


Sign in / Sign up

Export Citation Format

Share Document