scholarly journals Violence Detection Using Spatiotemporal Features with 3D Convolutional Neural Network

Sensors ◽  
2019 ◽  
Vol 19 (11) ◽  
pp. 2472 ◽  
Author(s):  
Fath U Min Ullah ◽  
Amin Ullah ◽  
Khan Muhammad ◽  
Ijaz Ul Haq ◽  
Sung Wook Baik

The worldwide utilization of surveillance cameras in smart cities has enabled researchers to analyze a gigantic volume of data to ensure automatic monitoring. An enhanced security system in smart cities, schools, hospitals, and other surveillance domains is mandatory for the detection of violent or abnormal activities to avoid any casualties which could cause social, economic, and ecological damages. Automatic detection of violence for quick actions is very significant and can efficiently assist the concerned departments. In this paper, we propose a triple-staged end-to-end deep learning violence detection framework. First, persons are detected in the surveillance video stream using a light-weight convolutional neural network (CNN) model to reduce and overcome the voluminous processing of useless frames. Second, a sequence of 16 frames with detected persons is passed to 3D CNN, where the spatiotemporal features of these sequences are extracted and fed to the Softmax classifier. Furthermore, we optimized the 3D CNN model using an open visual inference and neural networks optimization toolkit developed by Intel, which converts the trained model into intermediate representation and adjusts it for optimal execution at the end platform for the final prediction of violent activity. After detection of a violent activity, an alert is transmitted to the nearest police station or security department to take prompt preventive actions. We found that our proposed method outperforms the existing state-of-the-art methods for different benchmark datasets.

2021 ◽  
Vol 10 (6) ◽  
pp. 3137-3146
Author(s):  
Malik A. Alsaedi ◽  
Abdulrahman Saeed Mohialdeen ◽  
Baraa Munqith Albaker

Human activity recognition (HAR) is recently used in numerous applications including smart homes to monitor human behavior, automate homes according to human activities, entertainment, falling detection, violence detection, and people care. Vision-based recognition is the most powerful method widely used in HAR systems implementation due to its characteristics in recognizing complex human activities. This paper addresses the design of a 3D convolutional neural network (3D-CNN) model that can be used in smart homes to identify several numbers of activities. The model is trained using KTH dataset that contains activities like (walking, running, jogging, handwaving handclapping, boxing). Despite the challenges of this method due to the effectiveness of the lamination, background variation, and human body variety, the proposed model reached an accuracy of 93.33%. The model was implemented, trained and tested using moderate computation machine and the results show that the proposal was successfully capable to recognize human activities with reasonable computations.


Sensors ◽  
2021 ◽  
Vol 21 (14) ◽  
pp. 4916
Author(s):  
Ali Usman Gondal ◽  
Muhammad Imran Sadiq ◽  
Tariq Ali ◽  
Muhammad Irfan ◽  
Ahmad Shaf ◽  
...  

Urbanization is a big concern for both developed and developing countries in recent years. People shift themselves and their families to urban areas for the sake of better education and a modern lifestyle. Due to rapid urbanization, cities are facing huge challenges, one of which is waste management, as the volume of waste is directly proportional to the people living in the city. The municipalities and the city administrations use the traditional wastage classification techniques which are manual, very slow, inefficient and costly. Therefore, automatic waste classification and management is essential for the cities that are being urbanized for the better recycling of waste. Better recycling of waste gives the opportunity to reduce the amount of waste sent to landfills by reducing the need to collect new raw material. In this paper, the idea of a real-time smart waste classification model is presented that uses a hybrid approach to classify waste into various classes. Two machine learning models, a multilayer perceptron and multilayer convolutional neural network (ML-CNN), are implemented. The multilayer perceptron is used to provide binary classification, i.e., metal or non-metal waste, and the CNN identifies the class of non-metal waste. A camera is placed in front of the waste conveyor belt, which takes a picture of the waste and classifies it. Upon successful classification, an automatic hand hammer is used to push the waste into the assigned labeled bucket. Experiments were carried out in a real-time environment with image segmentation. The training, testing, and validation accuracy of the purposed model was 0.99% under different training batches with different input features.


2021 ◽  
Vol 13 (4) ◽  
pp. 628
Author(s):  
Liang Ye ◽  
Tong Liu ◽  
Tian Han ◽  
Hany Ferdinando ◽  
Tapio Seppänen ◽  
...  

Campus violence is a common social phenomenon all over the world, and is the most harmful type of school bullying events. As artificial intelligence and remote sensing techniques develop, there are several possible methods to detect campus violence, e.g., movement sensor-based methods and video sequence-based methods. Sensors and surveillance cameras are used to detect campus violence. In this paper, the authors use image features and acoustic features for campus violence detection. Campus violence data are gathered by role-playing, and 4096-dimension feature vectors are extracted from every 16 frames of video images. The C3D (Convolutional 3D) neural network is used for feature extraction and classification, and an average recognition accuracy of 92.00% is achieved. Mel-frequency cepstral coefficients (MFCCs) are extracted as acoustic features, and three speech emotion databases are involved. The C3D neural network is used for classification, and the average recognition accuracies are 88.33%, 95.00%, and 91.67%, respectively. To solve the problem of evidence conflict, the authors propose an improved Dempster–Shafer (D–S) algorithm. Compared with existing D–S theory, the improved algorithm increases the recognition accuracy by 10.79%, and the recognition accuracy can ultimately reach 97.00%.


2020 ◽  
Vol 12 (1) ◽  
pp. 39-55
Author(s):  
Hadj Ahmed Bouarara

In recent years, surveillance video has become a familiar phenomenon because it gives us a feeling of greater security, but we are continuously filmed and our privacy is greatly affected. This work deals with the development of a private video surveillance system (PVSS) using regression residual convolutional neural network (RR-CNN) with the goal to propose a new security policy to ensure the privacy of no-dangerous person and prevent crime. The goal is to best meet the interests of all parties: the one who films and the one who is filmed.


2019 ◽  
Author(s):  
Jinhyeong Bae ◽  
Jane Stocks ◽  
Ashley Heywood ◽  
Youngmoon Jung ◽  
Lisanne Jenkins ◽  
...  

AbstractDementia of Alzheimer’s Type (DAT) is associated with a devastating and irreversible cognitive decline. As a pharmacological intervention has not yet been developed to reverse disease progression, preventive medicine will play a crucial role for patient care and treatment planning. However, predicting which patients will progress to DAT is difficult as patients with Mild Cognitive Impairment (MCI) could either convert to DAT (MCI-C) or not (MCI-NC). In this paper, we develop a deep learning model to address the heterogeneous nature of DAT development. Structural magnetic resonance imaging was utilized as a single biomarker, and a three-dimensional convolutional neural network (3D-CNN) was developed. The 3D-CNN was trained using transfer learning from the classification of Normal Control and DAT scans at the source task. This was applied to the target task of classifying MCI-C and MCI-NC scans. The model results in 82.4% classification accuracy, which outperforms current models in the field. Furthermore, by implementing an occlusion map approach, we visualize key brain regions that significantly contribute to the prediction of MCI-C and MCI-NC. Results show the hippocampus, amygdala, cerebellum, and pons regions as significant to prediction, which are consistent with current understanding of disease. Finally, the model’s prediction value is significantly correlated with rates of change in clinical assessment scores, indicating the model is able to predict an individual patient’s future cognitive decline. This information, in conjunction with the identified anatomical features, will aid in building a personalized therapeutic strategy for individuals with MCI. This model could also be useful for selection of participants for clinical trials.


2019 ◽  
Author(s):  
Zini Jian ◽  
Xianpei Wang ◽  
Jingzhe Zhang ◽  
Xinyu Wang ◽  
Youbin Deng

Abstract Background: Clinically, doctors obtain the left ventricular posterior wall thickness (LVPWT) mainly by observing ultrasonic echocardiographic video stream to capture a single frame of images with diagnostic significance, and then mark two key points on both sides of the posterior wall of the left ventricle with their own experience for computer measurement. In the actual measurement, the doctor's selection point is subjective, which is not only time-consuming and laborious, but also difficult to accurately locate the edge, which will bring errors to the measurement results. Methods: In this paper, a convolutional neural network model of left ventricular posterior wall positioning was built under the TensorFlow framework, and the target region images were obtained after the positioning results were processed by non-local mean filtering and opening operation. Then the edge detection algorithm based on threshold segmentation is used. After the contour was extracted by adjusting the segmentation threshold through prior analysis and the OTSU algorithm, the design algorithm completed the computer selection point measurement of the thickness of the posterior wall of the left ventricle. Results: The proposed method can effectively extract the left ventricular posterior wall contour and measure its thickness. The experimental results show that the relative error between the measurement result and the hospital measurement value is less than 15%, which is less than 20% of the acceptable repeatability error in clinical practice. Conclusions: Therefore, the method proposed in this paper not only has the advantage of less manual intervention, but also can reduce the workload of doctors.


Sign in / Sign up

Export Citation Format

Share Document