scholarly journals Unsupervised Learning from Videos for Object Discovery in Single Images

Symmetry ◽  
2020 ◽  
Vol 13 (1) ◽  
pp. 38
Author(s):  
Dong Zhao ◽  
Baoqing Ding ◽  
Yulin Wu ◽  
Lei Chen ◽  
Hongchao Zhou

This paper proposes a method for discovering the primary objects in single images by learning from videos in a purely unsupervised manner—the learning process is based on videos, but the generated network is able to discover objects from a single input image. The rough idea is that an image typically consists of multiple object instances (like the foreground and background) that have spatial transformations across video frames and they can be sparsely represented. By exploring the sparsity representation of a video with a neural network, one may learn the features of each object instance without any labels, which can be used to discover, recognize, or distinguish object instances from a single image. In this paper, we consider a relatively simple scenario, where each image roughly consists of a foreground and a background. Our proposed method is based on encoder-decoder structures to sparsely represent the foreground, background, and segmentation mask, which further reconstruct the original images. We apply the feed-forward network trained from videos for object discovery in single images, which is different from the previous co-segmentation methods that require videos or collections of images as the input for inference. The experimental results on various object segmentation benchmarks demonstrate that the proposed method extracts primary objects accurately and robustly, which suggests that unsupervised image learning tasks can benefit from the sparsity of images and the inter-frame structure of videos.

2021 ◽  
Vol 2021 ◽  
pp. 1-16
Author(s):  
Haisheng Hui ◽  
Xueying Zhang ◽  
Zelin Wu ◽  
Fenlian Li

For the segmentation task of stroke lesions, using the attention U-Net model based on the self-attention mechanism can suppress irrelevant regions in an input image while highlighting salient features useful for specific tasks. However, when the lesion is small and the lesion contour is blurred, attention U-Net may generate wrong attention coefficient maps, leading to incorrect segmentation results. To cope with this issue, we propose a dual-path attention compensation U-Net (DPAC-UNet) network, which consists of a primary network and auxiliary path network. Both networks are attention U-Net models and identical in structure. The primary path network is the core network that performs accurate lesion segmentation and outputting of the final segmentation result. The auxiliary path network generates auxiliary attention compensation coefficients and sends them to the primary path network to compensate for and correct possible attention coefficient errors. To realize the compensation mechanism of DPAC-UNet, we propose a weighted binary cross-entropy Tversky (WBCE-Tversky) loss to train the primary path network to achieve accurate segmentation and propose another compound loss function called tolerance loss to train the auxiliary path network to generate auxiliary compensation attention coefficient maps with expanded coverage area to perform compensate operations. We conducted segmentation experiments using the 239 MRI scans of the anatomical tracings of lesions after stroke (ATLAS) dataset to evaluate the performance and effectiveness of our method. The experimental results show that the DSC score of the proposed DPAC-UNet network is 6% higher than the single-path attention U-Net. It is also higher than the existing segmentation methods of the related literature. Therefore, our method demonstrates powerful abilities in the application of stroke lesion segmentation.


2000 ◽  
Author(s):  
Masaaki Okuma ◽  
Ward Heylen ◽  
Hisayoshi Matsuoka ◽  
Paul Sas

Abstract This paper presents the result of using an experimental spatial matrix identification method to predict the dynamics of a frame structure under a different boundary condition. The single-input-multiple-output frequency response functions of the test structure under the free-free boundary condition are measured by hammer testing. Using the FRFs, a set of spatial matrices is determined to represent its structural dynamic characteristics by the method. Then, using the identified spatial matrices, the dynamic characteristics of the test structure under the boundary condition of clamping 4 points is predicted. The prediction is practically accurate. The result of the prediction demonstrates that the spatial matrices identified by the method can be used for structural modification and substructure synthesis in the field of computer aided mechanical engineering.


2001 ◽  
Vol 123 (3) ◽  
pp. 390-394 ◽  
Author(s):  
Masaaki Okuma ◽  
Ward Heylen ◽  
Hisayoshi Matsuoka ◽  
Paul Sas

This paper presents the results of using an experimental spatial matrix identification method to predict the dynamics of a frame structure under various boundary conditions. The single-input-multiple-output frequency response functions (FRFs) of the test structure under the free-free boundary condition are measured by hammer testing. Using the FRFs, a set of spatial matrices is constructed in order to represent the structural dynamic characteristics of the system by the new method. Using the spatial matrices, the dynamic characteristics of the test structure under the boundary condition of clamping 4 points is predicted. The prediction is adequately accurate for practical application. The results of the prediction demonstrate that the spatial matrices identified by this method can be used for structural modification and substructure synthesis in the field of computer-aided mechanical engineering.


1994 ◽  
Vol 03 (02) ◽  
pp. 157-185 ◽  
Author(s):  
DERSHUNG YANG ◽  
LARRY A. RENDELL ◽  
JULIE L. WEBSTER ◽  
DORIS S. SHAW ◽  
JAMES H. GARRETT

A new neural network called AUGURS is designed to assist a user of a Computer-Aided Design system in utilizing standard graphic symbols. With AUGURS, the CAD user can avoid searching for standard symbols in a large library and rely on AUGURS to automatically retrieve those symbols resembling the user’s drawing. More specifically, AUGURS inputs a bitmap image normalized with respect to location, size, and orientation, and outputs a list of standard symbols ranked by its assessment of the similarity between the symbol and the input image. Only the top ranked symbols are presented to the user for selection. AUGURS encodes geometric knowledge into its network structure and carefully balances its discriminant power and noise tolerance. The encoded knowledge enables AUGURS to learn reasonably well despite the limited number of training examples, the most serious challenge for the CAD domain. We have compared AUGURS with the Zipcode Net, a traditional layered feed-forward network with an unconstrained structure, and a network that inputs either Zernike or pseudo-Zernike moments. The experimental results conclude that AUGURS can achieve the best recognition performance among all networks being compared with reasonable recognition and learning efficiency.


2021 ◽  
Author(s):  
Yuchen Yue ◽  
Hua Li ◽  
Jianhua Luo

Establishing structured reconstruction models and efficient reconstruction algorithms according to practical engineering needs is of great concern in the applied research of Compressed Sensing (CS) theory. Targeting problems during high-speed video capture, the paper proposes a set of video CS scheme based on intra-frame and inter-frame constraints and Genetic Algorithm (GA). Firstly, it employs the intra-frame and inter-frame correlation of the video signals as the priori information, creating a video CS reconstruction model on the basis of temporal and spatial similarity constraints. Then it utilizes overcomplete dictionary of Ridgelet to divide the video frames into three structures, smooth, single-oriented, or multijointed. Video frames cluster according to the structure using Affinity Propagation (AP) algorithm, and finally clusters are reconstructed using evolutionary algorithm. It is proved efficient in terms of reconstruction result in the experiment.


Author(s):  
V. V. Moskalenko ◽  
M. O. Zaretsky ◽  
A. S. Moskalenko ◽  
A. O. Panych ◽  
V. V. Lysyuk

Context. A model and training method for observational context classification in CCTV sewer inspection vide frames was developed and researched. The object of research is the process of detection of temporal-spatial context during CCTV sewer inspections. The subjects of the research are machine learning model and training method for classification analysis of CCTV video sequences under the limited and imbalanced training dataset constraint. Objective. Stated research goal is to develop an efficient context classifier model and training algorithm for CCTV sewer inspection video frames under the constraint of the limited and imbalanced labeled training set. Methods. The four-stage training algorithm of the classifier is proposed. The first stage involves training with soft triplet loss and regularisation component which penalises the network’s binary output code rounding error. The next stage is needed to determine the binary code for each class according to the principles of error-correcting output codes with accounting for intra- and interclass relationship. The resulting reference vector for each class is then used as a sample label for the future training with Joint Binary Cross Entropy Loss. The last machine learning stage is related to decision rule parameter optimization according to the information criteria to determine the boundaries of deviation of binary representation of observations for each class from the corresponding reference vector. A 2D convolutional frame feature extractor combined with the temporal network for inter-frame dependency analysis is considered. Variants with 1D Dilated Regular Convolutional Network, 1D Dilated Causal Convolutional Network, LSTM Network, GRU Network are considered. Model efficiency comparison is made on the basis of micro averaged F1 score calculated on the test dataset. Results. Results obtained on the dataset provided by Ace Pipe Cleaning, Inc confirm the suitability of the model and method for practical use, the resulting accuracy equals 92%. Comparison of the training outcome with the proposed method against the conventional methods indicated a 4% advantage in micro averaged F1 score. Further analysis of the confusion matrix had shown that the most significant increase in accuracy in comparison with the conventional methods is achieved for complex classes which combine both camera orientation and the sewer pipe construction features. Conclusions. The scientific novelty of the work lies in the new models and methods of classification analysis of the temporalspatial context when automating CCTV sewer inspections under imbalanced and limited training dataset conditions. Training results obtained with the proposed method were compared with the results obtained with the conventional method. The proposed method showed 4% advantage in micro averaged F1 score. It had been empirically proven that the use of the regular convolutional temporal network architecture is the most efficient in utilizing inter-frame dependencies. Resulting accuracy is suitable for practical use, as the additional error correction can be made by using the odometer data.


Author(s):  
Debanjan Konar ◽  
Suman Kalyan Kar

This chapter proposes a quantum multi-layer neural network (QMLNN) architecture suitable for handwritten character recognition in real time, assisted by quantum backpropagation of errors calculated from the quantum-inspired fuzziness measure of network output states. It is composed of three second-order neighborhood-topology-based inter-connected layers of neurons represented by qubits known as input, hidden, and output layers. The QMLNN architecture is a feed forward network with standard quantum backpropagation algorithm for the adjustment of its weighted interconnection. QMLNN self-organizes the quantum fuzzy input image information by means of the quantum backpropagating errors at the intermediate and output layers of the architecture. The interconnection weights are described using rotation gates. After the network is stabilized, a quantum observation at the output layer destroys the superposition of quantum states in order to obtain true binary outputs.


Sign in / Sign up

Export Citation Format

Share Document