scholarly journals Multiple Instance Choquet Integral for multiresolution sensor fusion

2017 ◽  
Author(s):  
◽  
Xiaoxiao Du

Imagine you are traveling to Columbia, MO for the first time. On your flight to Columbia, the woman sitting next to you recommended a bakery by a large park with a big yellow umbrella outside. After you land, you need directions to the hotel from the airport. Suppose you are driving a rental car, you will need to park your car at a parking lot or a parking structure. After a good night's sleep in the hotel, you may decide to go for a run in the morning on the closest trail and stop by that recommended bakery under a big yellow umbrella. It would be helpful in the course of completing all these tasks to accurately distinguish the proper car route and walking trail, find a parking lot, and pinpoint the yellow umbrella. Satellite imagery and other geo-tagged data such as Open Street Maps provide effective information for this goal. Open Street Maps can provide road information and suggest bakery within a five-mile radius. The yellow umbrella is a distinctive color and, perhaps, is made of a distinctive material that can be identified from a hyperspectral camera. Open Street Maps polygons are tagged with information such as "parking lot" and "sidewalk." All these information can and should be fused to help identify and offer better guidance on the tasks you are completing. Supervised learning methods generally require precise labels for each training data point. It is hard (and probably at an extra cost) to manually go through and label each pixel in the training imagery. GPS coordinates cannot always be fully trusted as a GPS device may only be accurate to the level of several pixels. In many cases, it is practically infeasible to obtain accurate pixel-level training labels to perform fusion for all the imagery and maps available. Besides, the training data may come in a variety of data types, such as imagery or as a 3D point cloud. The imagery may have different resolutions, scales and, even, coordinate systems. Previous fusion methods are generally only limited to data mapped to the same pixel grid, with accurate labels. Furthermore, most fusion methods are restricted to only two sources, even if certain methods, such as pan-sharpening, can deal with different geo-spatial types or data of different resolution. It is, therefore, necessary and important, to come up with a way to perform fusion on multiple sources of imagery and map data, possibly with different resolutions and of different geo-spatial types with consideration of uncertain labels. I propose a Multiple Instance Choquet Integral framework for multi-resolution multisensor fusion with uncertain training labels. The Multiple Instance Choquet Integral (MICI) framework addresses uncertain training labels and performs both classification and regression. Three classifier fusion models, i.e. the noisy-or, min-max, and generalized-mean models, are derived under MICI. The Multi-Resolution Multiple Instance Choquet Integral (MR-MICI) framework is built upon the MICI framework and further addresses multiresolution in the fusion sources in addition to the uncertainty in training labels. For both MICI and MR-MICI, a monotonic normalized fuzzy measure is learned to be used with the Choquet integral to perform two-class classifier fusion given bag-level training labels. An optimization scheme based on the evolutionary algorithm is used to optimize the models proposed. For regression problems where the desired prediction is real-valued, the primary instance assumption is adopted. The algorithms are applied to target detection, regression and scene understanding applications. Experiments are conducted on the fusion of remote sensing data (hyperspectral and LiDAR) over the campus of University of Southern Mississippi - Gulfpark. Clothpanel sub-pixel and super-pixel targets were placed on campus with varying levels of occlusion and the proposed algorithms can successfully detect the targets in the scene. A semi-supervised approach is developed to automatically generate training labels based on data from Google Maps, Google Earth and Open Street Map. Based on such training labels with uncertainty, the proposed algorithms can also identify materials on campus for scene understanding, such as road, buildings, sidewalks, etc. In addition, the algorithms are used for weed detection and real-valued crop yield prediction experiments based on remote sensing data that can provide information for agricultural applications.

Author(s):  
C. Xiao ◽  
R. Qin ◽  
X. Huang ◽  
J. Li

<p><strong>Abstract.</strong> Individual tree detection and counting are critical for the forest inventory management. In almost all of these methods that based on remote sensing data, the treetop detection is the most important and essential part. However, due to the diversities of the tree attributes, such as crown size and branch distribution, it is hard to find a universal treetop detector and most of the current detectors need to be carefully designed based on the heuristic or prior knowledge. Hence, to find an efficient and versatile detector, we apply deep neural network to extract and learn the high-level semantic treetop features. In contrast to using manually labelled training data, we innovatively train the network with the pseudo ones that come from the result of the conventional non-supervised treetop detectors which may be not robust in different scenarios. In this study, we use multi-view high-resolution satellite imagery derived DSM (Digital Surface Model) and multispectral orthophoto as data and apply the top-hat by reconstruction (THR) operation to find treetops as the pseudo labels. The FCN (fully convolutional network) is adopted as a pixel-level classification network to segment the input image into treetops and non-treetops pixels. Our experiments show that the FCN based treetop detector is able to achieve a detection accuracy of 99.7<span class="thinspace"></span>% at the prairie area and 66.3<span class="thinspace"></span>% at the complicated town area which shows better performance than THR in the various scenarios. This study demonstrates that without manual labels, the FCN treetop detector can be trained by the pseudo labels that generated using the non-supervised detector and achieve better and robust results in different scenarios.</p>


2021 ◽  
Vol 970 (4) ◽  
pp. 54-64
Author(s):  
S.A. Yamashkin ◽  
A.A. Yamashkin ◽  
V.V. Zanozin ◽  
A.N. Barmin

The authors propose their solving the task of improving the accuracy of remote sensing data classification under conditions of labeled data scarcity through using a geosystem approach that involves analyzing the genetic uniformity of various-scale territorially adjacent formations and hierarchical levels. The advantage of the proposed GeoSystemNet model is a great number of freedom degrees, which enables flexible configuration of the model based on the task being solved. Testing the GeoSystemNet model for classifying the EuroSAT set, algorithmically expanded from the perspective of the geosystem approach, showed the possibility of increasing the classification accuracy under the conditions of training data scarcity within 9 %, as well as approaching the accuracy of the deep ResNet50 and GoogleNet models. The authors note that the use of the geosystem approach according to the methodology proposed in the article for solving the above-mentioned problem requires an individual project approach to the formation of the data for analysis.


2020 ◽  
Vol 86 (6) ◽  
pp. 383-392
Author(s):  
Liguo Wang ◽  
Xiaoyi Wang ◽  
Qunming Wang

Spatiotemporal fusion is an important technique to solve the problem of incompatibility between the temporal and spatial resolution of remote sensing data. In this article, we studied the fusion of Landsat data with fine spatial resolution but coarse temporal resolution and Moderate Resolution Imaging Spectroradiometer (MODIS) data with coarse spatial resolution but fine temporal resolution. The goal of fusion is to produce time-series data with the fine spatial resolution of Landsat and the fine temporal resolution of MODIS. In recent years, learning-based spatiotemporal fusion methods, in particular the sparse representation-based spatiotemporal reflectance fusion model (SPSTFM), have gained increasing attention because of their great restoration ability for heterogeneous landscapes. However, remote sensing data from different sensors differ greatly on spatial resolution, which limits the performance of the spatiotemporal fusion methods (including SPSTFM) to some extent. In order to increase the accuracy of spatiotemporal fusion, in this article we used existing 250-m MODISbands (i.e., red and near-infrared bands) to downscale the observed 500-m MODIS bands to 250 m before SPTSFM-based fusion of MODIS and Landsat data. The experimental results show that the fusion accuracy of SPTSFM is increased when using 250-m MODIS data, and the accuracy of SPSTFM coupled with 250-m MODIS data is greater than the compared benchmark methods.


2018 ◽  
Vol 5 (3) ◽  
pp. 247-258 ◽  
Author(s):  
Merry Crowson ◽  
Eleanor Warren‐Thomas ◽  
Jane K. Hill ◽  
Bambang Hariyadi ◽  
Fahmuddin Agus ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document