scholarly journals Changes to itk::PathToImageFilter.

2006 ◽  
Author(s):  
Zachary Pincus

I have made some changes to make itk::PathToImageFilter to make it a usable base class. There were some problems with the CVS version, including: (1) this ase classhad actual filter functionality, (2) the mechanism for specifying the size, spacing, origin, etc., of the output image was incomplete, contrary to ITK convention, and hard to extend, and (3) the documentation specified that the filter would calculate the size of the output image from the path, but it did not.I solved these problems by (1) moving the filter functionality to itk::TracePathOnImageFilter, (2) requiring that filter->GetOutput()->Set…(…) should be used to set the information for the output image, and (3) providing methods to calculate the correct image region a path inhabits, and a boolean flag to specify whether that calculation should be done (default true).

2012 ◽  
Vol 31 (6) ◽  
pp. 1628-1630
Author(s):  
Jia-jia OU ◽  
Bi-ye CAI ◽  
Bing XIONG ◽  
Feng LI

Author(s):  
Kholilatul Wardani ◽  
Aditya Kurniawan

 The ROI (Region of Interest) Image Quality Assessment is an image quality assessment model based on the SSI (Structural Similarity Index) index used in the specific image region desired to be assessed. Output assessmen value used by this image assessment model is 1 which means identical and -1 which means not identical. Assessment model of ROI Quality Assessment in this research is used to measure image quality on Kinect sensor capture result used in Mobile HD Robot after applied Multiple Localized Filtering Technique. The filter is applied to each capture sensor depth result on Kinect, with the aim to eliminate structural noise that occurs in the Kinect sensor. Assessment is done by comparing image quality before filter and after filter applied to certain region. The kinect sensor will be conditioned to capture a square black object measuring 10cm x 10cm perpendicular to a homogeneous background (white with RGB code 255,255,255). The results of kinect sensor data will be taken through EWRF 3022 by visual basic 6.0 program periodically 10 times each session with frequency 1 time per minute. The results of this trial show the same similar index (value 1: identical) in the luminance, contrast, and structural section of the edge region or edge region of the specimen. The value indicates that the Multiple Localized Filtering Technique applied to the noise generated by the Kinect sensor, based on the ROI Image Quality Assessment model has no effect on the image quality generated by the sensor.


1960 ◽  
Vol 48 (11) ◽  
pp. 1842-1846 ◽  
Author(s):  
F. Nicoll ◽  
Alan Sussman

Author(s):  
Lianli Gao ◽  
Pengpeng Zeng ◽  
Jingkuan Song ◽  
Yuan-Fang Li ◽  
Wu Liu ◽  
...  

To date, visual question answering (VQA) (i.e., image QA and video QA) is still a holy grail in vision and language understanding, especially for video QA. Compared with image QA that focuses primarily on understanding the associations between image region-level details and corresponding questions, video QA requires a model to jointly reason across both spatial and long-range temporal structures of a video as well as text to provide an accurate answer. In this paper, we specifically tackle the problem of video QA by proposing a Structured Two-stream Attention network, namely STA, to answer a free-form or open-ended natural language question about the content of a given video. First, we infer rich longrange temporal structures in videos using our structured segment component and encode text features. Then, our structured two-stream attention component simultaneously localizes important visual instance, reduces the influence of background video and focuses on the relevant text. Finally, the structured two-stream fusion component incorporates different segments of query and video aware context representation and infers the answers. Experiments on the large-scale video QA dataset TGIF-QA show that our proposed method significantly surpasses the best counterpart (i.e., with one representation for the video input) by 13.0%, 13.5%, 11.0% and 0.3 for Action, Trans., TrameQA and Count tasks. It also outperforms the best competitor (i.e., with two representations) on the Action, Trans., TrameQA tasks by 4.1%, 4.7%, and 5.1%.


1994 ◽  
Vol 3 (6) ◽  
pp. 868-872 ◽  
Author(s):  
Yian-Leng Chang ◽  
Xiaobo Li
Keyword(s):  

2021 ◽  
pp. 147387162110481
Author(s):  
Haijun Yu ◽  
Shengyang Li

Hyperspectral images (HSIs) have become increasingly prominent as they can maintain the subtle spectral differences of the imaged objects. Designing approaches and tools for analyzing HSIs presents a unique set of challenges due to their high-dimensional characteristics. An improved color visualization approach is proposed in this article to achieve communication between users and HSIs in the field of remote sensing. Under the real-time interactive control and color visualization, this approach can help users intuitively obtain the rich information hidden in original HSIs. Using the dimensionality reduction (DR) method based on band selection, high-dimensional HSIs are reduced to low-dimensional images. Through drop-down boxes, users can freely specify images that participate in the combination of RGB channels of the output image. Users can then interactively and independently set the fusion coefficient of each image within an interface based on concentric circles. At the same time, the output image will be calculated and visualized in real time, and the information it reflects will also be different. In this approach, channel combination and fusion coefficient setting are two independent processes, which allows users to interact more flexibly according to their needs. Furthermore, this approach is also applicable for interactive visualization of other types of multi-layer data.


Sign in / Sign up

Export Citation Format

Share Document