Research on Measuring Semantic Correlation Based on the Wikipedia Hyperlink Network

2013 ◽  
Vol 1 (3) ◽  
pp. 1-11 ◽  
Author(s):  
Feiyue Ye ◽  
Feng Zhang

As a free online encyclopedia with a large-scale of knowledge coverage, rich semantic information and quick update speed, Wikipedia brings new ideas to measure semantic correlation. In this paper, the authors present a new method for measuring the semantic correlation between words by mining rich semantic information that exists in Wikipedia. Unlike the previous methods that calculate semantic relatedness merely based on the page network or the category network, the authors' method not only takes into account the semantic information of the page network, it also combines the semantic information of the category network and it improves the accuracy of the results. Besides this, the authors analyze and evaluate the algorithm by comparing the calculation results with famous knowledge base (e.g., Hownet) and traditional methods based on Wikipedia on the same test set and prove its superiority.

2010 ◽  
Vol 30 (11) ◽  
pp. 2932-2936
Author(s):  
Ling-zhong ZHAO ◽  
Xue-song WANG ◽  
Jun-yan QIAN ◽  
Guo-yong CAI

Author(s):  
Sheng Zhang ◽  
Qi Luo ◽  
Yukun Feng ◽  
Ke Ding ◽  
Daniela Gifu ◽  
...  

Background: As a known key phrase extraction algorithm, TextRank is an analogue of PageRank algorithm, which relied heavily on the statistics of term frequency in the manner of co-occurrence analysis. Objective: The frequency-based characteristic made it a neck-bottle for performance enhancement, and various improved TextRank algorithms were proposed in the recent years. Most of improvements incorporated semantic information into key phrase extraction algorithm and achieved improvement. Method: In this research, taking both syntactic and semantic information into consideration, we integrated syntactic tree algorithm and word embedding and put forward an algorithm of Word Embedding and Syntactic Information Algorithm (WESIA), which improved the accuracy of the TextRank algorithm. Results: By applying our method on a self-made test set and a public test set, the result implied that the proposed unsupervised key phrase extraction algorithm outperformed the other algorithms to some extent.


2021 ◽  
Vol 13 (11) ◽  
pp. 2220
Author(s):  
Yanbing Bai ◽  
Wenqi Wu ◽  
Zhengxin Yang ◽  
Jinze Yu ◽  
Bo Zhao ◽  
...  

Identifying permanent water and temporary water in flood disasters efficiently has mainly relied on change detection method from multi-temporal remote sensing imageries, but estimating the water type in flood disaster events from only post-flood remote sensing imageries still remains challenging. Research progress in recent years has demonstrated the excellent potential of multi-source data fusion and deep learning algorithms in improving flood detection, while this field has only been studied initially due to the lack of large-scale labelled remote sensing images of flood events. Here, we present new deep learning algorithms and a multi-source data fusion driven flood inundation mapping approach by leveraging a large-scale publicly available Sen1Flood11 dataset consisting of roughly 4831 labelled Sentinel-1 SAR and Sentinel-2 optical imagery gathered from flood events worldwide in recent years. Specifically, we proposed an automatic segmentation method for surface water, permanent water, and temporary water identification, and all tasks share the same convolutional neural network architecture. We utilize focal loss to deal with the class (water/non-water) imbalance problem. Thorough ablation experiments and analysis confirmed the effectiveness of various proposed designs. In comparison experiments, the method proposed in this paper is superior to other classical models. Our model achieves a mean Intersection over Union (mIoU) of 52.99%, Intersection over Union (IoU) of 52.30%, and Overall Accuracy (OA) of 92.81% on the Sen1Flood11 test set. On the Sen1Flood11 Bolivia test set, our model also achieves very high mIoU (47.88%), IoU (76.74%), and OA (95.59%) and shows good generalization ability.


1979 ◽  
Vol 6 (2) ◽  
pp. 70-72
Author(s):  
T. A. Coffelt ◽  
F. S. Wright ◽  
J. L. Steele

Abstract A new method of harvesting and curing breeder's seed peanuts in Virginia was initiated that would 1) reduce the labor requirements, 2) maintain a high level of germination, 3) maintain varietal purity at 100%, and 4) reduce the risk of frost damage. Three possible harvesting and curing methods were studied. The traditional stack-pole method satisfied the latter 3 objectives, but not the first. The windrow-combine method satisfied the first 2 objectives, but not the last 2. The direct harvesting method satisfied all four objectives. The experimental equipment and curing procedures for direct harvesting had been developed but not tested on a large scale for seed harvesting. This method has been used in Virginia to produce breeder's seed of 3 peanut varieties (Florigiant, VA 72R and VA 61R) during five years. Compared to the stackpole method, labor requirements have been reduced, satisfactory levels of germination and varietal purity have been obtained, and the risk of frost damage has been minimized.


2019 ◽  
Vol 35 (14) ◽  
pp. i417-i426 ◽  
Author(s):  
Erin K Molloy ◽  
Tandy Warnow

Abstract Motivation At RECOMB-CG 2018, we presented NJMerge and showed that it could be used within a divide-and-conquer framework to scale computationally intensive methods for species tree estimation to larger datasets. However, NJMerge has two significant limitations: it can fail to return a tree and, when used within the proposed divide-and-conquer framework, has O(n5) running time for datasets with n species. Results Here we present a new method called ‘TreeMerge’ that improves on NJMerge in two ways: it is guaranteed to return a tree and it has dramatically faster running time within the same divide-and-conquer framework—only O(n2) time. We use a simulation study to evaluate TreeMerge in the context of multi-locus species tree estimation with two leading methods, ASTRAL-III and RAxML. We find that the divide-and-conquer framework using TreeMerge has a minor impact on species tree accuracy, dramatically reduces running time, and enables both ASTRAL-III and RAxML to complete on datasets (that they would otherwise fail on), when given 64 GB of memory and 48 h maximum running time. Thus, TreeMerge is a step toward a larger vision of enabling researchers with limited computational resources to perform large-scale species tree estimation, which we call Phylogenomics for All. Availability and implementation TreeMerge is publicly available on Github (http://github.com/ekmolloy/treemerge). Supplementary information Supplementary data are available at Bioinformatics online.


2014 ◽  
Vol 602-605 ◽  
pp. 3363-3366
Author(s):  
Yi Ming Sun ◽  
Chun Lei Han

In order to automatically identify the mobile phones' reviews that the users comment on the mobile phone on the internet and obtain valuable information from the reviews, this paper presents the process of constructing ontology for the mobile phones' reviews and preliminarily establish a domain ontology of the mobile phones' reviews. The ontology construction adopts the Protégé tool and the Seven Steps method of Stanford University research. The ontology can provide convenience for the semantic information mining on Web mobile phones' reviews, and it can provide a new method to effectively mine the use feelings of the phone from a large number of mobile phone users' reviews.


2018 ◽  
Vol 4 (4) ◽  
Author(s):  
Qiang Zhao ◽  
Yang Li ◽  
Zheng Zhang ◽  
Xiaoping Ouyang

The sputtering of graphite due to the bombardment of hydrogen isotopes is crucial to successfully using graphite in the fusion environment. In this work, we use molecular dynamics to simulate the sputtering using the large-scale atomic/molecular massively parallel simulator (lammps). The calculation results show that the peak values of the sputtering yield are between 25 eV and 50 eV. When the incident energy is greater than the energy corresponding to the peak value, a lower carbon sputtering yield is obtained. The temperature that is most likely to sputter is approximately 800 K for hydrogen, deuterium, and tritium. Below the 800 K, the sputtering yields increase with temperature. By contrast, above the 800 K, the yields decrease with increasing temperature. Under the same temperature and incident energy, the sputtering rate of tritium is greater than that of deuterium, which in turn is greater than that of hydrogen. When the incident energy is 25 eV, the sputtering yield at 300 K increases below an incident angle at 30 deg and remains steady after that.


2012 ◽  
Vol 166-169 ◽  
pp. 68-72
Author(s):  
Shu Tang Liu ◽  
Qi Liang Long

A new method tracing the load-deflection equilibrium path of a truss with doubly nonlinearity is proposed. The total global stiffness matrix equation has been formulated in terms of nodal coordinates, iteration formulations has been written through adopting a single control coordinate, so that an new method tracing the load-deflection equilibrium path has been proposed. Analysis results of Star dome truss and Schwedeler dome truss have shown that the proposed method is stable numerically, quick in convergence, high in degree of accuracy and easy in use. The proposed method can be used for large-scale truss structure.


1997 ◽  
Vol 6 (5) ◽  
pp. 547-564 ◽  
Author(s):  
David R. Pratt ◽  
Shirley M. Pratt ◽  
Paul T. Barham ◽  
Randall E. Barker ◽  
Marianne S. Waldrop ◽  
...  

This paper examines the representation of humans in large-scale, networked virtual environments. Previous work done in this field is summarized, and existing problems with rendering, articulating, and networking numerous human figures in real time are explained. We have developed a system that integrates together some well-known solutions along with new ideas. Models with multiple level of details, body-tracking technology and animation libraries to specify joint angles, efficient group representations to describe multiple humans, and hierarchical network protocols have been successfully employed to increase the number of humans represented, system performance, and user interactivity. The resulting system immerses participants effectively and has numerous useful applications.


Sign in / Sign up

Export Citation Format

Share Document