word relatedness
Recently Published Documents


TOTAL DOCUMENTS

19
(FIVE YEARS 1)

H-INDEX

7
(FIVE YEARS 0)

Author(s):  
Keisuke Inohara ◽  
Akira Utsumi

AbstractWe present a new Japanese dataset, Japanese Word Similarity and Association Norm (JWSAN), comprising human rating scores of similarity and association for 2145 word pairs, with a clear distinction between word similarity and word association. Computational models of human semantic memory or mental lexicon, such as distributed semantic models, must predict not only association but also similarity. People can distinguish between word similarity and association. However, although the SimLex-999 dataset is publicly available for English, there is no Japanese similarity dataset with a clear distinction between the two types of word relatedness. JWSAN is the first large Japanese dataset with similarity and association ratings, containing noun, verb, and adjective word pairs. It is also characterized by data collection from a sufficient number of age- and-gender-controlled assessors, with similarity and association ratings obtained via a web-based survey conducted of 6450 native speakers of Japanese. In addition, the effects of the gender and age of the raters were also examined; these factors were only given scant consideration in the past. This dataset can act as a benchmark for improving distributed semantic models in Japanese.


Author(s):  
Subhadra Dutta ◽  
Eric M. O’Rourke

Natural language processing (NLP) is the field of decoding human written language. This chapter responds to the growing interest in using machine learning–based NLP approaches for analyzing open-ended employee survey responses. These techniques address scalability and the ability to provide real-time insights to make qualitative data collection equally or more desirable in organizations. The chapter walks through the evolution of text analytics in industrial–organizational psychology and discusses relevant supervised and unsupervised machine learning NLP methods for survey text data, such as latent Dirichlet allocation, latent semantic analysis, sentiment analysis, word relatedness methods, and so on. The chapter also lays out preprocessing techniques and the trade-offs of growing NLP capabilities internally versus externally, points the readers to available resources, and ends with discussing implications and future directions of these approaches.


2019 ◽  
Vol 12 (2) ◽  
pp. 91
Author(s):  
Dinda Sigmawaty ◽  
Mirna Adriani

Queries and ranking with temporal aspects gain significant attention in field of Information Retrieval. While searching for articles published over time, the relevant documents usually occur in certain temporal patterns. Given a query that is implicitly time sensitive, we develop a temporal ranking using the important times of query by drawing from the distribution of query trend relatedness over time. We also combine the model with Dual Embedding Space Model (DESM) in the temporal model according to document timestamp. We apply our model using three temporal word embeddings algorithms to learn relatedness of words from news archive in Bahasa Indonesia: (1) QT-W2V-Rank using Word2Vec (2) QT-OW2V-Rank using OrthoTrans-Word2Vec (3) QT-DBE-Rank using Dynamic Bernoulli Embeddings. The highest score was achieved with static word embeddings learned separately over time, called QT-W2V-Rank, which is 66% in average precision and 68% in early precision. Furthermore, studies of different characteristics of temporal topics showed that QT-W2V-Rank is also more effective in capturing temporal patterns such as spikes, periodicity, and seasonality than the baselines.


IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 55261-55268
Author(s):  
Yongjing Yin ◽  
Jiali Zeng ◽  
Hongji Wang ◽  
Keqing Wu ◽  
Bin Luo ◽  
...  

2018 ◽  
Vol 34 (3) ◽  
pp. 939-966
Author(s):  
Rashadul Hasan Rakib ◽  
Aminul Islam ◽  
Evangelos Milios
Keyword(s):  

2017 ◽  
Author(s):  
Guy D. Rosin ◽  
Eytan Adar ◽  
Kira Radinsky
Keyword(s):  

2016 ◽  
Vol 216 ◽  
pp. 816-843 ◽  
Author(s):  
Mohamed Ben Aouicha ◽  
Mohamed Ali Hadj Taieb ◽  
Abdelmajid Ben Hamadou
Keyword(s):  

Author(s):  
Jie Mei ◽  
Xinxin Kou ◽  
Zhimin Yao ◽  
Andrew Rau-Chaplin ◽  
Aminul Islam ◽  
...  

2014 ◽  
Vol 40 (3) ◽  
pp. 539-562 ◽  
Author(s):  
Ahmed Hassan ◽  
Amjad Abu-Jbara ◽  
Wanchen Lu ◽  
Dragomir Radev

Automatically identifying the sentiment polarity of words is a very important task that has been used as the essential building block of many natural language processing systems such as text classification, text filtering, product review analysis, survey response analysis, and on-line discussion mining. We propose a method for identifying the sentiment polarity of words that applies a Markov random walk model to a large word relatedness graph, and produces a polarity estimate for any given word. The model can accurately and quickly assign a polarity sign and magnitude to any word. It can be used both in a semi-supervised setting where a training set of labeled words is used, and in a weakly supervised setting where only a handful of seed words is used to define the two polarity classes. The method is experimentally tested using a gold standard set of positive and negative words from the General Inquirer lexicon. We also show how our method can be used for three-way classification which identifies neutral words in addition to positive and negative words. Our experiments show that the proposed method outperforms the state-of-the-art methods in the semi-supervised setting and is comparable to the best reported values in the weakly supervised setting. In addition, the proposed method is faster and does not need a large corpus. We also present extensions of our methods for identifying the polarity of foreign words and out-of-vocabulary words.


2014 ◽  
Author(s):  
Nitish Aggarwal ◽  
Kartik Asooja ◽  
Paul Buitelaar
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document