scholarly journals Technical Paper Recommendation: A Study in Combining Multiple Information Sources

2001 ◽  
Vol 14 ◽  
pp. 231-252 ◽  
Author(s):  
C. Basu ◽  
H. Hirsh ◽  
W. W. Cohen ◽  
C. Nevill-Manning

The growing need to manage and exploit the proliferation of online data sources is opening up new opportunities for bringing people closer to the resources they need. For instance, consider a recommendation service through which researchers can receive daily pointers to journal papers in their fields of interest. We survey some of the known approaches to the problem of technical paper recommendation and ask how they can be extended to deal with multiple information sources. More specifically, we focus on a variant of this problem -- recommending conference paper submissions to reviewing committee members -- which offers us a testbed to try different approaches. Using WHIRL -- an information integration system -- we are able to implement different recommendation algorithms derived from information retrieval principles. We also use a novel autonomous procedure for gathering reviewer interest information from the Web. We evaluate our approach and compare it to other methods using preference data provided by members of the AAAI-98 conference reviewing committee along with data about the actual submissions.

2011 ◽  
pp. 972-985
Author(s):  
Ákos Hajnal ◽  
Tamás Kifor ◽  
Gergely Lukácsy ◽  
László Z. Varga

More and more systems provide data through web service interfaces and these data have to be integrated with the legacy relational databases of the enterprise. The integration is usually done with enterprise information integration systems which provide a uniform query language to all information sources, therefore the XML data sources of Web services having a procedural access interface have to be matched with relational data sources having a database interface. In this chapter the authors provide a solution to this problem by describing the Web service wrapper component of the SINTAGMA Enterprise Information Integration system. They demonstrate Web services as XML data sources in enterprise information integration by showing how the web service wrapper component integrates XML data of Web services in the application domain of digital libraries.


2000 ◽  
Vol 09 (04) ◽  
pp. 383-401 ◽  
Author(s):  
FRANÇOIS GOASDOUÉ ◽  
VÉRONIQUE LATTÈS ◽  
MARIE-CHRISTINE ROUSSET

PICSEL is an information integration system over sources that are distributed and possibly heterogeneous. The approach which has been chosen in PICSEL is to define an information server as a knowledge-based mediator in which CARIN is used as the core logical formalism to represent both the domain of application and the contents of information sources relevant to that domain. In this paper, we describe the way the expressive power of the CARIN language is exploited in the PICSEL information integration system, while maintaining the decidability of query answering. We illustrate it on examples coming from the tourism domain, which is the first real case that we have to consider in PICSEL, in collaboration with the travel agency Degriftour. see


2014 ◽  
Vol 533 ◽  
pp. 440-443
Author(s):  
Gang Huang ◽  
Xiu Ying Wu ◽  
Man Yuan

Due to information integration system is a need to focus on different periods independently designed data sources and a unified information system to provide their data to the end user, so it will inevitably encounter data changes over time to bring the knowledge of information contained, the concept will be certain changes in circumstances occur. This paper analyzes the semantic-oriented information integration systems and solutions proposed to consider the full range of semantic information integration problems at different stages of the primary purposes of information integration systems.


2014 ◽  
Vol 6 (1) ◽  
pp. 97-105 ◽  
Author(s):  
Irene Petrou ◽  
Marios Meimaris ◽  
George Papastefanatos

The number of open government initiatives and directives around the globe with focused interest on publishing large amounts of data on the Web as “open” is increasing rapidly in the recent years. Opening up data aims for citizens, scientists and organizations to easily access, discover and exploit the data and consequently to benefit out of them. As a result, there has been an emerging need of integrating and representing those data in transparent and reusable ways, with high degree of interoperability which will further facilitate the discovery of new connections and insights by linking data coming from disperse sources. Statistical data published either by government bodies or by national statistical authorities are used for policy and decision making purposes, as they present important socioeconomic indicators. In this paper, we present a generic methodology describing the basic steps and overall model to publish statistical data coming from tabular data sources or relational databases in the form of Linked Open Data.


Author(s):  
Ákos Hajnal ◽  
Tamás Kifor ◽  
Gergely Lukácsy ◽  
László Z. Varga

More and more systems provide data through web service interfaces and these data have to be integrated with the legacy relational databases of the enterprise. The integration is usually done with enterprise information integration systems which provide a uniform query language to all information sources, therefore the XML data sources of Web services having a procedural access interface have to be matched with relational data sources having a database interface. In this chapter the authors provide a solution to this problem by describing the Web service wrapper component of the SINTAGMA Enterprise Information Integration system. They demonstrate Web services as XML data sources in enterprise information integration by showing how the web service wrapper component integrates XML data of Web services in the application domain of digital libraries.


2013 ◽  
Vol 455 ◽  
pp. 434-437
Author(s):  
Jing Tao Zhou

Master-slave P2P mapping principle proposed in our previous work [ is a semantic P2P mapping paradigm with modularity and loosely coupled characteristics. The intent of this paper is to define a common case study of this paradigm for the semantic information integration. The domain of the case study is a semantic P2P information integration system called SGII[, i.e., system that help in information coordinating and interoperating by orchestrating the content and formalization expression of master-slave P2P mapping between elements from different peer node models which represent the data exposed (shared) by data sources. Furthermore, an illustrative example of master-slave P2P mapping paradigm is given to explain how the mappings are implemented and to demonstrate the paradigm can hence be applied in semantic information integration scenarios.


2020 ◽  
pp. 193896552097358
Author(s):  
Saram Han ◽  
Christopher K. Anderson

As consumers increasingly research and purchase hospitality and travel services online, new research opportunities have become available to hospitality academics. There is a growing interest in understanding the online travel marketplace among hospitality researchers. Although many researchers have attempted to better understand the online travel market through the use of analytical models, experiments, or survey collection, these studies often fail to capture the full complexity of the market. Academics often rely upon survey data or experiments owing to their ease of collection or potentially to the difficulty in assembling online data. In this study, we hope to equip hospitality researchers with the tools and methods to augment their traditional data sources with the readily available data that consumers use to make their travel choices. In this article, we provide a guideline (and Python code) for how to best collect/scrape publicly available online hotel data. We focus on the collection of online data across numerous platforms, including online travel agents, review sites, and hotel brand sites. We outline some exciting possibilities regarding how these data sources might be utilized, as well as discuss some of the caveats that have to be considered when analyzing online data.


Author(s):  
Heiko Paulheim ◽  
Christian Bizer

Linked Data on the Web is either created from structured data sources (such as relational databases), from semi-structured sources (such as Wikipedia), or from unstructured sources (such as text). In the latter two cases, the generated Linked Data will likely be noisy and incomplete. In this paper, we present two algorithms that exploit statistical distributions of properties and types for enhancing the quality of incomplete and noisy Linked Data sets: SDType adds missing type statements, and SDValidate identifies faulty statements. Neither of the algorithms uses external knowledge, i.e., they operate only on the data itself. We evaluate the algorithms on the DBpedia and NELL knowledge bases, showing that they are both accurate as well as scalable. Both algorithms have been used for building the DBpedia 3.9 release: With SDType, 3.4 million missing type statements have been added, while using SDValidate, 13,000 erroneous RDF statements have been removed from the knowledge base.


Author(s):  
Nirmit Singhal ◽  
Amita Goel, ◽  
Nidhi Sengar ◽  
Vasudha Bahl

The world generated 52 times the amount of data in 2010 and 76 times the number of information sources in 2022. The ability to use this data creates enormous opportunities, and in order to make these opportunities a reality, people must use data to solve problems. Unfortunately, in the midst of a global pandemic, when people all over the world seek reliable, trustworthy information about COVID-19 (Coronavirus). Tableau plays a key role in this scenario because it is an extremely powerful tool for quickly visualizing large amounts of data. It has a simple drag-and-drop interface. Beautiful infographics are simple to create and take little time. Tableau works with a wide variety of data sources. COVID-19 (Coronavirus)analytics with Tableau will allow you to create dashboards that will assist you. Tableau is a tool that deals with big data analytics and generates output in a visualization technique, making it more understandable and presentable. Data blending, real-time reporting, and data collaboration are one of its features. Ultimately, this paper provides a clear picture of the growing COVID19 (Coronavirus) data and the tools that can assist more effectively, accurately, and efficiently. Keywords: Data Visualization, Tableau, Data Analysis, Covid-19 analysis, Covid-19 data


Sign in / Sign up

Export Citation Format

Share Document