Information Integration Seedling for Data Integration and exploitation System that Learns (DIESEL)

2009 ◽  
Author(s):  
Pedro M. Domingos
2014 ◽  
Vol 667 ◽  
pp. 112-117
Author(s):  
Yun Sheng Huang ◽  
Xian Feng Xu ◽  
Wen Bo Zhao

Marketing and Distribution Information Integration Platform (MDIIP) is the key point of informatization work for power companies. On this basis, the pater proposes a solution including overall architecture, function composition and system features etc. for Power Consumption Dispatching Management System (PCDMS) based on Transmission, Distribution and Marketing Data Integration Platform (TDMDIP). The system realizes the automatical combination of transmission, distribution and marketing model (TDM Model) and longitudinal interconnection of network topology, and effectively interacts real-time monitoring analysis features with daily operation management businesses. It not only expands the dispatcher’s view, but also improves intellectualization level of dispatching daily business.


2013 ◽  
Vol 347-350 ◽  
pp. 1935-1939
Author(s):  
Guo Feng Yang ◽  
Hong Yan Zheng ◽  
Hong Ouyang ◽  
Jia Kui Zhao ◽  
Ting Shun Li ◽  
...  

Power grid is a complicated and huge system, many links are involved during its life-cycle, and serious information exchange and sharing problems can be found between these links, they restrict the level of power grids informatization construction. With the presentation of building information modeling (BIM) and the application of IFC standard, the research for building different life cycle stages of information exchange and sharing is strongly pushed forward as well as gained good results. The successful application of BIM in construction business provides great chance for grid whole life data information integration and sharing. After intensively studying every links business and data characteristics of power grid, cooperated with the BIM theory, this paper comes up with a method that applies BIM theory into power system data integration and sharing, with an electric power data integration platform construction scheme based on BIM, by which grid full life cycle of data integration and sharing problem can be successfully solved.


Author(s):  
Manjunath Ramachandra

In an information supply chain, data gets pumped over the web from several distinct sources in various forms calling for efforts towards data integration. The integration happens at various levels by means of different algorithms. This chapter provides a framework for the interoperability of the data from different sources. The use case of integration in data warehouses is provided.


Author(s):  
Juan M. Gómez ◽  
Ricardo Colomo ◽  
Marcos Ruano ◽  
Ángel García

Technological advances in high-throughput techniques and efficient data gathering methods, coupled computational biology efforts, have resulted in a vast amount of life science data often available in distributed and heterogeneous repositories. These repositories contain information such as sequence and structure data, annotations for biological data, results of complex computations, genetic sequences and multiple bio-datasets. However, the heterogeneity of these data, have created a need for research in resource integration and platform independent processing of investigative queries, involving heterogeneous data sources. When processing huge amounts of data, information integration is one of the most critical issues, because it’s crucial to preserve the intrinsic semantics of all the merged data sources. This integration would allow the proper organization of data, fostering the analysis and access the information to accomplish critical tasks, such as the processing of micro-array data to study protein function and medical researches in making detailed studies of protein structures to facilitate drug design (Ignacimuthu, 2005). Furthermore, DNA micro-array research community urgently requires technology to allow up-to-date micro-array data information to be found, accessed and delivered in a secure framework (Sinnot, 2007). Several research disciplines, such as Bioinformatics, where information integration is critical, could benefit from harnessing the potential of a new approach: the Semantic Web (SW). The SW term was coined by Berners-Lee, Hendler and Lassila (2001) to describe the evolution of a Web that consisted of largely documents for humans to read towards a new paradigm that included data and information for computers to manipulate. The SW is about adding machine-understandable and machine-processable metadata to Web resource through its key-enabling technology: ontologies (Fensel, 2002). Ontologies are a formal explicit and shared specification of a conceptualization. The SW was conceived as a way to solve the need for data integration on the Web. This article expounds SAMIDI, a Semantics-based Architecture for Micro-array Information and Data Integration. The most remarkable innovation offered by SAMIDI is the use of semantics as a tool for leveraging different vocabularies and terminologies and foster integration. SAMIDI is composed of a methodology for the unification of heterogeneous data sources from the analysis of the requirements of the unified data set and a software architecture.


In information integration systems, duplicate records bring problems in data processing and analysis. To represent the similarity between two records from different data sources with different schema, the optimal bipartite graph matching is adopted on the attributes of them, and the similarity is measured as the weight of such matching. Based on similarity estimation, the basic idea in this chapter is to estimate the range of the records similarity and to determine whether they are duplicate records according to the estimation. When data integration is performed on XML data, there are many problems because of the flexibility of XML. One of the current implementations is to use Data Exchange to carry out the above operations. This chapter proposes the concept of quality assurance mechanisms besides the data integrity and reliability.


2013 ◽  
Vol 310 ◽  
pp. 605-608 ◽  
Author(s):  
Xiao Bin Wang ◽  
Qing Jun Wang ◽  
Ming Yu Bao

Modern enterprises have established many information management systems based on management of enterprise information. But any of the systems can only manage information of a department, and even on different task directions in the same department there are many information management systems. Between these systems, it is hard to realize mutual contact or data sharing, not even coordinated work. How to establish an information integration mechanism to make these systems share data for coordinated work and values as 1+1>2 becomes the problem to be solved by modern enterprises in an earnest status. As an effective method to reach mutual communication between data of the isomeric systems, the data integration system can shield off the isomerism of systems it covers and unify the data modes of these systems. Then, mode shifting is made between different systems to make these systems have the same mode on the data integration layer, to provide convenience for mutual communication between these systems, to reduce the coupling of the whole system and to provide operation function of the enterprise.


Author(s):  
Seán O’Riain ◽  
Andreas Harth ◽  
Edward Curry

With increased dependence on efficient use and inclusion of diverse corporate and Web based data sources for business information analysis, financial information providers will increasingly need agile information integration capabilities. Linked Data is a set of technologies and best practices that provide such a level of agility for information integration, access, and use. Current approaches struggle to cope with multiple data sources inclusion in near real-time, and have looked to Semantic Web technologies for assistance with infrastructure access, and dealing with multiple data formats and their vocabularies. This chapter discusses the challenges of financial data integration, provides the component architecture of Web enabled financial data integration and outlines the emergence of a financial ecosystem, based upon existing Web standards usage. Introductions to Semantic Web technologies are given, and the chapter supports this with insight and discussion gathered from multiple financial services use case implementations. Finally, best practice for integrating Web data based on the Linked Data principles and emergent areas are described.


2012 ◽  
Vol 532-533 ◽  
pp. 555-559
Author(s):  
Xi Li ◽  
Hui Sheng Ye ◽  
Qi Wang ◽  
Jin Liu ◽  
Xiao Xiao Liu

The fact that development and construction of electric energy information systems are based on different time periods and business needs results in the distribution of electric energy information in geographically dispersed databases, but in many cases there is need to browse, query, analyze and do other operations on all of the electrical energy information. Therefore, the integration of electric energy information is a key issue in the current power marketing work. This paper, based on current popular data integration methods, designs a distributed electric energy information integration and management model utilizing Oracle products to meet different business needs. It also gives a detailed description of the model structure and various components therein. Finally a brief analysis is given on the advantages and significance of this proposed platform.


2001 ◽  
Vol 10 (03) ◽  
pp. 237-271 ◽  
Author(s):  
DIEGO CALVANESE ◽  
GIUSEPPE DE GIACOMO ◽  
MAURIZIO LENZERINI ◽  
DANIELE NARDI ◽  
RICCARDO ROSATI

Information integration is one of the most important aspects of a Data Warehouse. When data passes from the sources of the application-oriented operational environment to the Data Warehouse, possible inconsistencies and redundancies should be resolved, so that the warehouse is able to provide an integrated and reconciled view of data of the organization. We describe a novel approach to data integration in Data Warehousing. Our approach is based on a conceptual representation of the Data Warehouse application domain, and follows the so-called local-as-view paradigm: both source and Data Warehouse relations are defined as views over the conceptual model. We propose a technique for declaratively specifying suitable reconciliation correspondences to be used in order to solve conflicts among data in different sources. The main goal of the method is to support the design of mediators that materialize the data in the Data Warehouse relations. Starting from the specification of one such relation as a query over the conceptual model, a rewriting algorithm reformulates the query in terms of both the source relations and the reconciliation correspondences, thus obtaining a correct specification of how to load the data in the materialized view.


Sign in / Sign up

Export Citation Format

Share Document