scholarly journals XML schema validation using parsing expression grammars

Author(s):  
Kimio Kuramitsu ◽  
Shin'ya Yamaguchi

Schema validation is an integral part of reliable information exchange on the Web. However, implementing an efficient schema validation tool is not easy. We highlight the use of parsing expression grammars (PEGs), a recognition-based foundation for describing syntax, and apply it to the XML/DTD validation. This paper shows that structural schema constraints in document type definitions (DTDs) can be validated by the converted PEGs with the linear time and constant space consumption. We study the performance of several existing PEG-based tools, and then confirm that the converted PEGs achieve a practical and even competitive level of performance under existing standard XML/DTD validators.

2015 ◽  
Author(s):  
Kimio Kuramitsu ◽  
Shin'ya Yamaguchi

Schema validation is an integral part of reliable information exchange on the Web. However, implementing an efficient schema validation tool is not easy. We highlight the use of parsing expression grammars (PEGs), a recognition-based foundation for describing syntax, and apply it to the XML/DTD validation. This paper shows that structural schema constraints in document type definitions (DTDs) can be validated by the converted PEGs with the linear time and constant space consumption. We study the performance of several existing PEG-based tools, and then confirm that the converted PEGs achieve a practical and even competitive level of performance under existing standard XML/DTD validators.


2013 ◽  
Vol 10 (1) ◽  
pp. 79-104
Author(s):  
Guillem Rull ◽  
Carles Farré ◽  
Ernest Teniente ◽  
Toni Urpí

With the emergence of the Web and the wide use of XML for representing data, the ability to map not only flat relational but also nested data has become crucial. The design of schema mappings is a semi-automatic process. A human designer is needed to guide the process, choose among mapping candidates, and successively refine the mapping. The designer needs a way to figure out whether the mapping is what was intended. Our approach to mapping validation allows the designer to check whether the mapping satisfies certain desirable properties. In this paper, we focus on the validation of mappings between nested relational schemas, in which the mapping assertions are either inclusions or equalities of nested queries. We focus on the nested relational setting since most XML?s Document Type Definitions (DTDs) can be represented in this model. We perform the validation by reasoning on the schemas and mapping definition. We take into account the integrity constraints defined on both the source and target schema. We consider constraints and mapping?s queries which may contain arithmetic comparisons and negations. This class of mapping scenarios is significantly more expressive than the ones addressed by previous work on nested relational mapping validation. We encode the given mapping scenario into a single flat database schema, so we can take advantage of our previous work on validating flat relational mappings, and reformulate each desirable property check as a query satisfiability problem.


Author(s):  
Chiara Degano

This chapter focuses on computer mediated communication from a linguistic perspective, exploring aspects of textuality which have been impacted by the pervasive spread of the hypertext. Central features in the construction of texts are the notions of cohesion and coherence, originally tailored on linear time-based modes of communication, where both the elements and their sequentiality – fully controlled by the author – contribute to meaning making. In light of the disruption of linear sequentiality brought by the space-based logic of the hypertext, this chapter aims to understand how cohesion and coherence work in the website environment, with specific regard to genres characterised by an argumentative drive, which potentially suffer more than other text types from the loss of the author’s control on the linear dispositio of arguments. The analysis identifies different patterns for the construction of cohesion and coherence in argumentative websites, which accommodate traditional standards of textuality into the new environment.


Author(s):  
Béatrice Bouchou ◽  
Denio Duarte ◽  
Mírian Halfeld Ferrari ◽  
Martin A. Musicante

The XML Messaging Protocol, a part of the Web service protocol stack, is responsible for encoding messages in a common XML format (or type), so that they can be understood at either end of a network connection. The evolution of an XML type may be required in order to reflect new communication needs, materialized by slightly different XML messages. For instance, due to a service evolution, it might be interesting to extend a type in order to allow the reception of more information, when it is available, instead of always disregarding it. The authors’ proposal consists in a conservative XML schema evolution. The framework is as follows: administrators enter updates performed on a valid XML document in order to specify new documents expected to be valid, and the system computes new types accepting both such documents and previously valid ones. Changing the type is mainly changing regular expressions that define element content models. They present the algorithm that implements this approach, its properties and experimental results.


2013 ◽  
Vol 07 (03) ◽  
pp. 237-255 ◽  
Author(s):  
CRISTOBAL VERGARA-NIEDERMAYR ◽  
FUSHENG WANG ◽  
TONY PAN ◽  
TAHSIN KURC ◽  
JOEL SALTZ

XML is ubiquitously used as an information exchange platform for web-based applications in healthcare, life sciences, and many other domains. Proliferating XML data are now managed through latest native XML database technologies. XML data sources conforming to common XML schemas could be shared and integrated with syntactic interoperability. Semantic interoperability can be achieved through semantic annotations of data models using common data elements linked to concepts from ontologies. In this paper, we present a framework and software system to support the development of semantic interoperable XML based data sources that can be shared through a Grid infrastructure. We also present our work on supporting semantic validated XML data through semantic annotations for XML Schema, semantic validation and semantic authoring of XML data. We demonstrate the use of the system for a biomedical database of medical image annotations and markups.


Author(s):  
Norman Walsh

DocBook is a general purpose XML schema particularly well suited to books and papers about computer hardware and software (though it is by no means limited to these applications). DocBook has been under active maintenance for more than 20 years,; it began life as an SGML document type definition. Because it is a large and robust schema, and because its main structures correspond to the general notion of what constitutes a “book,” DocBook has been adopted by a large and growing community of authors writing books of all kinds. DocBook is supported “out of the box” by a number of commercial tools, and there is rapidly expanding support for it in a number of free software environments. These features have combined to make DocBook a generally easy to understand, widely useful, and very popular schema. Dozens of organizations are using DocBook for millions of pages of documentation, in various print and online formats, worldwide.


2018 ◽  
Vol 26 (1) ◽  
pp. 1-6 ◽  
Author(s):  
Mathias Glatz ◽  
Hermann Maurer ◽  
Muhammad Tanvir Azfal
Keyword(s):  

Author(s):  
Krzysztof Juszczyszyn

The World Wide Web (WWW) is a global, ubiquitous, and fundamentally dynamic environment for information exchange and processing. By connecting vast numbers of individuals, the Web enables creation of virtual communities, and during the last 10 years, became a universal collaboration infrastructure. The so-called Semantic Web, a concept proposed by Tim Berners-Lee, is a new WWW architecture that enhances content with formal semantics (Berners-Lee, Hendler, & Lassila, 2001). Hence, the Web content is made suitable for machine processing (i.e., it is described by the associated metadata), as opposed to HTML documents available only for human consumption. Languages such as Resource Description Framework (RDF) and Ontology Web Language (OWL) along with well-known XML are used for description of Web resources. In other words, the Semantic Web is a vision of the future Web in which information is given explicit meaning. This will enable autonomous software agents to reason about Web content and produce intelligent responses to events (Staab, 2002). The ultimate goal of the next generation’s Web is to support the creation of virtual communities which will be composed of software agents and humans cooperating within the same environment. Sharing knowledge within such a community requires a shared conceptual vocabularies—ontologies, which represent the formal common agreement about the meaning of data (Gomez-Perez & Corcho, 2002). Artificial intelligence defines ontologies as explicit, formal specification of a shared conceptualization (Studer, Benjamins, & Fensel, 1998). In this case, a conceptualization stands for an abstract model of some concept from the real world; explicit means that the type of concept used is explicitly defined. Formal refers to the fact that an ontology should be machine readable; and finally shared means that ontology expresses knowledge that is accepted by all the subjects. In short, an ontology defines the terms used to describe and represent an area of knowledge. However, the shared ontologies must be first constructed by using information from many sources which may be of arbitrary quality. Thus, it is necessary to find a way to seamlessly combine the knowledge from many sources, maybe diverse and heterogeneous. The resultant ontologies enable virtual communities and teams to manage and exchange their knowledge. It should be noted, that the word ontology has been used to describe notions with different degrees of structure—from taxonomies (e.g., Yahoo hierarchy), metadata schemes (e.g., Dublin Core), to logical theories. The Semantic Web needs ontologies with a significant degree of structure. These should allow the specification of at least the following kinds of things: • Concepts (which identify the classes of things like cars or birds) from many domains of interest • The relationships that can exist among concepts • The properties (or attributes) those concepts may have


Author(s):  
JOSÉ BORGES ◽  
MARK LEVENE

In this paper, we study the complexity of a data mining algorithm for extracting patterns from user web navigation data that was proposed in previous work.3 The user web navigation sessions are inferred from log data and modeled as a Markov chain. The chain's higher probability trails correspond to the preferred trails on the web site. The algorithm implements a depth-first search that scans the Markov chain for the high probability trails. We show that the average behaviour of the algorithm is linear time in the number of web pages accessed.


2006 ◽  
Vol 36 (4) ◽  
pp. 327-338 ◽  
Author(s):  
Silvio Antonio Carro ◽  
Jacob Scharcanski

Sign in / Sign up

Export Citation Format

Share Document