scholarly journals The Glossaryfication Web Service – an automated glossary creation tool to support One Health communication

2021 ◽  
Vol 4 ◽  
Author(s):  
Estibaliz Lopez de Abechuco ◽  
Nazareno Scaccia ◽  
Taras Günther ◽  
Matthias Filter

Efficient communication and collaboration across sectors is an important precondition for true One Health Surveillance (OHS) activities. Despite the overall willingness to embrace the One Health paradigm, it is still challenging to accomplish this in day-to-day practice due to the differences in terminology and interpretation of sector-specific terms. In this sense, simple interventions like the inclusion of integrative glossaries in OHS documents (e.g. reports, research papers and guidelines) would help to reduce misunderstandings and could significantly improve the written communication in OHS. Here, we present the Glossaryfication Web Service that generates a document-specific glossary for any text file provided by the user. The web service automatically adds the available definitions with their corresponding references for the words in the document that match with terms in the user-selected glossaries. The Glossaryfication Web Service was developed to provide added value to the OHEJP Glossary that was developed within the OHEJP project ORION. The OHEJP Glossary improves the communication and collaboration among OH sectors by providing an online resource that lists relevant OH terms and sector-specific definitions. The Glossaryfication Web Service supports the practical use of the curated OHEJP Glossary and can also source information from other glossaries relevant for OH professionals (currently supporting the online CDC, WHO and EFSA glossaries). The Glossaryfication Web Service was created using the open-source software KNIME and the KNIME Text Processing extension (https://www.knime.com/knime-text-processing). The Glossaryfication KNIME workflow is deployed on BfR’s KNIME Server infrastructure providing an easy-to-use web interface where the users can upload their documents (any text-type file e.g. PDF, Word, Excel) and select the desired glossary to compare with. The Glossaryfication KNIME workflow reads in the document provided via the web interface and applies natural language processing (e.g. text cleaning, stemming), transforming (bag-of-words generation) and information retrieval methods to identify the matching terms in the selected glossaries. The Glossaryfication Web Service generates as an output a table containing all the terms that match with the selected glossaries. It also provides the available definitions, corresponding references and additional meta-information, e.g. the term frequency, i.e., how often each term appears in the given text, and the sectoral classification (only for the OHEJP Glossary terms). Furthermore, the workflow generates a tag cloud where the terms are categorized as: (i) exact match when the term in the text matches exactly with the entry of this term in the glossary; (ii) inexact match when the term appears in the text slightly modified (e.g. plural forms or suffixes) and (iii) non-matching that corresponds to all the other words appearing in the text that do not match with any glossary term. Through the user interface, the users can then choose if they want to download the whole list of terms, select only the exact/inexact matching terms, or just choose those terms and definitions that match with the meaning intended for this term in the user-provided document. The resulting table of terms can be downloaded as an Excel file and added to the user’s document as a document-specific glossary. The Glossaryfication Web Service provides an easy-to-adopt solution to enrich documents and reports with more comprehensive and unambiguous glossaries. Furthermore, it improves the referentiality of terms and definitions from different OH sectors. An additional feature provided by the Glossaryfication Web Service is the possibility of extending its use to other glossaries from other national or international institutions allowing the user to customize this glossary creation service.

2021 ◽  
Vol 7 ◽  
Author(s):  
Nazareno Scaccia ◽  
Taras Günther ◽  
Estibaliz Lopez de Abechuco ◽  
Matthias Filter

In many interdisciplinary research domains, the creation of a shared understanding of relevant terms is considered the foundation for efficient cross-sector communication and interpretation of data and information. This is also true for the domain of One Health (OH) where many One Health Surveillance (OHS) documents rarely contain glossaries with a list of terms for which their specific meaning in the context of the given document is defined (Cornelia et al. 2018, Buschhardt et al. 2021). The absence of glossaries within these documents may lead to misinterpretation of surveillance results due to the wrong interpretation of terminology specifically when term definitions differ across OH sectors. Under the One Health EJP project ORION, the OHEJP Glossary was recently created. The OHEJP Glossary is a tool to improve communication and collaboration amongst OH sectors by providing an easy-to-use online resource that lists relevant OH terms and sector-specific definitions. To improve the accessibility of content from the OHEJP Glossary and support the creation of integrative glossaries in future OHS-related documents, the OHEJP Glossaryfication Web Service was created. This service can support the practical use of the OHEJP Glossary and other relevant online glossaries by OH professionals. The Glossaryfication Web Service (GWS) is an application that automatically identifies terms in any uploaded text-based document and creates a document-specific list of matching definitions in selected online glossaries. This auto-generated document-specific glossary can easily be adjusted by the user, for example, by selecting the desired definition in case multiple definitions were found for a specific term. The document-specific glossary could then be downloaded, manually adjusted and finally included into the original document where it supports the correct interpretation of terminology used. Especially in sector-specific reports, such as from animal health or public health authorities, this can be beneficial to ensure the correct interpretation by other OH sectors in the future. The GWS was developed with the open-source desktop software KNIME Analytics Platform and runs as a web service on a KNIME Web Server infrastructure. The core data processing functionality in the GWS is based on KNIME’s Text Processing extension. KNIME's JavaScript nodes provided the basis for an interactive user interface where users can easily upload their files and select between different reference glossaries, such as the OHEJP Glossary, the CDC Glossary, the WHO Glossary or the EFSA Glossary. After retrieval of the user input settings, the GWS tags words within the provided document and maps these tagged words with matching entries in the selected glossaries. As the main output, the user receives a downloadable list of matching terms with their corresponding definitions, sectorial assignments and references, which can then be added by the user to the original document. The GWS is freely accessible via this link as well as the underlying KNIME workflow.


2005 ◽  
Vol 6 (1-2) ◽  
pp. 86-93 ◽  
Author(s):  
Henk Harkema ◽  
Ian Roberts ◽  
Rob Gaizauskas ◽  
Mark Hepple

Recent years have seen a huge increase in the amount of biomedical information that is available in electronic format. Consequently, for biomedical researchers wishing to relate their experimental results to relevant data lurking somewhere within this expanding universe of on-line information, the ability to access and navigate biomedical information sources in an efficient manner has become increasingly important. Natural language and text processing techniques can facilitate this task by making the information contained in textual resources such as MEDLINE more readily accessible and amenable to computational processing. Names of biological entities such as genes and proteins provide critical links between different biomedical information sources and researchers' experimental data. Therefore, automatic identification and classification of these terms in text is an essential capability of any natural language processing system aimed at managing the wealth of biomedical information that is available electronically. To support term recognition in the biomedical domain, we have developed Termino, a large-scale terminological resource for text processing applications, which has two main components: first, a database into which very large numbers of terms can be loaded from resources such as UMLS, and stored together with various kinds of relevant information; second, a finite state recognizer, for fast and efficient identification and mark-up of terms within text. Since many biomedical applications require this functionality, we have made Termino available to the community as a web service, which allows for its integration into larger applications as a remotely located component, accessed through a standardized interface over the web.


2019 ◽  
Vol 9 (3) ◽  
pp. 594 ◽  
Author(s):  
Hye-Jeong Song ◽  
Ji-Eun Choi ◽  
Yoon-Kyoung Lee ◽  
Ji Yoon ◽  
Jong-Dae Kim ◽  
...  

Speaking is a way for humans to communicate with others using language. The ability to speak according to the speaker is very diverse. In general, language skills improve as intelligence improves. Therefore, it is known that the analysis of a speaker’s utterances is a good tool to evaluate the intellectual maturity of the speaker. Until recently, these evaluations have been done manually based on the experience of a handful of experts, but this approach is not only time consuming and costly, but also highly subjective. In this paper, we propose a Korean automatic speech analysis system based on Natural Language Processing (NLP) and web service to solve this problem. For this study, we constructed a web service based on Django to respond to the requests of various users. When a user delivered a transcription file of utterances to the server via the web, the server analyzed the speech ability of the speaker based on various indicators. The server compared the transcription file with the language ability indicators of persons of the same age as the speaker and displayed the result immediately to the user. In this study, we used KoNLPy, a Korean language-processing tool. The automatic speech analysis service analyzed not only the overall language ability of the speaker but also individual domains such as sentence completion ability and vocabulary ability. In addition, a faster and immediate service was made possible without sacrificing accuracy as compared to human analysis.


2014 ◽  
Vol 2014 ◽  
pp. 1-18 ◽  
Author(s):  
Miguel Ángel Rodríguez-García ◽  
Alejandro Rodríguez-González ◽  
Ricardo Colomo-Palacios ◽  
Rafael Valencia-García ◽  
Juan Miguel Gómez-Berbís ◽  
...  

Precise, reliable and real-time financial information is critical for added-value financial services after the economic turmoil from which markets are still struggling to recover. Since the Web has become the most significant data source, intelligent crawlers based on Semantic Technologies have become trailblazers in the search of knowledge combining natural language processing and ontology engineering techniques. In this paper, we present the SONAR extension approach, which will leverage the potential of knowledge representation by extracting, managing, and turning scarce and disperse financial information into well-classified, structured, and widely used XBRL format-oriented knowledge, strongly supported by a proof-of-concept implementation and a thorough evaluation of the benefits of the approach.


Author(s):  
Tata Sutabri ◽  
Yohanes Bowo Widodo ◽  
Sondang Sibuea ◽  
Ismi Rajiani ◽  
Yaziz Hasan

The success of aquascape maintenance lies in certain critical factors, including temperatures, lights and water filtration, which should be constantly monitored; but at some point, these keys are not well monitored, causing correctional action being taken too late and creating damage. Using an online control system as an IoT device can assist in control and automation functions, while an online feature provides monitoring, remote access, and remote override as other benefits that would help maintain an aquascape. This research’s subject involves designing an IoT device as an online control system that would automatically control temperatures, lights and water filtration using an Arduino-compatible board based on an ATMega328/P microcontroller connected to web services and a web portal for a monitoring and management interface. This control system uses Wi-Fi as a communication line with the web service. API is a bridge from the control system and the web service, and the portal’s front end and back end are custom-built using PHP and MySQL. The overall process was developed using a combination of the spiral model and prototyping paradigm modelling. The result is a fully working prototype of an online control system complete with a web interface, which helps aquascapers maintain their tank as human factors could be minimized by this device.


Author(s):  
Y. Tefera ◽  
F. Poiesi ◽  
D. Morabito ◽  
F. Remondino ◽  
E. Nocerino ◽  
...  

This paper presents a web-based 3D imaging pipeline, namely 3Dnow, that can be used by anyone without the need of installing additional software other than a browser. By uploading a set of images through the web interface, 3Dnow can generate sparse and dense point clouds as well as mesh models. 3D reconstructed models can be downloaded with standard formats or previewed directly on the web browser through an embedded visualisation interface. In addition to reconstructing objects, 3Dnow offers the possibility to evaluate and georeference point clouds. Reconstruction statistics, such as minimum, maximum and average intersection angles, point redundancy and density can also be accessed. The paper describes all features available in the web service and provides an analysis of the computational performance using servers with different GPU configurations.


2019 ◽  
Vol 16 (3) ◽  
Author(s):  
Yaoyong Li ◽  
Elisa Aguilar-Martinez ◽  
Andrew D. Sharrocks

AbstractThe interconversion of sequences that constitute the genome and the proteome is becoming increasingly important due to the generation of large amounts of DNA sequence data. Following mapping of DNA segments to the genome, one fundamentally important task is to find the amino acid sequences which are coded within a list of genomic sections. Conversely, given a series of protein segments, an important task is to find the genomic loci which code for a list of protein regions. To perform these tasks on a region by region basis is extremely laborious when a large number of regions are being studied. We have therefore implemented an R package geno2proteo which performs the two mapping tasks and subsequent sequence retrieval in a batch fashion. In order to make the tool more accessible to users, we have created a web interface of the R package which allows the users to perform the mapping tasks by going to the web page http://sharrocksresources.manchester.ac.uk/tofigaps and using the web service.


2005 ◽  
Vol 8 (1) ◽  
pp. 16-18
Author(s):  
Howard F. Wilson
Keyword(s):  

2019 ◽  
Vol 54 (6) ◽  
Author(s):  
Sawsan Ali Hamid ◽  
Rana Alauldeen Abdalrahman ◽  
Inam Abdullah Lafta ◽  
Israa Al Barazanchi

Recently, web services have presented a new and evolving model for constructing the distributed system. The meteoric growth of the Web over the last few years proves the efficacy of using simple protocols over the Internet as the basis for a large number of web services and applications. Web service is a modern technology of web, which can be defined as software applications with a programmatic interface based on Internet protocol. Web services became common in the applications of the web by the help of Universal, Description, Discovery and Integration; Web Service Description Language and Simple Object Access Protocol. The architecture of web services refers to a collection of conceptual components in which common sets of standard can be defined among interoperating components. Nevertheless, the existing Web service's architecture is not impervious to some challenges, such as security problems, and the quality of services. Against this backdrop, the present study will provide an overview of these issues. Therefore, it aims to propose web services architecture model to support distributed system in terms of application and issues.


2020 ◽  
Vol 7 (3) ◽  
pp. 471-494
Author(s):  
Katsumi NITTA ◽  
Ken SATOH

AbstractArtificial intelligence (AI) and law is an AI research area that has a history spanning more than 50 years. In the early stages, several legal-expert systems were developed. Legal-expert systems are tools designed to realize fair judgments in court. In addition to this research, as information and communication technologies and AI technologies have progressed, AI and law has broadened its view from legal-expert systems to legal analytics and, recently, a lot of machine-learning and text-processing techniques have been employed to analyze legal information. The research trends are the same in Japan as well and not only people involved with legal-expert systems, but also those involved with natural language processing as well as lawyers have become interested in AI and law. This report introduces the history of and the research activities on applying AI to the legal domain in Japan.


Sign in / Sign up

Export Citation Format

Share Document