Inside the Web: A Look at Digital Libraries and the Invisible/Deep Web

2008 ◽  
Vol 37 (1) ◽  
pp. 71-82 ◽  
Author(s):  
Mila C. Su
Author(s):  
F. J. CABRERIZO ◽  
J. LÓPEZ-GIJÓN ◽  
A. A. RUÍZ ◽  
E. HERRERA-VIEDMA

The Web is changing the information access processes and it is one of the most important information media. Thus, the developments on the Web are having a great influence over the developments on others information access instruments as digital libraries. As the development of digital libraries is to satisfy user need, user satisfaction is essential for the success of a digital library. The aim of this paper is to present a model based on fuzzy linguistic information to evaluate the quality of digital libraries. The quality evaluation of digital libraries is defined using users' perceptions on the quality of digital services provided through their Websites. We assume a fuzzy linguistic modeling to represent the users' perception and apply automatic tools of fuzzy computing with words based on the LOWA and LWA operators to compute global quality evaluations of digital libraries. Additionally, we show an example of application of this model where three Spanish academic digital libraries are evaluated by fifty users.


2018 ◽  
Vol 12 (1) ◽  
pp. 114
Author(s):  
Mohammad Niqresh

The study aims at identifying the concept of digital library, it also tries to shed the light on the most significant intellectual issues by presenting its definition, development, functions (selection and acquisition of information resources from the web, sources indexing, communication and management of intellectual property rights, production of electronic resources and its availability, and digital resources maintaining), characteristics, and the purpose of turning into digital library, passed by the proposed stages of digital library transition, Types of Intellectual Property (Copyright, Patents, Trademarks, Commercial Secrets), it also discusses copyrights and intellectual property, the problems and challenges of digital library, and finally the future of digital library. Many researchers agree that the main objective of the digital library is to accomplish all the functions of the traditional library, but in the form of electronic digital libraries which are only an extension for jobs that are performed and the resources that are accessible in digital library. The study concluded that digital libraries emerged as an obligatory result of revolution of the third millennium which is called the communications revolution, as libraries are able to prove that they are able to stand and cope with all the modern technology, where there is no conflict between the new and modern trends in libraries issue, but it also benefits from both of them concerning their evolution instruments in service for beneficiaries in every time and place.


Author(s):  
Punam Bedi ◽  
Neha Gupta ◽  
Vinita Jindal

The World Wide Web is a part of the Internet that provides data dissemination facility to people. The contents of the Web are crawled and indexed by search engines so that they can be retrieved, ranked, and displayed as a result of users' search queries. These contents that can be easily retrieved using Web browsers and search engines comprise the Surface Web. All information that cannot be crawled by search engines' crawlers falls under Deep Web. Deep Web content never appears in the results displayed by search engines. Though this part of the Web remains hidden, it can be reached using targeted search over normal Web browsers. Unlike Deep Web, there exists a portion of the World Wide Web that cannot be accessed without special software. This is known as the Dark Web. This chapter describes how the Dark Web differs from the Deep Web and elaborates on the commonly used software to enter the Dark Web. It highlights the illegitimate and legitimate sides of the Dark Web and specifies the role played by cryptocurrencies in the expansion of Dark Web's user base.


2011 ◽  
pp. 972-985
Author(s):  
Ákos Hajnal ◽  
Tamás Kifor ◽  
Gergely Lukácsy ◽  
László Z. Varga

More and more systems provide data through web service interfaces and these data have to be integrated with the legacy relational databases of the enterprise. The integration is usually done with enterprise information integration systems which provide a uniform query language to all information sources, therefore the XML data sources of Web services having a procedural access interface have to be matched with relational data sources having a database interface. In this chapter the authors provide a solution to this problem by describing the Web service wrapper component of the SINTAGMA Enterprise Information Integration system. They demonstrate Web services as XML data sources in enterprise information integration by showing how the web service wrapper component integrates XML data of Web services in the application domain of digital libraries.


Author(s):  
Giorgio Poletti

An analysis of the reality surrounding us clearly reveals the great amount of information, available in different forms and through different media. Volumes of information available in real time and via the Web are concepts perceived as closely related. This perception is supported by the remark that the objective of the Web was the definition and construction of a universal archive, a virtual site in which the access to documents was possible with no limits of time or space. In this digital library, documents have to be equipped with logical connections making possible for each user the definition of a reading map that expands according to the demand for knowledge gradually built up. This perspective is pointing now in the direction of the Semantic Web, a network satisfying our requests while understanding them, not by some magic telepathic communication between browser and navigator, but rather a data warehouse in which documents are matched to meta-data,1 letting specialized software to distinguish fields, importance, and correlation between documents. Semantic Web and library terms have an ever increasing close relationship, fundamental for the progress and the didactic efficiency in knowledge society.


2011 ◽  
Vol 1 (3) ◽  
pp. 54-70 ◽  
Author(s):  
Abdullah Wahbeh ◽  
Mohammed Al-Kabi ◽  
Qasem Al-Radaideh ◽  
Emad Al-Shawakfa ◽  
Izzat Alsmadi

The information world is rich of documents in different formats or applications, such as databases, digital libraries, and the Web. Text classification is used for aiding search functionality offered by search engines and information retrieval systems to deal with the large number of documents on the web. Many research papers, conducted within the field of text classification, were applied to English, Dutch, Chinese, and other languages, whereas fewer were applied to Arabic language. This paper addresses the issue of automatic classification or classification of Arabic text documents. It applies text classification to Arabic language text documents using stemming as part of the preprocessing steps. Results have showed that applying text classification without using stemming; the support vector machine (SVM) classifier has achieved the highest classification accuracy using the two test modes with 87.79% and 88.54%. On the other hand, stemming has negatively affected the accuracy, where the SVM accuracy using the two test modes dropped down to 84.49% and 86.35%.


2018 ◽  
Vol 52 (2) ◽  
pp. 266-277 ◽  
Author(s):  
Hyo-Jung Oh ◽  
Dong-Hyun Won ◽  
Chonghyuck Kim ◽  
Sung-Hee Park ◽  
Yong Kim

Purpose The purpose of this paper is to describe the development of an algorithm for realizing web crawlers that automatically collect dynamically generated webpages from the deep web. Design/methodology/approach This study proposes and develops an algorithm to collect web information as if the web crawler gathers static webpages by managing script commands as links. The proposed web crawler actually experiments with the algorithm by collecting deep webpages. Findings Among the findings of this study is that if the actual crawling process provides search results as script pages, the outcome only collects the first page. However, the proposed algorithm can collect deep webpages in this case. Research limitations/implications To use a script as a link, a human must first analyze the web document. This study uses the web browser object provided by Microsoft Visual Studio as a script launcher, so it cannot collect deep webpages if the web browser object cannot launch the script, or if the web document contains script errors. Practical implications The research results show deep webs are estimated to have 450 to 550 times more information than surface webpages, and it is difficult to collect web documents. However, this algorithm helps to enable deep web collection through script runs. Originality/value This study presents a new method to be utilized with script links instead of adopting previous keywords. The proposed algorithm is available as an ordinary URL. From the conducted experiment, analysis of scripts on individual websites is needed to employ them as links.


Sign in / Sign up

Export Citation Format

Share Document