scholarly journals Deep Web Search Interface Identification: A Semi-Supervised Ensemble Approach

Information ◽  
2014 ◽  
Vol 5 (4) ◽  
pp. 634-651 ◽  
Author(s):  
Hong Wang ◽  
Qingsong Xu ◽  
Lifeng Zhou
2013 ◽  
Vol 5 (4) ◽  
pp. 1278-1283
Author(s):  
Hong Wang ◽  
Qingsong Xu ◽  
Youyang Chen ◽  
Jinsong Lan
Keyword(s):  
Deep Web ◽  

2009 ◽  
pp. 784-788 ◽  
Author(s):  
Kevin C. Chang
Keyword(s):  

2010 ◽  
Vol 39 (1) ◽  
pp. 33-40 ◽  
Author(s):  
Ritu Khare ◽  
Yuan An ◽  
Il-Yeol Song
Keyword(s):  

The Dark Web ◽  
2018 ◽  
pp. 359-374
Author(s):  
Dilip Kumar Sharma ◽  
A. K. Sharma

ICT plays a vital role in human development through information extraction and includes computer networks and telecommunication networks. One of the important modules of ICT is computer networks, which are the backbone of the World Wide Web (WWW). Search engines are computer programs that browse and extract information from the WWW in a systematic and automatic manner. This paper examines the three main components of search engines: Extractor, a web crawler which starts with a URL; Analyzer, an indexer that processes words on the web page and stores the resulting index in a database; and Interface Generator, a query handler that understands the need and preferences of the user. This paper concentrates on the information available on the surface web through general web pages and the hidden information behind the query interface, called deep web. This paper emphasizes the Extraction of relevant information to generate the preferred content for the user as the first result of his or her search query. This paper discusses the aspect of deep web with analysis of a few existing deep web search engines.


Author(s):  
Dilip Kumar Sharma ◽  
A. K. Sharma

ICT plays a vital role in human development through information extraction and includes computer networks and telecommunication networks. One of the important modules of ICT is computer networks, which are the backbone of the World Wide Web (WWW). Search engines are computer programs that browse and extract information from the WWW in a systematic and automatic manner. This paper examines the three main components of search engines: Extractor, a web crawler which starts with a URL; Analyzer, an indexer that processes words on the web page and stores the resulting index in a database; and Interface Generator, a query handler that understands the need and preferences of the user. This paper concentrates on the information available on the surface web through general web pages and the hidden information behind the query interface, called deep web. This paper emphasizes the Extraction of relevant information to generate the preferred content for the user as the first result of his or her search query. This paper discusses the aspect of deep web with analysis of a few existing deep web search engines.


Sign in / Sign up

Export Citation Format

Share Document