scholarly journals RSS FEED VISUALIZATION

Author(s):  
JYOTSNA BAGRET ◽  
PRASANNA MUNDADA ◽  
SABAH TAZEEN ◽  
TANUJA MULLA

This paper describes how the web content visualization can be greatly improved using the modeling technique. Web content visualization is the outcome of effort made to avail an improved 3D visualization unlike the 2D web content visualization at present. Web page navigation in this case will be depicted by a 2D graph and the web content will be visualized in the form of 3D graph. Also the RSS feeds will be visualized in the form of 3D graph. In normal browser we type name of the URL in the address bar and that URL is downloaded. But the 3D browser takes any URL as an input and generates a 3D graph of the whole website. When we type the URL, a root node of this URL is created. And then this URL goes to the Parser. The parser, parse this web page and gives output in the form of the set of the hyperlinks. Corresponding to each link we create a node and it is attached to the root node. In this way the whole 3D graph of the website is generated. Different color schemes are used for the nodes of different links e.g. text links, image links, video links etc. Advanced search facility is also provided. Moreover as the graph is 3D in nature, the user can rotate the graph as per his requirement.

Author(s):  
Jie Zhao ◽  
Jianfei Wang ◽  
Jia Yang ◽  
Peiquan Jin

Company acquisition relation reflects a company's development intent and competitive strategies, which is an important type of enterprise competitive intelligence. In the traditional environment, the acquisition of competitive intelligence mainly relies on newspapers, internal reports, and so on, but the rapid development of the Web introduces a new way to extract company acquisition relation. In this paper, the authors study the problem of extracting company acquisition relation from huge amounts of Web pages, and propose a novel algorithm for company acquisition relation extraction. The authors' algorithm considers the tense feature of Web content and classification technology of semantic strength when extracting company acquisition relation from Web pages. It first determines the tense of each sentence in a Web page, which is then applied in sentences classification so as to evaluate the semantic strength of the candidate sentences in describing company acquisition relation. After that, the authors rank the candidate acquisition relations and return the top-k company acquisition relation. They run experiments on 6144 pages crawled through Google, and measure the performance of their algorithm under different metrics. The experimental results show that the algorithm is effective in determining the tense of sentences as well as the company acquisition relation.


Author(s):  
Shailesh Shivakumar ◽  
Venkata Suresh Pachigolla

Segregating the web page content into logical chunks is one of the popular techniques for modular organization of web page. While chunk-based approach works well for public web scenarios, in case of mobile-first personalization cases, chunking strategy would not be as effective for performance optimization due to dynamic nature of the Web content and due to the nature of content granularity. In this paper, the authors propose a novel framework Micro chunk based Web Delivery Framework which proposes and uses a novel concept of “micro chunk”. The micro chunk based Web Delivery framework aims to address the performance challenges posed by regular chunk in a personalized web scenario. The authors will look at the methods for creating micro chunk and they will discuss the advantages of micro chunk when compared to a regular chunk for a personalized mobile web scenario. They have created a prototype application implementing the Micro chunk based Web Delivery Framework and benchmarked it against a regular personalized web application to quantify the performance improvements achieved by micro chunk design.


Author(s):  
Kai-Hsiang Yang

This chapter will address the issues of Uniform Resource Locator (URL) correction techniques in proxy servers. The proxy servers are more and more important in the World Wide Web (WWW), and they provide Web page caches for browsing the Web pages quickly, and also reduce unnecessary network traffic. Traditional proxy servers use the URL to identify their cache, and it is a cache-miss when the request URL is non-existent in its caches. However, for general users, there must be some regularity and scope in browsing the Web. It would be very convenient for users when they do not need to enter the whole long URL, or if they still could see the Web content even though they forgot some part of the URL, especially for those personal favorite Web sites. We will introduce one URL correction mechanism into the personal proxy server to achieve this goal.


2017 ◽  
Vol 1 (1) ◽  
pp. 35-51
Author(s):  
S.D. Scott ◽  
Y.H. Koh

The majority of web-pages are unsuitable for viewing on PDAs, WAP phones and similar devices without first being adapted. However, little empirical work has been done on what actually constitutes a good PDA or WAP web-page. This paper ranks a number of PDA web-pages from different categories empirically and correlates the result against the design metrics present. The findings are then compared against a similar set of experiments for PC web-pages. The results of this comparison suggest that, as well as omitting, summarizing and converting individual multimedia objects in the web-page to a less resource intensive form, the design metrics need to be changed during adaptation to enhance the presentation of web-content on non-PC devices. The paper concludes by investigating the effect of applying some suitable changes to the design metrics on web=page content chunks, which form the basic units in automatic content adaptation systems.


2013 ◽  
Vol 347-350 ◽  
pp. 2575-2579
Author(s):  
Wen Tao Liu

The web data collection is the process of collecting the semi-structured, large-scale and redundant data which include web content, web structure and web usage in the web by the crawler and it is often used for the information extraction, information retrieval, search engine and web data mining. In this paper, the web data collection principle is introduced and some related topics are discussed such as page download, coding problem, updated strategy, static and dynamic page. The multithread technology is described and multithread mode for the web data collection is proposed. The web data collection with multithread can get better resource utilization, better average response time and better performance.


2014 ◽  
Vol 651-653 ◽  
pp. 1806-1810
Author(s):  
Long Li ◽  
An Min Zhou ◽  
Yong Fang ◽  
Liang Liu ◽  
Qian Wu
Keyword(s):  
Web Page ◽  

The paper studies the VIPS algorithm, and improves VIPS which has the deficiency with complex rules and low performance, according that the Web page has the feature of DIV structure in Web2.0, and combines the method based statistics information, introduces a DVIPS algorithm of extracting web main content.


2020 ◽  
pp. 151-156
Author(s):  
A. P. Korablev ◽  
N. S. Liksakova ◽  
D. M. Mirin ◽  
D. G. Oreshkin ◽  
P. G. Efimov

A new species list of plants and lichens of Russia and neighboring countries has been developed for Turboveg for Windows, the program, intended for storage and management of phytosociological data (relevés), is widely used all around the world (Hennekens, Schaminée, 2001; Hennekens, 2015). The species list is built upon the database of the Russian website Plantarium (Plantarium…: [site]), which contains a species atlas and illustrated an online Handbook of plants and lichens. The nomenclature used on Plantarium was originally based on the following issues: vascular plants — S. K. Cherepanov (1995) with additions; mosses — «Flora of mosses of Russia» (Proect...: [site]); liverworts and hornworts — A. D. Potemkin and E. V. Sofronova (2009); lichens — «Spisok…» G. P. Urbanavichyus ed. (2010); other sources (Plantarium...: [site]). The new species list, currently the most comprehensive in Turboveg format for Russia, has 89 501 entries, including 4627 genus taxa compare to the old one with 32 020 entries (taxa) and only 253 synonyms. There are 84 805 species and subspecies taxa in the list, 37 760 (44.7 %) of which are accepted, while the others are synonyms. Their distribution by groups of organisms and divisions are shown in Table. A large number of synonyms in the new list and its adaptation to work with the Russian literature will greatly facilitate the entry of old relevé data. The ways of making new list, its structure as well as the possibilities of checking taxonomic lists on Internet resources are considered. The files of the species list for Turboveg 2 and Turboveg 3, the technique of associating existing databases with a new species list (in Russian) are available on the web page https://www.binran.ru/resursy/informatsionnyye-resursy/tekuschie-proekty/species_list_russia/.


2009 ◽  
Author(s):  
Mirko Luca Lobina ◽  
Davide Mula
Keyword(s):  
Web Page ◽  

2021 ◽  
Vol 13 (2) ◽  
pp. 50
Author(s):  
Hamed Z. Jahromi ◽  
Declan Delaney ◽  
Andrew Hines

Content is a key influencing factor in Web Quality of Experience (QoE) estimation. A web user’s satisfaction can be influenced by how long it takes to render and visualize the visible parts of the web page in the browser. This is referred to as the Above-the-fold (ATF) time. SpeedIndex (SI) has been widely used to estimate perceived web page loading speed of ATF content and a proxy metric for Web QoE estimation. Web application developers have been actively introducing innovative interactive features, such as animated and multimedia content, aiming to capture the users’ attention and improve the functionality and utility of the web applications. However, the literature shows that, for the websites with animated content, the estimated ATF time using the state-of-the-art metrics may not accurately match completed ATF time as perceived by users. This study introduces a new metric, Plausibly Complete Time (PCT), that estimates ATF time for a user’s perception of websites with and without animations. PCT can be integrated with SI and web QoE models. The accuracy of the proposed metric is evaluated based on two publicly available datasets. The proposed metric holds a high positive Spearman’s correlation (rs=0.89) with the Perceived ATF reported by the users for websites with and without animated content. This study demonstrates that using PCT as a KPI in QoE estimation models can improve the robustness of QoE estimation in comparison to using the state-of-the-art ATF time metric. Furthermore, experimental result showed that the estimation of SI using PCT improves the robustness of SI for websites with animated content. The PCT estimation allows web application designers to identify where poor design has significantly increased ATF time and refactor their implementation before it impacts end-user experience.


2012 ◽  
Vol 241-244 ◽  
pp. 2779-2782
Author(s):  
Heng Yao Tang ◽  
Xiao Yan Zhan

On the problems existing in the realization of current accessibility website, we design a web designing architecture, using the web log mining technique to extract user interests and access priority sequence and adopting the dynamic web page information to fill the web page commonly used structure, realize the intelligent , personalized accessibility.


Sign in / Sign up

Export Citation Format

Share Document