CFP last date
20 May 2024
Reseach Article

A Parametric Layered Approach to Perform Web Page Ranking

by Ratika Goel, Anchal Garg
International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
Volume 67 - Number 14
Year of Publication: 2013
Authors: Ratika Goel, Anchal Garg
10.5120/11467-7251

Ratika Goel, Anchal Garg . A Parametric Layered Approach to Perform Web Page Ranking. International Journal of Computer Applications. 67, 14 ( April 2013), 44-48. DOI=10.5120/11467-7251

@article{ 10.5120/11467-7251,
author = { Ratika Goel, Anchal Garg },
title = { A Parametric Layered Approach to Perform Web Page Ranking },
journal = { International Journal of Computer Applications },
issue_date = { April 2013 },
volume = { 67 },
number = { 14 },
month = { April },
year = { 2013 },
issn = { 0975-8887 },
pages = { 44-48 },
numpages = {9},
url = { https://ijcaonline.org/archives/volume67/number14/11467-7251/ },
doi = { 10.5120/11467-7251 },
publisher = {Foundation of Computer Science (FCS), NY, USA},
address = {New York, USA}
}
%0 Journal Article
%1 2024-02-06T21:24:51.574709+05:30
%A Ratika Goel
%A Anchal Garg
%T A Parametric Layered Approach to Perform Web Page Ranking
%J International Journal of Computer Applications
%@ 0975-8887
%V 67
%N 14
%P 44-48
%D 2013
%I Foundation of Computer Science (FCS), NY, USA
Abstract

Web crawling is the foremost step to perform the effective and efficient web content search so that the user will get the specific web pages initially in an indexed form. Web crawling is not only used for searching a webpage over the web but also to order them according to user interest. There are number of available search engines and the crawlers that accept the user query and provide the page search. But, there is still the requirement and scope to improve the search mechanism. In this present work, dynamic and user interest evolution based parametric approach is defined to perform the web crawling and to arrange the web pages in more definite way. In this work a layered approach is defined, in which the initial indexing will be performed based on the keyword oriented content match and later on the indexing will be modified based on user recommendation. The presented work will provide an recommendation based web page indexing so that effective web crawling will be performed.

References
  1. Avanish K. Singh,"Novel Architecture of Web Crawler for URL Distribution", International Journal of Computer Science and Technology, Vol 2, Issue 3, Sept 2011, pp 42-45
  2. P. Srinivasan,"Web Crawling Agents for Retrieving Biomedical Information", ACM, NETTAB 2002 Bologna, Italy, pp 1-8
  3. J. Cho,"Finding replicated web collections", In proceedings of the 2000 ACM international conference of Management of Data (SIGMOD) 2000 pp 355-366
  4. J. Cho," Parallel Crawlers", In proceedings of WWW2002, Honolulu, hawaii, USA, May 7-11, 2002. ACM 1-58113-449-5/02/005
  5. O. Brandman, "Crawler-Friendly Web Servers", In Workshop on Performance and Architecture of Web Servers ( PAWS), June 2000, pp 1-16
  6. Charu C. Aggarwal,"On Learning Strategies for Topic Specific Web Crawling", WI '06 Proceedings of the 2006 IEEE/WIC/ACM International Conference on Web Intelligence
  7. V. Shkapenyuk,"Design and Implementation of a High-Performance Distributed Web Crawler", In Proceedings of the 18th International Conference on Data Engineering (ICDE 2002), pp. 357–368
  8. B. Polverini,"Levels of Awareness: Design Considerations for Web Crawlers and Censorware Detection", White paper, Princeton University, May 2011
  9. Hussein Issa Rutgers Business School, Rutgers University "Application of Duplicate Records detection Techniques to Duplicate Payments in a Real Business Environment"
  10. Hani Khoshdel Nikkhoo "The Impact of Near Duplicate Documents on Information Retrieval Evaluation" by A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Mathematics in Computer Science Waterloo, Ontario, Canada, 2010
  11. J Prasanna KumarProfessor, "Duplicate and Near Duplicate Documents Detection: A Review". European Journal of Scientific Research ISSN 1450-216X Vol. 32 No. 4 (2009), pp. 514-527
  12. Cho, Junghoo; Hector Garcia-Molina (2000). "Synchronizing a database to improve freshness". Proceedings of the 2000 ACM SIGMOD international conference on Management of data. Dallas, Texas, UnitedStates: ACM. pp. 117, 2009
Index Terms

Computer Science
Information Sciences

Keywords

Crawling Indexing Recommender system