Similarity Measures of Web Repositories constructed by Web-Scrapping from Specific Web Resource

Alagappa Institute of Skill Development & Computer Centre,Alagappa University, Karaikudi, India.15 -16 February 2017. IT Skills Show & International Conference on Advancements In Computing Resources (SSICACR-2017)

Format: Volume 5, Issue 1, No 17, 2017

Copyright: All Rights Reserved ©2017

Year of Publication: 2017

Author: J.Tamilselvan,Dr. A.Senthilrajan


View PDF Format


Information extraction systems apply machine learning to the task. These systems differ in how the IE problem characterized and in the style of text that they handle. The most important tasks in information extraction from the web are understanding webpage structure and its organization as many web sites contain large collections of pages displayed using a common template or layout,which makes it increasingly difficult to discover relevant data about a specific topic. Extracting data from such template pages has become an important issue in recent days as the number of web pages available on the Internet has growing in day by day. Tools and protocols to extract all this information have now come in demand as researchers and web surfers want to discover new knowledge at an ever increasing rate. A web crawler also known as, a robot or a spider is a system for the bulk downloading of web pages, whereas the goal of a focused crawler is seeking pages that are relevant to a pre-defined set of topics from a specific web resource. Collecting and indexing those accessible web documents,which can answer all ad-hoc queries, a focused crawler analyzes its crawl boundary to find the links that are likely to be most relevant for the crawl, and avoids irrelevant regions of the Web. Since all search engines take their data fed using crawlers, it is critical to improve its working ability. As the size of data is huge,Common Crawlers are no longer applicable in real life. So there is need to develop a domain specific crawler builds on stock of existing algorithms. This led to considerable savings in hardware and network resources, and helps keep the crawl more up-to-date. This paper proposed a novel framework called SWNLP, which enables bidirectional integration of page structure understanding and text understanding in an iterative manner. We have applied the proposed framework to the judgments information system to extract text of judgments and relate the similarity measures.


[1]Gautam Pant, Padmini Srinivasan1, FilippoMenczer, ―Crawling the Web‖, Department of Management Sciences, The University of Iowa, Iowa City IA 52242, USA. [2]DebajyotiMukhopadhyay, Arup Biswas, SukantaSinha, ―A New Approach to Design Domain Specific Ontology Based Web Crawler‖, West Bengal University of Technology, pp.70091. [3]Chakrabarti, Soumen, Martin van den Berg, and Byron Dom. ―Focused crawling: a new approach to topic-specificWeb resource discovery‖, Elsevier, 1999. [4]Sk.AbdulNabi, Dr.P.Premchand, ―Effective Performance of Information Retrieval by using Domain Based Crawler‖, Vol. 4, No.7, 2013. [5] Scott Deerwester, Susan T. Dumais, George W. Furnas and Thomas K. Landauer, Richard Harshman, ―Indexing by Latent Semantic Analysis‖, 41(6):391-407, 1990. [6]Radhika Gupta, AP Nidhi, ―Focused Crawling System based in Improved LSI‖, Volume 2 Issue 9, September 2013. [7] Hong-Wei Hao, Cui-Xia Mu, Xu-Cheng Yin, Shen Li, Zhi-Bin Wang, ―An Improved Topic Relevance Algorithm for focused Crawling‖. [8] Ali Pesaranghader, Ahmad Pesaranghader, Norwati Mustapha, NurfadhlinaMohdSharef, ―Improving Multi-term Topics Focused Crawling by Introducing Term Frequencyy-Information Content (TF-IC) Measure‖, September 2013.


Web Crawler, Text Extraction, Structured Web Data, Deep Web.

This work is licensed under a Creative Commons Attribution 3.0 Unported License.   

Facebook IconYouTube IconTwitter IconVisit Our Blog