Sciweavers

203 search results - page 26 / 41
» Conceptual-Model-Based Data Extraction from Multiple-Record ...
Sort
View
WWW
2009
ACM
14 years 8 months ago
Sitemaps: above and beyond the crawl of duty
Comprehensive coverage of the public web is crucial to web search engines. Search engines use crawlers to retrieve pages and then discover new ones by extracting the pages' o...
Uri Schonfeld, Narayanan Shivakumar
WWW
2010
ACM
14 years 2 months ago
Not so creepy crawler: easy crawler generation with standard xml queries
Web crawlers are increasingly used for focused tasks such as the extraction of data from Wikipedia or the analysis of social networks like last.fm. In these cases, pages are far m...
Franziska von dem Bussche, Klara A. Weiand, Benedi...
WWW
2004
ACM
14 years 8 months ago
Link fusion: a unified link analysis framework for multi-type interrelated data objects
Web link analysis has proven to be a significant enhancement for quality based web search. Most existing links can be classified into two categories: intra-type links (e.g., web h...
Wensi Xi, Benyu Zhang, Zheng Chen, Yizhou Lu, Shui...
DEBU
1998
95views more  DEBU 1998»
13 years 7 months ago
What can you do with a Web in your Pocket?
The amount of information available online has grown enormously over the past decade. Fortunately, computing power, disk capacity, and network bandwidth have also increased dramat...
Sergey Brin, Rajeev Motwani, Lawrence Page, Terry ...
KDD
2007
ACM
155views Data Mining» more  KDD 2007»
14 years 8 months ago
Mining templates from search result records of search engines
Metasearch engine, Comparison-shopping and Deep Web crawling applications need to extract search result records enwrapped in result pages returned from search engines in response ...
Hongkun Zhao, Weiyi Meng, Clement T. Yu