Sciweavers

31 search results - page 5 / 7
» Detecting near-duplicates for web crawling
Sort
View
WWW
2009
ACM
14 years 8 months ago
Detecting soft errors by redirection classification
A soft error redirection is a URL redirection to a page that returns the HTTP status code 200 (OK) but has actually no relevant content to the client request. Since such redirecti...
Taehyung Lee, Jinil Kim, Jin Wook Kim, Sung-Ryul K...
SICHERHEIT
2008
13 years 8 months ago
Monkey-Spider: Detecting Malicious Websites with Low-Interaction Honeyclients
Abstract: Client-side attacks are on the rise: malicious websites that exploit vulnerabilities in the visitor's browser are posing a serious threat to client security, comprom...
Ali Ikinci, Thorsten Holz, Felix C. Freiling
WWW
2010
ACM
13 years 11 months ago
Time is of the essence: improving recency ranking using Twitter data
Realtime web search refers to the retrieval of very fresh content which is in high demand. An effective portal web search engine must support a variety of search needs, including ...
Anlei Dong, Ruiqiang Zhang, Pranam Kolari, Jing Ba...
EDBTW
2010
Springer
13 years 6 months ago
Using visual pages analysis for optimizing web archiving
Due to the growing importance of the World Wide Web, archiving it has become crucial for preserving useful source of information. To maintain a web archive up-to-date, crawlers ha...
Myriam Ben Saad, Stéphane Gançarski
WWW
2011
ACM
13 years 2 months ago
Prophiler: a fast filter for the large-scale detection of malicious web pages
Malicious web pages that host drive-by-download exploits have become a popular means for compromising hosts on the Internet and, subsequently, for creating large-scale botnets. In...
Davide Canali, Marco Cova, Giovanni Vigna, Christo...