Synchronizing a database to improve freshness
SIGMOD '00 Proceedings of the 2000 ACM SIGMOD international conference on Management of data
Breadth-first crawling yields high-quality pages
Proceedings of the 10th international conference on World Wide Web
Introduction to algorithms
Inside C#
The indexable web is more than 11.5 billion pages
WWW '05 Special interest tracks and posters of the 14th international conference on World Wide Web
Beginning PHP 5 and MySQL 5: From Novice to Professional, Second Edition (Beginning: from Novice to Professional)
Effective change detection using sampling
VLDB '02 Proceedings of the 28th international conference on Very Large Data Bases
Hi-index | 0.00 |
Search engines save copies of crawled web pages to provide instant search results, saved pages may become old and un-updated as original pages change providing new information and new links, and most of websites don't submit these new changes to search engines so search engines don't depend mainly on websites techniques of submitting changes. Keeping pages fresh and updated in search engine is important for giving real page ranks and for providing real time information. Techniques were invented to improve the page update process by search engine. In this paper the author combines two of good known techniques and implements the new one via experiments that improve better results in different experiment cases.