Trawling the Web for emerging cyber-communities
WWW '99 Proceedings of the eighth international conference on World Wide Web
Authoritative sources in a hyperlinked environment
Journal of the ACM (JACM)
A random graph model for massive graphs
STOC '00 Proceedings of the thirty-second annual ACM symposium on Theory of computing
Hubs, authorities, and communities
ACM Computing Surveys (CSUR)
Proceedings of the 9th international World Wide Web conference on Computer networks : the international journal of computer and telecommunications netowrking
Breadth-first crawling yields high-quality pages
Proceedings of the 10th international conference on World Wide Web
The degree sequence of a scale-free random graph process
Random Structures & Algorithms
Extracting Large-Scale Knowledge Bases from the Web
VLDB '99 Proceedings of the 25th International Conference on Very Large Data Bases
Fast Algorithms for Mining Association Rules in Large Databases
VLDB '94 Proceedings of the 20th International Conference on Very Large Data Bases
Using PageRank to Characterize Web Structure
COCOON '02 Proceedings of the 8th Annual International Conference on Computing and Combinatorics
Stochastic models for the Web graph
FOCS '00 Proceedings of the 41st Annual Symposium on Foundations of Computer Science
Proceedings of the 13th international conference on World Wide Web
Partitioning of Web graphs by community topology
WWW '05 Proceedings of the 14th international conference on World Wide Web
A survey of models of the web graph
CAAN'04 Proceedings of the First international conference on Combinatorial and Algorithmic Aspects of Networking
Genealogical trees on the web: a search engine user perspective
Proceedings of the 17th international conference on World Wide Web
Networks, fractals and complexity: web-based information patterns
Proceedings of the International Conference on Management of Emergent Digital EcoSystems
A hierarchical adaptive probabilistic approach for zero hour phish detection
ESORICS'10 Proceedings of the 15th European conference on Research in computer security
Hi-index | 0.00 |
This paper proposes a random Web crawl model. A Web crawl is a (biased and partial) image of the Web. This paper deals with the hyperlink structure, i.e. a Web crawl is a graph, whose vertices are the pages and whose edges are the hypertextual links. Of course a Web crawl has a very special structure; we recall some known results about it. We then propose a model generating similar structures. Our model simply simulates a crawling, i.e. builds and crawls the graph at the same time. The graphs generated have lot of known properties of Web crawls. Our model is simpler than most random Web graph models, but captures the sames properties. Notice that it models the crawling process instead of the page writing process of Web graph models.