The anatomy of a large-scale hypertextual Web search engine
WWW7 Proceedings of the seventh international conference on World Wide Web 7
Trawling the Web for emerging cyber-communities
WWW '99 Proceedings of the eighth international conference on World Wide Web
Authoritative sources in a hyperlinked environment
Proceedings of the ninth annual ACM-SIAM symposium on Discrete algorithms
Proceedings of the 9th international World Wide Web conference on Computer networks : the international journal of computer and telecommunications netowrking
Extracting Large-Scale Knowledge Bases from the Web
VLDB '99 Proceedings of the 25th International Conference on Very Large Data Bases
Proceedings of the 27th International Conference on Very Large Data Bases
A large-scale study of the evolution of web pages
WWW '03 Proceedings of the 12th international conference on World Wide Web
Efficient URL caching for world wide web crawling
WWW '03 Proceedings of the 12th international conference on World Wide Web
Stochastic models for the Web graph
FOCS '00 Proceedings of the 41st Annual Symposium on Foundations of Computer Science
Towards Compressing Web Graphs
DCC '01 Proceedings of the Data Compression Conference
Random Evolution in Massive Graphs
FOCS '01 Proceedings of the 42nd IEEE symposium on Foundations of Computer Science
The web as a graph: measurements, models, and methods
COCOON'99 Proceedings of the 5th annual international conference on Computing and combinatorics
A methodology for the evaluation of web graph models and a test case
Proceedings of the 38th conference on Winter simulation
Hi-index | 0.00 |
Valid models of the WWW are important for creating WWW-like representations upon which new algorithms and applications for searching, indexing, compression etc. can be tested, and mostly for predicting the evolution of the web and the emergence of important new phenomena. Researchers have viewed the WWW as a graph, the so-called web graph. We present a brief review of the most typical random graph models for the web and introduce a validation process for web graph models. We evaluate the behaviour of the Exponential Growth Copying (EGC) model, which has been explicitly designed to model the WWW, and analyse the effect of individual parameters on its effectiveness through simulation modelling. Specifically, we derive the in and out degree distributions of the resulting graphs for various parameter values and measure them against the empirical analytical results from the real web (i.e. power laws for in and out degrees). Finally, we suggest appropriate values to improve EGC effectiveness and deliver a realistic model of the web graph.