Towards compressing Web graphs
We consider the problem of compressing graphs of the link structure of the World Wide Web. We provide efficient algorithms for such compression that are motivated by random graph models for describing the Web. The algorithms are based on reducing the compression problem to the problem of finding a m...
Saved in:
Published in | Proceedings DCC 2001. Data Compression Conference pp. 203 - 212 |
---|---|
Main Authors | , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
2001
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | We consider the problem of compressing graphs of the link structure of the World Wide Web. We provide efficient algorithms for such compression that are motivated by random graph models for describing the Web. The algorithms are based on reducing the compression problem to the problem of finding a minimum spanning free in a directed graph related to the original link graph. The performance of the algorithms on graphs generated by the random graph models suggests that by taking advantage of the link structure of the Web, one may achieve significantly better compression than natural Huffman-based schemes. We also provide hardness results demonstrating limitations on natural extensions of our approach. |
---|---|
ISBN: | 0769510310 9780769510316 |
ISSN: | 1068-0314 |
DOI: | 10.1109/DCC.2001.917151 |