We investigate the use of autonomically created small-world graphs as a framework for the long term storage of digital objects on the Web in a potentially hostile environment. We ...
Consider a scenario where one desires to simulate the execution of some graph algorithm on random input graphs of huge, perhaps even exponential size. Sampling and storing these h...
Bidimensionality theory was introduced by Demaine et al. [JACM 2005 ] as a framework to obtain algorithmic results for hard problems on minor closed graph classes. The theory has ...
This paper proposes a random Web crawl model. A Web crawl is a (biased and partial) image of the Web. This paper deals with the hyperlink structure, i.e. a Web crawl is a graph, w...
We describe a new model for studying intermittently connected mobile networks, based on Markovian random temporal graphs, that captures the influence of message size, maximum tol...
John Whitbeck, Vania Conan, Marcelo Dias de Amorim