[Top][All Lists]

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
## [igraph] betweenness running time estimate and

**From**: |
David Hunkins |

**Subject**: |
[igraph] betweenness running time estimate and |

**Date**: |
Sun, 13 Jul 2008 14:16:13 -0700 |

Thank you Gabor, Tamas and others very much for the igraph library.

`I am analyzing some online photo sharing data donated to me by a large
``photo sharing site.
`

`In my first dataset I have about 5 million vertices and 5 million
``edges. In the next dataset I'll get about 50 million vertices and 200
``million edges. I am currently running analysis on a 2GB/2.0GHz Intel
``Core Duo Macbook Pro. I am conditioning and loading the weighted,
``directed graph data in Pajek format.
`

`A couple of questions with which your user community may have some
``experience.
`

`First, I am having a trouble determining how long I should expect
``various operations to take even with the 5M/5M dataset, and I am
``wondering whether there is a simple metric (such as density of the
``graph) that can be computed and then used to calculate order-of-
``magnitude running times for the various other functions of interest.
``Yesterday, for example I ran betweenness.estimate(graph, cutoff=2) and
``it hasn't completed after 24 hours (and the Macbook is not paging at
``all, just getting a little 'warm').
`

`Second, do you know anybody who has used commercial (or otherwise)
``scalable computing resources to run analyses on datasets of this size?
``I fear that, even if I can melt my laptops to solve the problem for
``the 5M/5M dataset, I will need other resources when I go to 50M/200M
``dataset.
`
Thanks for any assistance you and the community can provide. Kösz!
David Hunkins

**[igraph] betweenness running time estimate and**,
*David Hunkins* **<=**