spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ted Yu <yuzhih...@gmail.com>
Subject Re: Is spark suitable for large scale pagerank, such as 200 million nodes, 2 billion edges?
Date Thu, 15 Jan 2015 18:23:04 GMT
Have you seen http://search-hadoop.com/m/JW1q5pE3P12 ?

Please also take a look at the end-to-end performance graph on
http://spark.apache.org/graphx/

Cheers

On Thu, Jan 15, 2015 at 9:29 AM, txw <txw@outlook.com> wrote:

> Hi,
>
>
> I am run PageRank on a large dataset, which include 200 million nodes and
> 2 billion edges?
>
> Is spark suitable for large scale pagerank? How many cores and MEM do I
> need and how long will it take?
>
>
> Thanks
>
>
> Xuewei Tang
>

Mime
View raw message