spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Alexander Czech <>
Subject Loading a large parquet file how much memory do I need
Date Mon, 27 Nov 2017 09:56:29 GMT
I want to load a 10TB parquet File from S3 and I'm trying to decide what
EC2 instances to use.

Should I go for instances that in total have a larger memory size than
10TB? Or is it enough that they have in total enough SSD storage so that
everything can be spilled to disk?


View raw message