spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Rishi Shah <rishishah.s...@gmail.com>
Subject [pyspark 2.4.0] write with partitionBy fails due to file already exits
Date Mon, 01 Jul 2019 11:17:58 GMT
Hi All,

I have a simple partition write like below:

df = spark.read.parquet('read-location')
df.write.partitionBy('col1').mode('overwrite').parquet('write-location')

this fails after an hr with "file already exists (in .staging directory)"
error. Not sure what am I doing wrong here..

-- 
Regards,

Rishi Shah

Mime
View raw message