It's good 
Thanks for your reply Michael.

On Thu, Aug 20, 2015 at 11:03 PM, Michael Armbrust <michael@databricks.com> wrote:
We will probably fix this in Spark 1.6

https://issues.apache.org/jira/browse/SPARK-10040

On Thu, Aug 20, 2015 at 5:18 AM, Aram Mkrtchyan <aram.mkrtchyan.87@gmail.com> wrote:
We want to migrate our data (approximately 20M rows) from parquet to postgres, when we are using dataframe writer's jdbc method the execution time is very large,  we have tried the same with batch insert it was much effective. 
Is it intentionally implemented in that way?