hive-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "David Mollitor (JIRA)" <>
Subject [jira] [Updated] (HIVE-21466) Increase Default Size of SPLIT_MAXSIZE
Date Tue, 19 Mar 2019 01:59:00 GMT


David Mollitor updated HIVE-21466:
    Attachment: HIVE-21466.1.patch

> Increase Default Size of SPLIT_MAXSIZE
> --------------------------------------
>                 Key: HIVE-21466
>                 URL:
>             Project: Hive
>          Issue Type: Improvement
>          Components: Configuration
>    Affects Versions: 4.0.0, 3.2.0
>            Reporter: David Mollitor
>            Assignee: David Mollitor
>            Priority: Minor
>         Attachments: HIVE-21466.1.patch, HIVE-21466.1.patch
> {code:java}
>  MAPREDMAXSPLITSIZE(FileInputFormat.SPLIT_MAXSIZE, 256000000L, "", true),
> {code}
> []
> This field specifies a maximum size for each MR (maybe other?) splits.
> This number should be a multiple of the HDFS Block size. The way that this maximum is
implemented, is that each block is added to the split, and if the split grows to be larger
than the maximum allowed, the split is submitted to the cluster and a new split is opened.
> So, imagine the following scenario:
>  * HDFS block size of 16 bytes
>  * Maximum size of 40 bytes
> This will produce a split with 3 blocks. (2x16) = 32; another block will be inserted,
(3x16) = 48 bytes in the split. So, while many operators would assume a split of 2 blocks,
the actual is 3 blocks. Setting the maximum split size to a multiple of the HDFS block size
will make this behavior less confusing.
> The current setting is ~256MB and when this was introduced, the default HDFS block size
was 64MB. That is a factor of 4x. However, now HDFS block sizes are 128MB by default, so I
propose setting this to 4x128MB.  The larger splits (fewer tasks) should give a nice performance
boost for modern hardware.

This message was sent by Atlassian JIRA

View raw message