hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Kihwal Lee (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HADOOP-12861) RPC client fails too quickly when server connection limit is reached
Date Tue, 01 Mar 2016 19:47:18 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-12861?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15174305#comment-15174305

Kihwal Lee commented on HADOOP-12861:

Due to the way the IPC Client does SASL negotiation, once a connection is established, an
immediately following connection reset is handled by {{handleSaslConnectionFailure()}}.  This
will throw an {{IOException}}, which {{FailoverOnNetworkExceptionRetry.shouldRetry}} prescribes
{{RetryAction.FAILOVER_AND_RETRY}} for, without any delay. The client can burn though retries
rather quickly and have a permanent failure.

> RPC client fails too quickly when server connection limit is reached
> --------------------------------------------------------------------
>                 Key: HADOOP-12861
>                 URL: https://issues.apache.org/jira/browse/HADOOP-12861
>             Project: Hadoop Common
>          Issue Type: Bug
>          Components: ipc
>    Affects Versions: 2.7.0
>            Reporter: Daryn Sharp
>            Assignee: Daryn Sharp
> The NN's rpc server immediately closes new client connections when a connection limit
is reached. The client rapidly retries a small number of times with no delay which causes
clients to fail quickly. If the connection is refused or timedout, the connection retry policy
tries with backoff. Clients should treat a reset connection as a connection failure so the
connection retry policy is used.

This message was sent by Atlassian JIRA

View raw message