spark-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Apache Spark (Jira)" <j...@apache.org>
Subject [jira] [Assigned] (SPARK-30821) Executor pods with multiple containers will not be rescheduled unless all containers fail
Date Fri, 01 May 2020 16:50:02 GMT

     [ https://issues.apache.org/jira/browse/SPARK-30821?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Apache Spark reassigned SPARK-30821:
------------------------------------

    Assignee: Apache Spark

> Executor pods with multiple containers will not be rescheduled unless all containers
fail
> -----------------------------------------------------------------------------------------
>
>                 Key: SPARK-30821
>                 URL: https://issues.apache.org/jira/browse/SPARK-30821
>             Project: Spark
>          Issue Type: Improvement
>          Components: Kubernetes
>    Affects Versions: 3.1.0
>            Reporter: Kevin Hogeland
>            Assignee: Apache Spark
>            Priority: Major
>
> Since the restart policy of launched pods is Never, additional handling is required for
pods that may have sidecar containers. The executor should be considered failed if any containers
have terminated and have a non-zero exit code, but Spark currently only checks the pod phase.
The pod phase will remain "running" as long as _any_ pods are still running. Kubernetes sidecar
support in 1.18/1.19 does not address this situation, as sidecar containers are excluded from
pod phase calculation.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org


Mime
View raw message