diff --git a/core/src/main/scala/org/apache/spark/internal/config/package.scala b/core/src/main/scala/org/apache/spark/internal/config/package.scala index 324ef701c4266..6d51424f0baff 100644 --- a/core/src/main/scala/org/apache/spark/internal/config/package.scala +++ b/core/src/main/scala/org/apache/spark/internal/config/package.scala @@ -1023,8 +1023,7 @@ package object config { private[spark] val MAX_EXECUTOR_FAILURES = ConfigBuilder("spark.executor.maxNumFailures") .doc("The maximum number of executor failures before failing the application. " + - "This configuration only takes effect on YARN, or Kubernetes when " + - "`spark.kubernetes.allocation.pods.allocator` is set to 'direct'.") + "This configuration only takes effect on YARN and Kubernetes.") .version("3.5.0") .intConf .createOptional @@ -1032,8 +1031,8 @@ package object config { private[spark] val EXECUTOR_ATTEMPT_FAILURE_VALIDITY_INTERVAL_MS = ConfigBuilder("spark.executor.failuresValidityInterval") .doc("Interval after which executor failures will be considered independent and not " + - "accumulate towards the attempt count. This configuration only takes effect on YARN, " + - "or Kubernetes when `spark.kubernetes.allocation.pods.allocator` is set to 'direct'.") + "accumulate towards the attempt count. This configuration only takes effect on YARN " + + "and Kubernetes.") .version("3.5.0") .timeConf(TimeUnit.MILLISECONDS) .createOptional diff --git a/docs/configuration.md b/docs/configuration.md index 6957ca9a03d23..4a85c4f256a95 100644 --- a/docs/configuration.md +++ b/docs/configuration.md @@ -565,8 +565,7 @@ of the most common options to set are: numExecutors * 2, with minimum of 3 The maximum number of executor failures before failing the application. - This configuration only takes effect on YARN, or Kubernetes when - spark.kubernetes.allocation.pods.allocator is set to 'direct'. + This configuration only takes effect on YARN and Kubernetes. 3.5.0 @@ -576,8 +575,7 @@ of the most common options to set are: Interval after which executor failures will be considered independent and not accumulate towards the attempt count. - This configuration only takes effect on YARN, or Kubernetes when - spark.kubernetes.allocation.pods.allocator is set to 'direct'. + This configuration only takes effect on YARN and Kubernetes. 3.5.0