@@ -24,16 +24,16 @@ private[spark] object Config extends Logging {
24
24
25
25
val KUBERNETES_NAMESPACE =
26
26
ConfigBuilder (" spark.kubernetes.namespace" )
27
- .doc(" The namespace that will be used for running the driver and executor pods. When using" +
28
- " spark-submit in cluster mode, this can also be passed to spark-submit via the" +
29
- " --kubernetes-namespace command line argument." )
27
+ .doc(" The namespace that will be used for running the driver and executor pods. When using " +
28
+ " spark-submit in cluster mode, this can also be passed to spark-submit via the " +
29
+ " --kubernetes-namespace command line argument." )
30
30
.stringConf
31
31
.createWithDefault(" default" )
32
32
33
33
val EXECUTOR_DOCKER_IMAGE =
34
34
ConfigBuilder (" spark.kubernetes.executor.docker.image" )
35
- .doc(" Docker image to use for the executors. Specify this using the standard Docker tag" +
36
- " format." )
35
+ .doc(" Docker image to use for the executors. Specify this using the standard Docker tag " +
36
+ " format." )
37
37
.stringConf
38
38
.createOptional
39
39
@@ -56,10 +56,10 @@ private[spark] object Config extends Logging {
56
56
57
57
val KUBERNETES_SERVICE_ACCOUNT_NAME =
58
58
ConfigBuilder (s " $APISERVER_AUTH_DRIVER_CONF_PREFIX.serviceAccountName " )
59
- .doc(" Service account that is used when running the driver pod. The driver pod uses" +
60
- " this service account when requesting executor pods from the API server. If specific" +
61
- " credentials are given for the driver pod to use, the driver will favor" +
62
- " using those credentials instead." )
59
+ .doc(" Service account that is used when running the driver pod. The driver pod uses " +
60
+ " this service account when requesting executor pods from the API server. If specific " +
61
+ " credentials are given for the driver pod to use, the driver will favor " +
62
+ " using those credentials instead." )
63
63
.stringConf
64
64
.createOptional
65
65
@@ -68,9 +68,9 @@ private[spark] object Config extends Logging {
68
68
// based on the executor memory.
69
69
val KUBERNETES_EXECUTOR_MEMORY_OVERHEAD =
70
70
ConfigBuilder (" spark.kubernetes.executor.memoryOverhead" )
71
- .doc(" The amount of off-heap memory (in megabytes) to be allocated per executor. This" +
72
- " is memory that accounts for things like VM overheads, interned strings, other native" +
73
- " overheads, etc. This tends to grow with the executor size. (typically 6-10%)." )
71
+ .doc(" The amount of off-heap memory (in megabytes) to be allocated per executor. This " +
72
+ " is memory that accounts for things like VM overheads, interned strings, other native " +
73
+ " overheads, etc. This tends to grow with the executor size. (typically 6-10%)." )
74
74
.bytesConf(ByteUnit .MiB )
75
75
.createOptional
76
76
@@ -117,7 +117,7 @@ private[spark] object Config extends Logging {
117
117
.intConf
118
118
.checkValue(value => value > 0 , " Maximum attempts of checks of executor lost reason " +
119
119
" must be a positive integer" )
120
- .createWithDefault(5 )
120
+ .createWithDefault(10 )
121
121
122
122
val KUBERNETES_NODE_SELECTOR_PREFIX = " spark.kubernetes.node.selector."
123
123
}
0 commit comments