Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Spark Issue: Max Number of Executor Failures Reached

Things on this page are fragmentary and immature notes/thoughts of the author. Please read with your own judgement!

Symptom

21/06/01 15:03:28 INFO ApplicationMaster: Final app status: FAILED, exitCode: 11, (reason: Max number of executor failures (6) reached)

Possible Causes

The option spark.yarn.max.executor.failures is set to a value which is too small. In my case, I believe the Hadoop team misconfigured the option when they update the version of Spark.

Possible Solutions

By default, spark.yarn.max.executor.failures is set to numExecutors×2numExecutors \times 2. A simple fix is to manually set spark.yarn.max.executor.failures to numExecutors×2numExecutors \times 2.