Ben Chuanlong Du's Blog

It is never too late to learn.

Spark Issue: Max Number of Executor Failures Reached

Things on this page are fragmentary and immature notes/thoughts of the author. Please read with your own judgement!

Symptom

21/06/01 15:03:28 INFO ApplicationMaster: Final app status: FAILED, exitCode: 11, (reason: Max number of executor failures (6) reached)

Possible Causes

The option spark.yarn.max.executor.failures is set to a value which is too small. In my case, I believe the Hadoop team misconfigured the option when they update the version of Spark.

Possible Solutions

By default, spark.yarn.max.executor.failures is set to \(numExecutors \times 2\). A simple fix is to manually set spark.yarn.max.executor.failures to \(numExecutors \times 2\).

Comments