How to optimize shuffle spill in Apache Spark application
Learning to performance-tune Spark requires quite a bit of investigation and learning. There are a few good resources including this video. Spark 1.4 has some better diagnostics and visualisation in the interface which can help you.
In summary, you spill when the size of the RDD partitions at the end of the stage exceed the amount of memory available for the shuffle buffer.
You can:
- Manually
repartition()
your prior stage so that you have smaller partitions from input. - Increase the shuffle buffer by increasing the memory in your executor processes (
spark.executor.memory
) - Increase the shuffle buffer by increasing the fraction of executor memory allocated to it (
spark.shuffle.memoryFraction
) from the default of 0.2. You need to give backspark.storage.memoryFraction
. - Increase the shuffle buffer per thread by reducing the ratio of worker threads (
SPARK_WORKER_CORES
) to executor memory
If there is an expert listening, I would love to know more about how the memoryFraction settings interact and their reasonable range.
To add to the above answer, you may also consider increasing the default number (spark.sql.shuffle.partitions) of partitions from 200 (when shuffle occurs) to a number that will result in partitions of size close to the hdfs block size (i.e. 128mb to 256mb)
If your data is skewed, try tricks like salting the keys to increase parallelism.
Read this to understand spark memory management:
https://0x0fff.com/spark-memory-management/
https://www.tutorialdocs.com/article/spark-memory-management.html