Specifying default configuration overrides for Spark ensures your clusters are consistent and large jobs will not fail. Amazon Web Services pro Frank Kane shows you how to set these overrides for your Apache Spark/Elastic MapReduce (EMR) cluster.

Learn more about running Spark for big data analysis on the Amazon Elastic MapReduce service (EMR) with video training from Frank Kane.

Article image: Screenshot from "How do I configure Apache Spark on an Amazon Elastic MapReduce (EMR) cluster?" (source: O'Reilly).