After a brief summer break, I started exploring the Spark UI. The first steps I took are exploring changing Spark configuration parameters such as spark.sql.shuffle.partitions.

I can update the default value by adding a config() statement in the Scala code:

val spark = SparkSession
    .builder()
    .appName("optimization")
    .config("spark.sql.shuffle.partitions", "300")  // <-- Here
    .master("local[*]")
    .getOrCreate()

Querying the value of this parameters, shows that it’s indeed 300.