PySpark SparkConf – Attributes and Applications

1. Objective

In our last Pyspark tutorial, we saw Pyspark Serializers. Today, we will discuss PySpark SparkConf. Moreover, we will see attributes in PySpark SparkConf and running Spark Applications. Also, we will learn PySpark SparkConf example. As we need to set a few configurations and parameters, to run a Spark application on the local/cluster for that we use SparkConf. So, to learn to run SparkConf using PySpark, this document will help. 
So, let’s start PySpark SparkConf.

PySpark SparkConf

PySpark SparkConf – Attributes and Applications

Test how much you learned about PySpark 

2. What is PySpark SparkConf?

We need to set a few configurations and parameters, to run a Spark application on the local/cluster, this is what SparkConf helps with. Basically,  to run a Spark application, it offers configurations.

  • Code

For PySpark, here is the code block which has the details of a SparkConf class:

class pyspark.SparkConf (
 loadDefaults = True,
 _jvm = None,
 _jconf = None

Basically, with SparkConf() we will create a SparkConf object first. So, that will load the values from spark. Even Java system properties. Hence, by using the SparkConf object, now we can set different parameters and their parameters will take priority over the system properties.
However, there are better methods, which support chaining, in a SparkConf class. Let’s say, we can write conf.setAppName(“PySpark App”).setMaster(“local”). Though, it cannot be modified by any user once we pass a SparkConf object to Apache Spark.
Have a look at 5 best PySpark books

If these professionals can make a switch to Big Data, so can you:
Rahul Doddamani Story - DataFlair
Rahul Doddamani
Java → Big Data Consultant, JDA
Follow on
Mritunjay Singh Success Story - DataFlair
Mritunjay Singh
PeopleSoft → Big Data Architect, Hexaware
Follow on
Rahul Doddamani Success Story - DataFlair
Rahul Doddamani
Big Data Consultant, JDA
Follow on
I got placed, scored 100% hike, and transformed my career with DataFlair
Enroll now
Deepika Khadri Success Story - DataFlair
Deepika Khadri
SQL → Big Data Engineer, IBM
Follow on
DataFlair Web Services
You could be next!
Enroll now

3. Attributes of PySpark SparkConf

Thus here are the most commonly used attributes of SparkConf:

Attributes of PySpark SparkConf

Attributes of PySpark SparkConf

i. set(key, value)

It helps to set a configuration property.

ii. setMaster(value)

In order to set the master URL, we use it.

Join DataFlair on Telegram

iii. setAppName(value)

We use it to set an application name.

iv. get(key, defaultValue=None)

It helps to get a configuration value of a key.
Read PySpark Broadcast and Accumulator With Examples

v. setSparkHome(value)

In order to set Spark installation path on worker nodes, we use it.
In the following code, we can use to create SparkConf and SparkContext objects as part of our applications. Also, using sbt console on base directory of our application we can validate:

from pyspark import SparkConf,SparkContext
conf = SparkConf().setAppName("Spark Demo").setMaster("local")
sc = SparkContext(conf=conf)

4. Running Spark Applications Using SparkConf

In addition, here are some different contexts in which we can run spark applications:

  • local – conf

SparkConf.setAppName(“Spark Demo”).setMaster(“local”)

  • yarn-client – conf

SparkConf.setAppName(“Spark Demo”).setMaster(“yarn-client”)

  • mesos URL
  • spark URL – conf

SparkConf.setAppName(“Spark Demo”).setMaster(“spark master URL”)

  • Code snippet to get all the properties

for i in sc.getConf.getAll: print(i)
So, this was all about Pyspark SparkConf. Hope you like our explanation.
Explore PySpark Pros and Cons

5. Conclusion

Hence, we have learned all about PySpark SparkConf, including its code which will help to create one. Moreover, we discussed different attributes of PySpark SparkConf and also running Spark applications. Still, if any doubt, comment below.
See also –
PySpark RDD with Operations and Commands
For reference

Leave a Reply

Your email address will not be published. Required fields are marked *

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.