Adding Custom Schema to Spark Dataframe

Updated On February 12, 2021 | By Mahesh Mogal

In the last blog, we have loaded our data to Spark Dataframe. We have also used "inferschema" option to let spark figure out the schema of the Dataframe on its own. But in many cases, you would like to specify a schema for Dataframe. This will give you much better control over column names and especially data types. Let us see how we can add our custom schema while reading data in Spark.

Adding Custom Schema

In spark, schema is array StructField of type StructType. Each StructType has 4 parameters.

  • Column Name
  • Data type of that column
  • Boolean value indication if values in this column can be null or not
  • Metadata column - this is optional column which can be used to add additional information about column

Let us write our custom schema for our flights data.

Now our data frame has column names and data types which we have specified. If you want to print schema for any dataframe you can use below function.

Using Metadata With Custom Schema

We can add extra information about columns using the metadata filed. This filed takes key-value pairs and we can choose any number of keys and values depending on our needs.

We can check our data frame and its schema now.

custom schema with metadata
Custom schema with Metadata

If you want to check schema with its metadata then we need to use following code. We can read all of schema with this function or also read schema for one column as well.

This is how we can add a custom schema to our dataframes. I hope this helps. See you in the next blog.

Adding Custom Schema to Spark Dataframe

Mahesh Mogal

I am passionate about Cloud, Data Analytics, Machine Learning, and Artificial Intelligence. I like to learn and try out new things. I have started blogging about my experience while learning these exciting technologies.

Stay Updated with Latest Blogs

Get latest blogs delivered to your mail directly.

Recent Posts

Where and Filter in Spark Dataframes

In this blog, we will learn how to filter rows from spark dataframe using Where and Filter functions.

Where and Filter in Spark Dataframes
Read More
Distinct Rows and Distinct Count from Spark Dataframe

Getting distinct values from columns or rows is one of most used operations. We will learn how to get distinct values as well as count of distinct values.

Distinct Rows and Distinct Count from Spark Dataframe
Read More
Sorting in Spark Dataframe

In this blog, we will learn how to sort rows in spark dataframe based on some column values.

Sorting in Spark Dataframe
Read More

Leave a Reply

Your email address will not be published. Required fields are marked *

linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram
Share via
Copy link
Powered by Social Snap