Yelp Data Processing Using Spark And Hive Part 1

Yelp Data Processing Using Spark And Hive Part 1

In this big data project, we will continue from a previous hive project "Data engineering on Yelp Datasets using Hadoop tools" and do the entire data processing using spark.


Each project comes with 2-5 hours of micro-videos explaining the solution.

Code & Dataset

Get access to 50+ solved projects with iPython notebooks and datasets.

Project Experience

Add project experience to your Linkedin/Github profiles.

Customer Love

Read All Reviews

Arvind Sodhi

VP - Data Architect, CDO at Deutsche Bank

I have extensive experience in data management and data processing. Over the past few years I saw the data management technology transition into the Big Data ecosystem and I needed to follow suit. I... Read More

James Peebles

Data Analytics Leader, IQVIA

This is one of the best of investments you can make with regards to career progression and growth in technological knowledge. I was pointed in this direction by a mentor in the IT world who I highly... Read More

What will you learn

Introduction to the dataset and objectives of this project
What are JSON types of file and the data schema of JSON
Reading the Data and transforming it into Hive table
What is Ingestion and How to do Ingestion of data using Spark
How to create data storing and distribution center
Creating customized HDFS and saving data
Various ways of integrating Hive with Spark
Saving a file and building a Hive table on output
What are Normalization and Denormalization and its use
Normalizing and denormalizing dataset into hive tables
Joining different datasets
Various complex data structures in Hive through spark
Transforming the size of a table
Writing customized Query in Hive performing Self joins among tables
Various Complex Data Structure in Hive through Spark
Understanding the arrays and designing the final analysis

Project Description

Data engineering is the science of acquiring, aggregating or collection, processing and storage of data either in batch or in real time as well as providing variety of means of serving these data to other users which could include a data scientist. It involves software engineering practises on big data.

In this big data project for beginners, we will continue from a previous hive project on "Data engineering on Yelp Datasets using Hadoop tools" where we applied some data engineering principles to the Yelp Dataset in the areas of processing, storage and retrieval. Like in that session, We will not include data ingestion since we are already downloading the data from the yelp challenge website. But unlike that session, we will focus on doing the entire data processing using spark.

Similar Projects

In this PySpark project, you will simulate a complex real-world data pipeline based on messaging. This project is deployed using the following tech stack - NiFi, PySpark, Hive, HDFS, Kafka, Airflow, Tableau and AWS QuickSight.

In this hadoop project, we are going to be continuing the series on data engineering by discussing and implementing various ways to solve the hadoop small file problem.

Learn to write a Hadoop Hive Program for real-time querying.

Curriculum For This Mini Project

Introduction to the Yelp dataset
Objectives of this project
Introduction to the JSON schema
Read the data and transform to Hive parquet table
Ingest Json data using Spark
Write to HDFS
Integrate Hive with Spark
Understanding Normalizing and Denormalizing
Normalizing and Denormalizing datasets into Hive tables
Transform the table and write in a single line
Query to find users with more followers than their friends
Error troubleshooting
Initial import of data
Exploring various data structures
Exploring arrays
Designing the analysis