Each project comes with 2-5 hours of micro-videos explaining the solution.
Get access to 50+ solved projects with iPython notebooks and datasets.
Add project experience to your Linkedin/Github profiles.
Recently I became interested in Hadoop as I think its a great platform for storing and analyzing large structured and unstructured data sets. The experts did a great job not only explaining the... Read More
I have extensive experience in data management and data processing. Over the past few years I saw the data management technology transition into the Big Data ecosystem and I needed to follow suit. I... Read More
Clicksteam data records the flow or trail of a user when he/she visits a website. For example, if you have pages A-Z and want to see how many people land on Page G and then go to Page B - you can analyze this data and see the clickstream pattern of your visitors. This data is stored in semi structured web logs. Often you will hear the term web log analysis - this is the same as analyzing clickstream data. Segmenting, and analyzing this clickstream data will give you a more refined look at your customer's behavior patterns - from the time they land on your website till the time they either buy your product or leave without buying.
You have built a wonderful website and your transaction page has all the information that is required for someone to know before buying the product. Still you see that a huge number of your website visitors leave before buying a single product. This is because of a broken link or path somewhere which prevents users to quickly and easily buy your product. Hadoop helps you to extract, store and analyze the clickstream data or web log data and merge it with the traditional customer data - in order to get better insights into the behavior of the visitor and optimize the path to product buying. Hive is the easiest of the Hadoop tools to learn. If you are from a data warehousing background and know SQL well - it will be a breeze to work on Hive. Hive is a data warehouse infrastructure built on top of Hadoop and is quite versatile in its usage, as it supports different storage types such as plain text, RCFile, Amazon S3, HBase, ORC, etc. Hive has its own SQL like language called HiveQL with schemas - which transparently converts queries to MapReduce or Apache Spark jobs.
You will be working on solving these business problems for the end-user in this Hadoop Hive Project:
Optimizing the click through path of the users
Which is the most optimum path for a user to follow in order to buy the product?
After how many clicks does a user lose interest to buy a product?
Which products do users usually buy together?
Where can the website resources be allocated to provide the best user experience to a visitor to make him return again?
In this PySpark project, you will simulate a complex real-world data pipeline based on messaging. This project is deployed using the following tech stack - NiFi, PySpark, Hive, HDFS, Kafka, Airflow, Tableau and AWS QuickSight.
In this big data project, we will look at how to mine and make sense of connections in a simple way by building a Spark GraphX Algorithm and a Network Crawler.
In this project, we will walk through all the various classes of NoSQL database and try to establish where they are the best fit.