Recap of Hadoop News for July

Recap of Hadoop News for July

News on Hadoop-July 2016

Hadoop News for July

Driven 2.2 allows enterprises to monitor large scale Hadoop and Spark applications. July 4, 2016.

Driven Inc., a leader in Application Performance Monitoring (APM) for big data applications has launched its next version – Driven 2.2. Driven Cloud is a component of Driven 2.2 and it is the first SaaS offering that manages big data applications running on Hadoop as a Service (HaaS)


For the complete list of big data companies and their salaries- CLICK HERE

Hortonworks is pulling up its pants, gets pragmatic in its approach. July 12, 2016.

Hortonworks has come a long way in its 5-year journey as a Hadoop vendor. It is catching up to its rivals – Cloudera and MapR in giving vendor specific products. Hortonworks is revamping its data warehouse optimization use cases through an OEM arrangement with AtScale, Pivotal and Syncsort. They are also reselling AtScale to improve BI query and reporting in Hadoop.


Big Data and Hadoop Training Certification

If you would like more information about Big Data careers, please click the orange "Request Info" button on top of this page.

Wayne State U Deploys Analytics with SAS and Hadoop. July 14, 2016.

With the intent to improve the reporting system for student data and finance, Wayne State University in Michigan has implemented a SAS Visual Analytics platform with Hadoop. The new big data analytics platform built on Hadoop will be used by the staff and administrators to analyse information like student performance trends, revenue, admission trends and spending to facilitate smart decision-making.

(Source: )

Commvault stomps out with protection for Hadoop, Big Data analytics systems. July 15,

Commvault, a bridge between product systems and data aims to provide data protection environment on Hadoop. Commvault provides analytics needed to access data on a large group of servers running hadoop jobs with each server having its own stored part of the overall dataset. Commvault envisages an agent running on each server that sends data to be stored in a single central target, or on the cloud or on-premises.

(Source: )

Global Hadoop market predictions show positive growth till 2015. July 18, 2016.

As per Statistics MRC, the Global Hadoop market is expected to increase at a CAGR of 53.7%. This growth is augmented by the rapid increase of structured and unstructured data, consumer consumption data and the growing need of Big Data analytics.


Securing the World with Hadoop. July 20, 2016.

The leading provider of cyber security solutions, Dell SecureWorks uses Hadoop Ecosystem to process 185 billion logs per day with data volume growing at a rate of 66% every year. Dell SecureWorks plans to adopt newest technologies in the growing hadoop ecosystem to help customers stay secure by using HBase to enhance enrichment capabilities, Kafka for messaging, Spark MLib for machine learning and Spark Streaming for stream processing.

(Source: )

Teradata bolsters Hadoop consulting biz with new UK acquisition. July 25, 2016.

 Teradata Corp. acquired a UK based consultancy named “Big Data Partnership Ltd.” For an undisclosed amount. Big Data Partnership firm is strong enough to take on analytics initiative any stage of development whether it is  a production hadoop cluster in need of modernization or a half-baked internal demo. Teradata plans to integrate Big Data Partnership capabilities into its Think Big professional services to provide support and employee training so that they make the best out of the data infrastructure.

(Source: )

Cloudera provides an easy path for Analytics on Hadoop. July 26, 2016.

Cloudera one of the leading Hadoop vendors, has released Cloudera Navigator Optimizer and Cloudera Enterprise 5.8 to make analytics on Hadoop easier. Anyone looking to augment their database platform with Hadoop or optimize their analytics solutions can use these tools to do so.


Big Data Certification and Hadoop Training




Relevant Projects

Data Warehouse Design for E-commerce Environments
In this hive project, you will design a data warehouse for e-commerce environments.

Real-Time Log Processing in Kafka for Streaming Architecture
The goal of this apache kafka project is to process log entries from applications in real-time using Kafka for the streaming architecture in a microservice sense.

Yelp Data Processing using Spark and Hive Part 2
In this spark project, we will continue building the data warehouse from the previous project Yelp Data Processing Using Spark And Hive Part 1 and will do further data processing to develop diverse data products.

Hive Project - Visualising Website Clickstream Data with Apache Hadoop
Analyze clickstream data of a website using Hadoop Hive to increase sales by optimizing every aspect of the customer experience on the website from the first mouse click to the last.

Airline Dataset Analysis using Hadoop, Hive, Pig and Impala
Hadoop Project- Perform basic big data analysis on airline dataset using big data tools -Pig, Hive and Impala.

Design a Hadoop Architecture
Learn to design Hadoop Architecture and understand how to store data using data acquisition tools in Hadoop.

Online Hadoop Projects -Solving small file problem in Hadoop
In this hadoop project, we are going to be continuing the series on data engineering by discussing and implementing various ways to solve the hadoop small file problem.

Data Mining Project on Yelp Dataset using Hadoop Hive
Use the Hadoop ecosystem to glean valuable insights from the Yelp dataset. You will be analyzing the different patterns that can be found in the Yelp data set, to come up with various approaches in solving a business problem.

Implementing Slow Changing Dimensions in a Data Warehouse using Hive and Spark
Hive Project- Understand the various types of SCDs and implement these slowly changing dimesnsion in Hadoop Hive and Spark.

Tough engineering choices with large datasets in Hive Part - 2
This is in continuation of the previous Hive project "Tough engineering choices with large datasets in Hive Part - 1", where we will work on processing big data sets using Hive.