Do you need to understand big data and how it will impact your business? This Specialization is for you. You will gain an understanding of what insights big data can provide through hands-on experience with the tools and systems used by big data scientists and engineers. Previous programming experience is not required! You will be guided through the basics of using Hadoop with MapReduce, Spark, Pig and Hive. By following along with provided code, you will experience how one can perform predictive modeling and leverage graph analytics to model problems. This specialization will prepare you to ask the right questions about data, communicate effectively with data scientists, and do basic exploration of large, complex datasets. In the final Capstone Project, developed in partnership with data software company Splunk, you’ll apply the skills you learned to do basic analyses of big data.
What Will I Learn?
Understand different technology trends, salary trends, Big Data market and different job roles in Big Data
Understand what Hadoop is for, and how it works
Understand complex architectures of Hadoop and its component
Hadoop installation on your machine
Understand how MapReduce, Hive and Pig can be used to analyze big data sets
High quality documents
Demos: Running HDFS commands, Hive queries, Pig queries
Sample data sets and scripts (HDFS commands, Hive sample queries, Pig sample queries, Data Pipeline sample queries)
Start writing your own codes in Hive and Pig to process huge volumes of data
Design your own data pipeline using Pig and Hive
Understand modern data architecture: Data Lake
Practice with Big Data sets