Course level

Intermediate

Duration

Last update

April 14, 2021


About the author

Suyog Nagaokar


Suyog has 8+ years experience in Data Engineering, providing automated and optimized solutions to businesses based on Hadoop, Spark and Streaming frameworks, thus helping them generate value from data.

He has experience with Telecom and Banking domains focusing on Customer Genomics, Fraud Analytics, Digital Banking and Machine Learning for Telecom.

Suyog has also mentored industry professionals with 0-15 years experience and Engineering students on Big Data in renowned institutes like EduPristine, IMS ProSchool and DataVision.

Connect with Suyog on LinkedIn

Description

Hadoop tutorial with HDFS, YARN, MapReduce, Hive and Sqoop!


What will I learn?

In this comprehensive course, you will learn and master the Hadoop architecture and its components like HDFS, YARN, MapReduce, Hive and Sqoop!

With this Big Data training, you'll understand the detailed concepts of Hadoop Eco-system along with hands-on labs and learn how to use each component to solve real business problems! 

  • Understand Hadoop architecture and commands
  • Install and work with a real Hadoop installation right on your desktop with Cloudera Quickstart VM
  • Store and query your data with Sqoop, Hive, MySQL.
  • Write Hive queries to analyze data on Hadoop


By the end of the course you will understand:
  • Hadoop architecture and commands
  • HDFS, YARN, MapReduce frameworks
  • Concepts and practical implementation of Sqoop, Hive and MySQL


And you will be able to:
  • Write Hadoop commands.
  • Manage big data on a cluster with HDFS and MapReduce
  • Work with a real Hadoop installation right on your desktop with Cloudera Quickstart VM.
  • Store and query data with Sqoop, Hive, MySQL.
  • Write Hive queries to analyze data on Hadoop.
  • Learn how your cluster is managed with YARN and Hue


What's included?
  • Videos
  • Source Codes
  • Web Links
  • Certificate of Graduation for successful completion of the course


Requirements
  • Access to a PC running 64-bit Windows or Linux with an Internet connection
  • At least 8GB of *free* (not total) RAM, if you want to participate in the hands-on activities and exercises. If your PC does not meet these requirements, you can still follow along in the course without doing hands-on activities.
  • Some activities will require some prior programming experience, preferably in Python.
  • A basic familiarity with the Linux command line will be very helpful.




Course Curriculum


  What is Big Data and Hadoop (Coming soon)
Available in days
days after you enroll
  Hadoop Distributions and Setup (Coming soon)
Available in days
days after you enroll
  Data Warehousing with Apache Hive (Coming soon)
Available in days
days after you enroll
  Import/Export Data with Apache Sqoop (Coming soon)
Available in days
days after you enroll