Beginner
160 hours
80 hours
Course Description
What is Big Data ?. Gartner defines Big Data as high volume,velocity and variety information assets that demand cost-effective,innovative forms of enhanced insight and decision making. According to IBM, 80% of data captured today is unstructured,from sensors used to gather climate information,posts to social media sites,digital pictures and videos,purchase transaction records and cell phone GPS signals,to name a few. All of this unstructured data is also Big Data.
What you'll learn ? In this course we will go through BigData use case examples and also learn how to manage BIGDATA using Hadoop and related technologies. We will learn how to install and configure a single-node Hadoop cluster and perform Hadoop Distributed File System (HDFS) and Hadoop Map-Reduce operations.
Hadoop Essentials teaches the fundamentals of setting up a Hadoop cluster, as well as the "soup" of related technologies like Hive, Pig and Oozie. In addition, you will learn how to write MapReduce programs using Java. Learn how to use Apache Spark as an alternative to traditional MapReduce processing
This BigData training will be suitable for software developers, architects, IT Services, deployment engineers, IT Support and development managers.
Course Prerequisites
NA.
Course Includes