Big Data using Hadoop

Duration: 3 Days

Training Fee: RM 2650.00

HRDF Claimable

About this course

In this Big Data training attendees will gain practical skill set on Hadoop in detail, including its core and eco system components. This course focuses on case study approach for learning various tools and completely industry relevant training and a great blend of analytics and technology.

Candidates will be awarded Big Data Analytics using Hadoop Attendance Certificate on successful completion of projects that are provided as part of the training.

After the completion of this course, you will be able to:

  • Knowing BigData
  • Drawing BigData ideas towards your company
  • Work on the concepts of HDFS and MapReduce framework.
  • Learn data loading techniques
  • Perform data analytics using Hive and YARN.
  • Implement Spark applications on YARN (Hadoop).
  • Stream data and Streaming API.
  • Analyze Hive and Spark SQL architecture.
  • Implement Spark SQL queries to perform several computations.

Target Audience

Software engineers and programmers who want to understand the BigData and its concept with larger Hadoop ecosystem, and use it to store, analyze, and vend “big data” at scale. Project, program, or product managers who want to understand the lingo and high-level architecture of Hadoop. Data analysts and database administrators who are curious about Hadoop and how it relates to their work. System architects who need to understand the components available in the Hadoop ecosystem, and how they fit together.

Prerequisites

All participants must be familiar with the fundamentals of Programming and Web technologies a basic familiarity with the Linux command line will be very helpful.

Course Content

Introduction and Understanding Big Data

  • What is Big Data Definition
  • The Network System and Tick Bandwidth
  • LAN and W AN Data flow Data packet and block
  • Server Clustering

 

The Hadoop & Ecosystem Installation

  • Install Hadoop
  • Hadoop Overview and History
  • The Ecosystem

 

Hadoop core components- HDFS

  • HDFS: What it is, and how it works
  • Install the Sample dataset into HDFS using the UI
  • Install the Sample dataset into HDFS using the command line.

 

Hadoop core components- Mapreduce (YARN)

  • MapReduce: What it is, and how it works
  • How MapReduce distributes processing

 

Hadoop Data Analysis Tools: Hadoop-Spark

  • Why Spark? The Resilient Distributed Dataset (RDD)?
  • Dataset and Spark.

 

Hadoop Data Analysis Tools: Hadoop-Hive

  • What is Hive? How Hive Works?
  • MySQL and Hadoop Integration
  • Hadoop Projects and Exercise

Training Methods

Public Class Training

Private Class Training

In-House Class Training

Schedule
10 - 12 Jan 2019
7 - 9 Mar 2019
9 - 11 May 2019
11 - 13 July 2019
5 - 7 Sept 2019
7 - 9 Nov 2018

Please contact us if you need more information about Private or In-House Class Training – click here

Training Methods

  • Public Class Training

  • Your Information

  • Participant List

  • NameContact NumberDesignation 
    Add a new row

Any Questions?

  • Contact Us

Call Us : 03-21165778