Follow us on: Like us on Facebook     Follow us on Twitter     View our profile on LinkedIn     Find us on Google+     Visit our blog

Big Data & Hadoop (2.0) Developer Training

Dates: 30th, 31st Jan & 1st feb 2015 at Bangalore

3 Days Instructor Led Hands-On Training

Hurry Register Now!

KnowledgeWorks Event Album: Click Here

Apache Hadoop enables organizations to analyze massive volumes of structured and unstructured data and is currently very hot trend across the software tech industry. Hadoop will be adopted as default enterprise data hub by most of the enterprise soon. Hence Hadoop is being tagged by many as one of the most desired tech skills for 2014 and coming years.

This course will provide you an excellent kick start in building your fundamentals in developing big data solutions using hadoop platform and its ecosystem tools. The course is well balanced between theory and hands-on lab (more than 15 lab exercises) spread on real world uses cases like retail data analysis, sentiment analysis, log analysis, real time trend analysis etc.


What participants will learn?
The attendees will learn below topics through lectures and hands-on exercises
   – Understand Big Data, Hadoop 2.0 architecture and it’s Ecosystem
   – Deep Dive into HDFS and YARN Architecture
   – Writing map reduce algorithms using java APIs
   – Advanced Map Reduce features & Algorithms
   – How to leverage Hive & Pig for structured and unstructured data analysis
   – Data import and export using Sqoop and Flume and create workflows using Oozie
   – Hadoop Best Practices, Sizing and capacity planning
   – Creating reference architectures for big data solutions

Intended Audience: Architects and developers, who wish to write, build and maintain Apache Hadoop jobs.


Course Prerequisites: The participants should have basic knowledge of java, SQL and Linux. It is advised to refresh these skills to obtain maximum benefit from this workshop.

Faculty Profile:
He has about 15+ years of industry experience working on enterprise Java, SOA and Cloud computing platforms. He has worked with TCS, HP, and iGATE Patni and worked on large scale projects for customers like Motorola, Home Depot, CKWB Bank, P&G in the roles of solution and technical architect. He is a freelance who provides consulting and training on Cloud Computing, Big data & Hadoop. He has been teaching Hadoop for 2 years and has trained more than 500 people in Hadoop from large MNCs like EMC, CISCO, HP, YODLEE, YAHOO, SAMSUNG, VeriSign, Success Factors etc.

Course Content:
  What is Big Data & Why Hadoop?

    • Big Data Characteristics, Challenges with traditional system

  Hadoop Overview & it’s Ecosystem

    • Anatomy of Hadoop Cluster, Installing and Configuring Hadoop

    • Setting up hadoop cluster (Single Node)

  HDFS  and YARN

    • HDFS Architecture, Name Nodes, Data Nodes and Secondary Name Node

    • Understanding HDFS HA and Federation architecture

    • YARN Architecture, Resource Manager, Node Manager and Application Master

    • Hands-On Exercise

  Map Reduce Anatomy (MR2)

    • How Map Reduce Works?

    • Writing Mapper, Reducer and Driver using Java APIs,

    • Understanding Hadoop Data Type, Input& Output Formats

    • Hands On Exercises

  Developing Map Reduce Programs

    • Setting up Eclipse Development Environment, Creating Map Reduce Projects, Debugging and Unit Testing

    • Developing a map reduce algorithm on real world scenario

    • Hands On Exercises

  Advanced Map Reduce Concepts

    • Combiner, Partitioner, Counter, Setup and cleanup, Distributed Cache

    • Passing parameters, Multiple Inputs, Chaining multiple jobs

    • Applying Compression, Speculative Execution, Zero Reducers

    • Handling small files and bad records, Handling Binary data like images, documents etc.

    • Map and Reduce Side Joins, data partitioning


  Sqoop & Flume

    • Importing and Exporting data from RDBMS using Sqoop

    • Importing and Exporting data from non-RDBMS sources using Flume

    • Hands On Exercise using Sqoop

  Structured Data Analysis using Hive

    • Hive Architecture, Internal & External Tables, Partitioning, Buckets

    • Writing queries – Joins, Union, Dynamic partitioning, Sampling

    • Writing UDFs, reading different data formats

    • Hands On Exercise

  Semi or Unstructured Data Analysis using Pig

    • Pig Basics, Loading data files

    • Writing queries – SPLIT, FILTER, JOIN, GROUP, SAMPLE, ILLUSTRATE etc.

    • Writing UDFs

    • Hands On Exercise – Tweets Analysis

  Orchestrating data workflows using Oozie

     Understanding Oozie workflow definitions

    • Hands On Exercise – Writing an workflow

  Hadoop Best Practices, Advanced Tips & Techniques

    • Managing HDFS and YARN

    • Hadoop Cluster sizing, capacity planning and optimization

    • Hadoop Deployment options

Fee Details:

Rs. 19,000.00 + 12.36% Service Tax, Per Participant

Subject to availability of seats. Registration is first come first serve basis.

Terms & Conditions


Hurry Register Now!

Time: 09:30am to 05:30pm


Venue Details:

KnowledgeWorks IT Consulting Pvt. Ltd.,

No: 65, Sri Vinayaka Towers, 3rd Floor, 8th B Main,

27th Cross, Jayanagar 4th Block,

Bangalore – 560011

Payment Options:

Account Name: KnowledgeWorks IT Consulting Pvt. Ltd.,

Bank Name: HDFC Bank

Bank  Account Number: 02612020000021

Account Type: Current Account (CA)

Beneficiary Bank Address: Jayanagar Branch, Bangalore

RTGS / NEFT / IFSC Code: HDFC0000261

For any clarifications, Please contact:

Mr. Sudhindra D N: +91 9886221314 | T: +91 80 26630622, 22459941, 41533451

E: | W:

2015. KnowledgeWorks. All rights Reserved