20130328

Big Data & Hadoop Administrator Training


Dates: 6th & 7th Apr 2013 at Bangalore
2 Days Hands-On Instructor Led Training












Overview:
Apache Hadoop,
 the open source data management software that helps organizations analyze massive volumes of structured and unstructured data, is a very hot topic across the tech industry. Employed by such big named websites as eBay, Facebook, and Yahoo, Hadoop is being tagged by many as one of the most desired tech skills for 2012 and coming years along with Cloud Computing.
What participants will learn?
The attendees will learn below topics through lectures and hands-on exercises
  – Understand Big Data & Hadoop Ecosystem
  – Hadoop Distributed File System – HDFS
  – Planning, designing and deploying a fully distributed cluster
  – Managing and monitoring HDFS, Map Reduce Components
  – Periodic and regular maintenance activities for the cluster
  – Best practices for cluster optimization
  – Diagnosing, tuning and solving Apache Hadoop issues
  – Populating HDFS using Sqoop
Intended Audience:
Architects, developers, administrators who wish to design, deploy and manage Hadoop clusters.
Course Prerequisites:
The participants should have basic understanding or knowledge of Linux.
Faculty Profile:
He has about 14+ years of industry experience working on enterprise java, SOA and Cloud computing platforms. He has worked with TCS, HP, Patni and worked on large scale projects for customers like Motorola, Home Depot, CKWB Bank, P&G in the roles of solution and technical architect. He provides consulting and training on Cloud Computing, Big data & Hadoop, Google App Engine, and Amazon Web Services.
Course Outline:
What is Big Data & Why Hadoop?
• Big Data Characteristics, Challenges with traditional system
Hadoop Overview & it’s Ecosystem
• HDFS – Hadoop Distributed File System
• Name Nodes, Secondary Name Nodes and Data Nodes
• Map Reduce Anatomy - How Map Reduce Works?
• Job Tracker, Task Tracker
• Planning and designing a Hadoop Cluster
Hands-On Exercise – Setting up a distributed Hadoop Cluster ( 3 Nodes )
Hands-On Exercise – Basic HDFS Operations
Populating HDFS from RDBMS
Hands On Exercises - Using Sqoop for importing and exporting data
Managing and Scheduling jobs
• Starting and stopping Map Reduce jobs
• Overview of various schedulers for scheduling jobs
• FIFO, Fair and Capacity Scheduler
Hands On Exercises – Managing and scheduling jobs
Managing HDFS
• Understand Name Node and Secondary Name Node Files
• Checking HDFS health
• Rebalancing nodes in cluster
• Backing up name node metadata
• Commissioning and decommissioning Map Reduce nodes
Hands On Exercises – Adding and Removing Nodes
Hands On Exercises – NameNode Recovery
Monitoring Hadoop Cluster
• Checking counters, metrics & Log Files
• Using the Name Node and Job Tracker Web UIs
• Monitoring with Ganglia
Hands On Exercises – Using Web UIs and HDFS Health Check
Hands On Exercises – Configuring Ganglia for Monitoring
Benchmarking and optimizing a Cluster
• Various Configuration Parameters & Best Practices
• Benchmarking a Cluster
Hands On Exercises – Running Benchmarking operations
Best Practices
Fee Details:
Rs. 17,800.00 + 12.36% Service Tax Per Participant
Subject to availability of seats. Terms & Conditions
Registration is first come first serve basis.

Payment Options:
Account Name:KnowledgeWorks IT Consulting Pvt. Ltd.,
Bank Name: HDFC Bank
Bank  Account Number: 02612020000021
Account TypeCurrent Account (CA)
Beneficiary Bank AddressJayanagar Branch, Bangalore
RTGS / NEFT / IFSC CodeHDFC0000261
Time: 09:30am to 05:30pm 
Venue:
Will be Confirmed to the Registered Participants
For any clarifications, Please contact:
Mr. Sudhindra D N (+91 9886221314)
T: +91 80 26630622, 22459941 
E: sudhi@knowledgeworks.co.in