• <More on Intel.com
Masthead Light

Intel® Distribution for Apache Hadoop* Software

Part of the Intel® Datacenter Software family

Intel® Datacenter Software  |  Intel Distribution for Apache Hadoop Overview 

Solutions   |   Where to Buy   |   Training   |   Support


Developer Training

Develop Applications on the Apache Hadoop* platform

Apache Hadoop* Developer Training provides an essential understanding of how to write applications on Hadoop. Using the Intel® Distribution for Apache Hadoop* as the platform for instruction, developers new to Hadoop can quickly learn to take advantage of the MapReduce* framework through technical sessions and hands-on labs. By the end of this three day Hadoop training, you will be prepared to handle real world challenges that you may come across.

Hadoop Training Course Objectives

  • Using the Hadoop and HDFS platform
  • Loading data into HDFS
  • Introduction to MapReduce
  • Writing and debugging MapReduce jobs
  • Implementing common algorithms on Hadoop
  • Using Mahout for advanced data mining
  • Benchmarking and optimizing performance

Course Details

  • Format: Classroom training
  • Length: 3 days

Audience
This course is recommended for:

  • Java* developer who wants to write Hadoop applications using the MapReduce API
  • Programmers looking to develop new data-driven services on a Hadoop platform

Prerequisites

  • Knowledge of object oriented programming and proficiency in Java

Certification
Students who complete the Hadoop training course are then eligible for the Intel Certification for Apache Hadoop* Professional Exam. As a trusted technology provider to the industry, the Intel Certification carries the assurance that your Hadoop expertise is backed by a brand recognized around the world.

   Day 1: Foundation

   Day 2: Implementation

   Day 3: Advanced

   - Why Apache Hadoop?

   - Apache Hadoop
     Concepts

   - Apache Hadoop
     Ecosystem

   - MapReduce, YARN, and
     HDFS

   - Developing MR jobs

   - MapReduce Examples

   - Hadoop API version

   - MR and HDFS Internals

 

   - Review Day 1

   - Invoicing with Hadoop

   - Invoicing from multiple
     logs

   - Enterprise data
     integration

   - Integration with RDBMS

   - Importing data with
     Sqoop

   - Importing data with
     Flume

   - Hands-on HBase

   - Hands-on Hive

   - Hands-on Pig

 

   - Review Day 2

   - Scaling for enterprise
     use

   - Optimizing application
     performance

   - Best practices and
     anti-patterns

   - Lessons from the real
     world

 

Why Intel and Apache Hadoop?

We believe that every organization and individual should have the ability to generate value from all the data that they can access. The value of data depends on the ability to collect, store, analyze, visualize, and act on it. The Intel® Distribution is built from the hardware up to enable the widest range of use cases on Apache Hadoop by delivering the performance, security, and scalability that enterprises need. Intel offers a robust platform on which the ecosystem can innovate in next-generation analytics. Intel is committed to open source and supports the Apache developer community by contributing in code and in kind.

Success Stories

View More