Hadoop is an open source, Java-based programming framework that supports the processing and storage of extremely large data sets in a distributed computing environment. It provides a software framework for distributed storage and processing of big data using the MapReduce programming model.

The core of Apache Hadoop consists of a storage part, known as Hadoop Distributed File System (HDFS), and a processing part which is a MapReduce programming model.

The Hadoop framework itself is mostly written in the Java programming language, with some native code in C and command line utilities written as shell scripts. Though MapReduce Java code is common, any programming language can be used with “Hadoop Streaming” to implement the “map” and “reduce” parts of the user’s program.

  • Undergraduates
  • Graduates
  • Working Professionals
  • Engineers
  • Introduction
  • Use case walkthrough
  • NoSQL Introduction
  • Hbase Introduction
  • Hbase Architecture
  • Hbase Schema Design
  • Hbase Java API
  • Hbase Operations, cluster management
  • MapReduce for Developers
  • Hadoop in the Enterprise
  • Architecture
  • Hadoop CLI
  • MapReduce Programming
  • MapReduce Formats
  • MapReduce Features
  1. Week Days: 8 WKS
  2. Week Ends: 7 WKS
  3. Tailored: Can be tailored as per batch requirement subject to possibility of time adjustment.
  • Knowledge of back-end programming
  • Knowledge of java
  • Basic knowledge of Linux
  • Will to learn advance concepts and software technologies