What is Hadoop?
Hadoop is the Apache Software Foundation top-level project
that holds the various Hadoop subprojects that graduated from the Apache
Incubator. The Hadoop project provides and sup-ports the envelopment of open
source software that supplies a framework for the development of highly
scalable distributed computing applications. The Hadoop framework handles the
processing details, leaving developers free to focus on application logic.Hadoop Training in india
The Apache Hadoop project develops open-source software for
reliable, scalable, distributed computing, including:
·
Hadoop Core, our flagship sub-project, provides
a distributed file system (HDFS) and support for the Map Reduce distributed
computing metaphor.
·
HBase builds on Hadoop Core to provide a Scalable, distributed database.
·
Pig is a high-level data-flow language and
execution framework for parallel computation. It is built on top of Hadoop
Core.Hadoop online training in hyderabad
·
Zookeeper is a highly available and reliable
coordination system. Distributed applications use Zookeeper to store and
mediate updates for critical shared state.
·
Hive is a data warehouse infrastructure built on
Hadoop Core that provides data sum-marization, adhoc querying and analysis of
data sets.
Course Content
