In this page, we describes about audiences, objective, prerequisites and the topics that are covered as part of tutorial.
Apache Flume is tool/Service for efficiently collecting, combining and moving large amounts of log data. Flume is a highly distributed, reliable and available tool/service. Flume is an intermediate tool/service for data collection from external data sources and sends to centralized data storage like Hadoop HDFS or HBase.
In this tutorial, all the topics are covered with in-depth information and can understandable by anyone. Readers can be anyone who has a basic knowledge on orgnizational controlling activities.
The tutorial is intended for the Hadoop developers who are having minimum Hadoop and java knowledge. And can also understood by readers who are having no Hadoop knowledge, But takes a little bit time to understand.
Readers who are having the Hadoop and java knowledge can understand the concepts a bit easily. Reader without having Hadoop and java knowledge may need to go through the topic more than once to understand clearly.