Sei sulla pagina 1di 6

Hadoop Online Training In

Hyderabad
Hadoop course
content
COURSE DETAILS IN A NUTSHELL
1. In-detail explanation on the concepts of HDFS & MapReduce Programming
frameworks languages: Java & Scala
2. What is Hadoop 2.X Architecture & How to set up Hadoop Frame works: Hadoop
Cluster Distributed File
3. How to write complex MapReduce Programs
System (HDFS) &
4. In-detail explanation on how to load data using tools like
MapReduce, spark
Sqoop & Flume,solr
Loading Tools: Sqoop &
5. How to perform data analysis using tools like PIG, HIVE &
Flume
YARN
Analytical Tools: Pig,
6. How to implement & integrate HBASE & MapReduce
7. How to execute Advanced Usage and Indexing Hive and YARN
8. How to schedule jobs using Oozie Scheduling Tools:
9. What are the best practices for overall Hadoop development Oozie
10. RTAs on Data Analytics
11. What is Spark & brief about its ecosystem & how to work on
RDD Using Spark

CURRICULUM for HADOOP 2.X


S Concepts Syllabus Objectives Topics
No
The syllabus for this lecture would brief 1. Big Data, Limitations and Solutions of
about: existing Data Analytics Architecture,
Understan 1. Big Data 2. Hadoop,
ding Big 2. Big Data problems & solutions, their 3. Hadoop Features,
1 Data and limitations 4. Hadoop Ecosystem,
Hadoop 3. HADOOP’s solutions that handle Big 5. Hadoop 2.x core components,
Data issue 6. Hadoop Storage: HDFS,
4. Common Hadoop Ecosystem and its 7. Hadoop Processing
Architecture 8. MapReduce Framework,
5. Introduction to HDFS 9. Hadoop Different Distributions.
6. What is a file and how to write & read
7. Brief on MapReduce Framework and
its working style.
The syllabus for this lecture would brief 10. Linux commands
about: 30 Essential Linux Basic Commands
Pre-requisites to learn hadoop You Must Know

11. vmware
 Basics
Hadoop  Installations
2 requirem  Backups
e nts
12. sql basics
 Introduction to SQL
 MySQL Essentials
 Database Fundamentals

13. Hands on exercise and Assignments


The syllabus for this lecture would brief 14. Hadoop 2.x Cluster Architecture
Hadoop about: 15. Federation and High Availability,
Architect 1. What is Hadoop Cluster Architecture 16. A Typical Production Hadoop Cluster,
u re and 2. What are the important Configuring 17. Hadoop Cluster Modes,
3 HDFS files in a Hadoop Cluster 18. Common Hadoop Shell Commands,
3. What are the various Data loading 19. Hadoop 2.x Configuration Files,
techniques 20. Single node cluster and Multi node
4. What are Single node and Multi cluster set up Hadoop
nodes Administration.
and their setups 21. Hands on exercise and Assignments
The syllabus for this lecture would brief 22. MapReduce Use Cases,
about: 23. Traditional way Vs MapReduce way,
1. In-depth analysis on Hadoop 24. Why MapReduce,
MapReduce Framework 25. Hadoop 2.x MapReduce Architecture,
Hadoop 2. How MapReduce works on data stored 26. Hadoop 2.x MapReduce Components,
MapRedu in HDFS. 27. YARN MR Application Execution
ce 3. What are Splits, Combiner & Flow,
4 Framewor
Partitioner. 28. YARN Workflow,
k
4. How to work on MapReduce 29. Anatomy of MapReduce Program,
using 30. Demo on MapReduce.
different data sets 31. Input Splits,
32. Relation between Input Splits and
HDFS Blocks,
33. MapReduce Combiner &
Partitioner,
34. Hands on exercise and
Assignments
The syllabus for this lecture would brief 35. About Pig,
Pig about: 36. MapReduce Vs Pig,
5 1. What is PIG & types of use, demo case 37. Pig Use Cases,
2. How to couple PIG with MapREduce 38. Programming Structure in Pig,
3. What are PIG Latin Scripting 39. Pig Running Modes,
4. What are PIG running Modes PIG 40. Pig components,
UDF, Pig Streaming, Testing PIG 41. Pig Execution,
Scripts. 42. Pig Latin Program,
The syllabus for this lecture would brief 43. Data Models in Pig,
about: 44. Pig Data Types,
1. What are HIVE concepts 45. Shell and Utility Commands,
2. What are HIVE data types 46. Pig Latin Relational Operators,
3. What are Loading & Querying in 47. File Loaders,
HIVE, 48. Group Operator,
4. How to run HIVE scripts 49. COGROUP Operator,
5. What are Hive UDF 50. Joins and COGROUP,
Hive 51. Union,
. 52. Diagnostic Operators,
53. Specialized joins in Pig,
54. Hands on exercise and Assignments
.
55. Hive Background,
56. Hive Use Case,
57. About Hive,
58. Hive Vs Pig,
59. Hive Architecture and Components,
60. Metastore in Hive,
61. Limitations of Hive,
62. Comparison with Traditional Database,
63. Hive Data Types and Data Models,

6 64. Partitions and Buckets,


65. Hive Tables(Managed Tables and
External Tables),
66. Importing Data,
67. Querying Data,
68. Managing Outputs,
69. Hive Script,
70. Hive UDF,
71. Retail use case in Hive,
72. Hands on exercise and
Assignments
The syllabus for this lecture would brief 73. Hive QL: Joining Tables,
about: 74. Dynamic Partitioning,
1. What are Advanced HIVE concepts 75. Custom Map/Reduce Scripts,
2. What are UDF, Dynamic Partitioning, 76. Hive Indexes and views
HIVE indexes & Views 77. Hive query optimizers,
Advanced 3. What are Optimizations in HIVE 78. User Defined Functions,
Hive and 4. In-depth analysis on HBase, its 79. HBase:
7 HBase Architecture, components and its 80. Introduction to NoSQL
running modes 81. Databases and HBase,
82. HBase v/s RDBMS,
83. HBase Components,
84. HBase Architecture,
85. Run Modes & Configuration,
86. HBase Cluster Deployment.
87. Hands on exercise and Assignments
8 Advanced The syllabus for this lecture would brief 88. HBase Data Model,
HBase about: 89. HBase Shell,
1. What are Advanced HBase Concepts 90. HBase Client API,
2. How to perform bulk loading 91. Data Loading Techniques,
3. What are filters 92. ZooKeeper
4. What is Zookeeper and how it helps in 93. Demos on Bulk Loading,
Cluster monitoring. 94. Getting and Inserting Data,
5. Why HBase utilizes Zookeeper 95. Filters in HBase.
96. Hands on exercise and
Assignments

The syllabus for this lectureould brief 97. Introduction.


about: 98. Import Data.
1.Import data from other databases to hdfs 99. Export Data.
9 Sqoop 2.Import data from other databases to hive 100. Sqoop Syntax.
3.export data from hadoop to other databses 101. Databases connection.
102.Hands on exercise and
Assignments
The syllabus for this lecture would brief 103..Introduction to Impala
about: 104..Impala Configuration
Impala 105..Comparison between Hive and
10 Impala Impala
106..Impala Commands
107.Hands on exercise and
Assignments
The syllabus for this lecture would brief 108. What is Apache Spark,
about: 109. Spark Ecosystem,
Processin
g 1. What is Spark Ecosystem 110. Spark Components,
Distribute 2. What is Scala and its utility in Spark 111. History of Spark
d Data 3. What is SparkContext 112. Spark Versions/Releases,
11 with 4. How to work on RDD in Spark 113. What is Scala?,
Apache 5. How to run a Spark Cluster 114. Why Scala?,
Spark 6. Comparison of MapReduce vs Spark 115. SparkContext,
116. Spark Sql
117. Hands on exercise and
Assignments.
The syllabus for this lecture would brief 118. Introduction.
about: 119. Configuration and Setup
Flume and solr 120. Flume Sink with example
121. Channel
12 Flume & 122. Flume Source with example
solr 123. Complex flume architecture
124. Streaming data storing into solr
125. customization of solr
126. Hands on exercise and
Assignments
127. Introduction to Hue
The syllabus for this lecture would brief 128. Advantages of Hue
about: 129. Hue Web Interface
13 Hue Hue 130. Ecosystems in Hue
131. Hands on exercise and
Assignments

The syllabus for this lecture would brief 132.Oozie,


about: 133.Oozie Components,
1. How multiple Hadoop ecosystem 134.Oozie Workflow,
components work 135.Scheduling with Oozie,
2. How they should be implemented to 136.Demo on Oozie Workflow,
Oozie solve Big Data Issues 137.Oozie Co-ordinator,
14 138.Oozie Commands,
139.Oozie Web Console,
140.Oozie for MapReduce,
141.PIG, Hive, and Sqoop,
142.Combine flow of MR, PIG, Hive
in Oozie
143.Hands on exercise and
Assignments
The syllabus for this lecture would brief .
about: 144. Tableau Fundamentals
Tableau 145. Tableau Analytics.
1 Tableau
146. Visual Analytics.
147. Hands on exercise and
Assignments

PROJECTS:

Hadoop -Tableau live integration 1.Hadoop Integration with Tableau


Hadoop
Topics : This is a project that gives you
1. Project
opportunity to work on retail data analytics.

:
Multi-node cluster setup  Running a Hadoop multi-node
using a 4 node cluster
Topics : This is a project that gives you  Deploying of MapReduce job on
opportunity to work on real world Hadoop the Hadoop cluster
Hadoop multi-node cluster setup in a distributed  You will get a complete
environment. demonstration of working with
2.
Project2 various Hadoop cluster master and
slave nodes, installing Java as a
prerequisite for running Hadoop,
installation of Hadoop and
mapping the nodes in the Hadoop
cluster.
 Streaming Twitter data
Social media analytics
 Store data into hadoop
 Process social media data
Hadoop
3.  Sentiment analysis on twitter data
Project3 Topics : This is a project that gives you
 Final result store in table
opportunity to work on social media
 Connect BI Tool.
Analytics.

<a href=http://airistrainings.com/hadoop-online-training-hyderbad.php>Hadoop Online Training In


Kukatpally</a>

<a href=http://airistrainings.com/hadoop-online-training-hyderbad.php>Best Hadoop Online


Training In Madhapur</a>

Potrebbero piacerti anche