Apache Storm Introduction - API and Topology
Overview/Description
Target Audience
Prerequisites
Expected Duration
Lesson Objectives
Course Number
Expertise Level
Overview/Description
Storm makes it easy to stream massive amounts of unbounded streams of data while providing a fault-tolerant system. This system accompanied with other integrations can take any business to the next level. In this course, you will be introduced to Trident and its relationship to Storm, along with its integration. Many other integrations will be explored including Hadoop, Kafka, JMX, Ganglia, and even automation using Puppet as well as monitoring and analytics tools. In this course, you will also learn how to deploy the Storm architecture.
Target Audience
Individuals interested in how to perform an installation and configure it, as well as design and create basic Storm topologies
Prerequisites
None
Expected Duration (hours)
3.1
Lesson Objectives Apache Storm Introduction - API and Topology
start the course
use Trident for a simple topology
describe topology state management with Trident
describe the different types of Trident spouts available for implementing fault-tolerant Trident state management
describe the different Trident State APIs available for implementing fault-tolerant Trident state management
describe distributed RPC model and how it is used with Apache Storm
describe DRPC modes of operation and topology types
deploy a Trident topology to a Storm cluster
describe the Storm UI home page
launch a Storm topology to a local cluster and view cluster activity in the Storm UI
analyze a Storm topology using the Storm UI
describe the process of using the Nimbus Thrift client for obtaining Storm cluster metrics
set up a Maven project in Eclipse IDE that can be used to write Java client code for connecting to a Nimbus Thrift server
write Java client code that connects to a Nimbus Thrift server and retrieves Storm cluster statistics
describe the general architecture of Apache Kafka
describe Kafka components and data model
produce and consume a Kafka topic
consume Kafka messages in a Storm topology
describe some options for using Storm's Core APIs to implement micro-batching in a Storm Core topology
describe Apache Hadoop's use with Storm
download and install Apache Hadoop on a development machine
describe how Apache Storm applications can be run on Hadoop YARN clusters to leverage YARN resource management
describe the Puppet architecture and some key framework components
describe how JMX and Ganglia can be integrated and used to monitor Storm clusters
describe how HBase and Redis can be integrated and used as datastores with Apache Storm
integrate and use JMX in Storm to obtain Storm Nimbus and Supervisor metrics
demonstrate increased knowledge of configuring and installing Apache Storm
Course Number: df_apsi_a02_it_enus
Expertise Level
Beginner