Hortonworks is sponsoring a quick, hands-on introduction to key Apache projects. Come and listen to a short technical introduction and then get hands-on with your personal machine, ask questions, and leave with a working environment to continue your journey.

Streaming Analytics

Introduction: This workshop will provide a hands on introduction to stream processing  Sandbox on students’ personal machines.
Format: A short introductory lecture on Streaming Analytics Manager, which will be followed by a demo, lab exercises and a Q&A session. The lecture will be followed by lab time to work through the lab exercises and ask questions.
Objective: To provide a quick and short hands-on introduction to Stream Processing without coding. In the lab you will use SAM to connect, aggregate and process real-time events.. You will learn how to connect and consume streaming sensor data, filter and transform the data and persist to multiple data sources.
Pre-requisites: Registrants must bring a laptop that has the latest VirtualBox installed and an image for HDF will be provided.

View Details

Apache Spark and Apache Hive processing on the Cloud

Introduction: This workshop will provide an overview to cloud-based solutions and a hands on session with HDCloud on AWS.

Format: A short introductory lecture about general Apache Hadoop platforms on the cloud.  This is followed by a walk through and lab leveraging HDCloud and hive

Objective: To provide a quick and short hands-on introduction to Hadoop on the cloud. This lab will use the following Apache Hadoop components: Apache Hadoop HDFS, Apache Hadoop YARN, Apache Hive, Apache ORC, and Apache Ambari User Views. You will learn how to move data into HDFS/s3 storage using create Apache Hive tables.
Pre-requisites: Registrants must bring a laptop that can run the Hortonworks Data Cloud.  Please setup an HDCloud account prior to this session starting​

If you would like to spin up your own cluster, make sure to have an AWS account and follow the instructions below.

Instructions: http://hortonworks.github.io/hdp-aws/index.html

Short how-to videos

Part 1 of 3: How to set up a cloud controller:

https://www.youtube.com/watch?v=Q5ovR8YTFSg

Part 2 of 3: How to create a cluster:

https://www.youtube.com/watch?v=yhFU-D0Uijw

Part 3 of 3: How to launch Ambari & Zeppelin:

https://www.youtube.com/watch?v=YrEwOqbw7Is&t=1s

 

View Details

Data Science

Introduction: This workshop will provide a hands on introduction to basic Machine Learning techniques with Spark ML using a Sandbox on students’ personal machines.
Format: A short introductory lecture on a select important supervised and unsupervised Machine Learning techniques followed by a demo, lab exercises and a Q&A session. The lecture will be followed by lab time to work through the lab exercises and ask questions.
Objective: To provide a quick and short hands-on introduction to Machine Learning with Spark ML. In the lab, you will use the following components: Apache Zeppelin (a “Modern Data Science Toolbox”) and Apache Spark. You will learn how to analyze the data, structure the data, train Machine Learning models and apply them to answer real-world questions.
Pre-requisites: Registrants must bring a laptop that can run the Hortonworks Data Cloud.

At this Crash Course everyone will have a cluster assigned to them to try several workloads using Machine Learning, Spark and Zeppelin on HDCloud.

If you would like to spin up your own cluster, make sure to have an AWS account and follow the instructions below.

Instructions: http://hortonworks.github.io/hdp-aws/index.html

Short how-to videos

Part 1 of 3: How to set up a cloud controller:

https://www.youtube.com/watch?v=Q5ovR8YTFSg

Part 2 of 3: How to create a cluster:
https://www.youtube.com/watch?v=yhFU-D0Uijw

Part 3 of 3: How to launch Ambari & Zeppelin:

https://www.youtube.com/watch?v=YrEwOqbw7Is&t=1s

View Details

Apache NiFi

Introduction: This workshop will provide a hands on introduction to simple event data processing and data flow processing using a Sandbox on students’ personal machines.
Format: A short introductory lecture to Apache NiFi and computing used in the lab followed by a demo, lab exercises and a Q&A session. The lecture will be followed by lab time to work through the lab exercises and ask questions.
Objective: To provide a quick and short hands-on introduction to Apache NiFi. In the lab, you will install and use Apache NiFi to collect, conduct and curate data-in-motion and data-at-rest with NiFi. You will learn how to connect and consume streaming sensor data, filter and transform the data and persist to multiple data sources.
Pre-requisites: Registrants must bring a laptop that has the latest VirtualBox installed and an image for HDF will be provided.

View Details

Apache Spark

Introduction: This workshop will provide a hands-on introduction to Apache Spark using the HDP Sandbox on students’ personal machines.
Format: A short introductory lecture about Apache Spark components used in the lab followed by a demo, lab exercises and a Q&A session. The lecture will be followed by lab time to work through the lab exercises and ask questions.
Objective: To provide a quick and short hands-on introduction to Apache Spark. This lab will use the following Spark and Apache Hadoop components: Spark, Spark SQL, Apache Hadoop HDFS, Apache Hadoop YARN, Apache ORC, and Apache Ambari User Views. You will learn how to move data into HDFS using Spark APIs, create Apache Hive tables, explore the data with Spark and Spark SQL, transform the data and then issue some SQL queries.
Pre-requisites: Registrants must bring a laptop that can run the Hortonworks Data Cloud.

At this Crash Course everyone will have a cluster assigned to them to try several workloads using Machine Learning, Spark and Zeppelin on HDCloud.

If you would like to spin up your own cluster, make sure to have an AWS account and follow the instructions below.

Instructions: http://hortonworks.github.io/hdp-aws/index.html

Short how-to videos

Part 1 of 3: How to set up a cloud controller:

https://www.youtube.com/watch?v=Q5ovR8YTFSg

Part 2 of 3: How to create a cluster:

https://www.youtube.com/watch?v=yhFU-D0Uijw

Part 3 of 3: How to launch Ambari & Zeppelin:

https://www.youtube.com/watch?v=YrEwOqbw7Is&t=1s

View Details

Apache Hadoop

Introduction: This workshop will provide a hands on introduction to Apache Hadoop using the HDP Sandbox on students’ personal machines.

Format: A short introductory lecture about Apache Hadoop and a few key additional Apache projects in the extended ecosystem used in the lab followed by a demo, lab exercises and a Q&A session.

Objective: To provide a quick and short hands-on introduction to Hadoop. This lab will use the following Hadoop components: HDFS, YARN, Apache Pig, Apache Hive, Apache Spark, and Apache Ambari User Views. You will learn how to move data into HDFS, explore the data, clean the data, issue SQL queries and then build a report with Apache Zeppelin.

Pre-requisites: Registrants must bring a laptop and have the Hortonworks Sandbox installed.

View Details