Flume and Sqoop for Ingesting Big Data

Engineers who want to port data from legacy data stores to HDFS
  • Self-Paced Learning

    Learning Style
  • Intermediate

    Difficulty
  • 2 Hours

    Course Duration
Pricing
About Individual Course:
  • Individual course plan gives you access to this course
On Sale!
Now Only $10.00 Regular Price $49.00
Now Only $10.00 Regular Price $49.00
/ Each
When you subscribe, you get:
Learn Subscription Plan gives you access to this course, PLUS:

  • 620+ high impact technical, end user and leadership courses
  • Peer to peer learning and access to expert mentors
  • Learner and Manager Analytics
  • Access to Cognitive Learning Research platform to troubleshoot project issues
7-Day FREE Trial
On Sale!
Now Only $14.99 Regular Price $24.99
Now Only $14.99 Regular Price $24.99
/ Month
Team
Pricing
  • Buy 5-9 Enrollments And Save 15% ($12.74 monthly.)
  • Buy 10-19 Enrollments And Save 25% ($11.24 monthly.)
  • Buy 20-above Enrollments And Save 40% ($8.99 monthly.)
Engineers who want to port data from legacy data stores to HDFS

Import data : Flume and Sqoop play a special role in the Hadoop ecosystem. They transport data from sources like local file systems, HTTP, MySQL and Twitter which hold/produce data to data stores like HDFS, HBase and Hive. Both tools come with built-in functionality and abstract away users from the complexity of transporting data between these systems. 

Flume: Flume Agents can transport data produced by a streaming application to data stores like HDFS and HBase. 

Sqoop: Use Sqoop to bulk import data from traditional RDBMS to Hadoop storage architectures like HDFS or Hive. 

Course Objective:

Practical implementations for a variety of sources and data stores ..

  • Sources : Twitter, MySQL, Spooling Directory, HTTP
  • Sinks : HDFS, HBase, Hive

Flume features : 

Flume Agents, Flume Events, Event bucketing, Channel selectors, Interceptors

Sqoop features : 

Sqoop import from MySQL, Incremental imports using Sqoop Jobs

Audience:

  • Engineers building an application with HDFS/HBase/Hive as the data store
  • Engineers who want to port data from legacy data stores to HDFS

Prerequisite:

  • Knowledge of HDFS is a prerequisite for the course
  • HBase and Hive examples assume basic understanding of HBase and Hive shells
  • HDFS is required to run most of the examples, so you'll need to have a working installation of HDFS
More Information
Lab Access No
Learning Style Self-Paced Learning
Difficulty Intermediate
Course Duration 2 Hours
Language English
Write Your Own Review
Only registered users can write reviews. Please Sign in or create an account
Sales Support

Sales (866) 991-3924

Mon-Fri. 8am-6pm CST

Have Questions? Ask Us.

Why QuickStart

Turn Training Into A Personalized Learning Experience


  • Problem Solving through ExpertConnect & Peer-To-Peer Learning
  • Find The Quickest Path To Learn With Career Paths
  • Access All Courses With Master Subscription
  • Manage Your Team With Learning Analytics
  • Virtual Classroom Training & Self-Paced Learning
  • Integrate With Your LMS Through API's