Hadoop HDFS: Working with Files


Overview/Description
Expected Duration
Lesson Objectives
Course Number
Expertise Level



Overview/Description

In this Skillsoft Aspire course, learners will encounter basic Hadoop file system operations such as viewing the contents of directories and creating new ones. This 8-video course assumes good understanding of what Hadoop is, and how HDFS enables processing of big data in parallel by distributing large data sets across a cluster; learners should also be familiar with running commands from the Linux shell, with some fluency in basic Linux file system commands. Begin by working with files in various ways, including transferring files between a local file system and HDFS (Hadoop Distributed File System) and explore ways to create and delete files on HDFS. Then examine different ways to modify files on HDFS. After exploring the distributed computing concept, prepare to begin working with HDFS in a production setting. In the closing exercise, write a command to create a directory/data/products/files on HDFS, for which data/products may not exist; list two commands for two copy operations—one from local file system to HDFS, and another for reverse transfer, from HDFS to local host.



Expected Duration (hours)
0.8

Lesson Objectives

Hadoop HDFS: Working with Files

  • Course Overview
  • identify the different ways to use the ls and mkdir commands to explore and create directories on HDFS
  • transfer files from your local file system to HDFS using the copyFromLocal command
  • copy files from your local file system to HDFS using the put command
  • transfer files from HDFS to your local file system using the copyToLocal command
  • use the get and getmerge functions to retrieve one or multiple files from HDFS
  • work with the appendToFile and rm commands on the hdfs dfs shell
  • utilize HDFS commands to work with and manipulate files using the HDFS shell
  • Course Number:
    it_dshdfsdj_03_enus

    Expertise Level
    Beginner