Copy HDFS files to local Linux, which you can do in two ways. Those are GET and copyToLocal. There is a little difference between these two.
How to get File from Hadoop to Local
Either you can use GET or copyToLocal command to copy files to local from HDFS. Check here how it is.
1. The Get Command
The get command copies HDFS-based files to the local Linux file system. The get command is similar to copyToLocal, except that copyToLocal must copy to a local Linux file system based file.
[hadoop@hc1nn tmp]$ hdfs dfs -get /tmp/flume/agent2.cfg #Display the list of files [hadoop@hc1nn tmp]$ ls -l ./agent2.cfg -rwxr-xr-x. 1 hadoop hadoop 1343 Jul 26 20:23 ./agent2.cfg
This example copies the HDFS-based file agent2.cfg to the local Linux directory (” . “).
- copyToLocal, which is file to file of Linux
- GET command: You can use to copy HDFS files to local Linux directory.
2. The copyToLocal Command
hadoop fs -copyToLocal [-ignorecrc] [-crc] URI <localdst>
Similar to get command, except that the destination is restricted to a local file reference.
- When you have local file reference, in local LINUX, so you can copy the files from HDFS.
” Success is not final, failure is not fatal: it is the courage to continue that counts.”
5 Features of Distributed File System (HDFS)
The popularity of computer system arises due to nature of some applications. In such cases, it is necessary to facilitate sharing long-storage devices and their data to make system more user friendly.
The main functionality of DFS is transparency which means user would be unaware about data location, movement, access, etc.
The main feature of DFS is high availability. This feature states that if one server goes offline or failure, the data stored on its hard drives is still available for other nodes.
File name should not be changed when its physical location changes.
Access to file from anywhere or from any remote location.