There are a number of CLIs that you can run from the edge node, which is the gateway node to the Hadoop cluster consisting of master & slave (aka worker) nodes. Let’s look at the different CLIs (i.e Command Line Interfaces)
Most of the CLIs listed below are in /usr/bin
|
[root@quickstart /]# ls -ltr /usr/bin | grep 'hdfs\|impala-shell\|beeline\|hive\|spark-shell\|pyspark\|spark-submit' -rwxr-xr-x 1 root root 8144 Sep 22 2015 db_archive -rwxr-xr-x 1 root root 328 Mar 23 2016 hiveserver2 -rwxr-xr-x 1 root root 321 Mar 23 2016 hive -rwxr-xr-x 1 root root 324 Mar 23 2016 beeline -rwxr-xr-x 1 root root 504 Mar 23 2016 sqoop-create-hive-table -rwxr-xr-x 1 root root 1138056 Mar 23 2016 git-upload-archive -rwxr-xr-x 1 root root 141 Mar 23 2016 spark-submit -rwxr-xr-x 1 root root 140 Mar 23 2016 spark-shell -rwxr-xr-x 1 root root 185 Mar 23 2016 pyspark -rwxr-xr-x 1 root root 192 Mar 23 2016 hdfs -rwxr-xr-x 1 root root 1856 Mar 23 2016 impala-shell |
hdfs CLI
The following command will give you the commands you can use with “hdfs”
|
[root@quickstart /]# hdfs |
“dfs” – run a filesystem command on the file systems supported in Hadoop.…