Start Hadoop
Start-all.sh
Disable HADOOP
Stop-all.sh
View the file list
View the files in the/user/admin/aaron directory of hdfs.
Hadoop fs-ls/user/admin/aaron
List all files (including files in subdirectories) in the/user/admin/aaron directory of hdfs ).
Hadoop fs-lsr/user/admin/aaron
Create a file directory
Hadoop fs-mkdir/user/admin/aaron/newDir
Delete an object
Delete the needDelete file under the/user/admin/aaron directory in hdfs
Hadoop fs-rm/user/admin/aaron/needDelete
Delete the/user/admin/aaron directory in hdfs and all files in the directory
Hadoop fs-rmr/user/admin/aaron
Upload files
Hadoop fs-put/home/admin/newFile/user/admin/aaron/
Download files
Hadoop fs-get/user/admin/aaron/newFile/home/admin/newFile
View files
Hadoop fs-cat/home/admin/newFile
Create an empty file
Hadoop fs-touchz/user/new.txt
Rename a file on hadoop
Hadoop fs-mv/user/test.txt/user/OK .txt
Saves all content in the specified directory of hadoop as a file and goes down to the local directory.
Hadoop dfs-getmerge/user/home/t
Submit a MAPREDUCE JOB
H bin/hadoop jar/home/admin/hadoop/job. jar [jobMainClass] [jobArgs]
Killing a running JOB
Hadoop job-kill job_20100531_37_0053
More HADOOP commands
Hadoop
You can see the description of more commands:
Namenode-format the DFS filesystem
Secondarynamenode run the DFS secondary namenode
Namenode run the DFS namenode
Datanode run a DFS datanode
Dfsadmin run a DFS admin client
Fsck run a DFS filesystem checking utility
Fs run a generic filesystem user client
Balancer run a cluster balancing utility
Jobtracker run the MapReduce job Tracker node
Pipes run a Pipes job
Tasktracker run a MapReduce task Tracker node
Job manipulate MapReduce jobs
Queue get information regarding JobQueues
Version print the version
Jar <jar> run a jar file
Distcp <srcurl> <desturl> copy file or directories recursively
Archive-archiveName NAME <src> * <dest> create a hadoop archive
Daemonlog get/set the log level for each daemon
Or
CLASSNAME run the class named CLASSNAME
Most commands print help when invoked w/o parameters.
Note:
1. List all commands supported by Hadoop Shell
$ Bin/hadoop fs-help
2. display detailed information about a command
$ Bin/hadoop fs-help command-name
3. You can use the following command to view the historical log summary in the specified path to display the job details, failure and termination task details.
$ Bin/hadoop job-history output-dir
4. For more details about the job, such as the successful task and the number of attempts made to each task, run the following command.
$ Bin/hadoop job-history all output-dir
5. Format a New Distributed File System
$ Bin/hadoop namenode-format
6. On the assigned NameNode, run the following command to start HDFS and start the DataNode daemon on all listed slave instances.
$ Bin/start-dfs.sh
7. On the assigned JobTracker, run the following command to start Map/Reduce.
$ Bin/start-mapred.sh
8. On the assigned NameNode, run the following command to stop HDFS:
$ Bin/stop-dfs.sh
9. On the assigned JobTracker, run the following command to stop Map/Reduce:
$ Bin/stop-mapred.sh
DFSShell
10. Create a directory named/foodir.
$ Bin/hadoop dfs-mkdir/foodir
11. Create a directory named/foodir
$ Bin/hadoop dfs-mkdir/foodir
12. view the file content named/foodir/myfile.txt.
$ Bin/hadoop dfs-cat/foodir/myfile.txt
DFSAdmin
13. Place the cluster in Security Mode
$ Bin/hadoop dfsadmin-safemode enter
14. display the Datanode list
$ Bin/hadoop dfsadmin-report
15. Retire Datanode node datanodename
$ Bin/hadoop dfsadmin-decommission datanodename
16. The bin/hadoop dfsadmin-help command can list all currently supported commands. For example:
*-Report: Reports basic HDFS statistics. Some information can be seen on the NameNode Web Service Homepage.
*-Safemode: Generally, this mode is not required. The administrator can manually enter or exit the security mode.
*-FinalizeUpgrade: deletes the cluster backup created during the last upgrade.
17. explicitly place HDFS in Security Mode
$ Bin/hadoop dfsadmin-safemode
18. Before the upgrade, the administrator needs to use the (upgrade termination operation) command to delete the existing backup files.
$ Bin/hadoop dfsadmin-finalizeUpgrade
19. You need to know whether to perform the upgrade termination operation on a cluster.
$ Dfsadmin-upgradeProgress status
20. Use the-upgrade option to run the new version.
$ Bin/start-dfs.sh-upgrade
21. To return to the old version, you must stop the cluster and deploy Hadoop of the old version. Use the rollback option to start the cluster.
$ Bin/start-dfs.h-rollback
22. The following new commands or options are used to support quotas. The first two are administrator commands.
* Dfsadmin-setquota <N> <directory>... <directory>
Set the quota of each directory to N. This command will try on each directory. If N is not a positive long integer, the directory does not exist or the file name, or the directory exceeds the quota, an error report will be generated.
* Dfsadmin-clrquota <directory>... <director>
Delete a quota for each directory. This command will try on each directory. If the directory does not exist or is a file, an error report will be generated. If no quota is set in the directory, no error is returned.
* Fs-count-q <directory>... <directory>
The-q option is used to report the quota set for each directory and the remaining quota. If no quota is set for the directory, none and inf are reported.
23. Create a hadoop archive file
$ Hadoop archive-archiveName NAME <src> * <dest>
-ArchiveName NAME: NAME of the file to be created.
The path name of the src file system, which is the same as the regular expression.
The target directory where dest saves the file.
24. recursively copy files or directories
$ Hadoop distcp <srcurl> <desturl>
Srcurl source Url
Desturl target Url
25. Run the HDFS File System Check Tool (fsck tools)
Usage: hadoopfsck [GENERIC_OPTIONS] <path> [-move |-delete |-openforwrite] [-files [-blocks [-locations |-racks]