Hadoop Shell commands
Use bin/hadoop FS
1. cat
Description: outputs the content of the specified file in the path to stdout.
Usage: hadoop fs-cat URI [URI…]
Example:
hadoopfs-cathdfs://host1:port1/file1hdfs://host2:port2/file2
hadoopfs-catfile:///file3/user/
A. Common Hadoop commands1. The FS command for Hadoop#查看hadoop所有的fs命令Hadoop FS#上传文件 (both put and copyfromlocal are upload commands)Hadoop fs-put jdk-7u55-linux-i586.tar.gz hdfs://hucc01:9000/jdkhadoop fs-copyfromlocal jdk-7u55-li
region:#hbase> major_compact ‘r1‘, ‘c1‘#Compact a single column family within a table:#hbase> major_compact ‘t1‘, ‘c1‘
Configuration Management and node restart1) Modify the HDFs configurationHDFs Configuration Location:/etc/hadoop/conf
# 同步hdfs配置cat /home/hadoop/slaves|xargs -i -t scp /etc/hadoop/conf/hdfs-site.x
A. Common Hadoop commands1. The FS command for Hadoop#查看hadoop所有的fs命令Hadoop FS#上传文件 (both put and copyfromlocal are upload commands)Hadoop fs-put jdk-7u55-linux-i586.tar.gz hdfs://hucc01:9000/jdkhadoop fs-copyfromlocal jdk-7u55-li
To make it easier for you to review your memory, we will summarize the hadoop commands of today's experiment for later viewing.Note: The following commands are performed in hadoop/bin.1. hadoop FS-ls \-> View All the following directories.2.
Overview:
The file system (FS) shell contains commands for various classes of-shell, directly interacting with Hadoop Distributed File System (HDFS), and support for other file systems, such as: Local file system fs,hftp Fs,s3 FS, and others. Calls to the FS shell
FS ShellThe call file system (FS) shell command should use the form bin/hadoop FS . All of the FS shell commands use URI paths as parameters. The URI format is scheme://authority/path . For the HDFs file system, Scheme is HDFs , to the local file system, scheme is file. The scheme and authority parameters are optiona
FS Shell
Cat
Chgrp
chmod
Chown
Copyfromlocal
Copytolocal
Cp
Du
Dus
Expunge
Get
Getmerge
Ls
Lsr
Mkdir
Movefromlocal
Mv
Put
Rm
RMr
Setrep
Stat
Tail
Test
Text
Touchz
FS ShellThe call file system (FS) shell command should use the form Bin/
Original address: http://hadoop.apache.org/docs/r1.0.4/cn/hdfs_shell.html
FS Shell
Cat
Chgrp
chmod
Chown
Copyfromlocal
Copytolocal
Cp
Du
Dus
Expunge
Get
Getmerge
Ls
Lsr
Mkdir
Movefromlocal
Mv
Put
Rm
RMr
Setrep
Stat
Tail
Test
Text
Touchz
FS ShellThe call file system (FS)
Apache-->hadoop's official Website document Command learning:http://hadoop.apache.org/docs/r1.0.4/cn/hdfs_shell.html
FS Shell
The call file system (FS) shell command should use the bin/hadoop fs scheme://authority/path. For the HDFs file system, Scheme is HDFs, to the local file system, scheme is file. The scheme and authority parameters are optional,
Use bin/hadoop FS Scheme: // authority/path. For HDFS file systems, scheme isHDFSFor the local file system, scheme isFile. The scheme and authority parameters are optional. If not specified, the default scheme specified in the configuration will be used. An HDFS file or directory such/Parent/childCan be expressedHDFS: // namenode: namenodeport/parent/child, Or simpler/Parent/child(Assume that the default value in your configuration file isNamenode: na
Overview
The filesystem (FS) Shell is invoked by bin/hadoop FS Scheme: // autority/path. For HDFS the scheme isHDFS, And for the local filesystem the scheme isFile. The scheme and authority are optional. If not specified, the default scheme specified in the configuration is used. an HDFS file or directory such/Parent/childCan be specifiedHDFS: // namenodehost/parent/childOr simply/Parent/child(Given that yo
, file random modification a file can have only one writer, only support append.Data form of 3.HDFSThe file is cut into a fixed-size block, the default block size is 64MB, the size of the block can be configured, if the file size is less than 64MB, it is stored separately into a block. A file storage method is divided into blocks by size, stored on different nodes, with three replicas per block by default.HDFs Data Write Process: HDFs Data Read process: 4.MapReduce: Google's MapReduce open sou
FS ShellThe call file system (FS) shell command should use the form Bin/hadoop FS scheme://authority/path. For the HDFs file system, Scheme is HDFs, to the local file system, scheme is file. The scheme and authority parameters are optional, and if not specified, the default scheme specified in the configuration is used. An HDFs file or directory such as /parent/child can be represented as Hdfs://namenode:na
Hadoop FS: Use the widest range of surfaces to manipulate any file system.Hadoop DFS and HDFs DFS: can only operate on HDFs file system-related (including operations with local FS), which is already deprecated, typically using the latter.The following reference is from StackOverflowFollowing is the three commands which appears same but has minute differences
Ha
Hadoop FS: Use the widest range of surfaces to manipulate any file system.Hadoop DFS and HDFs DFS: can only operate on HDFs file system-related (including operations with local FS), which is already deprecated, typically using the latter.The following reference is from StackOverflowFollowing is the three commands which appears same but has minute differences
Ha
[jobMainClass] [jobArgs]
Killing a running JOB
Hadoop job-kill job_20100531_37_0053
More HADOOP commands
Hadoop
You can see the description of more commands:
Namenode-format the DFS filesystem
Secondarynamenode run the DFS secondary namenode
Namenode run the DFS namenode
Da
Hadoop shell command operation, enter the Hadoop carriage return can be found:The commonly used commands are:Hadoop Namenode-format This is the command used to format the file system before starting Hadoop.Hadoop Dfsadmin This is the management command for Hadoop, and we can
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.