This article provides a detailed analysis of some commonly used commands in hadoop. For more information, see Hadoop installation directory HADOOP_HOME:/home/admin/hadoop.
Start and closeStart Hadoop1. go to the HADOOP_HOME directory.
2. execute sh bin/start-all.sh
Disable Hadoop1. go to the HADOOP_HOME directory.2.
Assume that the installation directory for Hadoop is Hadoop_home/home/admin/hadoop.
Start and closeStart Hadoop1. Enter the Hadoop_home directory.
2. Execute SH bin/start-all.sh
Turn off Hadoop1. Enter the Hadoop_home directory.2. Execute SH bin/stop-all.shFile operationsHadoop uses HDFs, which is similar in functionality to the disk systems we use. and wildcar
Assume that the Hadoop installation directory HADOOP_HOME is/home/admin/hadoop.Start and closeStart Hadoop1. Go to the HADOOP_HOME directory.2. Execute sh bin/start-all.sh
Disable Hadoop1. Go to the HADOOP_HOME directory.2. Execute sh bin/stop-all.shFile OperationsHadoop uses HDFS to implement functions similar to the disk system we use. Wildcard characters are also supported, such *.
View the file listView
A. Common Hadoop commands1. The FS command for Hadoop#查看hadoop所有的fs命令Hadoop FS#上传文件 (both put and copyfromlocal are upload commands)Hadoop fs-put jdk-7u55-linux-i586.tar.gz hdfs://hucc0
Assume that the Hadoop installation directory HADOOP_HOME is/home/admin/hadoop.Start and closeStart Hadoop1. Go to the HADOOP_HOME directory.2. Execute sh bin/start-all.sh
Disable Hadoop1. Go to the HADOOP_HOME directory.2. Execute sh bin/stop-all.shFile OperationsHadoop uses HDFS to implement functions similar to the disk system we use. Wildcard characters are also supported, such *.
View the file listView
Suppose the installation directory for Hadoop is hadoop_home to/home/admin/hadoop.Start and closeStart Hadoop1. Enter the Hadoop_home directory.2. Perform SH bin/start-all.sh
Turn off Hadoop1. Enter the Hadoop_home directory.2. Perform SH bin/stop-all.shFile actionsHadoop uses HDFs, which is similar to the disk system we use. and supports wildcard characters, such as *.
View File ListView the files in the
region:#hbase> major_compact ‘r1‘, ‘c1‘#Compact a single column family within a table:#hbase> major_compact ‘t1‘, ‘c1‘
Configuration Management and node restart1) Modify the HDFs configurationHDFs Configuration Location:/etc/hadoop/conf
# 同步hdfs配置cat /home/hadoop/slaves
/ci_cuser_20141231141853691/* ' >ci_cusere_20141231141853691.csv echo $?~/.bash_profile: Each user can use this file to enter shell information dedicated to their own use, when the user logs on, theThe file is only executed once! By default, he sets some environment variables to execute the user's. bashrc file.Hadoop fs-cat ' $1$2/* ' >$3.csvMV $3.csv/home/ocdc/cocString command = "CD" + Ciftpinfo.getftppath () + "" +hadooppath+ "Hadoop fs-cat '/user
commands in the Hadoop process:1. View the process
Linux under View process: $ PS Linux View Java process: $ jps
2. View Hadoop-related commands
$ hadoop (using such as: $ Hadoop namenode–format)
3. Vi
Fsck commands in Hadoop
The fsck command in Hadoop can check the file in HDFS, check whether there is upt corruption or data loss, and generate the overall health report of the hdfs file system. Report content, including:Total blocks (Total number of blocks), Average block r
To make it easier for you to review your memory, we will summarize the hadoop commands of today's experiment for later viewing.Note: The following commands are performed in hadoop/bin.1. hadoop FS-ls \-> View All the following directories.2.
1. Test the speed of Hadoop writesWrite data to the HDFs file system, 10 files, 10MB per file, files stored in/benchmarks/testdfsio/io_dataHadoop jar Share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.6.0-tests.jar Testdfsio-write-nrfiles 10- FileSize 10MB2. Test the speed of
1.hadoopView the directory on HDFs: hadoop fs-ls/ Create a directory on HDFs: -mkdir/jiatest upload the file to HDFs Specify directory: -put test.txt /Jiatest upload jar package to Hadoop run: hadoop jar maven_test-
Create a table of contents
Hadoop dfs-mkdir/homeUploading files or directories to HDFs
Hadoop dfs-put Hello/Hadoop dfs-put hellodir//View Table of Contents
Hadoop Dfs-ls/Create an empty file
Hadoop Dfs-touchz/361wayDelete a fi
Preface: Well, it's just a little bit more comfortable without writing code, but we can't slack off. The hive operation's file needs to be loaded from hereSimilar to the Linux commands, the command line begins with the Hadoop FS -(dash) LS / list file or directory cat Hadoop FS -cat ./hello.txt/opt/old/ Htt/hello.txt View files can dump directories or
success:mysql-h172.16.77.15-uroot-p123 mysql-h host address-u user name-P user PasswordView Character SetsShow variables like '%char% ';To Modify a character set:VI/ETC/MY.CNF add Default-character-set=utf8 under [client]create sudo without password loginTo set the Aboutyun user with no password sudo permissions: Chmode u+w/etc/sudoersaboutyun all= (root) nopasswd:allchmod u-w/etc/sudoers test: sudo ifconfigUbuntu View Service List codesudo service--status-allsudo initctl listTo view the file s
1.1)vim/etc/udev/rules.d/ --persistent-Net.rulesVI/etc/sysconfig/network-scripts/ifcfg-Eth0type=Ethernetuuid=57d4c2c9-9e9c-48f8-a654-8e5bdbadafb8onboot=yesnm_controlled=YesBootproto = staticDefroute=Yesipv4_failure_fatal=Yesipv6init=NoNAME="System eth0"HWADDR=xx: 0c: in: -: E6:ecipaddr =172.16.53.100PREFIX= -gateway=172.16.53.2Last_connect=1415175123dns1=172.16.53.2The virtual machine's network card is using the virtual network cardSave Exit X or Wq2)Vi/etc/sysconfig/networkNetworking=yesHostnam
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.