/hadoop jar/home/admin/hadoop/job. jar [jobMainClass] [jobArgs]
Killing a running JobAssume that Job_Id is job_20100531_37_0053.1. Go to the HADOOP_HOME directory.2. Execute sh bin/hadoop job-kill job_20100531_37_0053
More Hadoop commandsThe Hadoop operation
job1. Enter the Hadoop_home directory.2. Execute SH bin/hadoop jar/home/admin/hadoop/job.jar [jobmainclass] [Jobargs]
Kill a running jobSuppose job_id is: job_201005310937_00531. Enter the Hadoop_home directory.2. Execute SH bin/hadoop job-kill job_201005310937_0053
more commands for HadoopThe operations
This article provides a detailed analysis of some commonly used commands in hadoop. For more information, see Hadoop installation directory HADOOP_HOME:/home/admin/hadoop.
Start and closeStart Hadoop1. go to the HADOOP_HOME directory.
2. execute sh bin/start-all.sh
Disable Hadoop1. go to the HADOOP_HOME directory.2.
Unix operating commands
Pan lei, Institute of Theoretical Physics, Chinese Emy of Sciences
Chapter 1 logon and exit
Before you log on, you must first apply for a user registration name from the system administrator. No matter which computer the user logs on to itpnet, the user will access the same file system. 1.1 When a login prompt appears on the screen, you can enter your User Registration
as the input of another command, you can use the pipeline symbol |. The following shows that the echo output is used as the input of sort using the pipeline.
[root@localhost ~]# echo -e "apple\npear\nbanana" | sortapplebananapear[root@localhost ~]#
Each command may have options or parameters. You can also use | to output the second command as the input of the Third Command, and so on. It is a common method in Linux/Unix to build a long command pipel
Two days ago, the "some little-known but interesting Unix/Linux commands" series were released.Article, Received praise from some friends. (I wanted to write "a lot of friends", but I thought Linux was really a small crowd, so brother D would not be forced! Haha) D. Will make persistent efforts. The links to the first two articles are as follows:
Some little-known but interesting
Article Title: Command-oriented Unix system commands in one go. Linux is a technology channel of the IT lab in China. Includes basic categories such as desktop applications, Linux system management, kernel research, embedded systems, and open source.
In Unix, there is a very practical function, namely the command-oriented function. Simply put, command-oriented me
A. Common Hadoop commands1. The FS command for Hadoop#查看hadoop所有的fs命令Hadoop FS#上传文件 (both put and copyfromlocal are upload commands)Hadoop fs-put jdk-7u55-linux-i586.tar.gz hdfs://hucc01:9000/jdkhadoop fs-copyfromlocal jdk-7u55-li
3. Accumulated Linux/UNIX operation commands (gzip, jar, tar, find, crontab) and gzipcrontab
In Linux/UNIX, users often use text interfaces to set up systems or operating systems. The author is constantly working on these commands, therefore, this article has been specially developed, prepared, and started. This articl
When using a git command that can has a large amount of output (like git log , git diff or git blame ), Git opens the command outpu T in our terminal "pager"; On most modern unix-based systems, the default pager would be ' less '. Learning a few less commands would help us deal with this git command output that opens in the pager. In this lesson, we show some of the very useful of the less
Grep command series: How to Use grep commands in Linux/UNIX
How can I use the grep command in Linux, Apple OS X, and other UNIX-like systems? Can you give me some simple examples?
The grep command is used to search for text, or to search for a file containing a given string or word in a row from a given file. In general, grep displays the matched rows. Use grep t
UNIX linux vi command 1. vi file editing commands have special characteristics, including editing mode and command mode (inconsistent with unix commands, such as ll and who). Therefore, when editing files under Vi commands, it is often necessary to switch various command mod
Classified communication of some LinuxUnix commands: ftp file transfer protocol login logs on to Unixmailx to read or send emails to rlogi
Summarizes the usage of some Linux/Unix commands.
Communication:
Ftp file transfer protocol
Login logs on to Unix
Mailx reads or sends emails
Rlogin logs on to remote
. Compare the command with the comm and uniq commands.
Dir
Purpose: Dir is used to list the contents of directories on some platforms.
For example, Dir (DIR has the same functions as LS, but it has fewer options. DIR is not suitable for all UNIX systems ).
Du
Purpose: Report disk usage by Du.
Example: Du. The following is the list output after the du command is executed.
Echo
Purpose: ECHO is used to displa
1.Xcode Operating principle
When we click Run, Xcode automatically does the following things for us:
Compile--->.o (target file)---> Links--->.out execution
2. Common UNIX Commands
Mac system uses UNIX file system, all files are placed in the root directory/below, so there is no concept of windows in the C-drive, D-Disk
Becau
Today, let's learn about Common Unix shell commands. We all know that Unix shell is not only a user interaction interface, but alsoProgramDesign Language: when the system registers, it will execute a Unix shell command file. profile, which will give a brief introduction to Common c
all connections on an application that supports SOCKET proxy to a remote machine. This is useful for more secure use of public wifi to browse networks and bypass strict firewalls.
5. Enter the password on the local machine once, and then use the ssh key proxy to log on to other remote machines with a security identity without entering your password again.
For more information, see the ssh manual page.
5. mdfind
The unique commands on this mac, of co
region:#hbase> major_compact ‘r1‘, ‘c1‘#Compact a single column family within a table:#hbase> major_compact ‘t1‘, ‘c1‘
Configuration Management and node restart1) Modify the HDFs configurationHDFs Configuration Location:/etc/hadoop/conf
# 同步hdfs配置cat /home/hadoop/slaves|xargs -i -t scp /etc/hadoop/conf/hdfs-site.x
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.