Learn about linux commands for hadoop administration, we have the largest and most updated linux commands for hadoop administration information on alibabacloud.com
People rely on search engines every day to find specific content from the massive amount of data on the Internet. But have you ever wondered how these searches are executed? One method is Apache Hadoop, which is a software framework that can process massive data in a distributed manner. An Application of Hadoop is to index Internet Web pages in parallel. Hadoop i
Setting up Hadoop cluster environment steps under Ubuntu 12.04I. Preparation before setting up the environment:My native Ubuntu 12.04 32bit as Maser, is the same machine that was used in the stand-alone version of the Hadoop environment, http://www.linuxidc.com/Linux/2013-01/78112.htmAlso in the KVM Virtual 4 machines, respectively named:Son-1 (Ubuntu 12.04 32bit
To make it easier for you to review your memory, we will summarize the hadoop commands of today's experiment for later viewing.Note: The following commands are performed in hadoop/bin.1. hadoop FS-ls \-> View All the following directories.2.
Apache-->hadoop's official Website document Command learning:http://hadoop.apache.org/docs/r1.0.4/cn/hdfs_shell.html
FS Shell
The call file system (FS) shell command should use the bin/hadoop fs scheme://authority/path. For the HDFs file system, Scheme is HDFs, to the local file system, scheme is file. The scheme and authority parameters are optional, and if not specified, the default scheme specified in the configuration is used. An HDFs
Fsck commands in Hadoop
The fsck command in Hadoop can check the file in HDFS, check whether there is upt corruption or data loss, and generate the overall health report of the hdfs file system. Report content, including:Total blocks (Total number of blocks), Average block replication (Average number of copies), upt blocks, number of lost blocks,... and so on.---
yourself, all of which are performed under that user's environment.4. Download and install JDK
sudo apt-get install Oracle-java8-installer #也可使用wget直接从oracle官网下载
After downloading, configure the JDK environment variables and execute Java, javac commands for testing.
5. Configure hosts
As described in the experimental environment above, the Hadoop cluster consists of a master node and two slave nodes, wh
Preface: Well, it's just a little bit more comfortable without writing code, but we can't slack off. The hive operation's file needs to be loaded from hereSimilar to the Linux commands, the command line begins with the Hadoop FS -(dash) LS / list file or directory cat Hadoop FS -cat ./hello.txt/opt/old/ Htt/h
This article is intended to provide the most basic, can be used in the production environment of Hadoop, HDFS distributed environment of the building, Self is a summary and collation, but also to facilitate the new learning to use.Installation and configuration of the base environment JDKIt is not easy to find JDK7 's installation packages directly to Oracle's official website (http://www.oracle.com/), as it is now officially recommended JDK8. Found a
From:http://www.2cto.com/database/201303/198460.htmlHadoop HDFs Common CommandsHadoop common commands:Hadoop FSView all commands supported by Hadoop HDFsHadoop fs–lslisting directory and file informationHadoop FS–LSRLoop lists directories, subdirectories, and file informationHadoop fs–put Test.txt/user/sunlightcsCopy the test.txt of the local file system to the/user/sunlightcs directory of the HDFs file sys
machine:
6.1 Install ssh
For example on Ubuntu Linux:
$ Sudo apt-get install ssh$ Sudo apt-get install rsync
Now check that you can ssh to the localhost without a passphrase:$ Ssh localhost
If you cannot ssh to localhost without a passphrase, execute the following commands:$ Ssh-keygen-t dsa-p'-f ~ /. Ssh/id_dsa$ Cat ~ /. Ssh/id_dsa.pub> ~ /. Ssh/authorized_keys
Then can ssh from master to slaves: scp ~ /.
mytable from Database mydb to the E:\MySQL\mytable.sql file.
c:\> mysqldump-h localhost-u root-p mydb mytable>e:\mysql\mytable.sql
Example 3: Export the structure of the database mydb to a e:\MySQL\mydb_stru.sql file.
c:\> mysqldump-h localhost-u root-p mydb--add-drop-table >e:\mysql\mydb_stru.sql
Note :-h localhost can be omitted, it is generally used on the virtual host.3) Export data structure onlyformat : mysqldump-u [Database user name]-p-t [the name of the data
localhost-u root-p mydb >e:\mysql\mydb.sql
Then enter the password, wait for an export to succeed, you can check the target file for success. Example 2: Export mytable from Database mydb to the E:\MySQL\mytable.sql file.
c:\> mysqldump-h localhost-u root-p mydb mytable>e:\mysql\mytable.sql
Example 3: Export the structure of the database mydb to a e:\MySQL\mydb_stru.sql file.
c:\> mysqldump-h localhost-u root-p mydb--add-drop-table >e:\mysql\mydb_stru.sql
I recently tried to build the environment for Hadoop, but I really don't know how to build it. The next hop was a step-by-step error. Answers from many people on the Internet are also common pitfalls (for example, the most typical is the case sensitivity of commands, for example, hadoop commands are in lower case, and
Two cyanEmail: [Email protected] Weibo: HTTP://WEIBO.COM/XTFGGEFWould like to install a single-node environment is good, and then after the installation of the total feel not enough fun, so today continue to study, to a fully distributed cluster installation. The software used is the same as the previous one-node installation of Hadoop, as follows:
Ubuntu 14.10-Bit Server Edition
Hadoop2.6.0
JDK 1.7.0_71
Ssh
Rsync
Prepare
Javax.security.auth.Subject.doAs (subject.java:396)At Org.apache.hadoop.ipc.server$handler.run (server.java:953)Workaround: Add the following in the Hdfs-site.xml
1234
HDFs common commands to create folders
1
./hadoop Fs–mkdir/usr/local/hadoop/godlike
Uploading files
1
./
Linux build Hadoop Environment 1, install JDK (1) Download and install JDK: Make sure the computer is networked after the command line enter the following command to install the JDK sudo apt-get install SUN-JAVA6-JDK (2) Configure the computer Java environment: Open/etc /profile, enter the following content at the end of the file export Java_home = (JAVA installation directory) export CLASSPATH = ".: $JAVA
1. Hadoop software delivered to virtual machines Or use WINSCP to put the Hadoop software installation package in the Linux downloads folder. 2. Select the installation directory Copy the Hadoop installation package into this installation directory, where we select the/usr/local directory in CentOS. 3. Unzip the instal
This article assumes that users have a basic understanding of Docker, Master Linux basic commands, and understand the general installation and simple configuration of Hadoop.Lab Environment: windows10+vmware WorkStation 11+linux.14.04 server+docker 1.7 windows 10 as the physical machine operating system, the network segment is: 10.41.0.0/24, the virtual machine u
A lot of people have written about the process of installing Hadoop under Virtual Machine Linux, but in the actual installation process, there are still a lot of other people are not involved but really need to pay attention to the part, so as to ensure the operation, but also as a small summary of their own.
First, about the Linux system settings
1. Set static
Hadoop version: hadoop-0.20.2Eclipse Version: eclipse-java-helios-sr2-linux-gtk.tar.gz
======================== installation eclipse=======================
1, first download eclipse. Not much.
2. Install Eclipse(1) to extract the eclipse-java-helios-sr2-linux-gtk.tar.gz into a directory, I unzipped to the/home/wangxin
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.