Full distributed model of Hadoop Environment building

Source: Internet
Author: User

First, this article explains:
This test is fully distributed on five virtual machine systems. Hadoop distributed mode is distributed on multiple machines, configuration is very similar to pseudo-distributed, the name node and data node are distributed on different hosts, each data node has the corresponding task scheduling.

Second, the environment description:
Virtual Software: VM10.0
Operating system: Cent OS 6.7
Hadoop version: hadoop-2.7.2
JDK version: jdk1.8_77
Note: Built on an independent model, refer to the article: the standalone model of the Hadoop environment.

Three, the basic Environment configuration
1. Set up different configuration folders for different modes under ${hadoop_home}/etc/
Standalone mode: Hadoop_alone (recommended)
Pseudo distribution mode: Hadoop_presudo (recommended)
Fully distributed mode: Hadoop_cluster (recommended)

2, the parameter configuration directory to specify the way
A, default mode: ${hadoop_home}/etc/hadoop/*.xml
B. Specify the configuration directory through the startup parameters
Start-dfs.sh–config=/path
C, through the environment variable Hadoop_conf_dir=/path (/etc/profile)
Note: Because of the full distribution pattern: I am here to format the name node and the B method used when starting Hadoop.
Iv. Environment Construction
1th Step: Build the Java and Hadoop infrastructure on the master machine.
Reference article: Building a Hadoop environment in standalone mode

2nd step: Configure SSH on the master machine
650) this.width=650; "src=" Http://www.itroad.org/uploadfile/2016/0414/20160414065505551.png "style=" height:370px; width:600px; "alt=" 20160414065505551.png "/>

3rd Step: Configure the hosts on the master machine
650) this.width=650; "src=" Http://www.itroad.org/uploadfile/2016/0414/20160414065538112.png "style=" HEIGHT:153PX; width:600px; "alt=" 20160414065538112.png "/>


4th Step: Clone 4 virtual machines and modify the configuration of the 4 network cards
Reference article: How to modify network adapter after cloning Linux

5th step: Configure ${hadoop_home}/etc/hadoop_ cluster/core-site.xml 650) this.width=650 on the master machine; "Src=" http://www.itroad.org/ Uploadfile/2016/0414/20160414065602348.png "style=" height:436px;width:600px; "alt=" 20160414065602348.png "/>


6th Step: Configure ${hadoop_home}/etc/hadoop_cluster/hdfs-site.xml
650) this.width=650; "src=" Http://www.itroad.org/uploadfile/2016/0414/20160414065621844.png "style=" HEIGHT:414PX; width:600px; "alt=" 20160414065621844.png "/>


7th Step, Configuration ${hadoop_home}/etc/hadoop_cluster/mapred-site.xml
650) this.width=650; "src=" Http://www.itroad.org/uploadfile/2016/0414/20160414070011229.png "style=" HEIGHT:464PX; width:600px; "alt=" 20160414070011229.png "/>



8th step, Configuration ${hadoop_home}/etc/hadoop_cluster/yarn-site.xml 650) this.width=650; "Src=" Http://www.itroad.org/uploadfile /2016/0414/20160414070045314.png "style=" height:479px;width:600px; "alt=" 20160414070045314.png "/>


9th step, delete the ${hadoop_home}/etc/hadoop_cluster on the other four hosts

The 10th step, the machine on the ${hadoop_home}/etc/hadoop_cluster remote copy to another four machines on the ${hadoop_home}/etc/
650) this.width=650; "src=" Http://www.itroad.org/uploadfile/2016/0414/20160414065737404.png "style=" HEIGHT:41PX; width:600px; "alt=" 20160414065737404.png "/>

etc...

11th step, format the HDFs file system on the master machine
650) this.width=650; "src=" Http://www.itroad.org/uploadfile/2016/0414/20160414070118629.png "style=" height:30px; width:600px; "alt=" 20160414070118629.png "/>


12th Step, start the daemon for HDFs and Yran (start-dfs.sh,start-yarn.sh)
start-dfs.sh--config/home/program/hadoop/etc/hadoop_cluster/
start-yarn.sh--config/home/program/hadoop/etc/hadoop_cluster/
Note the stop process is followed by yarn close then close HDFs, do not format the name node frequently

13th step, operation of the files in HDFs via FS
A. Create directory: Hadoop fs-mkdir-p/user/ubuntu
B. Display directory: Hadoop fs-ls-r/
......
14th step, visit: http://IP:50070
Visit: http://IP:8088
650) this.width=650; "src=" Http://www.itroad.org/uploadfile/2016/0414/20160414070141437.png "style=" HEIGHT:424PX; width:600px; "alt=" 20160414070141437.png "/>
650) this.width=650; "src=" Http://www.itroad.org/uploadfile/2016/0414/20160414070144910.png "style=" HEIGHT:207PX; width:600px; "alt=" 20160414070144910.png "/>


Full distributed model of Hadoop Environment building

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.