Spark Installation II: Hadoop cluster deployment

Source: Internet
Author: User
Tags xsl

First, Hadoop download

Use the 2.7.6 version, because the company production environment is this version

CD/optwget http://mirrors.hust.edu.cn/apache/hadoop/common/hadoop-2.7.6/ Hadoop-2.7.6.tar.gz

Second, the configuration file

Reference Document: https://hadoop.apache.org/docs/r2.7.6

Configuration of 7 files in the $hadoop_home/etc/hadoop directory

1.core-site.xml

<?XML version= "1.0" encoding= "UTF-8"?><?xml-stylesheet type= "text/xsl" href= "configuration.xsl "?><Configuration>    < Property>        <name>Fs.defaultfs</name>        <value>hdfs://pangu10:9000</value>        <Description>NameNode URI,HDFS Processing External ports</Description>    </ Property>    < Property>        <name>Hadoop.tmp.dir</name>        <value>/opt/hdfs/tmp</value>        <Description>When HDFs is reformatted (such as a new Datenode), this temporary directory needs to be deleted</Description>    </ Property></Configuration>
View Code

2.hdfs-site.xml

<?XML version= "1.0" encoding= "UTF-8"?><?xml-stylesheet type= "text/xsl" href= "configuration.xsl "?><Configuration>    < Property>        <name>Dfs.namenode.name.dir</name>        <value>File:/opt/hdfs/name</value>        <Description>HDFs namespace metadata stored on Namenode</Description>    </ Property>    < Property>        <name>Dfs.datanode.data.dir</name>        <value>File:/opt/hdfs/data</value>        <Description>Physical storage location of data blocks on Datanode</Description>    </ Property>    < Property>        <name>Dfs.replication</name>        <value>1</value>        <Description>Set the number of Dfs replicas, not set by default is 3</Description>    </ Property>    < Property>        <name>Dfs.namenode.secondary.http-address</name>        <value>pangu11:50090</value>        <Description>Setting the port for Secondname</Description>    </ Property></Configuration>
View Code

3.yarn-site.xml

<?XML version= "1.0"?><Configuration>    < Property>        <name>Yarn.resourcemanager.hostname</name>        <value>Pangu10</value>        <Description>Specify the hostname where the ResourceManager is located</Description>    </ Property>    < Property>        <name>Yarn.nodemanager.aux-services</name>        <value>Mapreduce_shuffle</value>        <Description>NodeManager run on the secondary service, need to be configured as Mapreduce_shuffle, to run the MapReduce program</Description>    </ Property>    < Property>        <name>Yarn.nodemanager.pmem-check-enabled</name>        <value>False</value>    </ Property>    < Property>        <name>Yarn.nodemanager.vmem-check-enabled</name>        <value>False</value>    </ Property></Configuration>
View Code

4.mapred-site.xml

<?XML version= "1.0"?><?xml-stylesheet type= "text/xsl" href= "configuration.xsl "?><Configuration>    < Property>        <name>Mapreduce.framework.name</name>        <value>Yarn</value>        <Description>Specify mapreduce using the yarn framework</Description>    </ Property></Configuration>
View Code

5.slaves

Pangu10pangu11pangu12

6.yarn-env.sh

Find Line 23rd

# Export JAVA_HOME=/HOME/Y/LIBEXEC/JDK1. 6.0/

Replaced by

Export JAVA_HOME=/OPT/JDK1. 8. 0_181/

7.hadoop-env.sh

25 Rows found

Export Java_home=${java_home}

Replaced by

Export JAVA_HOME=/OPT/JDK1. 8. 0_181/

Third, copy to Slave

Iv. format of HDFs

The shell executes the following command

Hadoop Namenode-format

Formatting succeeds if the following red log content appears

 -/Ten/ A  A: -: -INFO util. Gset:capacity =2^ the=32768Entries -/Ten/ A  A: -: -INFO Namenode. fsimage:allocated New blockpoolid:bp-1164998719-192.168.56.10-153936231358418/10/12 12:38:33 INFO Common. Storage:storage Directory/opt/hdfs/name has been successfully formatted. -/Ten/ A  A: -: -INFO Namenode. Fsimageformatprotobuf:saving imagefile/opt/hdfs/name/current/fsimage.ckpt_0000000000000000000 using no compression -/Ten/ A  A: -: -INFO Namenode. Fsimageformatprotobuf:imagefile/opt/hdfs/name/current/fsimage.ckpt_0000000000000000000 of size theBytes savedinch 0seconds. -/Ten/ A  A: -: -INFO Namenode. Nnstorageretentionmanager:going to retain1Images with Txid >=0 -/Ten/ A  A: -: -INFO util. Exitutil:exiting with status0 -/Ten/ A  A: -: -INFO Namenode. Namenode:shutdown_msg:/************************************************************shutdown_msg:shutting down NameNode at pangu10/ 192.168.56.10************************************************************/

V. Start Hadoop

Spark Installation II: Hadoop cluster deployment

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.