Want to know hadoop cluster configuration best practices? we have a huge selection of hadoop cluster configuration best practices information on alibabacloud.com
Add hard disks to the Hadoop cluster.
Hadoop worker nodes expand hard disk space
After receiving the task from the boss, the hard disk space in the Hadoop cluster is insufficient, and a machine is required to be added to the Hadoop
dns.hadoop.com.
Dns.hadoop.com. in a 192.168.1.230
H1.hadoop.com. in a 192.168.1.231
H2.hadoop.com. in a 192.168.1.20.
H3.hadoop.com. in a 192.168.1.233
H4.hadoop.com. in a 192.168.1.234
H5.hadoop.com. in a 192.168.1.235
H6.hadoop.com. in a 192.168.1.236
H7.hadoop.com. in a 192.168.1.237
H8.hadoop.com. in a 192.168.1.238
Configure reverse resolution File
Cp named. localhost named.192.168.1.zone
Add the following content:
$ TTL 1D
@ In soa dns.hadoop.com. grid.dns.hadoop.com (
0; serial
1D; ref
Virtual machine-based Hadoop cluster installation1. The software we needXshell, SSH secure, virtual machine, Linux centos64, Hadoop1.2.1 installation package2. Install the above software3, install Linux, there is no more elaboration4. Install the JDK firstMy path isjava_home=/usr/lib/jvm/jdk1.7.0_79Path=
PATH:
Java_home/binClasspath=
J AV AH OM E /LIb/d T.JaR:
Java_home/lib/t
In the home of two computers with VMware + RedHatLinuxAS6 + Hadoop-0.21.0 to build a 3 node Hadoop cluster, although it is already set up a similar cluster, I also ran Java API to operate HDFS and Map/reduce, but this time it was still challenged. Some small details and some omissions would be like a roller coaster. Th
/id_rsa.pub ~/.ssh/authorized_keys4) master native uses SSH localhost test:The first time you will be prompted whether "is you sure want to continue connecting (yes/no)?", enter Yes directly, the next time you enter SSH localhost will not be prompted.5) Modify the hosts for each node (MASTER,NODE1,NODE2,NODE3):Statistics add the following host list:The purpose is to use the SSH connection for the rear, without entering the IP, using the machine name.6) In order to ensure that master can automati
First, Hadoop downloadUse the 2.7.6 version, because the company production environment is this versionCD/optwget http://mirrors.hust.edu.cn/apache/hadoop/common/hadoop-2.7.6/ Hadoop-2.7.6.tar.gzSecond, the configuration fileReference Document: https://hadoop.apache.org/docs
file (clip)7. Delete a folder?Answer: RM-RF file (folder)8. Do I need to install zookeeper?The default value of HBASE_MANAGES_ZK in the Conf/hbase-env.sh configuration document is True, which indicates that HBase uses its own zookeeper instance. However, the instance can only serve hbase in standalone or pseudo-distributed mode. When installing full distribution mode, you need to configure your own zookeeper instance. After configuring the Hbase.zook
. Client:retrying Connect to server:hadoop-05/192.168.0.7:8032. Already tried 6 time (s); Retry policy is Retryuptomaximumcountwithfixedsleep (maxretries=10, sleeptime=1000 MILLISECONDS)
ResourceManager not connected. Check that the Yarn-site.xml are all configured.
However, the discovery port number is inconsistent with the default port number, so modify the
The configuration file changes to the following:
Rerun, the same error still occurs, and the
After the Hadoop cluster is set up, the Hadoop cluster is accessed locally via the Java API, as follows (see all node name information on the Hadoop cluster)
Import org.apache.hadoop.conf.Configuration;
Import Org.apache.hadoop.f
This article describes the configuration method for using the HDFs Java API.1, first solve the dependence, pomDependency> groupId>Org.apache.hadoopgroupId> Artifactid>Hadoop-clientArtifactid> version>2.7.2version> Scope>ProvidedScope> Dependency>2, configuration files, storage HDFs
-2.4.1.tar.gz-c/java/decompression hadoopls lib/native/See what files are in the extracted directory CD etc/hadoop/into the profile directory vim hadoop-env.sh Modify Profile environment variable (export java_home=/java/jdk/jdk1.7.0_65) *-site.xml*vim core-site.xml Modify configuration file (go to official website for parameter meaning) ./
If there is a place to look at the mask, take a look at the HDFs ha this articleThe official scheme is as follows
Configuration target:
Node1 Node2 Node3:3 Station ZookeeperNode1 Node2:2 sets of ResourceManager
First configure Node1, configure Etc/hadoop/yarn-site.xml:
Configuration etc/hadoop/mapred-site.xml:
Cop
command[Email protected]:~/.ssh# cat id_rsa.pub >> master_key [email protected]:~/.ssh# SCP master_key [email Pro tected]:/root/.ssh/Write Authorized_keys[Email protected]:~/.ssh# cat Master_key >> Authorized_keys"Note": Each of the two machines completes the public key write operation5 Main Ideas Install JDKDetailed installation steps reference how to install the Oracle Java JDK on Ubuntu LinuxInstall compiled software in Ubuntu, the general step is to unzip the installation package, modify th
Reprinted from: http://www.cnblogs.com/spark-china/p/3941878.html
Prepare a second, third machine running Ubuntu system in VMware;
Building the second to third machine running Ubuntu in VMware is exactly the same as building the first machine, again not repeating it.Different points from installing the first Ubuntu machine are:1th: We name the second to third Ubuntu machine for Slave1, Slave2, as shown in:There are three virtual machines in the created VMware:2nd: To simplify the
Most Hadoop clusters adopt Kerberos as the authentication protocolInstalling the KDC
Starting Kerberos authentication requires the installation of the KDC server and the necessary software. The command to install the KDC can be executed on any machine.
Yum-y Install krb5-server krb5-lib krb5-auth-dialog krb5-workstation
Next, install the Kerberos client and the command on the other nodes in the
Transferred from: http://blog.csdn.net/huyuxiang999/article/details/17691405First, the experimental environment:1, Hardware: 3 Dell Servers, CPU:2.27GHZ*16, Memory: 16GB, one for master, and the other 2 for slave.2, the system: all CentOS6.33, Hadoop version: CDH4.5, the use of the MapReduce version is not yarn, but Mapreduce1, the entire cluster under Cloudera Manager monitoring,
org.apache.hadoop.ipc.Client:Retrying Connect to server:0.0.0.0/0.0.0.0:8031. Already tried 7 time (s); Retry policy is Retryuptomaximumcountwithfixedsleep (maxretries=10, sleeptime=1000 MILLISECONDS) 2017-06-05 09:49:46,472 INFO org.apache.hadoop.ipc.Client:Retrying Connect to server:0.0.0.0/0.0.0.0:8031. Already tried 8 time (s); Retry policy is Retryuptomaximumcountwithfixedsleep (maxretries=10, sleeptime=1000 MILLISECONDS) 2017-06-05 09:49:47,474 INFO org.apache.hadoop.ipc.Client:Retrying C
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.