Hadoop development is divided into two components: the build of Hadoop clusters, the configuration of the Eclipse development environment. Several of the above articles have documented my Hadoop cluster setup in detail, A simple Hadoop-1.2.1 cluster consisting of a master an
Apache Hadoop configuration Kerberos Guide
Generally, the security of a Hadoop cluster is guaranteed using kerberos. After Kerberos is enabled, you must perform authentication. After verification, you can use the GRANT/REVOKE statement to control role-based access. This article describes how to configure kerberos in a CDH cluster.
1. KDC installation and
, there is an. SSH directory
Id_rsa private Key
Id_rsa.pub Public Key
Known_hosts via SSH link to this host, there will be a record here
2. Give the public key to the trusted host (native)
Enter the Ssh-copy-id host name at the command line
Ssh-copy-id Master
Ssh-copy-id slave1
Ssh-copy-id Slave2
The password for the trusted host needs to be entered during replication
3. Verify, enter in command line: SSH Trust host name
SSH Master
SSH slave1
SSH slave2
If you are not prompted to enter a passwor
/. ssh/directory on the host 192.168.1.106.
Scp./id_rsa.pub root@192.168.1.106:/root/. ssh/authorized_keys
3) copy the public key on the host 192.168.1.106 to the corresponding/root/. ssh/directory on the host 192.168.1.105.
Scp./id_rsa.pub root@192.168.1.105:/root/. ssh/authorized_keys
4) both machines enter the/root/. ssh directory and run cat id_rsa.pub> authorized_keys
5) After configuration, ssh cloud001 and ssh cloud002 on both hosts should be p
1. Install
Here the installation of hadoop-0.20.2 as an Example
Install Java first. Refer to this
Download hadoop
Extract
tar -xzf hadoop-0.20.2
2. Configuration
Modify Environment Variables
Vim ~ /. Bashrcexport hadoop_home =/home/RTE/
Hadoop Configuration class analysis, hadooptext classConfiguration is a public class of five components in hadoop. Therefore, it is placed under the core, org. apache. hadoop. conf. Configruration. This class is the Configuration information class of the job.
2.2 Hadoop Configuration Detailed
Instead of using the Java.util.Properties management profile and using the Apache Jakarta Commons configuration profile, Hadoop uses a unique set of configuration file management systems and provides its own API , which is to use org.apache.
network segment. However, different transmission channels can be defined within the same network segment. 2 Environment
Platform: ubuntu12.04
Hadoop: hadoop-1.0.4
Hbase: hbase-0.94.5.
Topology:
Figure 2 hadoop and hbase Topology
Software Installation: APT-Get 3. installation and deployment (unicast) 3.1 deployment Method
Monitoring node (gmond):
Record the configuration and description of Hadoop, and use the new configuration items to be replenished and updated on a regular basis. Divide by Profile Name
Take the Hadoop 1.x configuration as an example
Core-site.xml
other machines as Datanode, stand-alone mode This machine is datanode, so modify slaves configuration file for the local domain name. For example: The machine name is HADOOP11, then:
[Hadoop@hadoop11 ~]$ Cat Hadoop/conf/slaves
Hadoop11
When the configuration is complete, st
there may be problems when analyzing the file )# ntpdate 202.120.2.101 ( server of Shanghai Jiaotong University )Third, install Hadoop The official download site for Hadoop , you can choose the appropriate version download:http://hadoop.apache.org/releases.htmlPerform the following operations on three machines, respectively:# Tar XF hadoop-2.7.2.tar.gz# MV
Follow the Hadoop installation tutorial _ standalone/pseudo-distributed configuration _hadoop2.6.0/ubuntu14.04 (http://www.powerxing.com/install-hadoop/) to complete the installation of Hadoop, My system is hadoop2.8.0/ubuntu16.
Hadoop Installation Tutorial _ standalone/pseu
Hadoop cluster hardware standard configuration
When selecting hardware, we often need to consider the performance and expenditure of applications. To this end, we must find a perfect balance between meeting actual needs and being economically feasible. The following uses the Hadoop cluster application as an example to
Document directory
1. Map stage
3. Let's take a general look at the Running code of the job:
This series of hadoop learning notes is based on hadoop: the definitive guide 3th, which collects additional information on the Internet and displays hadoop APIs, and adds its ownPracticeIs mainly used to learn the features and functions of
pluginWindows->preference->hadoop Map/reduce, this document configures the Hadoop processing directory in D:\hadoop. It should be noted that the directory indicates the relevant jar packages required for subsequent compilation of the source program and the required library files (required by Windows compilation).3) Switching angle of viewWindows->open Perspectiv
Recently, I think we should take a closer look at the hadoop source code. I used to understand the basic architecture and use it. Recently, I am working on a system. I think many things can be used to learn from the scalability of mapreduce. However, when our system version 0.1 appeared, we found that our configuration was messy. So I took a look at the hadoop
The configuration is the common class of the five components of Hadoop, so put it under the core, org.apache.hadoop.conf.Configruration. This class is the configuration information class for the job, and any configuration information that can be used must be passed through configur
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.