1. Create a user groupsudo addgroup Hadoop2. Create usersudo adduser-ingroup Hadoop HadoopEnter the password after entering, enter the password you want to set and then go all the way.3. Add Permissions for Hadoop userssudo gedit/etc/sudoersThen save the exit.4. Switch User Hadoop login operating system5. Install SSHsudo apt-get Install Openssh-serverStart the SS
Hadoop-1.x installation and configuration
1. Install JDK and SSH before installing Hadoop.
Hadoop is developed in Java. MapReduce and Hadoop compilation depend on JDK. Therefore, JDK1.6 or later must be installed first (JDK 1.6 is generally used in the actual production envi
Ubuntu version 12.04.3 64-bitHadoop is run on a Java virtual machine, so you'll need to install the Jdk,jdk installation configuration method to install it under another blog post ubuntu12.04 jdk1.7SOURCE Package Preparation:I downloaded the hadoop-1.2.1.tar.gz, this version is relatively stable, can be provided to the official website of the image http://www.apache.org/dyn/closer.cgi/
1. Before installing Hadoop, you need to install the JDK and SSH first.Hadoop is developed in Java language, and the operation of MapReduce and the compilation of Hadoop depend on the JDK. Therefore, you must first install JDK1.6 or later (JDK1.6 is generally used in a real-world production environment, because some components of Hadoop do not support JDK1.7 and
Now there are a lot of articles on hadoop installation on the network. I also tried to install it according to their methods. Hey, this is not good. If there is no problem, I can only find Gu's teacher one by one, what Mr. Gu provided was messy, and finally it was installed. I wrote this article based on the name of the machine. It is for reference only.
Machine name IP Address
Master 10.64.79.153 namenode
I. Hadoop installation and Considerations1. To install the Hadoop environment, you must have a Java environment in your system.2. SSH must be installed, and some systems will be installed by default, if not installed manually.Can be installed with Yum install-y ssh or RPM-IVH ssh rpm packageTwo. Install and configure the Java environmentHadoop needs to run in a J
libsnappy.lalrwxrwxrwx 1 root root 7 11:56 libsnappy.so libsnappy.so.1.2.1lrwxrwxrwx 1 root root 7 11:56 libsnappy.so.1-libsnappy.so.1.2.1-rwxr-xr-x 1 root root 147758 7 11:56 libsnappy.so.1.2.1If an error is not encountered during the installation and the/usr/local/lib directory has the above file, the installation is successful.4, Hadoop-snappy source
"/>, down to master. hadoop/opt/
wgethttp://archive-primary.cloudera.com/cm5/cm/5/cloudera-manager-el6-cm5.1.1_x86_64.tar.gz
[[emailprotected]opt]$sudotar-xzvfcloudera-manager-el6-cm5.1.0_x86_64.tar.gz
Check after installation
[[emailprotected]opt]$lsclouderacloudera-manager-el6-cm5.1.1_x86_64.tar.gzcm-5.1.1
Delete the cloudera-SCM created between us
# Confirm the following content in/etc/passwd and disapp
first, system and software environment1. Operating systemCentOS Release 6.5 (Final)Kernel version:2.6.32-431.el6.x86_64master.fansik.com:192.168.83.118node1.fansik.com:192.168.83.119node2.fansik.com:192.168.83.1202.JDK version:1.7.0_753.Hadoop version:2.7.2second, pre-installation preparation1. Turn off firewall and SELinux# Setenforce 0# Service Iptables Stop2. Configuring the host file192.168.83.118 maste
Follow the Hadoop installation tutorial _ standalone/pseudo-distributed configuration _hadoop2.6.0/ubuntu14.04 (http://www.powerxing.com/install-hadoop/) to complete the installation of Hadoop, My system is hadoop2.8.0/ubuntu16.
Hadoop
Hadoop installation is not difficult, but requires a lot of preparation work. 1. JDK needs to be installed first. Centos can be installed directly through Yum install java-1.6.0-openjdk. The installation methods for different release versions may be different. 2. After setting SSH, you must set SSH as the key for logon authentication. If you do not have this step
Two cyanEmail: [Email protected] Weibo: HTTP://WEIBO.COM/XTFGGEFNow it's time to learn about Hadoop in a systematic way, although it may be a bit late, but you want to learn the hot technology, let's start with the installation environment. Official documentsThe software and version used in this article are as follows:
Ubuntu 14.10-Bit Server Edition
Hadoop2.6.0
JDK 1.7.0_71
Ssh
Rsy
Two cyanEmail: [Email protected] Weibo: HTTP://WEIBO.COM/XTFGGEFNow it's time to learn about Hadoop in a systematic way, although it may be a bit late, but you want to learn the hot technology, let's start with the installation environment. Official documentsThe software and version used in this article are as follows:
Ubuntu 14.10-Bit Server Edition
Hadoop2.6.0
JDK 1.7.0_71
Ssh
Rsy
original path to the target path Hadoop fs-cat/user/hadoop/a.txt View the contents of the A.txt file Hadoop fs-rm/user/hadoop/a.txt Delete US The A.txt file below the Hadoop folder under the ER folderHadoop fs-rm-r/user/hadoop/a.
On the network on how to install a single-machine mode of Hadoop article many, according to its steps down most of the failure, in accordance with its operation detours through a lot but after all, still solve the problem, so by the way, detailed record of the complete installation process.This article mainly describes how to set up a virtual machine environment in a Windows environment, as well as some
change!): 10969 DataNode11745 NodeManager11292 SecondaryNameNode10708 NameNode11483 ResourceManager13096 Jps n.b. The old jobtracker have been replaced by the ResourceManager.
Access Web interfaces:
Cluster status:http://localhost:8088
HDFS status:http://localhost:50070
Secondary NameNode status:http://localhost:50090
Test Hadoop:hadoop jar ~/hadoop/share/hadoop/mapredu
network segment. However, different transmission channels can be defined within the same network segment. 2 Environment
Platform: ubuntu12.04
Hadoop: hadoop-1.0.4
Hbase: hbase-0.94.5.
Topology:
Figure 2 hadoop and hbase Topology
Software Installation: APT-Get 3. in
Rhive is a package that extends R computing power through hive high-performance queries. It can be very easy to invoke HQL in the R environment, and also allows the use of R's objects and functions in hive. In theory, the data processing capacity can be infinitely extended hive platform, with the tool r environment of data mining, is a perfect big data analysis mining work environment.Resource bundle:Http://pan.baidu.com/s/1ntwzeTbInstallationFirst, the inst
authorized_keys of datanode (
192.168.1.107 node ):
A. Copy the id_dsa.pub file of namenode:
$ SCP id_dsa.pub root@192.168.1.108:/home/hadoop/
B. log on to 192.168.1.108 and run $ cat id_dsa.pub>. Ssh/authorized_keys.
Other datanode perform the same operation.
Note: If the configuration is complete and the namenode still cannot access datanode, you can modify
Authorized_keys: $ chmod 600 authorized_keys.
4. Disable the Firewall
$ Sudo UFW disable
Not
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.