Hadoop consists of two parts:
Distributed File System (HDFS)
Distributed Computing framework mapreduce
The Distributed File System (HDFS) is mainly used for the Distributed Storage of large-scale data, while mapreduce is built on the Distributed File System to perform distributed computing on the data stored in the distributed file system.
Describes the functions of nodes in detail.
Namenode:
1. There is only one namenode in the
Use yum source to install the CDH Hadoop Cluster
This document mainly records the process of using yum to install the CDH Hadoop cluster, including HDFS, Yarn, Hive, and HBase.This article uses the CDH5.4 version for installation,
Install and configure Sqoop for MySQL in the Hadoop cluster environment,
Sqoop is a tool used to transfer data from Hadoop to relational databases. It can import data from a relational database (such as MySQL, Oracle, and S) into Hadoop HDFS, you can also import HDFS data to
Tags: security config virtual machine Background decryption authoritative guide will also be thought also needTo learn more about Hadoop data analytics, the first task is to build a Hadoop cluster environment, simplifying Hadoop as a small software, and then running it as a Hadoop
This series of articles describes how to install and configure hadoop in full distribution mode and some basic operations in full distribution mode. Prepare to use a single-host call before joining the node. This article only describes how to install and configure a single node.
1. Install Namenode and JobTracker
Thi
, I found that it was not the cause of hbase, but I did not delete them in hbase. Therefore, whether it is necessary to copy them to hbase remains to be tested in person.
2. Configure lzo:
1. Add some properties to the core-site.xml and mapred-site.xml files in the conf directory under the hadoop directory:
VI core-site.xml:
VI mapred-site.xml:
2. Synchronize the configuration files of each node!
Iii. hadoop
Use Windows Azure VM to install and configure CDH to build a Hadoop Cluster
This document describes how to use Windows Azure virtual machines and NETWORKS to install CDH (Cloudera Distribution Including Apache Hadoop) to build a Hadoop
Install and configure Mahout-distribution-0.7 in the Hadoop Cluster
System Configuration:
Ubuntu 12.04
Hadoop-1.1.2
Jdk1.6.0 _ 45
Mahout is an advanced application of Hadoop. To run Mahout, you must install
Apache Ambari is a Web-based open-source project that monitors, manages, and manages Hadoop lifecycles. It is also a project that selects management for the Hortonworks data platform. Ambari supports the following management services:
Apache HBaseApache HCatalogApache Hadoop HDFSApache HiveApache Hadoop MapReduceApache OozieApache PigApache SqoopApache TempletonA
After the accumulation of the front, today finally realized the cluster environment to deploy Hadoop, and successfully ran the official example.
Work as follows:
Two machines:
Namenode: Internet Small, 3G memory, machine name: yp-x100e,ip:192.168.101.130.
Datanode: Virtual machine, Win7 download VMWare10 virtual UBUNTU14, virtual machine name: ph-v370,ip:192.168.101.110
Ensure that you can ping each ot
Rhadoop is an open source project initiated by Revolution Analytics, which combines statistical language R with Hadoop. Currently, the project consists of three R packages, the RMR that support the use of R to write MapReduce applications , Rhdfs for the R language to access HDFs, and for R language Access The rhbase of HBase . Download URL for https://github.com/RevolutionAnalytics/RHadoop/wiki/Downloads. Note: The following record is the summary a
Address: http://blog.cloudera.com/blog/2013/04/how-to-use-vagrant-to-set-up-a-virtual-hadoop-cluster/
Vagrant is a very useful tool that can be used to program and manage multiple virtual machines (VMS) on a single physical machine ). It supports native virtualbox and provides plug-ins for VMWare Fusion and Amazon EC2 Virtual Machine clusters.
Vagrant provides an easy-to-use ruby-based internal DSL that all
Main steps:
1. Install and update GCC and ant (if the system has been installed, skip the following steps)
Yum-y install GCC gcc-C ++ Autoconf automake
Wget
Http://labs.renren.com/apache-mirror//ant/binaries/apache-ant-1.8.2-bin.tar.gz
Tar-jxvf apache-ant-1.8.2-bin.tar.bz2
Export ant_home =/usr/local/Apache-ant-1.8.2
VI/etc/profile
Source/etc/profile
Export Path = $ path: $ ant_home/bin
2.
Make some simple introductions to the above roles:Namenode-The entire HDFs namespace management ServiceSecondarynamenode-a redundant service that can be viewed as NamenodeJobtracker-Job Management services for parallel computingNode Services for Datanode-hdfsTasktracker-Job execution services for parallel computingManagement Services for Hbase-master-hbaseHbase-regionserver-Provide services for client-side inserts, deletes, query data, etc.Zookeeper-server-zookeeper collaboration and Configura
Build a Hadoop Client-that is, access Hadoop from hosts outside the Cluster
Build a Hadoop Client-that is, access Hadoop from hosts outside the Cluster
1. Add host ing (the same as namenode ing ):
Add the last line
[Root @ localho
Datanode nodemanager server: 192.168.1.100 192.168.1.101 192.168.1.102
Zookeeper server cluster (for namenode high-availability automatic failover): 192.168.1.100 192.168.1.101
Jobhistory server (used to record mapreduce logs): 192.168.1.1
NFS for namenode HA: 192.168.1.100
Environment deployment 1. Add the YUM repository to CDH4 1. the best way is to put the cdh4 package in the self-built yum warehouse. For how to build a self-built yum warehou
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.