First of all, of course, is to download a spark source code, in the http://archive.cloudera.com/cdh5/cdh/5/to find their own source code, compiled their own packaging, about how to compile packaging can refer to my original written article:
http://blog.csdn.net/xiao_jun_0820/article/details/44178169
After execution you should be able to get a compressed package similar to SPARK-1.6.0-CDH5.7.1-BIN-CUSTOM-SP
Add Virtual Machine node balancing data operations to the CDH cluster (Tutorial), cdh Virtual Machine
Note: The premise is that a new Virtual Machine node has been installed and the corresponding cdh has been installed. You can modify the host name, ip address, mac, and other operations on your own. This article only adds the cluster balance data operation to t
is (http://www.cloudera.com/documentation/enterprise/latest/topics/cm_vd.html# Concept_mb3_sfz_3q_unique_1) If it is suselinux, download Zypper/yast SLES.
tar -xzf cloudera-manager*.tar.gz
Create a user on the CM server
useradd --system --home=/opt/cm-5.6.0/run/cloudera-scm-server --shell=/bin/false"Cloudera SCM User" cloudera-scm
Create local storage directory for CM server
mkdir /var/lib/cloudera-scm-servermkdir /va
or download the Word document: http://download.csdn.net/download/xfg0218/9747346
about CDH and Cloudera Manager
CDH (Cloudera's distribution, including Apache Hadoop), is one of the many branches of Hadoop, built from Cloudera maintenance, based on the stable version of Apache Hadoop, and integrates many patches, Can
/6/x86_64/cdh/cloudera-cdh4.repo or http://archive.cloudera.com/cdh5/redhat/6/x86_64/cdh/cloudera-cdh5.repo, Please modify the path to the baseurl in the file according to your installed cdh version.
You can download the cdh4 warehouse archive or the cdh5 warehouse archive from here.
Because I am using the centos opera
Three cluster nodes
192.168.1.170 CDH-Master
Cdh-slave-1 192.168.1.171
Cdh-slave-2 192.168.1.171
1. Install centos6.5 (64-bit) and set up the basic environment, including:
(1) Add sudo Permissions
(2) modify the host name, gateway, static IP address, and DNS
(3) Disable SELinux and Firewall
Refer to the article
(4) modify the system time zone and configure the NT
levels increase in difficulty, but the benefits increase in turn. For CDH 4.2.0, it only achieves the job-level recovery granularity and is the simplest and least profitable implementation method.
References:
(1) CDH JobTracker HA introduction and installation solution:
Https://ccp.cloudera.com/display/CDH4DOC/Configuring+High+Availability+for+the+JobTracker+ (MRv1)
(2)
Cloudera Manager and CDH 5.14.0 Installation Process in CentOS 7
As we all know, the configuration of Apache Hadoop is cumbersome and fragmented. For this reason, Cloudera provides the Clouder Manager tool and encapsulates Apache Hadoop, flume, spark, hive, hbase and other big data products form CDH products with their own characteristics, and then use CM for installation. This facilitates cluster construct
: Master node ssh other node ...; If not successful, then the other nodes in the other node to do their own password-free login: On the node to use the command ssh-keygen-t dsa-p "-F ~/.SSH/ID_DSAAnd then repeat the above operation 3. Turn off the firewallTemporary shutdown:Service Iptables StopPermanently closed (after reboot):Chkconfig iptables off 4. Turn off SELINUXTemporary shutdown:Setenforce 0Modify configuration file/etc/selinux/config (restart effective):Change Selinux=enforcing to Seli
Tags: man manual enter row tar.gz err 1.4 for maximumHue: Https://github.com/cloudera/hue Hue Study document address : http://archive.cloudera.com/cdh5/cdh/5/hue-3.7.0-cdh5.3.6/manual.html I'm currently using hue-3.7.0-cdh5.3.6. Hue (Hue=hadoop User Experience) Hue is an open-source Apache Hadoop UI system that evolved from Cloudera desktop and finally cloudera the company's contribution to the Apache Foundation's Hadoop community, which is based on t
machines:SCP ~/.ssh/authorized_keys [Email protected]:~/.ssh/Now log on to other machines without a password.3 Installing JavaBecause CDH4 support JAVA7, consider CDH5 only support JAVA7, decisive on. (later MySQL also used the latest 5.6.16, later found that the tragedy, do not know which reason, so the JDK has changed to the official recommendation version, or not, and the MySQL back to the 5.1.X version, the final can be.) Personal guess JDK can still use 7, MySQL can only use 5.5, and then
Cloudermanager installation package Address: http://archive.cloudera.com/cm5/cm/5/cloudera-manager-el6-cm5.3.0_x86_64.tar.gz,The red part is the version number, the red part can be changed to get the different versions of the installation packageCDH
First of all, to ask, what is CDH?To install a Hadoop cluster that deploys 100 or even 1000 servers, package I including hive,hbase,flume ... Components, a day to build the complete, there is to consider the system after the update asked questions, then need to CDH
Advantages of the CDH version:Clear Version DivisionFaster version updateSupport for Kerberos secur
1. Create the lib121 directory under the hive0.13.1 version
Cd/opt/cloudera/parcels/cdh/lib/hive;mkdir lib1212. Download the hive1.2.1 version and copy all files from this version of Lib to lib121
3. Modify the Hive_lib variable in/opt/cloudera/parcels/cdh/lib/hive/bin/hive
hive_lib=${hive_home}/lib121
4. Update the JLine jar package on Hadoop and remove the ol
I. Installation PROTOBUFUbuntu system1 Create a file in the/etc/ld.so.conf.d/directory libprotobuf.conf write the content/usr/local/lib otherwise the error will be reported while loading shared libraries:libprotoc.so .8:cannot Open Shared obj2../configure Makemake Install2. Verify that the installation is completeProtoc--versionLibprotoc 2.5.0Two. Install the Snappy local libraryHttp://www.filewatcher.com/m/snappy-1.1.1.tar.gz.1777992-0.htmlDownload snappy-1.1.1.tar.gzUnzip./configuremake Makein
Original address: Http://blog.selfup.cn/1631.html?utm_source=tuicoolutm_medium=referral
Spit Groove
Recently "idle" to have nothing to do, through the CM to vcores use situation to look at a glance, found that no matter how many tasks in the cluster running, the allocated vcores will never exceed 120. The available vcores for the cluster are 360 (15 machines x24 virtual cores). That's equivalent to 1/3 of CPU resources, and as a semi-obsessive-compulsive disorder, this is something that can nev
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.