cmd, switch directory to/zookeeper1/bin/, execute command zkserver.cmd (the error log will be printed at this time, do not worry, this is the heartbeat check connection other ZK services, such as the start of more than half of the group of ZK Services, no error)Go to cmd, switch directory to/zookeeper2/bin/, execute command zkserver.cmdGo to cmd, switch directory to/zookeeper3/bin/, execute command zkserver.cmdSince then, the Windows pseudo-Cluster s
worker,supervisor is a working node, and no UI is only worker:CommonNimbusSupervisor (number of worker per supervisor can be adjusted by increasing or decreasing the number of slots):Note: The configuration information do not shelf write, or the startup will be error when the property value is not found.Seventh step: Copy the configured storm directory to the other server via "scp-r" (Note: If your current server is configured as Nimbus, Other servers are configured with the Supervisor configur
nodes inside the directory 7003 7004 70058. Copy the redis.conf to these three directories 9. Start modifying the redis.conf configuration fileVim 7003/redis.conf (the bottom of the picture is someone else's blog, the younger brother borrowed it)Modify the following properties to10. Modification Complete Start 11. See if the startup is successfulPs-ef|grep Redis Netstat-tnle|grep Redis12. Create a clusterRedis officially provides the REDIS-TRIB.RB tool, which is in the SRC directory of the ext
Spark Cluster Setup
1 Spark Compilation
1.1 Download Source code
git clone git://github.com/apache/spark.git-b branch-1.6
1.2 Modifying the pom file
Add cdh5.0.2 related profiles as follows:
1.3 Compiling
Build/mvn-pyarn-pcdh5.0.2-phive-phive-thriftserver-pnative-dskiptests Package
The above command, due to foreign maven.twttr.com by the wall, added hosts,199.16.156.89 maven.twttr.com, executed a
use the source command to make the configuration work after configuration is complete.Modifying the path in/etc/environmentEnter the Conf directory for Spark:The first step is to modify the slaves file to open the file first:We have modified the contents of the slaves file to:Step Two: Configure spark-env.shFirst copy the spark-env.sh.template to the spark-env.sh:Open the "spark-env.sh" fileAdd the following to the end of the fileSlave1 and slave2 Use the same spark installation configuration a
change the default page display of Nginx, the method mentioned twice before;Rs1: Echo "Rs1rs1" >/usr/share/nginx/html/index.htmlRS2: Echo "Rs2rs2" >/usr/share/nginx/html/index.htmlCan be seen in the browser, but the total problem, not recommended; Go to the command line on Windows or find another server ping 192.168.11.100Downtime Rs1:ifdown eth0 or/etc/init.d/nginx stopView Dir:ipvsadm-ln or IP addr you'll find 192.168.11.20 missing.And ping, will always link rs2, does not show rs1 connection;
original path to the target path Hadoop fs-cat/user/hadoop/a.txt View the contents of the A.txt file Hadoop fs-rm/user/hadoop/a.txt Delete US The A.txt file below the Hadoop folder under the ER folderHadoop fs-rm-r/user/hadoop/a.txt recursive deletions, folders and filesThe Hadoop fs-copyfromlocal/local Path/destination path is similar to the Hadoop fs-put feature. Hadoop fs-movefromlocal localsrc DST uploads local files to HDFs while deleting local files. Hadoop fs-chown User name: User group
VMware-based virtual Linux cluster setup-lvs + keepalivedBuild a virtual Linux cluster based on VMware-lvs + keepalived this article uses keepalived to achieve load balancing between the dual-machine hot backup of the lvs server and the Real Server. There are many blogs in this regard, but the environments for each person to build a
VMware-based virtual Linux cluster setup-lvs + keepalived,-lvskeepalivedBuild a virtual Linux cluster based on VMware-lvs + keepalived this article uses keepalived to achieve load balancing between the dual-machine hot backup of the lvs server and the Real Server. There are many blogs in this regard, but the environments for each person to build a
zlib-devel OpenSSL openssl-devel pcre pcre-develStep three: Enter the unzip directory./configure--with-http_ssl_moduleFourth step: Make make install You can refer to the article (http://www.cnblogs.com/skynet/p/4146083.html)3.2Nginx responsible for balanced configuration 3.2.1 Find the installation path of Nginx (default):/usr/local/nginx 3.2.2 Modify the configuration file conf/nginx.conf #设定负载均衡的服务器列表ip_hash Upstream Backend { server 127.0.0.1:8080; Server 127.0.0.1:8081;
primary node:
1
storm nimbus
Or
1
storm nimbus >/dev/null
4. Start from node supervisorExecute the command from the node (it can actually be started on the master node):
1
storm supervisor >/dev/null
5. Start the Storm UI
1
storm ui >/dev/null
Port default is 8080, Access address: http://vm1:8080/You can view the status of the clus
,Find 112 rows or soIF ["%kafka_jvm_performance_opts%"] EQU [""] ( set KAFKA_JVM_PERFORMANCE_OPTS=-SERVER-XX:+USECOMPRESSEDOOPS-XX: +useparnewgc-xx:+useconcmarksweepgc-xx:+cmsclassunloadingenabled-xx:+cmsscavengebeforeremark-xx:+ Disableexplicitgc-djava.awt.headless=true)Remove -xx:+usecompressedoopsTest Cluster(1) Create a topicKafka-topics.bat--create--zookeeper 127.0.0.1:2181--replication-factor 1--partitions 1--topic test(2) To see if the creatio
hbase-1.2.4jdk1.8.0_101The first step, download the latest version from the Apache FoundationHTTPS://mirrors.tuna.tsinghua.edu.cn/apache/hbase/1.2.4/hbase-1.2.4-bin.tar.gzStep two , unzip to the serverTAR-ZXVF hbase-1.2. 4The third step is to configure the HBase cluster to modify 3 files (first the ZK cluster is already installed) Note: Since HBase final data is stored in HDFs, Hadoop's hdfs-site.xml and c
protocol, shutting it down can disconnectIptables-a input-p icmp-j DROPHost View:Ifconfig PS aux |grep nginx tail/var/log/ha_logFrom Machine view:Ifconfig PS aux |grep nginx tail/var/log/ha_logHost Open ICMP protocoliptables-d input-p icmp-j drop-d DeleteAnd on the same way to view, will find the host and the user's services automatically switched back;You can also stop the host's Heartbeat service Test View:/etc/init.d/heartbeat stopIn fact, the configuration of a variety of highly available s
Tomcat/webapps directory, remember not to change the name of the War package2. Add the slf4j,commons-loggin.jar to the apache-tomcat-7.0.42\webapps\solr\web-inf\lib . The 2 jar packages are missing from Solr 's own war package3. Configure Solr.home. add such a section to Web. XML below apache-tomcat-7.0.42\webapps\solr\web-infChange the E:\solr_home to your own directory, and then copy the solr-4.4.0 \EXAMPLE\SOLR to your ${solr_home}4. do the same for all nodes.5. Modify the ${solr_home}/solr
0 Preface
Starting from this article, the author will introduce how to build an ArcGIS 10.1 for server cluster in a Linux environment. Because the cluster involves Server Load balancer processing and requests need to be distributed among different machines, domain name resolution is required. In addition, machines in the cluster need to share the site configurati
example, if you create a web app and deploy it to Tomcat (multi-application), the default rule is to add a prefix to the project name in the URI to access it. For example, we assume that the project is named shopping, giving it a separate domain name: shopping.website.com. Then the default URL needs shopping.website.com/shopping to be OK, now it needs to be set to shopping.website.com. How to use Nginx configuration, to achieve our goal? In the nginx.conf configuration, add the rewr
Tags: command modify port number Basedir Err instance localhost file amp 127.0.0.1MySQL Cluster tutorial (iii) MySQL multi-instance (multiple database) setup Multi-Instance Overview: MySQL multi-instance refers to the installation of MySQL, we can start a Linux server at the same time multiple MySQL database (instance), do not need to install multiple MySQL; If there are more than one Linux server, then we
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.