(0.00sec) mysql>insertintot2 VALUES (1, ' Lisan '); queryok,1rowaffected (0.00sec) mysql>select*from t2;+------+-------+|id| name|+------+-------+|1|lisan|+------+-------+1 rowinset (0.00NBSP;SEC)View the T2 table on 204:Mysql> SELECT * FROM t2;+------+-------+|ID |Name |+------+-------+|1 | Lisan |+------+-------+1 row in Set (0.00 sec)When you see the results above, the distributed MySQL data is successfully synchronized.Problems encountered during installation:1, unable to connect with conne
failures or to recover automatically from failures without the need for operator intervention. By moving applications on a failed server to a backup server, the cluster system can increase uptime to more than 99.9%, significantly reducing server and application downtime.High manageability: System administrators can remotely manage one or even a group of clusters as if they were in a standalone system."Disadvantage"We know that the application in the
Redis's master-slave copy strategy is realized through its persistent RDB file, the process is to dump out Rdb file, Rdb file to Slave, and then synchronize the operation of the dump to slave in real time. The following is an article on the principle of Redis reproduction, the author of the article for Sina Weibo Tianqi classmate (@ Rocking Bach).
This article discusses the replication capabilities of the Redis and the advantages and disadvantages of the Redis replication mechanism itself, as we
the error, install a higher version of the Can, can refer to https://www.cnblogs.com/PatrickLiu/p/8454579.html thank bloggers 2) Next Run the REDIS-TRIB.RB 4. Create a cluster/USR/LOCAL/REDIS/SRC/REDIS-TRID.RB Create--replicas 1 0.0.0.0:7000 0.0.0.0:7001 0.0.0.0:7002 0.0.0.1:7003 0.0.0.1:70 04 0.0.0.1:7005 Here's a place to watch.1.redis cluster port, need to correspond to release plus 10000 port, for
cmd, switch directory to/zookeeper1/bin/, execute command zkserver.cmd (the error log will be printed at this time, do not worry, this is the heartbeat check connection other ZK services, such as the start of more than half of the group of ZK Services, no error)Go to cmd, switch directory to/zookeeper2/bin/, execute command zkserver.cmdGo to cmd, switch directory to/zookeeper3/bin/, execute command zkserver.cmdSince then, the Windows pseudo-Cluster s
This article discusses the replication capabilities of Redis and the pros and cons of the Redis replication mechanism itself, as well as cluster setup issues. Overview of the Redis replication processThe Redis replication feature is based on a memory-snapshot-based persistence strategy that we discussed earlier, which means that no matter what your persistence strategy chooses, if you use the Redis replicat
nodes inside the directory 7003 7004 70058. Copy the redis.conf to these three directories 9. Start modifying the redis.conf configuration fileVim 7003/redis.conf (the bottom of the picture is someone else's blog, the younger brother borrowed it)Modify the following properties to10. Modification Complete Start 11. See if the startup is successfulPs-ef|grep Redis Netstat-tnle|grep Redis12. Create a clusterRedis officially provides the REDIS-TRIB.RB tool, which is in the SRC directory of the ext
Spark Cluster Setup
1 Spark Compilation
1.1 Download Source code
git clone git://github.com/apache/spark.git-b branch-1.6
1.2 Modifying the pom file
Add cdh5.0.2 related profiles as follows:
1.3 Compiling
Build/mvn-pyarn-pcdh5.0.2-phive-phive-thriftserver-pnative-dskiptests Package
The above command, due to foreign maven.twttr.com by the wall, added hosts,199.16.156.89 maven.twttr.com, executed a
;; *) Echo "Usage: $ [OPTIONS] " ;;Esac4. Launch Tomcat and HTTPD:Tomcat1/bin/startup. SH tomcat2/bin/startup. SH apache2/bin/run. SH startAfter booting, access to the http://localhost:8080/test/test.jsp through the browser, the page appears normal:So let's find a machine to test the load Balancer performance, the test results are as follows: [[emailprotected] ~/test]$for ((I= 0 ; I1000 ; I++ > do > wget http:// 192.168.1.100:8080/test/test.jsp 2>/dev/null > done [[ema
VMware-based virtual Linux cluster setup-lvs + keepalivedBuild a virtual Linux cluster based on VMware-lvs + keepalived this article uses keepalived to achieve load balancing between the dual-machine hot backup of the lvs server and the Real Server. There are many blogs in this regard, but the environments for each person to build a
use the source command to make the configuration work after configuration is complete.Modifying the path in/etc/environmentEnter the Conf directory for Spark:The first step is to modify the slaves file to open the file first:We have modified the contents of the slaves file to:Step Two: Configure spark-env.shFirst copy the spark-env.sh.template to the spark-env.sh:Open the "spark-env.sh" fileAdd the following to the end of the fileSlave1 and slave2 Use the same spark installation configuration a
change the default page display of Nginx, the method mentioned twice before;Rs1: Echo "Rs1rs1" >/usr/share/nginx/html/index.htmlRS2: Echo "Rs2rs2" >/usr/share/nginx/html/index.htmlCan be seen in the browser, but the total problem, not recommended; Go to the command line on Windows or find another server ping 192.168.11.100Downtime Rs1:ifdown eth0 or/etc/init.d/nginx stopView Dir:ipvsadm-ln or IP addr you'll find 192.168.11.20 missing.And ping, will always link rs2, does not show rs1 connection;
original path to the target path Hadoop fs-cat/user/hadoop/a.txt View the contents of the A.txt file Hadoop fs-rm/user/hadoop/a.txt Delete US The A.txt file below the Hadoop folder under the ER folderHadoop fs-rm-r/user/hadoop/a.txt recursive deletions, folders and filesThe Hadoop fs-copyfromlocal/local Path/destination path is similar to the Hadoop fs-put feature. Hadoop fs-movefromlocal localsrc DST uploads local files to HDFs while deleting local files. Hadoop fs-chown User name: User group
VMware-based virtual Linux cluster setup-lvs + keepalived,-lvskeepalivedBuild a virtual Linux cluster based on VMware-lvs + keepalived this article uses keepalived to achieve load balancing between the dual-machine hot backup of the lvs server and the Real Server. There are many blogs in this regard, but the environments for each person to build a
zlib-devel OpenSSL openssl-devel pcre pcre-develStep three: Enter the unzip directory./configure--with-http_ssl_moduleFourth step: Make make install You can refer to the article (http://www.cnblogs.com/skynet/p/4146083.html)3.2Nginx responsible for balanced configuration 3.2.1 Find the installation path of Nginx (default):/usr/local/nginx 3.2.2 Modify the configuration file conf/nginx.conf #设定负载均衡的服务器列表ip_hash Upstream Backend { server 127.0.0.1:8080; Server 127.0.0.1:8081;
primary node:
1
storm nimbus
Or
1
storm nimbus >/dev/null
4. Start from node supervisorExecute the command from the node (it can actually be started on the master node):
1
storm supervisor >/dev/null
5. Start the Storm UI
1
storm ui >/dev/null
Port default is 8080, Access address: http://vm1:8080/You can view the status of the clus
protocol, shutting it down can disconnectIptables-a input-p icmp-j DROPHost View:Ifconfig PS aux |grep nginx tail/var/log/ha_logFrom Machine view:Ifconfig PS aux |grep nginx tail/var/log/ha_logHost Open ICMP protocoliptables-d input-p icmp-j drop-d DeleteAnd on the same way to view, will find the host and the user's services automatically switched back;You can also stop the host's Heartbeat service Test View:/etc/init.d/heartbeat stopIn fact, the configuration of a variety of highly available s
Tomcat/webapps directory, remember not to change the name of the War package2. Add the slf4j,commons-loggin.jar to the apache-tomcat-7.0.42\webapps\solr\web-inf\lib . The 2 jar packages are missing from Solr 's own war package3. Configure Solr.home. add such a section to Web. XML below apache-tomcat-7.0.42\webapps\solr\web-infChange the E:\solr_home to your own directory, and then copy the solr-4.4.0 \EXAMPLE\SOLR to your ${solr_home}4. do the same for all nodes.5. Modify the ${solr_home}/solr
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.