Lab Environment: Two hosts: centos6.5+httpd2.4+php5.5 the basic environment that makes up the web, and the Web page is accessed properly, and ensures that the HTTPD24 service does not boot up. Node1.mylinux.com 10.204.80.79 Node2.mylinux.com 10.204.80.80 I am here to enable ansible to facilitate the management of two nodes, enable a host as a management node, ip:10.204.80.71, in the hosts of the three hosts to add the corresponding name resolution, management node to node1,node2 two nodes enable
High Availability of servers through HA, that is, high availability cluster of mysql servers through corosync + drbd + pacemaker. Main steps to implement the case application: 1. Preparations: 2. install and configure DRBD. 3. mysq
High Availability of servers through HA, that is, high availability cluster of mysql servers through corosync + drbd + pacemaker. Main steps to implement the case application: 1.
Ansible + Corosync + Pacemaker + NFS for http High Availability
Directory:
(1) experiment environment(2) Preparations(3) Configure basic configurations for node1 and node2(4) deploying nfs using ansible(5) deploying corosync and pacemaker using ansible(6) Use ansible to install the crmsh Tool(7) Use crmsh to configure http High Availability(8) Verification(9) Notes
(1) experiment environment1.1 environment
The search algorithm for the Niang, as well as the bidding rankings, I just want to say that I bought a watch last year.
One, MySQL replication master configuration
Please refer to: MySQL replication master (master-slave) Sync http://www.111cn.net/database/mysql/83904.htm
Second, Corosync pacemaker installation configuration
Please refer to: Corosync pacemaker Nginx High-availability installation configurat
node2
192.168.3.20.node2Modify/etc/ha. d/ha. cf to add the node to the file, and modify the specific configuration node as needed:Keepalive 2Deadtime 20Inittime 120Udpport 694Bcast eth0 (your broadcast Nic, usually eth0)Auto_failback onNode node1Node node2Modify/etc/ha. d/authkeys to modify the encryption mode:Auth 11 crcModify the/etc/ha. d/authkeys attribute to 600.Modify/etc/ha. d/haresources to add an ip address for automatic drift processing:
Node1 Ipaddr: 192.168.3.169/24/eth0 mysql httpd
This article briefly describes how to configure a Fence_vmware_soap type of stonith device (test-only learning) in Rhel 7 pacemaker.Stonith is the abbreviation for Shoot-the-other-node-in-the-head, and it protects the data from corruption due to node anomalies or simultaneous access.A node unresponsive does not mean that it does not have access to data, and if it wants to be sure that the data is secure, it needs to use Stonith to isolate the node to ensure that the data is properly accessed by
1. Preparation before installation High-availability clusters generally need to configure time synchronization, based on host name Mutual communication, SSH mutual trust communication, Corosync Pacemaker only need to configure time synchronization, based on host name to communicate with each other: yuminstallchrony #安装chronyvim /etc/chrony.confserver172.16.0.1iburst #设置为可用的ntp服务器systemctl restartchronyd #启动服务vim /etc/ hosts #将hosts文件分发给每个节点 172.16
managing master-slave DRBD hosts. The main use of drbdadm, because it more in line with the user's habits. Drbdsetup and Drbdmeta are less used as they are closer to the underlying device.DRBD's working features: real-time, transparent, set data synchronization typeThere are three types of data synchronization:A: Asynchronous, good performance, poor reliabilityB: Semi-synchronous performance reliability TradeoffC: Good synchronization reliability, poor performanceResource type for DRBD:Resource
Lab Environment:Virtual Machine 2 installs Corosync+pacemaker, the operating system used is CentOS6.5.Virtual Machine ip:172.16.103.2, 172.16.103.3Experimental steps:Prerequisite Preparation: Two virtual machine time synchronization, can use/etc/hosts file based on host name to communicate.First, install Corosync + pacemaker:# yum Install corosync# yum install pacemakerSecond, install the DRBD package:The D
DBRD can achieve data synchronization, DRBD is generally a master from, all read and write operations, Mount can only be performed on the master node. DRBD Master node can be swapped.The Corosync+pacemaker implementation of the DRBD master-slave point can be automatically switched to the slave node, and the failover continues to provide service.MySQL data can therefore be placed on a DRBD block.DRBD:DRBD: (Distributed replication block device) is a di
Pacemaker, is a cluster resource manager. It enables node and resource level failure detection and recovery of the maximum Availability Cluster service (also known as resource management) using the messaging and membership capabilities provided by your preferred cluster infrastructure (Openais or Heaerbeat).It can do clusters of any size and comes with a robust dependency model that enables administrators to accurately express the relationships betwee
Os:rhel 6.5 64bitcorosync:1.4.7--yum Mode installationpacemaker:1.1.2--Automatic installation as a Corosync dependency packPacemaker is the product of heartbeat development to 3.0 independent, Red Hat 6.0 series, using Yum to install Corosync, the default is to install Pacemaker as CRM.Pacemaker Common Configuration tool: Crmsh PCsCrmsh need to install RPM packages independentlyMain configuration files:/etc/corosync/corosync.conf/etc/crm/crm.confExper
resource management function and cannot form a cluster independently.2, Corosync IntroductionCorosync was originally used only to demonstrate an application of the Openais Cluster Framework interface specification, which can be said to be part of the Openais, but the latter is clearly beyond the official original idea, and more and more vendors are trying to use Corosync as a cluster solution. The RHCs cluster suite, such as Redhat, is based on the Corosync implementation.Corosync only provides
Corosync and pacemaker implement httpd and mysql dual ClustersI. Environment Introduction: Three are dual NIC: openstack-control.example.com openstack-control eth0: 172.16.171.100 eth1: 10.1.1.100 openstack-nova.example.com openstack-nova eth0: 172.16.171.110 eth1: Drawing openstack-neutron eth0: 172.16.171.120 eth1: 10.1.1.120 II. The configuration steps of corosync and pacemaker are as follows: 1. Configu
Lab environment:
Install: Yum install corosync pacemaker-y
Copy the configuration file
CP corosync. conf. Example corosync. conf
Vim corosync. conf
Edit the configuration file:
Compatibility: whitetankCompatible with corosync 0.8Totem {Defines the heartbeat information transfer layer between clustersVersion: 2 ---- versionSecauth: On --- whether to enable Security AuthenticationThreads: 0 --- How many threads are enabled to process heartbeat inf
After MySQL migrated to Azure, due to the nature of the cloud, the method of ha in the self-built data center of MySQL could not be deployed much on the cloud.This is mainly because the public cloud currently does not support: 1. shared storage; 2. Multicast;3. Vip.
Shared storage, the Azure file service can partially resolve this issue, but given the performance issues, this scenario does not use File service;
The main reason for multicast is that the cluster software needs multica
This article is from DBAnotes. Original article title: DRBD and Pacemaker
If someone asks you if a PC server can reach 99.99% high availability, what should you do? Maybe no machine can "Make sure" to achieve this availability. Of course, there may be no problems in a certain period of time, but it must be luck, however, high availability is basically impossible to achieve the goal through one machine. More often, we design solutions to ensure that t
Corosync and pacemaker implementation of the high availability of the difficulties in the CRM command, subcommand, more parameters, and so on, the following some common examples.One, CRM has two kinds of working methods
1, batch mode, which is to enter commands directly on the shell command line2, interactive mode (CRM (live) #) enters into Crmsh for interactive execution
Second, the order detailed
1, First ORDER subcommand
The code is
One: Experimental environment
Node
OS
Ip
Drbd_ip
DRBD with HDD
Vip
Web1
centos5.10
192.168.10.11
172.16.1.1
/dev/sdb
192.168.10.100
Web2
centos5.10
192.168.10.12
172.16.1.2
/dev/sdb
1. Where the two-node IP address is set as above:
2. Two nodes have been configured with mutual SSH trust and have done time synchronization
Two: Install related software (both nodes 1 and 2 are installed)
1. Insta
Tags: corosync pacemakerFirst, the environment introduction:Three are dual network cards:Openstack-control.example.com Openstack-controleth0:172.16.171.100eth1:10.1.1.100Openstack-nova.example.com Openstack-novaeth0:172.16.171.110eth1:10.1.1.110Openstack-neutron.example.com Openstack-neutroneth0:172.16.171.120eth1:10.1.1.120Second, Corosync and pacemaker configuration steps are as follows:1. Configure time zones and sync times2. Configure the cluster
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.