) clusters
High Performance Computing Cluster hp:high-performance (HPC) clusters
Grid Compute Grid Computing
Iii. detailed definitions of various clusters
⑴, load Balanced cluster-lb
When a load-balanced cluster is run, the workload is distributed to a set of servers on the back end, typically through one or more front-end load balancers, to achieve an entir
is not the root server, but direct access to our LDNS (Local DNS Server), typically maintained by network operators.The first load balancing is implemented by building local DNS server, the implementation is simple and easy to understand, for the same host name to assign multiple mappings, you can use round-robin, random and other ways to allocate requests. Ther
very important relay in the client send request, his role is to map the URL of the user request to a specific IP address, the world has 13 root servers, but usually for our domain name resolution is not the root server, but direct access to our LDNS (Local DNS Server), typically maintained by network operators.The first load balancing is implemented by building local DNS server, the implementation is simpl
requests.
If site traffic increases, NLB can easily add more WEB servers to the NLB cluster to increase capacity for additional load. To extend a Web server farm, more than 32 nodes, NLB can be used in conjunction with other load balancing technologies, such as Round-robin DNS (RRDNS). In fact, many organizations have
18.6 Load Balancing Cluster introduction
Main open source software LVs, keepalived, Haproxy, Nginx, etc.
The LVS belong to 4 layer (network OSI 7 layer model), Nginx belongs to 7 layer, Haproxy can be considered as 4 layer, can be used as 7 layer
The Keepalived load balancing function is actually the L
Ifcfg-bond0 File
#vi ifcfg-bond0
The original configuration file is as follows
After modification, you need to add a static IP address (NIC bindings do not recommend using dynamic IP), subnet masks, and gateways are as follows
(Note: In addition to BOND0 set IP and other information outside the real card can not set IP address)
Edit Ifcfg-eth0 File#vi Ifcfg-eth0The original configuration file is as follows
After modification
Three network adapters are prepared.
This environment uses Cent
rotation analysis
A method based on client scheduling access
Scheduling method based on application-layer system load
A scheduling method based on IP addressIn the IP-based load scheduling algorithm. IP load Balancing technology is the most efficient
Four, LVS 10 scheduling algorithm1. Static Dispatch
First, Brief introduction LVS is a shorthand for Linux Virtual server and is a highly scalable and highly available cluster solution on Linux servers. Developed by Dr. Zhangwensong, it is now integrated into the Linux kernel module and implements a load balancing schedule of four layers (transport layer) at the kernel level. Principle:ii. working mode of LVSThere are three different modes of operation dep
, listen, frontend, and backend are proxy segments. frontend is used to match the domain name or URI of the client request. backend defines the backend server cluster. listen is the set of frontend and backend, sometimes only listen can be used to replace frontend and backend. The following shows an HTTP Proxy case for listening to port 80. The proxy server forwards data requests to a single backend server 127.0.0.1: 8000;
The syntax format and description of the HAProxy main configuration file
1. Polling
Polling is round Robin, which distributes the client's Web requests to different back-end servers sequentially, according to the order in the Nginx configuration file.Examples of configurations are as follows:
http{
upstream SampleApp {
Server
Only 1 DNS portals were inserted into the upstream section, the SampleApp, which was also recalled in the Proxy_pass section later.
2, the least connection
Web requests are forw
Bind two network cards, share one IP, realize the redundancy effect. In fact, there are 7 types of bindings for Linux dual NICs:
The mode=0 indicates that load balancing (round-robin) is balanced, and both NICs work.
mode=1 says Fault-tolerance (active-backup) provides redundancy, working in a master-slave way, which means that only one NIC works by
implement simple load balancing.However, it is obvious that this round robin is not very uniform, and the pressure on one server is always high. The other is relatively close, but after removing the high pressure from the dns list, the next instance in the list is under high pressure,Then add the previous one back to the dns list, and the pressure will go back to the first stage. The possibility of attack
Use Apache as Web Proxy
The apusic application server can integrate with Apache in the following two ways:4.9.1. HTTP Proxy
The mod_proxy_http module provided by Apache can be used to integrate with the apusic application server, and the apusic application server does not require any settings. Mod_proxy_http provides the proxy HTTP Request function. Supports HTTP/0.9, HTTP/1.0, and HTTP/1.1 standards.
Configuration in standalone EnvironmentPerform the following configuration in % apache_home
Lvswith theNginxload Balancing scheduling algorithm1th Chapter LVSof the10Type scheduling algorithm1.1Static Algorithms: 1.1.1 RR (round robin):Polling scheduling algorithm:the principle of polling scheduling algorithm is to put the user's access request , the average allocation to each Web service node , starting from 1 , to the end of the last server node , and then at the beginning of a new round of loo
analysis of RR and Ip_hash strategy in Nginx load Balancing
Nginx upstream currently supports the distribution of load balancing methods 1, RR (default)
Each request is assigned to a different backend server in chronological order, and can be automatically removed if the backend server is down.
For example:
Generally, the load balancing method can be classified into two types: static load Balancing algorithm and dynamic load balancing algorithm. Static load
Recently looking at a book on Javaweb implementation of the Web site distributed architecture, it mentions that through load balancing, you can avoid issues such as server paralysis caused by massive access or the inability of users to access due to a single server failure. There are many ways to achieve load balancing
When the vertical expansion of the Web server becomes expensive or difficult, we need to consider the level of server expansion, namely load balancing technology. There are many techniques for load balancing, which we'll cover here.1.HTTP redirectionWe can do this at the code level by setting access to specific pages s
message;Note: The first three types are standard, the fourth is the post-add type, the kernel may not be supported by default, and the kernel must be self-compiled(1) The VIP is the public network address, RIP and dip is the private network address, and can not be in the same IP network, but need to communicate with each other through routing;(2) The source IP of the request message received by RS is a dip, so its response message will be sent to the dip;(3) The request message and the response
client, which greatly reduces the data size. generally, the value of gzip_comp_level does not differ much in terms of performance. Set it to 4-5. simply increasing it is meaningless, just a waste of CPU cycles.You can also reduce the file size by using some JavaScript and CSS reduction tools, but these are not very relevant to nginx, so I believe you can get more information through Google.Category: nginx tags: nginx Server Load balancer July 17, 201
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.