Apache HTTP Server is selected as the front-end Server Load balancer, and two Tomcat clusters are selected at the backend. The selected configuration method is session sticky (sticky session ), this method forwards requests from the same user to a specific Tomcat server to avoid session replication in the cluster. The disadvantage is that the user only communicat
ngx_process_change party. Update Time X/Timer = ngx_event_find_timer (); flags = ngx_update_time; # If (ngx_threads) if (timer = ngx_timer_infinite | timer> 500) {timer = 500 ;} # endif}/* ngx_use_accept_mutex indicates whether to lock the accept to solve the problem. When the number of nginx worker processes is greater than 1 and accept_mutex is enabled in the configuration file, this flag is set to 1 */If (ngx_use_accept_mutex) {/* ngx_accept_disab
In "high concurrency, massive data, distributed, nosql, cloud computing ...... I believe many of my friends have heard of and even often talked about "clusters and Server Load balancer". But not all of them have the opportunity to really come into use these technologies, not everyone really understands these technical terms. The following is a brief explanation.
Cluster)
A cluster is a loosely coupled mult
(Load balancer), which is used to distribute the user's requests, as well as many servers that handle user requests (Real Sever, RS)
LVS NAT Mode
This model is implemented using the Iptables NAT table
After the user's request to the dispatcher, the requested packet is forwarded to the back-end RS via a preset iptables rule.
RS needs to set the gateway as the Distributor's intranet IP
Set up nginx high-availability server Load balancer cluster instances
I. nginx high-availability cluster topology architecture
Node1: 17216.100.67Node2: 170.16.100.68VIP: 172.16.100.89RS1: 172.16.100.6Rs2: 172.16.100.69
Note: The Kernel Parameters arp_ignore and arp_announce of all Rs in nginx high-availability cluster architecture are 0.
Ii. nginx high-availability cluster setup steps1. Install the nginx
This article tries to explain the operation steps and simple configuration of using keepalived + Nginx + Tomcat to build a high-availability load balancing environment in Ubuntu Server environment, which does not involve performance tuning. Let's talk about their respective roles:
tomcat– Application Server
nginx– reverse proxy server, as a load
Apache implements the Load balancer parameter "proxypass/balancer://proxy/", where "Proxypass" is the command that configures the virtual server, "/" represents the URL prefix for sending Web requests, such as: http://myserver/Or HTTP://MYSERVER/NODE1, these URLs will conform to the above filter conditions, "balancer:/
, different Ribbon clients can use different configurations.
Create Ribbon configuration class
/*** The configuration class * should not be scanned by ComponentScan */@ Configurationpublic class RibbonConfiguration {@ Bean public IRule ribbonRule () {// Configure the server Load balancer rules, change to random return
: This article mainly introduces the Nginx + IIS + Web front-end (SpringMVC)-server load balancer (1). If you are interested in PHP tutorials, refer to it. Introduction
When developing a large Web project, if our web is published on the IIS of a server, when a large number of requests are sent to the IIS service, the computer will crash easily. So we thought, why don't we put our services on multiple compu
I. Introduction to the Basic overviewIi. types and principles of LVSThree, LVS scheduling algorithmIv. using DR and Nat to achieve web load balancingI. Introduction to the Basic overviewLVS is a load balancing software that works at the transport level and consists of two components of the IPVSADM and kernel space of the user space Ipvs. The Ipvsadm is a command-line tool for user space, primarily for manag
I.DemandNginx as a load balancer server, the user requests to reach nginxFirst, and then by Nginx According to the load configuration to forward the request to Tomcat server. Nginx Load Balancer server:192.168.101.3TOMCAT1 server
1) Open the "httpd.conf" file in the "/usr/local/apache2/conf" directory and add the following configuration item at the end of the file, as shown in Figure 4-2-1.
Proxyrequests OFF
proxypass/balancer://mycluster/
Balancermember ajp://localhost:10009 ROUTE=TOMCAT1
Balancermember ajp://localhost:20009 ROUTE=TOMCAT2
Figure 4-2-1
Description: Where "Mycluster" is the name of the cluster, "ajp://localhost:100
Considering the shortcomings of LVS and Nginx (because LVS uses synchronous request forwarding policy and Nginx is the asynchronous forwarding policy, combined with the disadvantage of both: as the Load Balancer server nginx and LVS processing the same request, all requests and response traffic will go through the Nginx server, However, when using LVS, only request traffic through the LVS network, the respo
memcached access to achieve session sharing well, look at my nginx.conf configuration below: #运行nginx所在的用户名和用户组http://zyjustin9.iteye.com/blog/2017237,http://freeloda.blog.51cto.com/2033581/1288553 #user www www; #启动进程数 worker_processes 8; #全局错误日志及PID文件 Error_log/usr/local/nginx/logs/nginx_error.log Crit; Pid/usr/local/nginx/nginx.pid; #Specifies the value for maximum file descriptors the can is opened by th
serviceinstancechooser, and then we look at the interface definition:Public interface Serviceinstancechooser { /** * Choose a serviceinstance from the LoadBalancer for the specified s Ervice * @param serviceId the service ID to look up the LoadBalancer * @return A serviceinstance that matches the ServiceId * /Serviceinstance Choose (String serviceId);}From the above annotations, we can know that the Choose () method chooses a corresponding service instance from the
1 overviewIn order to guarantee the stable operation of the bastion host-instruction channel, it is necessary to implement the load balance of instruction channel, and study to determine the use of LVS software.LVS is the abbreviation for Linux virtual server, that is, the Linux server, the official site is: www.linuxvirtualserver.org2 LVS ArchitectureLVS set up server cluster to achieve load balancing has
The previous article written the Nginx load balancer, this article realizes the high Availability (HA). The overall design of the system is to use Nginx to do load balancing, if there is an nginx single-machine failure, the whole system will not function properly. For the high-availability requirements of system architecture design, we need to solve the requireme
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.