HAproxy offers a flexible, powerful loadbalancer solution for clients looking for High Availabillity and failover between multiple web or application servers.
With this quick and easy guide, you can be up and running in minutes!
First of all, you'll need a server deployed in the same cloud as your web servers. We recommend the CentOS 6.4 x64 Template with 2GB or 4GB RAM as the ideal starter Loadbalancer, depending on your traffic. Like all dediserve servers, you can easily and quickly scale on demand if you need more capacity!
For most distributions you can install haproxy using your distribution's package manager.
Run the commands:
[[email protected] ~]# yum -y install haproxy
Install a base config
Once installed backup the HAProxy config file and download the dediserve default config:
[[email protected] ~]# cp /etc/haproxy/haproxy.cfg /etc/haproxy/haproxy.cfg.bak [[email protected] ~]# wget http://dediserve.com/haproxy/haproxy.cfg -O /etc/haproxy/haproxy.cfg chkconfig haproxy on
Editing /etc/haproxy/haproxy.cfg - There are a number of items that need to be changed in order to get HAProxy functional. These will be outlined below. Keep in mind you need to edit these values to reflect the server's IP's.
First and foremost change
listen webfarm 0.0.0.0:80
listen webfarm 127.0.0.1:80
Edit 127.0.0.1 to reflect your load balancer's public IP.
Now you can add your web servers. In the following you will want to replace the 10.0.0.X IP address with that web servers
server web1 10.0.0.1:80 check # Active in rotation server web2 10.0.0.2:80 check # Active in rotation server web3 10.0.0.3:80 check # Active in rotation server backup1 10.0.0.4:80 check backup # Not active "sorry server" - this one comes live if all web heads are down
Above is an example of what a four server config would look like. Once you have completed this portion you can then start HAProxy and start serving pages(assuming your web servers are ready).
service haproxy start
Below is the default configuration template for haproxy.cfg:
#global options global #logging is designed to work with syslog facility's due to chrooted environment #log loghost local0 info - By default this is commented out #chroot directory chroot /usr/share/haproxy #user/group id uid 99 gid 99 #running mode daemon defaults #HTTP Log format mode http #number of connection retries for the session retries 3 #try another webhead if retry fails option redispatch #session settings - max connections, and session timeout values maxconn 10000 contimeout 10000 clitimeout 50000 srvtimeout 50000 #Define your farm #listen webfarm 0.0.0.0:80 - Pass only HTTP traffic and bind to port 80 listen webfarm 0.0.0.0:80 #HTTP Log format mode http #stats uri /haproxy - results in http://
/haproxy (shows load balancer stats) stats uri /haproxy #balance roundrobin - Typical Round Robin #balance leastconn - Least Connections #balance static-rr - Static Round Robin - Same as round robin, but weights have no effect balance roundrobin #cookie prefix - Used for cookie-based persistence cookie webpool insert #option httpclose - http connection closing option httpclose #option forwardfor - best stated as "Enable insertion of the X-Forwarded-For header to requests sent to the web heads" aka send EU IP option forwardfor #Web Heads (Examples) #server WEB1 10.0.0.1:80 check - passes http traffic to this server and checks if its alive #server WEB1 10.0.0.1:80 check port 81 - same as above but checks port 81 to see if its alive (helps to remove servers from rotation) #server WEB1 10.0.0.1:80 check port 81 weight 100 - same as the above with weight specification (weights 1-256 / higher number higher weight) #server WEB1 10.0.0.1:80 check backup - defines this server as a backup for the other web heads #Working Example: *USE THIS HOSTNAME FORMAT* server WWW1 10.0.0.1:80 cookie webpool_WWW1 check port 81 # Active in rotation server WWW2 10.0.0.2:80 cookie webpool_WWW2 check port 81 # Active in rotation server WWW3 10.0.0.3:80 check # Active in rotation server WWW4 10.0.0.4:80 check backup # Not active "sorry server" - this one comes live if all web heads are down #SSL farm example #listen https 0.0.0.0:443 # mode tcp # server WEB1 10.0.0.1:443 check
Session Persistence with SSL
If the you wish to also balance SSL traffic, you will need to set the balance mode to "source" This setting takes a hash of the client's IP address and the number of servers in rotation, sending traffic from one IP address to the same web server consistently. The persistence will be reset if the number of servers is changed.:
listen https 0.0.0.0:443 mode tcp balance source server WEB1 10.0.0.1:443 check