CentOS7 haproxy+keepalived实现高可用集群搭建

1.1 本地操作系统环境

CentOS7 64位



[root@lb03 ~]# cat /etc/centos-release
CentOS Linux release 7.5.1804 (Core)
 
 
       
 
 
[root@lb03 ~]# uname -r
3.10.0-862.el7.x86_64
[root@lb03 ~]# rpm -qa haproxy
haproxy-1.5.18-7.el7.x86_64
[root@lb03 ~]# nginx -V
nginx version: nginx/1.12.2
built by gcc 4.8.5 20150623 (Red Hat 4.8.5-16) (GCC)
built with OpenSSL 1.0.2k-fips  26 Jan 2017
TLS SNI support enabled

 



  后端负载主机:192.168.25.71   192.168.25.72   两台节点上安装rabbitmq服务

        Haproxy 也是安装在 192.168.25.71 和 192.168.25.72 上,用于对外提供 RabbitMQ 均衡

        Keepalived实现haproxy的主备,高可用(避免单点问题),192.168.25.71(主)192.168.25.72(备),虚拟地址VIP 192.168.166.29

 rabbitmq配置可以参看


第2章 HAproxy安装

2.1 HAproxy简介

1)HAProxy提供高可用性、负载均衡以及基于TCP和HTTP应用的代理,支持虚拟主机,它是免费、快速并且可靠的一种解决方案。

2)HAProxy特别适用于那些负载特大的web站点,这些站点通常又需要会话保持或七层处理。

3)HAProxy运行在当前的硬件上,完全可以支持数以万计的并发连接。并且它的运行模式使得它可以很简单安全的整合进您当前的架构中, 同时可以保护你的web服务器不被暴露到网络上。

4)HAProxy实现了一种事件驱动, 单一进程模型,此模型支持非常大的并发连接数。多进程或多线程模型受内存限制 、系统调度器限制以及无处不在的锁限制,很少能处理数千并发连接。事件驱动模型因为在有更好的资源和时间管理的用户空间(User-Space) 实现所有这些任务,所以没有这些问题。此模型的弊端是,在多核系统上,这些程序通常扩展性较差。这就是为什么他们必须进行优化以 使每个CPU时间片(Cycle)做更多的工作。


2.2 HAproxy安装配置

2.2.1 Haproxy代理rabbitmq 配置文件



[root@lb01 haproxy]# cat haproxy.cfg
###########全局配置#########
global
#    log /dev/log    local0 
#    log /dev/log    local1 notice
    log 127.0.0.1 local0 info
    chroot /var/lib/haproxy     # 改变当前工作目录
    stats socket /run/haproxy/admin.sock mode 660 level admin   # 创建监控所用的套接字目录
    pidfile  /var/run/haproxy.pid   # haproxy的pid存放路径,启动进程的用户必须有权限访问此文件 
    maxconn  4000                   # 最大连接数,默认4000
    user   haproxy                  # 默认用户
    group   haproxy                 # 默认用户组
    daemon                          # 创建1个进程进入deamon模式运行。此参数要求将运行模式设置为"daemon
 
    # Default SSL material locations
    ca-base /etc/ssl/certs
    crt-base /etc/ssl/private
 
    # Default ciphers to use on SSL-enabled listening sockets.
    # For more information, see ciphers(1SSL). This list is from:
    #  https://hynek.me/articles/hardening-your-web-servers-ssl-ciphers/
    ssl-default-bind-ciphers ECDH+AESGCM:DH+AESGCM:ECDH+AES256:DH+AES256:ECDH+AES128:DH+AES:ECDH+3DES:DH+3DES:RSA+AESGCM:RSA+AES:RSA+3DES:!aNULL:!MD5:!DSS
    ssl-default-bind-options no-sslv3

 

###########默认配置#########

defaults
    log global
    mode    http                                # 默认的模式mode { tcp|http|health },tcp是4层,http是7层,health只会返回OK
    option  httplog                             # 采用http日志格式
    option  dontlognull                         # 启用该项,日志中将不会记录空连接。所谓空连接就是在上游的负载均衡器
                                                # 或者监控系统为了探测该 服务是否存活可用时,需要定期的连接或者获取某
                                                # 一固定的组件或页面,或者探测扫描端口是否在监听或开放等动作被称为空连接;
                                                # 官方文档中标注,如果该服务上游没有其他的负载均衡器的话,建议不要使用
                                                # 该参数,因为互联网上的恶意扫描或其他动作就不会被记录下来
    timeout connect 5000                    # 连接超时时间
    timeout client  50000                   # 客户端连接超时时间
    timeout server  50000                   # 服务器端连接超时时间
    option  httpclose       # 每次请求完毕后主动关闭http通道 
    option  httplog         # 日志类别http日志格式 
    #option  forwardfor      # 如果后端服务器需要获得客户端真实ip需要配置的参数,可以从Http Header中获得客户端ip  
    option  redispatch      # serverId对应的服务器挂掉后,强制定向到其他健康的服务器
    timeout connect 10000   # default 10 second timeout if a backend is not found
    maxconn     60000       # 最大连接数
    retries     3           # 3次连接失败就认为服务不可用,也可以通过后面设置 
#    errorfile 400 /etc/haproxy/errors/400.http
#    errorfile 403 /etc/haproxy/errors/403.http
#    errorfile 408 /etc/haproxy/errors/408.http
#    errorfile 500 /etc/haproxy/errors/500.http
#    errorfile 502 /etc/haproxy/errors/502.http
#    errorfile 503 /etc/haproxy/errors/503.http
#    errorfile 504 /etc/haproxy/errors/504.http
####################################################################
listen http_front
        bind 0.0.0.0:1080  
        stats refresh 30s 
        stats uri /haproxy?stats   
        stats realm Haproxy Manager
        stats auth admin:admin    
        #stats hide-version      
 
#####################我把RabbitMQ的管理界面也放在HAProxy后面了###############################
listen rabbitmq_admin 
    bind 0.0.0.0:8004
    server node1 192.168.25.73:15672
    server node2 192.168.25.74:15672
    server node3 192.168.25.75:15672
####################################################################
listen rabbitmq_cluster 
    bind 0.0.0.0:80
    option tcplog
    mode tcp
    timeout client  3h
    timeout server  3h
    option          clitcpka
    balance roundrobin      
    #balance url_param userid
    #balance url_param session_id check_post 64
    #balance hdr(User-Agent)
    #balance hdr(host)
    #balance hdr(Host) use_domain_only
    #balance rdp-cookie
    #balance leastconn
    #balance source //ip
    server   node1 192.168.25.73:15672 check inter 5s rise 2 fall 3   
    server   node2 192.168.25.74:15672 check inter 5s rise 2 fall 3
    server   node3 192.168.25.75:15672 check inter 5s rise 2 fall 3
 
 
2.2.2 Haproxy代理nginx
1、安装haproxy
 # yum install haproxy -y 2、编辑配置文件
 
 
 
[root@lb02 ~]# grep -Ev '^$|^#' /etc/haproxy/haproxy.cfg 
 
global
    # to have these messages end up in /var/log/haproxy.log you will
    # need to:
    #
    # 1) configure syslog to accept network log events.  This is done
    #    by adding the '-r' option to the SYSLOGD_OPTIONS in
    #    /etc/sysconfig/syslog
    #
    # 2) configure local2 events to go to the /var/log/haproxy.log
    #   file. A line like the following can be added to
    #   /etc/sysconfig/syslog
    #
    #    local2.*                       /var/log/haproxy.log
    #
    log 127.0.0.1 local0 info
 
    #chroot      /var/lib/haproxy
    #pidfile     /var/run/haproxy.pid
    maxconn     4000
    user        haproxy
    group       haproxy
    daemon
 
    # turn on stats unix socket
    #stats socket /var/lib/haproxy/stats
 
defaults
    mode                    http
    log                     global
    option                  httplog
    option                  dontlognull
    option http-server-close
    option forwardfor       except 127.0.0.0/8
    option                  redispatch
    retries                 3
    timeout http-request    10s
    timeout queue           1m
    timeout connect         10s
    timeout client          1m
    timeout server          1m
    timeout http-keep-alive 10s
    timeout check           10s
    maxconn                 3000
 
frontend main
    bind *:80
    acl url_static       path_beg       -i /static /images /javascript /stylesheets
    acl url_static       path_end       -i .jpg .gif .png .css .js
 
    use_backend static          if url_static
    default_backend             nginx
 
backend static
    balance     roundrobin
    server      static 127.0.0.1:80 check
 
backend nginx
    balance     roundrobin
    server  nginx1 192.168.25.73:80 check inter 2000 fall 3 weight 30
    server  nginx2 192.168.25.74:80 check inter 2000 fall 3 weight 30
    server  nginx3 192.168.25.75:80 check inter 2000 fall 3 weight 30
 
 
3、启动
 # haproxy -f /etc/haproxy/haproxy.cfg
 4、重启动
 # service haproxy restart5、查看haproxy是否已经启动

6、haproxy启用监控页面
编辑haproxy.cfg  加上下面参数  
listen admin_stats
        stats   enable
        bind    *:9090    //监听的ip端口号
        mode    http    //开关
        option  httplog
        log     global
        maxconn 10
        stats   refresh 30s   //统计页面自动刷新时间
        stats   uri /admin    //访问的uri   ip:8080/admin
        stats   realm haproxy
        stats   auth admin:Redhat  //认证用户名和密码
        stats   hide-version   //隐藏HAProxy的版本号
        stats   admin if TRUE   //管理界面,如果认证成功了,可通过webui管理节点
保存退出后
重起service haproxy restart
然后访问 http://192.168.25.72:9090/admin          用户名:admin 密码:Redhat
 
 

---------------------------------------------------------------------------------------------
参数举例说明:【/usr/local/haproxy/haproxy.cfg】
###########全局配置#########
global
  log 127.0.0.1 local0 #[日志输出配置,所有日志都记录在本机,通过local0输出]
  log 127.0.0.1 local1 notice #定义haproxy 日志级别[error warringinfo debug]
  daemon #以后台形式运行harpoxy
  nbproc 1 #设置进程数量
  maxconn 4096 #默认最大连接数,需考虑ulimit-n限制
  #user haproxy #运行haproxy的用户
  #group haproxy #运行haproxy的用户所在的组
  #pidfile /var/run/haproxy.pid #haproxy 进程PID文件
  #ulimit-n 819200 #ulimit 的数量限制
  #chroot /usr/share/haproxy #chroot运行路径
  #debug #haproxy 调试级别,建议只在开启单进程的时候调试
  #quiet
 
########默认配置############
defaults
  log global
  mode http #默认的模式mode { tcp|http|health },tcp是4层,http是7层,health只会返回OK
  option httplog #日志类别,采用httplog
  option dontlognull #不记录健康检查日志信息
  retries 2 #两次连接失败就认为是服务器不可用,也可以通过后面设置
  #option forwardfor #如果后端服务器需要获得客户端真实ip需要配置的参数,可以从Http Header中获得客户端ip
  option httpclose #每次请求完毕后主动关闭http通道,haproxy不支持keep-alive,只能模拟这种模式的实现
  #option redispatch #当serverId对应的服务器挂掉后,强制定向到其他健康的服务器,以后将不支持
  option abortonclose #当服务器负载很高的时候,自动结束掉当前队列处理比较久的链接
  maxconn 4096 #默认的最大连接数
  timeout connect 5000ms #连接超时
  timeout client 30000ms #客户端超时
  timeout server 30000ms #服务器超时
  #timeout check 2000 #心跳检测超时
  #timeout http-keep-alive10s #默认持久连接超时时间
  #timeout http-request 10s #默认http请求超时时间
  #timeout queue 1m #默认队列超时时间
  balance roundrobin #设置默认负载均衡方式,轮询方式
  #balance source #设置默认负载均衡方式,类似于nginx的ip_hash
  #balnace leastconn #设置默认负载均衡方式,最小连接数
 
########统计页面配置########
listen stats
  bind 0.0.0.0:1080 #设置Frontend和Backend的组合体,监控组的名称,按需要自定义名称
  mode http #http的7层模式
  option httplog #采用http日志格式
  #log 127.0.0.1 local0 err #错误日志记录
  maxconn 10 #默认的最大连接数
  stats refresh 30s #统计页面自动刷新时间
  stats uri /stats #统计页面url
  stats realm XingCloud\ Haproxy #统计页面密码框上提示文本
  stats auth admin:admin #设置监控页面的用户和密码:admin,可以设置多个用户名
  stats auth Frank:Frank #设置监控页面的用户和密码:Frank
  stats hide-version #隐藏统计页面上HAProxy的版本信息
  stats admin if TRUE #设置手工启动/禁用,后端服务器(haproxy-1.4.9以后版本)
 
########设置haproxy 错误页面#####
#errorfile 403 /home/haproxy/haproxy/errorfiles/403.http
#errorfile 500 /home/haproxy/haproxy/errorfiles/500.http
#errorfile 502 /home/haproxy/haproxy/errorfiles/502.http
#errorfile 503 /home/haproxy/haproxy/errorfiles/503.http
#errorfile 504 /home/haproxy/haproxy/errorfiles/504.http
 
########frontend前端配置##############
frontend main
  bind *:80 #这里建议使用bind *:80的方式,要不然做集群高可用的时候有问题,vip切换到其他机器就不能访问了。
  acl web hdr(host) -i www.abc.com  #acl后面是规则名称,-i为忽略大小写,后面跟的是要访问的域名,如果访问www.abc.com这个域名,就触发web规则,。
  acl img hdr(host) -i img.abc.com  #如果访问img.abc.com这个域名,就触发img规则。
  use_backend webserver if web   #如果上面定义的web规则被触发,即访问www.abc.com,就将请求分发到webserver这个作用域。
  use_backend imgserver if img   #如果上面定义的img规则被触发,即访问img.abc.com,就将请求分发到imgserver这个作用域。
  default_backend dynamic #不满足则响应backend的默认页面
 
########backend后端配置##############
backend webserver #webserver作用域
  mode http
  balance roundrobin #balance roundrobin 负载轮询,balance source 保存session值,支持static-rr,leastconn,first,uri等参数
  option httpchk /index.html HTTP/1.0 #健康检查, 检测文件,如果分发到后台index.html访问不到就不再分发给它
  server web1 10.16.0.9:8085 cookie 1 weight 5 check inter 2000 rise 2 fall 3
  server web2 10.16.0.10:8085 cookie 2 weight 3 check inter 2000 rise 2 fall 3
  #cookie 1表示serverid为1,check inter 1500 是检测心跳频率 
  #rise 2是2次正确认为服务器可用,fall 3是3次失败认为服务器不可用,weight代表权重
 
backend imgserver
  mode http
  option httpchk /index.php
  balance roundrobin 
  server img01 192.168.137.101:80 check inter 2000 fall 3
  server img02 192.168.137.102:80 check inter 2000 fall 3
 
backend dynamic 
  balance roundrobin 
  server test1 192.168.1.23:80 check maxconn 2000 
  server test2 192.168.1.24:80 check maxconn 2000
 
 
listen tcptest 
  bind 0.0.0.0:5222 
  mode tcp 
  option tcplog #采用tcp日志格式 
  balance source 
  #log 127.0.0.1 local0 debug 
  server s1 192.168.100.204:7222 weight 1 
  server s2 192.168.100.208:7222 weight 1
---------------------------------------------------------------------------------------------
第3章 Keepalived安装配置
3.1 Keepalived介绍
keepalived是一个免费开源的,用C编写的类似于layer3, 4 & 7交换机制软件,具备我们平时说的第3层、第4层和第7层交换机的功能。主要提供loadbalancing(负载均衡)和 high-availability(高可用)功能,负载均衡实现需要依赖Linux的虚拟服务内核模块(ipvs),而高可用是通过VRRP协议实现多台机器之间的故障转移服务。 
 
3.2 配置防火墙
3.2.1 Centos7-firewalld解决keepalived的VIP问题
keepalived的VIP问题
 
 
firewall-cmd --direct --permanent --add-rule ipv4 filter INPUT 0 --in-interface eth0 --destination 224.0.0.18 --protocol vrrp -j ACCEPT
 
 
 
3.2.2 firewalld 防火墙常用操作
语法命令如下:启用区域端口和协议组合
firewall-cmd [--zone=<zone>] --add-port=<port>[-<port>]/<protocol> [--timeout=<seconds>]
 
此举将启用端口和协议的组合。 
端口可以是一个单独的端口 <port> 或者是一个端口范围 <port>-<port>。 
协议可以是 tcp 或 udp。
 
查看 firewalld 状态
systemctl status firewalld
 
开启 firewalld 
systemctl start firewalld
 
开放端口
// --permanent 永久生效,没有此参数重启后失效
firewall-cmd --zone=public --add-port=80/tcp --permanent
firewall-cmd --zone=public --add-port=9090/tcp --permanent
firewall-cmd --zone=public --add-port=1000-2000/tcp --permanent
 
keepalived的VIP问题
firewall-cmd --direct --permanent --add-rule ipv4 filter INPUT 0 --in-interface eth0 --destination 224.0.0.18 --protocol vrrp -j ACCEPT
 
重新载入
firewall-cmd --reload
 
查看
firewall-cmd --zone=public --query-port=80/tcp
 
删除
firewall-cmd --zone=public --remove-port=80/tcp --permanent
 
iptables 防火墙
也可以还原传统的管理方式使用 iptables
systemctl stop firewalld 
systemctl mask firewalld
 
安装 iptables-services
yum install iptables-services
 
设置开机启动
systemctl enable iptables
 
操作命令
systemctl stop iptables 
systemctl start iptables 
systemctl restart iptables 
systemctl reload iptables
 
保存设置
service iptables save
 
开放某个端口 在 /etc/sysconfig/iptables 里添加
-A RH-Firewall-1-INPUT -m state --state NEW -m tcp -p tcp --dport 8080 -j ACCEPT
3.3 安装keepalived 
 
 
yum install keepalived –y
systemctl enable keepalived
 
 
 
3.4 配置文件
3.4.1 Master节点
3.4.1.1  keepalived.conf
 
 
[root@lb01 ~]# cat /etc/keepalived/keepalived.conf
global_defs {
   router_id LB01
}
vrrp_script chk_haproxy
{
     script "/etc/keepalived/scripts/haproxy_check.sh"
     interval 2
     timeout 2
     fall 3
}
vrrp_instance haproxy {
    state MASTER 
    interface eth0 
    virtual_router_id 1
    priority  100    
    authentication {         
            auth_type PASS         
            auth_pass password     
    }   
    virtual_ipaddress { 
           192.168.25.229
    }
    track_script {
         chk_haproxy
    }
    notify_master "/etc/keepalived/scripts/haproxy_master.sh"
}
 
 
3.4.1.2  haproxy_check.sh 
 
 
[root@lb01 ~]# cat /etc/keepalived/scripts/haproxy_check.sh 
#!/bin/bash
LOGFILE="/var/log/keepalived-haproxy-state.log"
 date >>$LOGFILE 
 if [ `ps -C haproxy --no-header |wc -l` -eq 0 ];then
     echo "fail: check_haproxy status" >>$LOGFILE 
     exit 1
 else
     echo "success: check_haproxy status" >>$LOGFILE
     exit 0
fi
 
 
3.4.1.3  haproxy_master.sh
 
 
[root@lb01 ~]# cat /etc/keepalived/scripts/haproxy_master.sh 
   #!/bin/bash
     LOGFILE="/var/log/keepalived-haproxy-state.log"
     echo "Being Master ..." >> $LOGFILE
 
 
 
3.4.2 Backup节点
3.4.2.1  keepalived.conf
 
 
[root@lb02 ~]# cat /etc/keepalived/keepalived.conf
global_defs {
   router_id LB02
}
vrrp_script chk_haproxy
{
     script "/etc/keepalived/scripts/haproxy_check.sh"
     interval 2
     timeout 2
     fall 3
}
vrrp_instance haproxy {
    state BACKUP 
    interface eth0 
    virtual_router_id 1
    priority  50    
    authentication {         
          auth_type PASS         
          auth_pass password     
    }         
    virtual_ipaddress { 
          192.168.25.229
    }
    track_script {
         chk_haproxy
    }
    notify_master "/etc/keepalived/scripts/haproxy_master.sh"
}
 
[root@lb02 ~]#
 
 
3.4.2.2  haproxy_check.sh
 
 
[root@lb01 ~]# cat /etc/keepalived/scripts/haproxy_check.sh 
#!/bin/bash
LOGFILE="/var/log/keepalived-haproxy-state.log"
 date >>$LOGFILE 
 if [ `ps -C haproxy --no-header |wc -l` -eq 0 ];then
     echo "fail: check_haproxy status" >>$LOGFILE 
     exit 1
 else
     echo "success: check_haproxy status" >>$LOGFILE
     exit 0
fi
[root@lb01 ~]#
 
 
3.4.2.3haproxy_master.sh
 
 
[root@lb01 ~]# cat /etc/keepalived/scripts/haproxy_master.sh 
   #!/bin/bash
     LOGFILE="/var/log/keepalived-haproxy-state.log"
     echo "Being Master ..." >> $LOGFILE
[root@lb01 ~]#
 
 
3.5 启动keepalived
3.5.1 启动服务
#keepalived –D
 
 
systemctl start keepalived
 
 
3.5.2 查看keepalived是否启动
 
 
 
[root@lb01 ~]# ps -ef|grep kee
root     24290     1  0 10:59 ?        00:00:00 /usr/sbin/keepalived -D
root     24291 24290  0 10:59 ?        00:00:00 /usr/sbin/keepalived -D
root     24292 24290  0 10:59 ?        00:00:00 /usr/sbin/keepalived -D
root     28622 13717  0 11:17 pts/1    00:00:00 grep --color=auto kee
[root@lb01 ~]#
 
 
3.5.3 vip检查


3.6 切换测试
3.6.1 关闭主keepalived
Master关闭keepalived服务,VIP是否切换至备机,业务是否正常,恢复原状;(验证keepalived高可用)
Master
Keepalived_ 日志
 
 
 
[root@lb01 ~]# systemctl stop keepalived
Nov 22 11:19:49 lb01 systemd: Stopping LVS and VRRP High Availability Monitor...
Nov 22 11:19:49 lb01 Keepalived[24290]: Stopping
Nov 22 11:19:49 lb01 Keepalived_vrrp[24292]: VRRP_Instance(haproxy) sent 0 priority
Nov 22 11:19:49 lb01 Keepalived_vrrp[24292]: VRRP_Instance(haproxy) removing protocol VIPs.
Nov 22 11:19:49 lb01 Keepalived_healthcheckers[24291]: Stopped
Nov 22 11:19:50 lb01 Keepalived_vrrp[24292]: Stopped
Nov 22 11:19:50 lb01 Keepalived[24290]: Stopped Keepalived v1.3.5 (03/19,2017), git commit v1.3.5-6-g6fa32f2
Nov 22 11:19:50 lb01 systemd: Stopped LVS and VRRP High Availability Monitor.
 
 
 
 
Backup
 
 
[root@lb02 ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
    inet 127.0.0.1/8 scope host lo
       valid_lft forever preferred_lft forever
    inet6 ::1/128 scope host 
       valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP group default qlen 1000
    link/ether 52:54:00:0b:29:c6 brd ff:ff:ff:ff:ff:ff
    inet 192.168.25.72/16 brd 192.168.255.255 scope global eth0
       valid_lft forever preferred_lft forever
    inet 192.168.25.229/32 scope global eth0
       valid_lft forever preferred_lft forever
    inet6 fe80::5054:ff:fe0b:29c6/64 scope link 
       valid_lft forever preferred_lft forever
3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP group default qlen 1000
    link/ether 52:54:00:1a:83:4d brd ff:ff:ff:ff:ff:ff
    inet6 fe80::5054:ff:fe1a:834d/64 scope link 
       valid_lft forever preferred_lft forever
 
 
Keepalived_ 日志
 
 
Nov 22 11:19:51 lb02 Keepalived_vrrp[26670]: VRRP_Instance(haproxy) Entering MASTER STATE
Nov 22 11:19:51 lb02 Keepalived_vrrp[26670]: VRRP_Instance(haproxy) setting protocol VIPs.
Nov 22 11:19:51 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:19:51 lb02 Keepalived_vrrp[26670]: VRRP_Instance(haproxy) Sending/queueing gratuitous ARPs on eth0 for 192.168.25.229
Nov 22 11:19:51 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:19:51 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:19:51 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:19:51 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:19:56 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:19:56 lb02 Keepalived_vrrp[26670]: VRRP_Instance(haproxy) Sending/queueing gratuitous ARPs on eth0 for 192.168.25.229
Nov 22 11:19:56 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:19:56 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:19:56 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:19:56 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
 
 
3.6.2 关闭主HAproxy
关闭主HAproxy ,VIP是否切换至备机,业务是否正常,恢复原状;(验证HAproxy高可用)
Master 
系统日志
 
 
Nov 22 11:29:43 lb01 systemd: Stopping HAProxy Load Balancer...
Nov 22 11:29:43 lb01 systemd: haproxy.service: main process exited, code=exited, status=143/n/a
Nov 22 11:29:43 lb01 systemd: Stopped HAProxy Load Balancer.
Nov 22 11:29:43 lb01 systemd: Unit haproxy.service entered failed state.
Nov 22 11:29:43 lb01 systemd: haproxy.service failed.
Nov 22 11:29:44 lb01 Keepalived_vrrp[29295]: /etc/keepalived/scripts/haproxy_check.sh exited with status 1
Nov 22 11:29:46 lb01 Keepalived_vrrp[29295]: /etc/keepalived/scripts/haproxy_check.sh exited with status 1
Nov 22 11:29:48 lb01 Keepalived_vrrp[29295]: /etc/keepalived/scripts/haproxy_check.sh exited with status 1
Nov 22 11:29:48 lb01 Keepalived_vrrp[29295]: VRRP_Script(chk_haproxy) failed
Nov 22 11:29:48 lb01 Keepalived_vrrp[29295]: VRRP_Instance(haproxy) Entering FAULT STATE
Nov 22 11:29:48 lb01 Keepalived_vrrp[29295]: VRRP_Instance(haproxy) removing protocol VIPs.
Nov 22 11:29:48 lb01 Keepalived_vrrp[29295]: VRRP_Instance(haproxy) Now in FAULT state
Nov 22 11:29:50 lb01 Keepalived_vrrp[29295]: /etc/keepalived/scripts/haproxy_check.sh exited with status 1
Nov 22 11:29:52 lb01 Keepalived_vrrp[29295]: /etc/keepalived/scripts/haproxy_check.sh exited with status 1
Nov 22 11:29:54 lb01 Keepalived_vrrp[29295]: /etc/keepalived/scripts/haproxy_check.sh exited with status 1
 
 
 
Backup
系统日志
 
 
Nov 22 11:29:49 lb02 Keepalived_vrrp[26670]: VRRP_Instance(haproxy) Transition to MASTER STATE
Nov 22 11:29:50 lb02 Keepalived_vrrp[26670]: VRRP_Instance(haproxy) Entering MASTER STATE
Nov 22 11:29:50 lb02 Keepalived_vrrp[26670]: VRRP_Instance(haproxy) setting protocol VIPs.
Nov 22 11:29:50 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:29:50 lb02 Keepalived_vrrp[26670]: VRRP_Instance(haproxy) Sending/queueing gratuitous ARPs on eth0 for 192.168.25.229
Nov 22 11:29:50 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:29:50 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:29:50 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:29:50 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:29:55 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:29:55 lb02 Keepalived_vrrp[26670]: VRRP_Instance(haproxy) Sending/queueing gratuitous ARPs on eth0 for 192.168.25.229
Nov 22 11:29:55 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:29:55 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:29:55 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
Nov 22 11:29:55 lb02 Keepalived_vrrp[26670]: Sending gratuitous ARP on eth0 for 192.168.25.229
 
 
Vip检查
Master
 
 
[root@lb01 ~]# ip a|grep 192.168.25.229
[root@lb01 ~]#
 
 
Backup
 
 
[root@lb02 ~]#  ip a|grep 192.168.25.229
    inet 192.168.25.229/32 scope global eth0



3)关闭后台服务器nginx 01,业务是否正常。(验证HAproxy状态检查)



[root@lb03 ~]# systemctl stop nginx
 Nov 22 11:34:50 localhost haproxy[31563]: Server nginx/nginx1 is DOWN, reason: Layer4 connection problem, info: "Connection refused", check duration: 0ms. 2 active and 0 backup servers left. 0 sessions active, 0 requeued, 0 remaining in queue.
[root@lb03 ~]# systemctl start nginx
Nov 22 11:36:10 localhost haproxy[31563]: Server nginx/nginx1 is UP, reason: Layer4 check passed, check duration: 0ms. 3 active and 0 backup servers online. 0 sessions requeued, 0 total in queue.



keepalived参考:http://blog.51cto.com/lanlian/1303195

haproxy监控页面配置参考: