[squid-users] IPVS/LVS load balancing Squid servers, anyone did it?
FUSTE Emmanuel
emmanuel.fuste at thalesgroup.com
Thu Aug 27 11:22:53 UTC 2020
Hi,
To complement this, on modern kernel take the opportunity to try nftlb
instead of LVS too.
https://www.zevenet.com/knowledge-base/nftlb/what-is-nftlb/
Emmanuel.
Le 27/08/2020 à 06:35, Bruce Rosenberg a écrit :
> Hi Eliezer,
>
> We are running a couple of Squid proxies (the real servers) in front
> of a pair of LVS servers with keepalived and it works flawlessly.
> The 2 x Squid proxies are active / active and the LVS servers are
> active / passive.
> If a Squid proxy dies the remaining proxy takes all the traffic.
> If the active LVS server dies, keepalived running on the backup LVS
> (via VRRP) moves the VIP to itself and it takes all the traffic, so
> the only difference between the two is one has a higher priority so it
> gets the VIP first.
> I have included some sanitised snippets from a keepalived.conf file
> that should help you.
> You could easily scale this out if you need more than 2 Squid proxies.
>
> The config I provided is for LVS/DR (Direct Route) mode.
> This method rewrites the MAC address of forwarded packets to that of
> one of the real servers and is the most scalable way to run LVS.
> It does require the LVS and real servers be on the same L2 network.
> If that is not possible then consider LVS/TUN mode or LVS/NAT mode.
>
> As LVS/DR rewrites the MAC address, it requires each real server to
> have the VIP address plumbed on an interface and also requires the
> real servers to ignore ARP requests for the VIP address as the only
> device that should respond to ARP requests for the VIP is the active
> LVS server.
> We do this by configuring the VIP on the loopback interface on each
> real but there are other methods as well such as dropping the ARP
> responses using arptables, iptables or firewalld.
> I think back in the kernel 2.4 and 2.6 days people used the noarp
> kernel module which could be configured to ignore ARP requests for a
> particular IP address but you don't really need this anymore.
>
> More info on the loopback arp blocking method -
> https://www.loadbalancer.org/blog/layer-4-direct-routing-lvs-dr-and-layer-4-tun-lvs-tun-in-aws/
> More info on firewall type arp blocking methods -
> https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/7/html/load_balancer_administration/s1-lvs-direct-vsa
> More info about LVS/DR - http://kb.linuxvirtualserver.org/wiki/LVS/DR
>
> If you are using a RPM based distro then to set up the LVS servers you
> only need the ipvsadm and keepalived packages.
> Install squid on the reals and configure the VIP on each and disable ARP.
> Then build the keepalived.conf on both LVS servers and restart keepalived.
>
> The priority configuration stanza in the vrrp_instance section
> determines the primary VRRP node (LVS server) for that virtual router
> instance.
> The secondary LVS server needs a lower priority compared to the primary.
> You can configure one as the MASTER and the other as the BACKUP but
> our guys make them both BACKUP and let the priority sort the election
> of the primary out.
> I think this might be to solve a problem of bringing up a BACKUP
> without a MASTER but I can't confirm that.
>
>
> Good luck.
>
>
> $ cat /etc/keepalived/keepalived.conf
>
> global_defs {
>
> notification_email {
> # rootmail at example.com <mailto:rootmail at example.com>
> }
> notification_email_from keepalive-daemon at lvs01.example.com
> <mailto:keepalive-daemon at lvs01.example.com>
> smtp_server 10.1.2.3 # mail.example.com
> <http://mail.example.com>
> smtp_connect_timeout 30
> lvs_id lvs01.example.com <http://lvs01.example.com> # Name to
> mention in email.
> }
>
> vrrp_instance LVS_example {
>
> state BACKUP
> priority 150
> interface eth0
> lvs_sync_daemon_interface eth0
> virtual_router_id 5
> preempt_delay 20
>
> virtual_ipaddress_excluded {
>
> 10.10.10.10 # Squid proxy
> }
>
> notify_master "some command to log or send an alert"
> notify_backup "some command to log or send an alert"
> notify_fault "some command to log or send an alert"
> }
>
>
> # SQUID Proxy
> virtual_server 10.10.10.10 3128 {
>
> delay_loop 5
> lb_algo wrr
> lb_kind DR
> protocol TCP
>
> real_server 10.10.10.11 3128 { # proxy01.example.com
> <http://proxy01.example.com>
> weight 1
> inhibit_on_failure 1
> TCP_CHECK {
> connect_port 3128
> connect_timeout 5
> }
> }
>
> real_server 10.10.10.12 3128 { # proxy02.example.com
> <http://proxy02.example.com>
> weight 1
> inhibit_on_failure 1
> TCP_CHECK {
> connect_port 3128
> connect_timeout 5
> }
> }
> }
>
>
> On Thu, Aug 27, 2020 at 8:24 AM Eliezer Croitor <ngtech1ltd at gmail.com
> <mailto:ngtech1ltd at gmail.com>> wrote:
>
> Hey All,
>
> I am reading about LB and tried to find an up-to-date example or
> tutorial specific to squid with no luck.
>
> I have seen:
> http://kb.linuxvirtualserver.org/wiki/Building_Web_Cache_Cluster_using_LVS
>
> Which makes sense and also is similar or kind of identical to WCCP
> with gre.
>
> Anyone knows about a working Squid setup with IPVS/LVS?
>
> Thanks,
>
> Eliezer
>
> ----
>
> Eliezer Croitoru
>
> Tech Support
>
> Mobile: +972-5-28704261
>
> Email: ngtech1ltd at gmail.com <mailto:ngtech1ltd at gmail.com>
>
> _______________________________________________
> squid-users mailing list
> squid-users at lists.squid-cache.org
> <mailto:squid-users at lists.squid-cache.org>
> http://lists.squid-cache.org/listinfo/squid-users
>
>
> _______________________________________________
> squid-users mailing list
> squid-users at lists.squid-cache.org
> http://lists.squid-cache.org/listinfo/squid-users
More information about the squid-users
mailing list