ADC

Configuring layer 3 clustering

Understanding the L3 cluster

The demand to expand the high availability deployment and increase the scalability of the client traffic across different networks guided to establish the L3 cluster. The L3 cluster lets you group Citrix ADC appliances across individual subnets (L2 cluster).

L3 cluster is also referred to as “cluster in Independent Network Configuration (INC) mode”. In L3 cluster deployment, the cluster nodes in the same network are grouped to form a Node group. L3 cluster uses GRE tunneling to steer the packets across networks. The heartbeat messages across the L3 clusters are routed.

Architecture

The L3 cluster architecture comprises the following components:

  • Nodegroup - The cluster nodes from each network (n1, n2) and (n3, n4), as shown in the following figure, are grouped to form a Node group. These Node groups are terminated to the layer 3 switch on either side of the network.
    • The cluster communicates with the client through the physical connections between the cluster node and the client-side connecting device. The logical grouping of these physical connections is called the client data plane.
    • The cluster communicates with the server through the physical connections between the cluster node and the server side connecting device. The logical grouping of these physical connections is called the server data plane.
  • Backplane Switch - Cluster nodes within the same network communicate with each other by using the cluster backplane. The backplane is a set of interfaces in which one interface of each node is connected to a common switch, which is called the cluster backplane switch.
  • Tunnel - By default, the packets between nodes in a L3 cluster are exchanged over an unencrypted GRE tunnel that uses the NSIP addresses of the source and destination nodes for routing. The steering mechanism changes for nodes belonging to the different network. The packets are steered through a GRE tunnel to the node on the other subnet, instead of rewriting the MAC.

    Note:

    For a new deployment, we recommend using the UDP tunnel to steer the packets across networks. The advantages of using a UDP tunnel are:

    • The steering between the packet engines is based on 4-tuple, which provides better distribution.
    • You can avoid the skewness in CPU utilization.

    Refer to the following table to understand how packets are steered in different tunnel modes.

    Tunnel mode Same network Different network
    None Backplane is used to steer the packets. GRE tunnel is used to steer the traffic.
    GRE GRE tunnel is used to steer the packets. GRE tunnel is used to steer the packets.
    UDP UDP tunnel is used to steer the packets. UDP tunnel is used to steer the packets.

Cluster architecture

Example

Consider an example of an L3 cluster deployment consisting of the following:

  • Three Citrix ADC appliances (n1, n2, and n3) nodes are grouped into Nodegroup1.
  • Similarly, the nodes n4 and n5 are grouped in Nodegroup2. In the third network, there are two node groups. Nodegroup3 includes n6 and n7 and Nodegroup4 includes n8 and n9.
  • The Citrix ADC appliances that belong to the same network are combined to form a node group.

Cluster deployment

Points to consider before configuring the L3 cluster

Consider the following points before configuring the L3 cluster on a Citrix ADC appliance:

  • The backplane is not mandatory while configuring L3 subnets. If the backplane is not specified, the node does not go to the backplane fail state.

    Note:

    If you have more than one node in the same L2 network, it is mandatory to define the backplane interface. If the backplane interface is not mentioned, the nodes go to the backplane fail state.

  • L2 features and striped SNIPs are not supported in the L3 cluster.

  • The external traffic distribution in the L3 cluster supports only Equal Cost Multiple Path (ECMP).
  • The ICMP errors and fragmentation are not processed when steering is disabled in an L3 cluster deployment:
  • The networking entities (route, route6, pbr, and pbr6) must be bound to the configuration node group.
  • VLAN, RNAT, and IP tunnel cannot be bound to a configuration node group.
  • Configuration node group must always have property STRICT “YES.
  • The cluster nodes must not be added to a config node group via “add cluster node” command.
  • The add cluster instance -INC enabled command clears the networking entities (route, route6, PBR, pb6, RNAT, IP tunnel, ip6tunnel).
  • The clear config extended+ command does not clear the entities (route, route6, PBR, pb6, RNAT, IP tunnel, ip6tunnel) in an L3 cluster.

Configuring the L3 cluster

In an L3 cluster configuration, the cluster command has different attributes to configure that is based on nodes, and node groups. The L3 cluster configuration also includes an IPv6 profile apart from IPv4 profiles.

Configuring an L3 cluster on a Citrix ADC appliance consists of the following tasks:

  • Create a cluster instance
  • Create a node group in L3 cluster
  • Add a Citrix ADC appliance to the cluster and group with node group
  • Add cluster IP address to the node
  • Enable the cluster instance
  • Save the configuration
  • Add a node to an existing node group
  • Create a node group in L3 cluster
  • Group new nodes to the newly created node group
  • Join the node to the cluster

Configuring the following by using the CLI

  • To create a cluster instance

    add cluster instance <clid> -inc <ENABLED|DISABLED> -processLocal <ENABLED|DISABLED>

    Note:

    The “inc” parameter must be ENABLED for an L3 cluster.

  • To create a nodegroup in L3 cluster

    add cluster nodegroup <name>

  • To add a Citrix ADC appliance to the cluster and to associate with nodegroup

    add cluster node <nodeid> <nodeip> -backplane <interface_name> nodegroup <ng> –tunnelmode UDP

  • To add the cluster IP address on this node

    add ns ip <IPAddress> <netmask> -type clip

  • Enable the cluster instance

    enable cluster instance <clId>

  • Save the configuration

    save ns config

  • Warm reboot the appliance

    reboot -warm

  • To add a new node to an existing nodegroup

    add cluster node <nodeid> <nodeip> -nodegroup <ng> –tunnelmode UDP

  • To create a new nodegroup in L3 cluster

    add cluster nodegroup <ng>

  • To group new nodes to the newly created nodegroup

    add cluster node <nodeid> <nodeip> -nodegroup <ng> –tunnelmode UDP

  • To join the node to the cluster

    join cluster –clip <ip_addr> -password <password>

Example: The following is an example configuration for nodegroup1 and nodegroup2.

    > add cluster instance 1 –inc ENABLED –processLocal ENABLED

       Done

    > add cluster nodegroup ng1

       Done

    > add cluster node 0 1.1.1.1 –state ACTIVE -backplane 0/1/1 –nodegroup ng1 –tunnelmode UDP

       Done

    > add ns ip 1.1.1.100 255.255.255.255 –type clip

       Done

    > enable cluster instance 1

       Done

    > save ns config

       Done

    > add cluster node 1 1.1.1.2 –state ACTIVE –nodegroup ng1 –tunnelmode UDP

       Done

    > add cluster nodegroup ng2

       Done

    > add cluster node 4 2.2.2.1 –state ACTIVE –nodegroup ng2 –tunnelmode UDP

       Done

    > add cluster node 5 2.2.2.2 –state ACTIVE –nodegroup ng2 –tunnelmode UDP

       Done

    > join cluster -clip 1.1.1.100 -password nsroot
<!--NeedCopy-->

Advertising cluster IP address of a L3 cluster

Configure the cluster IP address to be advertised to the upstream router to make the cluster configuration accessible from any subnet. The cluster IP address is advertised as a kernel route by the dynamic routing protocols configured on a node.

Advertising the cluster IP address consists of the following tasks:

  • Enable the host route option of the cluster IP address. The host route option pushes the cluster IP address to a ZebOS routing table for kernel route redistribution through dynamic routing protocols.
  • Configuring a dynamic routing protocol on a node. A dynamic routing protocol advertises the cluster IP address to the upstream router. For more information on configuring a dynamic routing protocol, see Configuring Dynamic Routes.

To enable the host route option of the cluster IP address by using the CLI

  At the command prompt, type:

  • add nsip <IPAddress> <netmask> -hostRoute ENABLED
  • show nsip <IPAddress>
    > add ns ip 10.102.29.60 255.255.255.255 -hostRoute ENABLED

    > Done
<!--NeedCopy-->

Spotted, partially striped configurations on L3 cluster

The spotted and partially striped configurations on the L3 cluster slightly differ from the L2 cluster. The configuration might differ from node to node as the nodes reside on different subnets. The network configurations can be node specific in the L3 cluster, hence you have to configure the spotted or partially striped configurations based on the below-mentioned parameters.

To configure spotted, partially striped configurations on a Citrix ADC appliance over the L3 cluster, perform the following tasks:

  • Add a cluster owner group to an IPv4 static routing table
  • Add a cluster owner group to an IPv6 static routing table
  • Add a cluster owner group to an IPv4 policy based routing (PBR)
  • Add a cluster owner group to an IPv6 PBR
  • Add a VLAN
  • Bind a VLAN to a specific owner group of cluster node group

Configuring the following by using the CLI

  • To add a cluster ownergroup to an IPv4 static route table of the Citrix ADC appliance

    add route <network> <netmask> <gateway> -owner group <ng>

  • To add a cluster ownergroup to an IPv6 static route table of the Citrix ADC appliance appliance

    add route6 <network> -owner group <ng>

  • To add a cluster ownergroup to an IPv4 PBR

    add pbr <name> <action> -owner group <ng>

  • To add a cluster ownergroup to an IPv6 PBR

    add pbr6 <name> <action> -owner group <ng>

  • To add a VLAN

    add vlan <id>

  • To bind a VLAN to a specific ownergroup of cluster nodegroup

    bind vlan <id> -ifnum – [IPAddress <ipv4_addr | ipv6_addr> [-owner group <ng>]

    The following commands are sample examples of spotted and partially striped configurations which can be configured by using the CLI.

    > add route 10.102.29.0 255.255.255.0 10.102.29.2 –ownergroup ng2

        Done

    > add route6 fe80::9404:60ff:fedd:a464/64 –ownergroup ng1

        Done

    > add pbr pbr1 allow –ownergroup ng1

        Done

    > add pbr6 pbr2 allow –ownergroup ng2

        Done

    > add vlan 2

        Done

    > bind vlan 2 –ifnum 1/2 –[IPAddress 10.102.29.80 | fe80::9404:60ff:fedd:a464/64-ownergroup ng1

        Done
<!--NeedCopy-->

Configure node group

In an L3 cluster, to replicate the same set of configurations on more than one node group, the following commands are used:

Configuring the following by using the CLI

  • To add an IPv4 static route to the routing table of the Citrix ADC appliance

    add route <network> <netmask> <gateway> -ownerGroup <ng>

Sample Configuration:

add route 0 0 10.102.53.1 –ownerGroup ng1

add route 0 0 10.102.53.1 –ownerGroup ng2
<!--NeedCopy-->

You define a new node group ‘all’ to support the preceding configuration, and have to configure the following commands:

Configuring the following by using the CLI

  • To add a new nodegroup to cluster with strict parameter

    add cluster node group <name> -strict <YES | NO>

  • To bind a cluster node or an entity to the given nodegroup

    bind cluster nodegroup <name> -node <nodeid>

  • To add IPv4 static route to all ownergroup

    add route <network> <netmask> <gateway> -ownerGroup <ng>

Sample configuration:

add cluster nodegroup all –strict YES

bind cluster nodegroup all –node 1

bind cluster nodegroup all –node 2

add route 0 0 10.102.53.1 –ownerGroup all
<!--NeedCopy-->

Traffic distribution in a L3 cluster

In a cluster setup, external networks view the collection of Citrix ADC appliances as a single entity. So, the cluster must select a single node that must receive the traffic. In the L3 cluster this selection is done using the ECMP. The selected node is called the flow receiver.

Note

For an L3 cluster (nodes across different networks), only the ECMP traffic distribution can be used.

The flow receiver gets the traffic and then, using internal cluster logic determines the node that must process the traffic. This node is called the flow processor. The flow receiver steers the traffic to the flow processor over the backplane if the flow receiver and the flow processor are on the same network. The traffic is steered through the tunnel if the flow receiver and the flow processor are on different networks.

Note

  • The flow receiver and flow processor must be nodes capable of serving traffic.

  • From NetScaler 11 onwards, you can disable steering on the cluster backplane. For more information, see Disabling steering on the cluster backplane.

L3 cluster packet flow

The preceding figure shows a client request flowing through the cluster. The client sends a request to a virtual IP (VIP) address. A traffic distribution mechanism configured on the client data plane selects one of the cluster nodes as the flow receiver. The flow receiver receives the traffic, determines the node that must process the traffic, and steers the request to that node (unless the flow receiver selects itself as the flow processor). If the flow processor and flow receiver are in the same node group, the packet is steered over the backplane. And if the flow processor and flow receiver are in different node groups, then the packet is steered through the tunnel over the routed path.

The flow processor establishes a connection with the server. The server processes the request and sends the response to the subnet IP (SNIP) address that sent the request to the server. Since in the L3 cluster the SNIP is always a spotted SNIP, the node that owns the SNIP address receives the response from the server.

Configuring layer 3 clustering