Implementing Sticky Session Through Load Balancing

Concepts

Sticky sessions ensure continuity and consistency when you access applications. If a load balancer is deployed between a client and backend servers, connections may be forwarded to different servers for processing. Sticky sessions can resolve this issue. After sticky session is enabled, requests from the same client will be continuously distributed to the same backend server through load balancing.

For example, in most online systems that require user identity authentication, a user needs to interact with the server for multiple times to complete a session. These interactions require continuity. If sticky session is not configured, the load balancer may allocate certain requests to different backend servers. Since user identity has not been authenticated on other backend servers, interaction exceptions such as a user login failure may occur.

Therefore, select a proper sticky session type based on the application environment.

Table 1 Sticky session types

OSI Layer

Listener Protocol and Networking

Sticky Session Type

Stickiness Duration

Scenarios Where Sticky Sessions Become Invalid

Layer 4

TCP- or UDP-compliant Services

Source IP address: The source IP address of each request is calculated using the consistent hashing algorithm to obtain a unique hashing key, and all backend servers are numbered. The system allocates the client to a particular server based on the generated key. This allows requests from the same IP address are forwarded to the same backend server.

  • Default: 20 minutes

  • Maximum: 60 minutes

  • Range: 1 minute to 60 minutes

  • Source IP addresses of the clients have changed.

  • Requests from the clients exceed the session stickiness duration.

Layer 7

HTTP- or HTTPS-compliant ingresses

  • Load balancer cookie: The load balancer generates a cookie after receiving a request from the client. All subsequent requests with the cookie will be routed to the same backend server.

  • Application cookie: The application deployed on the backend server generates a cookie after receiving the first request from the client. All subsequent requests with the same cookie will be routed to the same backend server.

  • Default: 20 minutes

  • Maximum: 1440 minutes

  • Range: 1 minute to 1440 minutes

  • If requests sent by the clients do not contain a cookie, sticky sessions will not take effect.

  • Requests from the clients exceed the session stickiness duration.

Note

When creating a load balancer, configure sticky sessions by setting kubernetes.io/elb.lb-algorithm to ROUND_ROBIN or kubernetes.io/elb.lb-algorithm to LEAST_CONNECTIONS. If you set kubernetes.io/elb.lb-algorithm is to SOURCE_IP, source IP address-based sticky sessions are supported. In this case, you do not need to configure sticky sessions again.

Layer 4 Sticky Sessions for Services

In Layer 4 mode, source IP address-based sticky sessions can be enabled, where hash routing is performed based on the client IP address.

Enabling Layer 4 Sticky Session in a CCE Standard Cluster

In a CCE standard cluster, to enable source IP address-based sticky session for a Service, ensure the following conditions are met:

  1. Service Affinity of the Service must be set to Node-level, where the externalTrafficPolicy value of the Service must be Local.

  2. Anti-affinity has been enabled on the backend applications of the Service to prevent all pods from being deployed on the same node.

Procedure

  1. Create an Nginx workload.

    Set the number of pods to 3 and configure podAntiAffinity.

    kind: Deployment
    apiVersion: apps/v1
    metadata:
      name: nginx
      namespace: default
    spec:
      replicas: 3
      selector:
        matchLabels:
          app: nginx
      template:
        metadata:
          labels:
            app: nginx
        spec:
          containers:
            - name: container-0
              image: 'nginx:perl'
              resources:
                limits:
                  cpu: 250m
                  memory: 512Mi
                requests:
                  cpu: 250m
                  memory: 512Mi
          imagePullSecrets:
            - name: default-secret
          affinity:
            podAntiAffinity:                   # Pod anti-affinity
              requiredDuringSchedulingIgnoredDuringExecution:
                - labelSelector:
                    matchExpressions:
                      - key: app
                        operator: In
                        values:
                          - nginx
                  topologyKey: kubernetes.io/hostname
    
  2. Create a LoadBalancer Service, for example, using an existing load balancer. The following shows an example YAML file for configuring source IP address-based sticky sessions:

    apiVersion: v1
    kind: Service
    metadata:
      name: svc-example
      namespace: default
      annotations:
        kubernetes.io/elb.class: union
        kubernetes.io/elb.id: *****
        kubernetes.io/elb.lb-algorithm: ROUND_ROBIN      # Weighted round robin allocation policy
        kubernetes.io/elb.session-affinity-mode: SOURCE_IP    # Enable source IP address-based sticky session.
    spec:
      selector:
        app: nginx
      externalTrafficPolicy: Local    # Node level Service affinity
      ports:
        - name: cce-service-0
          targetPort: 80
          nodePort: 32633
          port: 80
          protocol: TCP
      type: LoadBalancer
    
  3. Log in to the ELB console and click the target load balancer. In the backend server group of the listener, check whether sticky session is enabled.

Layer 7 Sticky Sessions for Ingresses

In Layer 7 mode, sticky sessions can be enabled using HTTP cookies or application cookies.

Enabling Layer 7 Sticky Session in a CCE Standard Cluster

To enable cookie-based sticky session on an ingress, ensure the following conditions are met:

  1. Service Affinity of the ingress must be set to Node-level, where the externalTrafficPolicy value of the Service must be Local.

  2. Anti-affinity must be enabled for the ingress workload to prevent all pods from being deployed on the same node.

Procedure

  1. Create an Nginx workload.

    Set the number of pods to 3 and configure podAntiAffinity.

    kind: Deployment
    apiVersion: apps/v1
    metadata:
      name: nginx
      namespace: default
    spec:
      replicas: 3
      selector:
        matchLabels:
          app: nginx
      template:
        metadata:
          labels:
            app: nginx
        spec:
          containers:
            - name: container-0
              image: 'nginx:perl'
              resources:
                limits:
                  cpu: 250m
                  memory: 512Mi
                requests:
                  cpu: 250m
                  memory: 512Mi
          imagePullSecrets:
            - name: default-secret
          affinity:
            podAntiAffinity:                   # Pod anti-affinity
              requiredDuringSchedulingIgnoredDuringExecution:
                - labelSelector:
                    matchExpressions:
                      - key: app
                        operator: In
                        values:
                          - nginx
                  topologyKey: kubernetes.io/hostname
    
  2. Create a Service for the workload. This section uses a NodePort Service as an example.

    Configure sticky sessions during the creation of a Service. An ingress can access multiple Services, and each Service can have different sticky sessions.

    apiVersion: v1
    kind: Service
    metadata:
      name: nginx
      namespace: default
      annotations:
        kubernetes.io/elb.lb-algorithm: ROUND_ROBIN      # Weighted round robin allocation policy
        kubernetes.io/elb.session-affinity-mode: HTTP_COOKIE      # HTTP cookie
        kubernetes.io/elb.session-affinity-option: '{"persistence_timeout":"1440"}'   # Session stickiness duration, in minutes. The value ranges from 1 to 1440.
    spec:
      selector:
        app: nginx
      ports:
        - name: cce-service-0
          protocol: TCP
          port: 80
          targetPort: 80
          nodePort: 32633            # Custom node port
      type: NodePort
      externalTrafficPolicy: Local   # Node level Service affinity
    

    You can also select APP_COOKIE.

    apiVersion: v1
    kind: Service
    metadata:
      name: nginx
      namespace: default
      annotations:
        kubernetes.io/elb.lb-algorithm: ROUND_ROBIN      # Weighted round robin allocation policy
        kubernetes.io/elb.session-affinity-mode: APP_COOKIE     # Select APP_COOKIE.
        kubernetes.io/elb.session-affinity-option: '{"app_cookie_name":"test"}'  # Application cookie name
    ...
    
  3. Create an ingress and associate it with the Service.

    apiVersion: networking.k8s.io/v1
    kind: Ingress
    metadata:
      name: ingress-test
      namespace: default
      annotations:
        kubernetes.io/elb.class: union
        kubernetes.io/elb.port: '80'
        kubernetes.io/elb.id: *****
    spec:
      rules:
      - host: 'www.example.com'
        http:
          paths:
          - path: '/'
            backend:
              service:
                name: nginx     # Service name
                port:
                  number: 80
            property:
              ingress.beta.kubernetes.io/url-match-mode: STARTS_WITH
            pathType: ImplementationSpecific
      ingressClassName: cce
    
  4. Log in to the ELB console and click the target load balancer. In the backend server group of the listener, check whether sticky session is enabled.