Skip to main content
Version: v0.5

Capacity-based Scheduling

Overview

Using the policy-based scheduling mechanism, Nova provides for the capability to schedule a workload based on resource availability in a workload cluster. This means that a single Kubernetes resource or a group of Kubernetes resources will be placed on a target cluster that has sufficient capacity to host it.

Group Scheduling example

In this example we will see how Nova groups k8s objects into a ScheduleGroup and finds a workload cluster for the whole group. Let's say you have a group of microservices, which combine into an application. We will try to create two versions of the same application: microservices labeled color: blue and the same set of microservices labeled color: green. By adding .groupBy.labelKey to the SchedulePolicy spec, Nova will create two ScheduleGroups: one with all objects with color: blue and another one with color: green label.

apiVersion: policy.elotl.co/v1alpha1
kind: SchedulePolicy
metadata:
...
spec:
groupBy:
labelKey: color
...

Each group will be considered separately by Nova, and it is guaranteed that all objects in the group will run in the same workload cluster. In this tutorial, we will let Nova figure out which workload cluster has enough resources to host each group. This can be done by not setting .spec.clusterSelector. We will use GCP Microservice Demo App which includes 10 different microservices. Total resources requested in this app is 1570 millicores of CPU and 1368 Mi of memory.

Let's start with creating a namespace that we will use:

  1. kubectl --context=kind-workload-1 create namespace microsvc-demo

  2. kubectl --context=kind-workload-2 create namespace microsvc-demo

  3. KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl create namespace microsvc-demo

  4. KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl apply -f sample-group-scheduling/policy.yaml This policy is saying, for any objects with label microServicesDemo: "yes", group them based on the value of the "color" label and schedule a group to any cluster which has enough resources to run them.

  5. Now, let's create green and blue instances of our app:

    KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl apply -f sample-group-scheduling/blue-app.yaml -n microsvc-demo
    KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl apply -f sample-group-scheduling/green-app.yaml -n microsvc-demo
  6. Verifying whether the objects were assigned to the correct ScheduleGroup can be done by describing an object and looking at events:

    $ kubectl describe deployment frontend -n microsvc-demo
    Name: frontend
    Namespace: microsvc-demo
    CreationTimestamp: Wed, 01 Feb 2023 15:42:31 +0100
    Labels: color=blue
    microServicesDemo=yes
    nova.elotl.co/target-cluster=kind-workload-1
    ...
    Events:
    Type Reason Age From Message
    ---- ------ ---- ---- -------
    Normal AddedToScheduleGroup 17s nova-scheduler added to ScheduleGroup demo-policy-4f068569 which contains objects with groupBy.labelKey color=blue
    Normal SchedulePolicyMatched 17s nova-scheduler schedule policy demo-policy will be used to determine target cluster

  7. You can check if two ScheduleGroups were created: KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl get schedulegroups

    NAME                   AGE
    demo-policy-4f068569 9s
    demo-policy-f73297b2 4s
  8. novactl CLI provides a bit more context about schedulegroups: KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl nova get schedulegroups

     NAME                 NOVA WORKLOAD CLUSTER                   NOVA POLICY NAME    
    ------------------ -------------------------------------- --------------------------------------

    demo-policy-4f068569 kind-workload-1 demo-policy

    demo-policy-f73297b2 kind-workload-2 demo-policy
    ------------------ -------------------------------------- --------------------------------------
  9. From the output above, we can see which workload cluster is hosting each ScheduleGroup.

  10. Now, imagine you need to increase resource request or replica count on one of the microservices in the second app. In the meantime, there was other activity in the cluster and after your update there won't be enough resources in the cluster to satisfy your update. You can simulate this scenario using sample-group-scheduling/hog-pod.yaml manifest. You should edit it, so the hog-pod will take almost all resources in your cluster. Now, you can apply it to the same cluster where demo-policy-f73297b2 schedule group was scheduled (kind-workload-2 in my case). KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl apply -f sample-group-scheduling/hog-pod.yaml

  11. Now let's increase replica count in frontend-2 microservice (which is one of the microservices in green app) in the Nova control plane: KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl scale deploy/frontend-2 --replicas=5 -n microsvc-demo

  12. If there is enough resources to satisfy new schedule group requirements (existing resource request for 9 microservices + increased replica count of frontend-2), watching schedule group will show you schedule group being rescheduled to another cluster: KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl nova get schedulegroups

     NAME                 NOVA WORKLOAD CLUSTER                   NOVA POLICY NAME    
    ------------------ -------------------------------------- --------------------------------------

    demo-policy-4f068569 kind-workload-1 demo-policy

    demo-policy-f73297b2 kind-workload-1 demo-policy
    ------------------ -------------------------------------- --------------------------------------
  13. To understand why the ScheduleGroup was rescheduled, we can use kubectl describe schedulegroup <group-name> and see the event message:

        Name:                demo-policy-f73297b2 
    Labels: color=green
    nova.elotl.co/matching-policy=demo-policy
    nova.elotl.co/target-cluster=kind-workload-1
    API Version: policy.elotl.co/v1alpha1
    Kind: ScheduleGroup
    ...
    Type Reason Age From Message
    ---- ------ ---- ---- -------
    Warning ReschedulingTriggered 106s nova-agent deployment microsvc-demo/frontend-2 does not have minimum replicas available

  14. You can verify that green app is running by listing deployment in Nova Control Plane: KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl get deployments -n microsvc-demo -l color=green

    NAME                      READY   UP-TO-DATE   AVAILABLE   AGE
    adservice-2 1/1 1 1 2m1s
    cartservice-2 1/1 1 1 2m2s
    checkoutservice-2 1/1 1 1 2m3s
    currencyservice-2 1/1 1 1 2m2s
    emailservice-2 1/1 1 1 2m3s
    frontend-2 5/5 5 5 2m3s
    loadgenerator-2 1/1 1 1 2m2s
    paymentservice-2 1/1 1 1 2m3s
    productcatalogservice-2 1/1 1 1 2m2s
    recommendationservice-2 1/1 1 1 2m3s
    redis-cart-2 1/1 1 1 2m1s
    shippingservice-2 1/1 1 1 2m2s
  15. To remove all objects created for this demo, remove microsvc-demo namespace: KUBECONFIG=./nova-installer-output/nova-kubeconfig kubectl delete ns microsvc-demo