Talos Linux and Dynamic Resource Allocation (Beta)

I upgraded my Kubernetes cluster to 1.32 recently and the changelog mentioned that Dynamic Resource Allocation (DRA) had been graduated to Beta. I had been using the Intel GPU Device Plugin to schedule pods with Hardware Device requirements until now. This seemed like a good opportunity to try out and switch to DRA. Surprisingly, this was fairly simple! I initially ran into some speedbumps but now that I figured it out it’s fairly straightforward.

Since I use Talos Linux in my Homelab some of the following steps and requirements are Talos specific.

# Requirements

  • Containerd 1.7+
  • Kubernetes 1.32+
  • Container Device Interface (CDI) enabled and configured
  • DRA feature gates enabled in Kubernetes
  • Deploy DRA capable device driver

Let’s start with some straightforward steps.

# Containerd, Kubernetes and CDI

Update Talos to a minimum version of 1.9.0. Talos 1.9 contains Containerd 2.0 and has support for Kubernetes 1.32. Containerd 2.0 enabled CDI by default, so this fully covers the first three requirements. Almost fully, more on that later.

# Enable DRA Feature

The DRA beta feature is not enabled by default in Kubernetes 1.32. To enable the feature gates in Talos you need to apply the following configuration:

cluster:
  apiServer:
    extraArgs:
      feature-gates: DynamicResourceAllocation=true
      runtime-config: resource.k8s.io/v1beta1=true # Enables the API (CRD)
  controllerManager:
    extraArgs:
      feature-gates: DynamicResourceAllocation=true
  scheduler:
    extraArgs:
      feature-gates: DynamicResourceAllocation=true

machine:
  kubelet:
    extraArgs:
      feature-gates: DynamicResourceAllocation=true

References:

Apply the configuration to all nodes, then execute kubectl get deviceclasses. If the response is No resources found you have successfully enabled DRA!

# Configure CDI in Talos

CDI uses the following default host paths to store configuration for discovered devices (called cdi spec dirs): ["/etc/cdi", "/var/run/cdi"] This is a problem because /etc is read-only in Talos. To use CDI with Talos we need to modify these paths. Add the following Talos configuration:

machine:
  files:
  - path: /etc/cri/conf.d/20-customization.part
    op: create
    content: |
      # Set cdi dirs to /var/ because default locations are not writeable in talos
      [plugins."io.containerd.cri.v1.runtime"]
        cdi_spec_dirs = ["/var/cdi/static", "/var/cdi/dynamic"]      

# Deploy and use Device Driver

Now deploy your device driver with modified config for the cdi spec dirs. I used the Intel GPU Resource Driver Helm chart, which does not contain sufficient configuration options to modify the CDI Spec Dirs, so for now I apply these modifications by hand.

Once you’ve followed these steps you should be able to execute kubectl get resourceslices which should list all hardware devices your DRA enabled resource driver found in the cluster.

All that’s left is claiming the resource slice in a Pod. Similar to persistent volumes, resources can be claimed statically or dynamically. I have multiple of the same resource available in my cluster so I decided to use a dynamic ResourceClaimTemplate. The template creates the resource claim on Pod scheduling, and release the resource claim when the pod is descheduled.

apiVersion: resource.k8s.io/v1beta1
kind: ResourceClaimTemplate
metadata:
  name: i915
spec:
  spec:
    devices:
      requests:
      - name: i915
        deviceClassName: gpu.intel.com

I then requested and allocated the resource to my Pod by specifying the claim in my StatefulSet and mounting the claim to my Pod:

apiVersion: apps/v1
kind: StatefulSet
...
      containers:
        volumeMounts:
        - name: tmp
        resources: 
          claims:
          - name: i915
...
      volumes:
      - emptyDir: {}
        name: tmp
      resourceClaims:
      - name: i915
        resourceClaimTemplateName: i915

Deploy your app and if all goes well you should now be able to access the GPU device from inside the Pod. You can validate that the device is available inside the Pod by checking the following file paths:

  • /dev/dri/renderD128
  • /dev/dri/card0

And there you have it, you should now be able to use DRA and CDI to schedule your Pods with Hardware Device resource access.

References: