- A supported SRIOV hardware on the cluster nodes. Supported models can be found here.
- Kubernetes or Openshift cluster running on bare metal nodes.
- Multus-cni is deployed as default CNI plugin, and there is a default CNI plugin (flannel, openshift-sdn etc.) available for Multus-cni.
- On RedHat Enterprise Linux and Ubuntu operating systems, the
rdma-core
package must be installed to support RDMA resource provisioning. On RedHat CoreOS the package installation is not required.
Make sure to have installed the Operator-SDK, as shown in its install documentation, and that the binaries are available in your $PATH.
Clone this GitHub repository.
go get github.com/k8snetworkplumbingwg/sriov-network-operator
Deploy the operator.
If you are running an Openshift cluster:
make deploy-setup
If you are running a Kubernetes cluster:
make deploy-setup-k8s
Webhooks are disabled when deploying on a Kubernetes cluster as per the instructions above. To enable webhooks on Kubernetes cluster, there are two options:
-
Create certificates for each of the two webhooks using a single CA whose cert you provide through an environment variable.
For example, given
cacert.pem
,key.pem
andcert.pem
:kubectl create ns sriov-network-operator kubectl -n sriov-network-operator create secret tls operator-webhook-cert --cert=cert.pem --key=key.pem kubectl -n sriov-network-operator create secret tls network-resources-injector-cert --cert=cert.pem --key=key.pem export ADMISSION_CONTROLLERS_ENABLED=true export ADMISSION_CONTROLLERS_CERTIFICATES_OPERATOR_CA_CRT=$(base64 -w 0 < cacert.pem) export ADMISSION_CONTROLLERS_CERTIFICATES_INJECTOR_CA_CRT=$(base64 -w 0 < cacert.pem) make deploy-setup-k8s
-
Using https://cert-manager.io/, deploy it as:
kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.3.0/cert-manager.yaml
Define the appropriate Issuer and Certificates, as an example:
kubectl create ns sriov-network-operator cat <<EOF | kubectl apply -f - apiVersion: cert-manager.io/v1 kind: Issuer metadata: name: sriov-network-operator-selfsigned-issuer namespace: sriov-network-operator spec: selfSigned: {} --- apiVersion: cert-manager.io/v1 kind: Certificate metadata: name: operator-webhook-cert namespace: sriov-network-operator spec: secretName: operator-webhook-cert dnsNames: - operator-webhook-service.sriov-network-operator.svc issuerRef: name: sriov-network-operator-selfsigned-issuer --- apiVersion: cert-manager.io/v1 kind: Certificate metadata: name: network-resources-injector-cert namespace: sriov-network-operator spec: secretName: network-resources-injector-cert dnsNames: - network-resources-injector-service.sriov-network-operator.svc issuerRef: name: sriov-network-operator-selfsigned-issuer EOF
And then deploy the operator:
export ADMISSION_CONTROLLERS_ENABLED=true export ADMISSION_CONTROLLERS_CERTIFICATES_CERT_MANAGER_ENABLED=true make deploy-setup-k8s
By default, the operator will be deployed in namespace 'sriov-network-operator' for Kubernetes cluster, you can check if the deployment is finished successfully.
$ kubectl get -n sriov-network-operator all
NAME READY STATUS RESTARTS AGE
pod/sriov-network-config-daemon-bf9nt 1/1 Running 0 8s
pod/sriov-network-operator-54d7545f65-296gb 1/1 Running 0 10s
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
service/sriov-network-operator ClusterIP 10.102.53.223 <none> 8383/TCP 9s
NAME DESIRED CURRENT READY UP-TO-DATE AVAILABLE NODE SELECTOR AGE
daemonset.apps/sriov-network-config-daemon 1 1 1 1 1 kubernetes.io/os=linux,node-role.kubernetes.io/worker= 8s
NAME READY UP-TO-DATE AVAILABLE AGE
deployment.apps/sriov-network-operator 1/1 1 1 10s
NAME DESIRED CURRENT READY AGE
replicaset.apps/sriov-network-operator-54d7545f65 1 1 1 10s
You may need to label SR-IOV worker nodes using node-role.kubernetes.io/worker
label, if not already.
Note: By default, SR-IOV Operator will be deployed in namespace 'openshift-sriov-network-operator' for OpenShift cluster.
After the operator gets installed, you can configure it with creating the custom resource of SriovNetwork and SriovNetworkNodePolicy. But before that, you may want to check the status of SriovNetworkNodeState CRs to find out all the SRIOV capable devices in you cluster.
Here comes an example. As you can see, there are 2 SR-IOV NICs from Intel.
$ kubectl get sriovnetworknodestates.sriovnetwork.openshift.io -n sriov-network-operator node-1 -o yaml
apiVersion: sriovnetwork.openshift.io/v1
kind: SriovNetworkNodeState
spec: ...
status:
interfaces:
- deviceID: "1572"
driver: i40e
mtu: 1500
pciAddress: "0000:18:00.0"
totalvfs: 64
vendor: "8086"
- deviceID: "1572"
driver: i40e
mtu: 1500
pciAddress: "0000:18:00.1"
totalvfs: 64
vendor: "8086"
You can choose the NIC you want when creating SriovNetworkNodePolicy CR, by specifying the 'nicSelector'.
apiVersion: sriovnetwork.openshift.io/v1
kind: SriovNetworkNodePolicy
metadata:
name: policy-1
namespace: sriov-network-operator
spec:
nodeSelector:
feature.node.kubernetes.io/network-sriov.capable: "true"
resourceName: intelnics
priority: 99
mtu: 9000
numVfs: 2
nicSelector:
deviceID: "1572"
rootDevices:
- 0000:18:00.1
vendor: "8086"
deviceType: netdevice
After applying your SriovNetworkNodePolicy CR, check the status of SriovNetworkNodeState again, you should be able to see the NIC has been configured as instructed.
$ kubectl get sriovnetworknodestates.sriovnetwork.openshift.io -n sriov-network-operator node-1 -o yaml
...
- Vfs:
- deviceID: 1572
driver: iavf
pciAddress: 0000:18:02.0
vendor: "8086"
- deviceID: 1572
driver: iavf
pciAddress: 0000:18:02.1
vendor: "8086"
- deviceID: 1572
driver: iavf
pciAddress: 0000:18:02.2
vendor: "8086"
deviceID: "1583"
driver: i40e
mtu: 1500
numVfs: 3
pciAddress: 0000:18:00.0
totalvfs: 64
vendor: "8086"
...
At the same time, the SRIOV device plugin and CNI plugin has been provisioned to the worker node. You may check if resource name 'intel-nics' is reported by device plugin correctly.
$ kubectl get no -o json | jq -r '[.items[] | {name:.metadata.name, allocable:.status.allocatable}]'
[
{
"name": "minikube",
"allocable": {
"cpu": "72",
"ephemeral-storage": "965895780801",
"hugepages-1Gi": "0",
"hugepages-2Mi": "0",
"intel.com/intel-nics": "3",
"memory": "196706684Ki",
"openshift.io/sriov": "0",
"pods": "110"
}
}
]
Now you can create a SriovNetwork CR which refer to the 'resourceName' defined in SriovNetworkNodePolicy. Then a NetworkAttachmentDefinition CR will be generated by operator with the same name and namespace.
Here is an example:
apiVersion: sriovnetwork.openshift.io/v1
kind: SriovNetwork
metadata:
name: example-sriovnetwork
namespace: sriov-network-operator
spec:
ipam: |
{
"type": "host-local",
"subnet": "10.56.217.0/24",
"rangeStart": "10.56.217.171",
"rangeEnd": "10.56.217.181",
"routes": [{
"dst": "0.0.0.0/0"
}],
"gateway": "10.56.217.1"
}
vlan: 0
resourceName: intelnics
To remove the operator related resources.
If you are running an Openshift cluster:
make undeploy
If you are running a Kubernetes cluster:
make undeploy-k8s