Enable compute isolation
Kf Apps can be deployed on dedicated nodes in the cluster. This feature is required if you have the circumstances where you might want more control on a node where an App Pod lands. For example:
- If you are sharing the same cluster for different Apps but want dedicated nodes for a particular App.
- If you want dedicated nodes for a given organization (Kf Space).
- If you want to target a specific operating system like Windows.
- If you want to co-locate Pods from two different services that frequently communicate.
To enable compute isolation, Kf uses the Kubernetes nodeSelector. To use this feature, first add labels on the nodes or node pools where you want your App Pods to land and then add the same qualifying labels on the Kf Space. All the Apps installed in this Space then land on the nodes with matching labels.
Kf creates a Kubernetes pod to execute each Kf Build, the buildNodeSelector feature can be used to isolate compute resources to execute only the Build pods. One use case is to isolate Build pods to run on nodes with SSD, while running the App pods on other nodes. The BuildNodeSelectors feature provides compute resource optimization and flexibility in the cluster. Please refer to chapter ‘Configure BuildNodeSelectors and a build node pool’ on this page.
Configure nodeSelector in a Kf cluster
By default, compute isolation is disabled. Use the following procedure to configure labels and nodeSelector.
Add a label (
distype=ssd
) on the node where you want your application pods to land.kubectl label nodes nodeid disktype=ssd
Add the same label on the Kf Space. All Apps deployed in this Space will then land on the qualifying nodes.
kf configure-space set-nodeselector space-name disktype ssd
You can add multiple labels by running the same command again.
Check the label is configured.
kf configure-space get-nodeselector space-name
Delete the label from the space.
kf configure-space unset-nodeselector space-name disktype
Override nodeSelector for Kf stacks
Deployment of Kf Apps can be further targeted based
on what stack is being used to build and package the App. For
example, if you want your applications built with spaceStacksV2
to land on
nodes with Linux kernel 4.4.1., nodeSelector
values on a stack override the
values configured on the Space.
To configure the nodeSelector
on a stack:
Edit the
config-defaults
of your Kf cluster and add the labels.$ kubectl -n kf edit configmaps config-defaults
Add
nodeSelector
to the stacks definition...... ..... spaceStacksV2: | - name: cflinuxfs3 image: cloudfoundry/cflinuxfs3 nodeSelector: OS_KERNEL: LINUX_4.4.1 ..... .....
Configure BuildNodeSelectors and a Build node pool
Build node selectors are only effective at overriding the node selectors for the Build pods, they do not affect App pods. For example, if you specify both the node selectors on the Space and the Build node selectors in Kfsystem, App pods will have the Space node selectors while the Build pods will have the Build node selectors from Kfsystem; if node selectors are only specified in the Space, both the App and Build pods will have the node selector from the Space.
Add labels (
disktype:ssd
for example) on the nodes that you want your Build pods to be assigned to.kubectl label nodes nodeid disktype=ssd
Add/update Build node selectors (in the format of
key:value
pairs) by patching KfSystem CR.kubectl patch kfsystem kfsystem --type='json' -p='[{'op': 'replace', 'path': '/spec/kf/config/buildNodeSelectors', 'value': {<key>:<value>}}]'
For example, to add
disktype=ssd
as the Build node selector:kubectl patch kfsystem kfsystem --type='json' -p='[{'op': 'replace', 'path': '/spec/kf/config/buildNodeSelectors', 'value': {"disktype":"ssd"}}]'