![]() This adds an nfd-hook for the gpu-plugin, which will create labels for the GPUs that can then be used for POD deployment purposes or creation of GPU extended resources which allow then finer grained GPU resource management. The nfd-hook will install to the host system when the intel-gpu-initcontainer is run. It is added into the plugin deployment yaml. Signed-off-by: Ukri Niemimuukko <ukri.niemimuukko@intel.com> |
||
---|---|---|
.. | ||
labeler_test.go | ||
labeler.go | ||
main.go | ||
README.md |
Intel GPU NFD hook
This is the Node Feature Discovery binary hook implementation for the Intel GPUs. The intel-gpu-initcontainer which is built among other images can be placed as part of the gpu-plugin deployment, so that it copies this hook to the host system only in those hosts, in which also gpu-plugin is deployed.
When NFD worker runs this hook, it will add a number of labels to the nodes, which can be used for example to deploy services to nodes with specific GPU types. Selected numeric labels can be turned into kubernetes extended resources by the NFD, allowing for finer grained resource management for GPU-using PODs.
In the NFD deployment, the hook requires /host-sys -folder to have the host /sys -folder content mounted, and /host-dev to have the host /dev/ -folder content mounted. Write access is not necessary.
There is one supported environment variable named GPU_MEMORY_OVERRIDE, which is supposed to hold a numeric value. For systems with GPUs which do not support reading the GPU memory amount, the environment variable memory value is turned into a GPU memory amount label instead of a read value.