8000 nvidia-device-plugin doesn't schedule plugin Pods to workers · Issue #20 · NVIDIA/nvkind · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

nvidia-device-plugin doesn't schedule plugin Pods to workers #20

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
sfc-gh-lshi opened this issue Dec 12, 2024 · 2 comments · May be fixed by #36
Open

nvidia-device-plugin doesn't schedule plugin Pods to workers #20

sfc-gh-lshi opened this issue Dec 12, 2024 · 2 comments · May be fixed by #36

Comments

@sfc-gh-lshi
Copy link

With the currently documented steps, nvidia-device-plugin does not schedule the plugin DaemonSet to any of the cluster workers. This is because the worker nodes are missing required labels.

After installing the Helm chart, the worker(s) need to be labeled thusly:

kubectl label node "${KIND_CLUSTER_NAME}-worker" --overwrite nvidia.com/gpu.present=true

Only afterwards does kubectl --context=kind-${KIND_CLUSTER_NAME} get pod -n nvidia show any Pods.

@LogExE
Copy link
LogExE commented Dec 12, 2024

It looks like we have to pass --set gfd.enabled=true to helm install command

@jbdalido
Copy link

It looks like we have to pass --set gfd.enabled=true to helm install command

+1 @LogExE maybe open a PR to update the yaml?

@jacobtomlinson jacobtomlinson linked a pull request Mar 17, 2025 that will close this issue
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants
0