NVIDIA / k8s-dra-driver

Dynamic Resource Allocation (DRA) for NVIDIA GPUs in Kubernetes
Apache License 2.0
226 stars 41 forks source link

Bump github.com/NVIDIA/nvidia-container-toolkit from 1.15.1-0.20240419094620-0aed9a16addf to 1.16.0-rc.2 #140

Closed dependabot[bot] closed 1 month ago

dependabot[bot] commented 2 months ago

Bumps github.com/NVIDIA/nvidia-container-toolkit from 1.15.1-0.20240419094620-0aed9a16addf to 1.16.0-rc.2.

Release notes

Sourced from github.com/NVIDIA/nvidia-container-toolkit's releases.

v1.16.0-rc.2

What's Changed

  • Use relative path to locate driver libraries
  • Add RelativeToRoot function to Driver
  • Inject additional libraries for full X11 functionality
  • Extract options from default runtime if runc does not exist
  • Update libnvidia-container
  • Reduce logging for the NVIDIA Container runtime
  • Add Tracef to logger Interface
  • Add String function to oci.Runtime interface
  • Fix bug in argument parsing for logger creation
  • Use ref_name on release workflow

Changes in the Toolkit Container

  • Extract options from default runtime if runc does not exist
  • avoid using map pointers as maps are always passed by reference

Full Changelog: https://github.com/NVIDIA/nvidia-container-toolkit/compare/v1.16.0-rc.1...v1.16.0-rc.2

v1.16.0-rc.1

What's Changed

  • Support vulkan ICD files directly in a driver root. This allows for the discovery of vulkan files in GKE driver installations.
  • Increase priority of ld.so.conf.d config file injected into container. This ensures that injected libraries are preferred over libraries present in the container.
  • Set default CDI spec permissions to 644. This fixes permission issues when using the nvidia-ctk cdi transform functions.
  • Add dev-root option to nvidia-ctk system create-device-nodes command.
  • Fix location of libnvidia-ml.so.1 when a non-standard driver root is used. This enabled CDI spec generation when using the driver container on a host.
  • Recalculate minimum required CDI spec version on save.
  • Move nvidia-ctk hook commands to a separate nvidia-cdi-hook binary. The same subcommands are supported.
  • Use : as an nvidia-ctk config --set list separator. This fixes a bug when trying to set config options that are lists.

Changes in the Toolkit Container

  • Bump CUDA base image version to 12.5.0
  • Allow the path to toolkit.pid to be specified directly.
  • Remove provenance information from image manifests.
  • Add dev-root option when configuring the toolkit. This adds support for GKE driver installations.

Full Changelog: https://github.com/NVIDIA/nvidia-container-toolkit/compare/v1.15.0...v1.16.0-rc.1

Changelog

Sourced from github.com/NVIDIA/nvidia-container-toolkit's changelog.

v1.16.0-rc.2

  • Use relative path to locate driver libraries
  • Add RelativeToRoot function to Driver
  • Inject additional libraries for full X11 functionality
  • Extract options from default runtime if runc does not exist
  • Avoid using map pointers as maps are always passed by reference
  • Reduce logging for the NVIDIA Container runtime
  • Fix bug in argument parsing for logger creation

v1.16.0-rc.1

  • Support vulkan ICD files directly in a driver root. This allows for the discovery of vulkan files in GKE driver installations.

  • Increase priority of ld.so.conf.d config file injected into container. This ensures that injected libraries are preferred over libraries present in the container.

  • Set default CDI spec permissions to 644. This fixes permission issues when using the nvidia-ctk cdi transform functions.

  • Add dev-root option to nvidia-ctk system create-device-nodes command.

  • Fix location of libnvidia-ml.so.1 when a non-standard driver root is used. This enabled CDI spec generation when using the driver container on a host.

  • Recalculate minimum required CDI spec version on save.

  • Move nvidia-ctk hook commands to a separate nvidia-cdi-hook binary. The same subcommands are supported.

  • Use : as an nvidia-ctk config --set list separator. This fixes a bug when trying to set config options that are lists.

  • [toolkit-container] Bump CUDA base image version to 12.5.0

  • [toolkit-container] Allow the path to toolkit.pid to be specified directly.

  • [toolkit-container] Remove provenance information from image manifests.

  • [toolkit-container] Add dev-root option when configuring the toolkit. This adds support for GKE driver installations.

v1.15.0

  • Remove nvidia-container-runtime and nvidia-docker2 packages.

  • Use XDG_DATA_DIRS environment variable when locating config files such as graphics config files.

  • Add support for v0.7.0 Container Device Interface (CDI) specification.

  • Add --config-search-path option to nvidia-ctk cdi generate command. These paths are used when locating driver files such as graphics config files.

  • Use D3DKMTEnumAdapters3 to enumerate adpaters on WSL2 if available.

  • Add support for v1.2.0 OCI Runtime specification.

  • Explicitly set NVIDIA_VISIBLE_DEVICES=void in generated CDI specifications. This prevents the NVIDIA Container Runtime from making additional modifications.

  • [libnvidia-container] Use D3DKMTEnumAdapters3 to enumerate adpaters on WSL2 if available.

  • [toolkit-container] Bump CUDA base image version to 12.4.1

v1.15.0-rc.4

  • Add a --spec-dir option to the nvidia-ctk cdi generate command. This allows specs outside of /etc/cdi and /var/run/cdi to be processed.

  • Add support for extracting device major number from /proc/devices if nvidia is used as a device name over nvidia-frontend.

  • Allow multiple device naming strategies for nvidia-ctk cdi generate command. This allows a single CDI spec to be generated that includes GPUs by index and UUID.

  • Set the default --device-name-strategy for the nvidia-ctk cdi generate command to [index, uuid].

  • Remove libnvidia-container0 jetpack dependency included for legacy Tegra-based systems.

  • Add NVIDIA_VISIBLE_DEVICES=void to generated CDI specifications.

  • [toolkit-container] Remove centos7 image. The ubi8 image can be used on all RPM-based platforms.

  • [toolkit-container] Bump CUDA base image version to 12.3.2

... (truncated)

Commits


Dependabot compatibility score

Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
dependabot[bot] commented 1 month ago

Superseded by #142.