-
Notifications
You must be signed in to change notification settings - Fork 82
Requested AI conformance blog changes #757
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
n-boshnakov
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The changes look good to me, just a minor nit.
| ## What is Kubernetes AI Conformance? | ||
|
|
||
| As AI/ML applications become more prevalent, the need for a standardized environment to run them on Kubernetes has become critical. The CNCF's Kubernetes AI Conformance Working Group was established to address this need. It aims to define a clear, verifiable set of requirements that a Kubernetes distribution must meet to be considered "AI Conformant." | ||
| As AI/ML applications become more prevalent and crucial for business, the need for standardized environments [1] to run them has become critical. The CNCF's Kubernetes AI Conformance Working Group was established to address this need. It aims to define a clear, verifiable set of requirements that a Kubernetes distribution must meet to be considered "AI Conformant". In fact, equipped with these requirements CNCF established the [**Certified** Kubernetes AI Conformance Program](https://github.com/cncf/k8s-ai-conformance). |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| As AI/ML applications become more prevalent and crucial for business, the need for standardized environments [1] to run them has become critical. The CNCF's Kubernetes AI Conformance Working Group was established to address this need. It aims to define a clear, verifiable set of requirements that a Kubernetes distribution must meet to be considered "AI Conformant". In fact, equipped with these requirements CNCF established the [**Certified** Kubernetes AI Conformance Program](https://github.com/cncf/k8s-ai-conformance). | |
| As AI/ML applications become more prevalent and crucial for business, the need for standardized environments [1] to run them has become critical. The CNCF's Kubernetes AI Conformance Working Group was established to address this need. It aims to define a clear, verifiable set of requirements that a Kubernetes distribution must meet to be considered "AI Conformant". In fact, equipped with these requirements, CNCF established the [**Certified** Kubernetes AI Conformance Program](https://github.com/cncf/k8s-ai-conformance). |
| * **Workload Execution:** By passing the conformance tests, Gardener proves that it can reliably run sample AI/ML workloads that utilize GPU acceleration, confirming that the entire stack—from the operating system to the Kubernetes control plane—is functioning correctly. | ||
| * **GPU Discovery and Allocation:** Gardener-managed clusters correctly identify available GPUs on worker nodes and make them schedulable resources within Kubernetes. This allows users to simply request, for example, `nvidia.com/gpu` resources in their pod specifications. | ||
| * **Driver and Runtime Integrity:** The conformance verifies that the correct drivers and container runtimes are in place to expose GPUs to containers. Gardener’s managed approach guarantees that these components are correctly installed and versioned. | ||
| * **Workload Execution:** By passing the conformance tests, Gardener proves that it can reliably run sample AI/ML workloads that utilize GPU acceleration, confirming that the entire stack - from the operating system to the Kubernetes control plane - is functioning correctly. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Mega-nit: Let's use en-dashes for breaking the sentence
| * **Workload Execution:** By passing the conformance tests, Gardener proves that it can reliably run sample AI/ML workloads that utilize GPU acceleration, confirming that the entire stack - from the operating system to the Kubernetes control plane - is functioning correctly. | |
| * **Workload Execution:** By passing the conformance tests, Gardener proves that it can reliably run sample AI/ML workloads that utilize GPU acceleration, confirming that the entire stack – from the operating system to the Kubernetes control plane – is functioning correctly. |
| #### Meeting the Conformance Requirements | ||
| This ensures the correct drivers are installed and configured for your GPU nodes. Users no longer have to manually handle driver installations, version mismatches, or kernel module compilations. When you request a worker node with a GPU, the Operator ensures that it is ready for your AI workloads with the necessary drivers, software assets, and libraries, making the powerful hardware directly accessible to your Kubernetes pods. | ||
|
|
||
| [3]: In Apeiro, we enabled the [NVIDIA GPU Operator](./2025-08-25-garden-linux-enabling-ai-workloads-with-nvidia-gpus). Other GPU hardware will be supported in a similar fashion. Our goal is to extend this powerful, hands-off approach to a broader range of hardware accelerators, further strengthening the hardware sovereignty of our users. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think we don't have the blog entry present in our own repository. In any case, as this is the first appearance of the word "Apeiro" on this page, we need to ensure that we link out to the Apeiro website/docs.
| [3]: In Apeiro, we enabled the [NVIDIA GPU Operator](./2025-08-25-garden-linux-enabling-ai-workloads-with-nvidia-gpus). Other GPU hardware will be supported in a similar fashion. Our goal is to extend this powerful, hands-off approach to a broader range of hardware accelerators, further strengthening the hardware sovereignty of our users. | |
| [3]: In Apeiro, we enabled the [NVIDIA GPU Operator](https://documentation.apeirora.eu/blog/2025-08-25-garden-linux-enabling-ai-workloads-with-nvidia-gpus). Other GPU hardware will be supported in a similar fashion. Our goal is to extend this powerful, hands-off approach to a broader range of hardware accelerators, further strengthening the hardware sovereignty of our users. |
|
@vasu1124 You have pull request review open invite, please check |
|
@vasu1124 Hello, Vasu. Could you please take a look at the suggestions? |
|
The Gardener project currently lacks enough active contributors to adequately respond to all PRs.
You can:
/lifecycle stale |
What this PR does / why we need it:
@vasu1124 suggested various changes after this blog was already merged, based on similar changes in another blog published under ApeiroRA. This update shall catch up this blog to the new facts and statements.
Special notes for your reviewer:
@vasu1124 Please take over from here, make modifications or not, and either close or merge this PR as you see fit.