<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Containerized AI Execution on</title><link>https://docs.opennebula.io/7.2/solutions/ai_factory_blueprints/containerized_ai_execution/</link><description>Recent content in Containerized AI Execution on</description><generator>Hugo</generator><language>en</language><atom:link href="https://docs.opennebula.io/7.2/solutions/ai_factory_blueprints/containerized_ai_execution/index.xml" rel="self" type="application/rss+xml"/><item><title>Deployment of AI-ready Kubernetes</title><link>https://docs.opennebula.io/7.2/solutions/ai_factory_blueprints/containerized_ai_execution/ai_ready_k8s/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.opennebula.io/7.2/solutions/ai_factory_blueprints/containerized_ai_execution/ai_ready_k8s/</guid><description>&lt;p&gt;&lt;a id="ai_ready_k8s"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;Tools like Kubernetes provide robust orchestration for deploying AI workloads at scale, being able to manage isolation between cluster workloads and GPU resources for AI inference tasks. With the use of NVIDIA GPU Operator, you perform the provision of the necessary NVIDIA drivers and libraries for making GPU resources available to containers.
Kubernetes embraces multitenancy, supporting different isolated namespaces where the access from different teams or users are managed with Role Based Access Control (RBAC) and network policies. As an administrator, you can also enforce limits on the GPU usage or other resources consumed per namespace, ensuring fair resource allocation.&lt;/p&gt;</description></item><item><title>Deployment of NVIDIA Dynamo</title><link>https://docs.opennebula.io/7.2/solutions/ai_factory_blueprints/containerized_ai_execution/nvidia_dynamo/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.opennebula.io/7.2/solutions/ai_factory_blueprints/containerized_ai_execution/nvidia_dynamo/</guid><description>&lt;p&gt;&lt;a id="nvidia_dynamo"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;a href="https://docs.nvidia.com/dynamo/latest/index.html"&gt;NVIDIA® Dynamo&lt;/a&gt; is a high-performant inference framework for serving AI models in an agnostic way, such as across any framework, architecture or deployment scale, as well as in multi-node distributed environments. Being an agnostic inference engine, it supports different backends like TRT-LLM, vLLM, SGLang, etc. Dynamo also allows you to declare inference graphs which deploy different containerized components in a disaggregated way- like an API frontend, a prefill worker, a decode worker, a K/V cache, and others - and to let them interact to efficiently respond to the user queries.&lt;/p&gt;</description></item><item><title>Deployment of the NVIDIA KAI Scheduler</title><link>https://docs.opennebula.io/7.2/solutions/ai_factory_blueprints/containerized_ai_execution/nvidia_kai_scheduler/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.opennebula.io/7.2/solutions/ai_factory_blueprints/containerized_ai_execution/nvidia_kai_scheduler/</guid><description>&lt;p&gt;&lt;a id="nvidia_kai_scheduler"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;The &lt;a href="https://github.com/NVIDIA/KAI-Scheduler"&gt;NVIDIA® KAI Scheduler&lt;/a&gt; is an open source Kubernetes-native scheduler designed to optimize GPU resource allocation for AI and machine learning workloads at scale. It is capable of managing large GPU clusters and handling high-throughput demanding workload environments. KAI Scheduler targets both interactive jobs and large-scale training or inference tasks within the same cluster, orchestrating available resources across different users and teams. It also operates alongside other schedulers installed in a cluster.&lt;/p&gt;</description></item></channel></rss>