BT

Facilitating the Spread of Knowledge and Innovation in Professional Software Development

Write for InfoQ

Topics

Choose your language

InfoQ Homepage News CNCF Warns Kubernetes Alone Is Not Enough to Secure LLM Workloads

CNCF Warns Kubernetes Alone Is Not Enough to Secure LLM Workloads

Listen to this article -  0:00

A new blog from the Cloud Native Computing Foundation highlights a critical gap in how organizations are deploying large language models (LLMs) on Kubernetes: while Kubernetes excels at orchestrating and isolating workloads, it does not inherently understand or control the behavior of AI systems, creating a fundamentally different and more complex threat model.

The article argues that LLMs introduce a new class of risk because they operate on untrusted input and can dynamically decide actions, unlike traditional applications. In a typical deployment, such as exposing an LLM via an API or chat interface, Kubernetes ensures that pods are running and resources are stable, but it has no visibility into whether prompts are malicious, whether sensitive data is being exposed, or whether the model is interacting with internal systems in unsafe ways. This creates a scenario where infrastructure appears healthy while underlying risks go undetected.

The CNCF emphasizes that LLM-based systems must be treated as programmable, decision-making entities, not just compute workloads. By placing an LLM in front of internal tools, logs, APIs, or credentials, organizations are effectively introducing a new layer of abstraction that can be influenced through prompt input. This opens the door to risks such as prompt injection, unintended data exposure, and misuse of connected tools, threats that traditional Kubernetes security controls were not designed to handle.

This shift reflects a broader evolution in cloud-native systems, where Kubernetes is increasingly used to run AI and generative workloads. As adoption grows, the platform is being stretched beyond its original purpose of managing stateless microservices into orchestrating data-intensive, agent-driven, and inference-heavy systems. However, the security model has not fully caught up with these new use cases.

While Kubernetes provides strong primitives for scheduling, isolation, and resource management, it lacks built-in mechanisms to enforce application-level or semantic controls over AI systems. For example, it cannot determine whether a prompt should be executed, whether a response leaks sensitive information, or whether an LLM should have access to certain tools or APIs.

This limitation highlights the need for additional layers of control beyond infrastructure. Traditional Kubernetes security practices, such as RBAC, network policies, and container isolation, remain necessary but are insufficient on their own. Instead, organizations must consider AI-specific controls, including prompt validation, output filtering, tool access restrictions, and policy enforcement at the application layer.

The blog points to an emerging need for AI-aware platform engineering, where security is embedded across both infrastructure and application layers. This includes integrating frameworks such as the OWASP Top 10 for LLMs, applying policy-as-code, and introducing guardrails that govern how models interact with data and external systems.

Industry discussions increasingly frame this as a shift from traditional threat models to behavioral and context-aware security models, where the focus is not just on protecting infrastructure, but on controlling how intelligent systems behave within it. As LLMs evolve into autonomous or agentic systems capable of executing actions, these concerns become even more critical.

The CNCF's analysis serves as a warning for organizations rapidly adopting AI on Kubernetes: operational health does not equal security. A system can be fully compliant with Kubernetes best practices while still exposing significant risks through its AI layer.

Major technology and security vendors are converging on similar principles. Industry guidance increasingly recommends a multi-layered security model, combining runtime monitoring, human-in-the-loop controls, and strict policy enforcement around what AI systems are allowed to do. A consistent theme is that LLMs should never be treated as authoritative decision-makers: instead, they must operate within bounded contexts with explicit guardrails, continuous validation, and auditability.

As LLM adoption accelerates, the industry is being pushed to rethink long-standing assumptions about trust boundaries, workload isolation, and application behavior. The result is a new security paradigm, one where Kubernetes remains a foundational layer, but must be complemented by AI-specific governance, observability, and control mechanisms to ensure safe and reliable deployment of intelligent systems.

About the Author

Rate this Article

Adoption
Style

BT