Skip to main content

Your Platform Already Has AI Users

· 7 min read

Your Platform Already Has AI Users

Platform engineering matured around one thesis: reduce cognitive load for developers. The CNCF Platforms White Paper codified the pattern. Internal developer portals, golden paths, self-service infrastructure — all designed so human teams ship faster without drowning in complexity.

AI agents now consume the same infrastructure. They clone repos, call CI/CD APIs, create pull requests, query issue trackers. But they don't read your Confluence wiki or file support tickets. They parse tool schemas and call structured APIs.

The bottlenecks shift. Humans hit cognitive overload: too many tools, tribal knowledge, context switching. Agents hit context window limits: too many tokens, irrelevant information, unstructured access. Same problem class, different constraint.

The CNCF platform engineering model transfers directly. Here's the mapping, dimension by dimension.

Two Consumers, One Platform

The CNCF Platforms White Paper defines platform engineering across dimensions like self-service, golden paths, documentation, and guardrails. Each dimension has a direct analog for AI agents.

DimensionHuman TeamsAI Agents
Self-service goalReduce cognitive load, eliminate ticketsReduce context window waste, eliminate ambiguity
InterfacePortal, CLI, UI, docsMCP tools, APIs, structured schemas, tool descriptions
Golden pathsOpinionated templates, scaffolding, starter kitsSkills, agent prompts, workflow definitions, pre-built tool chains
DiscoveryService catalog, search, docs siteTool registries, ToolSearch, capability advertisement
DocumentationHow-tos, runbooks, READMEsStructured context (CLAUDE.md), tool schemas, few-shot examples in descriptions
Cognitive loadToo many choices, context switching, tribal knowledgeContext window limits, token budgets, irrelevant context noise
GuardrailsRBAC, policies, compliance gatesSandboxing, permission boundaries, tool allowlists, destructive action blocks
ToilManual deploys, ticket queues, config driftRe-discovering context, re-reading files, redundant searches, hallucinated paths
OnboardingOrientation docs, buddy system, ramp-up periodSystem prompts, CLAUDE.md, agent norms, skill loading
CollaborationSlack, PRs, pair programmingAgent channels, MCP messaging, shared memory stores, handoff protocols
ScalingHire more people, cross-trainSpawn more agents, share skills, parallelize
Failure modeBurnout, knowledge silos, bus factorContext overflow, stale memory, tool misuse, cascading bad decisions
Cost modelSalaries, time, opportunity costTokens, API calls, compute time, context window budget

The mapping isn't forced. These dimensions emerged because platform engineering solves a universal problem: making infrastructure accessible to its consumers. The consumer type changes the interface, not the architecture.

Where It Gets Concrete

Three dimensions show the mapping most clearly.

Self-service: from portals to tool registries

For humans, self-service means a portal where you click "Create Kubernetes Cluster" and get one in minutes. No tickets, no waiting.

For agents, self-service means discoverable tools with structured schemas. An agent calls ToolSearch("kubernetes cluster") and gets back a create_cluster function with typed parameters. No parsing documentation, no guessing at CLI flags.

The same backend serves both. The portal and the tool schema are two interfaces to the same provisioning API.

Golden paths: from templates to skills

Human golden paths are opinionated templates. create-react-app scaffolds a project. A Backstage template wires up CI/CD, monitoring, and deployment in one click.

Agent golden paths are skills — packaged workflows with clear inputs and outputs. A deploy-service skill encapsulates the same decisions the template made: which registry, which pipeline, which environment. The agent composes skills instead of following a wizard.

Documentation: from READMEs to structured context

Humans read READMEs, runbooks, and how-to guides. The documentation is narrative, written for comprehension.

Agents consume structured context: CLAUDE.md files with project conventions, tool schemas with parameter descriptions, few-shot examples embedded in tool descriptions. Narrative documentation wastes context window. Structured context is information-dense.

A platform that serves both invests in both formats — or better, generates one from the other.

Three Dimensions the CNCF Model Missed

The original model doesn't cover three concerns that matter only (or primarily) for AI agents.

Context management. Humans manage context through experience. An SRE who's been on-call for a year knows which alerts matter. That context is implicit and essentially free.

Agents have explicit context budgets. Every file read, every search result, every system prompt consumes tokens from a finite window. A platform at the "Provisional" level dumps entire files into the prompt. At "Optimizing," agents request exactly what they need and the platform delivers it pre-filtered.

Context management is the single highest-leverage investment for an AI agent platform. Bad context means wasted tokens, slower responses, and worse decisions.

Memory. Humans accumulate knowledge over time. Institutional memory lives in people's heads, in Slack history, in wikis.

Agents start fresh every session unless the platform provides memory infrastructure. At the "Provisional" level, there's no persistence — every session rediscovers the same project structure. At "Operational," sessions get annotated and handed off manually. At "Optimizing," agents extract insights automatically and share them across a federated knowledge store.

Authority. Human trust is earned through track record, code review, and organizational culture. A senior engineer merges without approval. A junior gets two reviewers.

Agent authority requires explicit boundaries. Permission systems define which tools an agent can call, which resources it can modify, which actions require human approval. At maturity, these boundaries become dynamic — an agent with a strong track record gets expanded permissions automatically.

The Combined Maturity Model

The CNCF Platform Engineering Maturity Model defines four levels: Provisional, Operational, Scalable, and Optimizing. Each level applies to both human and AI platform consumers. The table below shows the human level (top row) and AI agent equivalent (bottom row) for each dimension.

DimensionProvisionalOperationalScalableOptimizing
InvestmentHumanVoluntary, temporaryDedicated teamPlatform as productEnabled ecosystem
AIAd-hoc API keys, manual setupDedicated agent infra, budgeted tokensAgent platform as product, ROI per taskSelf-provisioning agent ecosystem
AdoptionHumanErraticExtrinsic pushIntrinsic pullParticipatory
AIHardcoded tool listsTool registries, skill loadingSelf-discovery, capability advertisementAgents contribute tools back
InterfacesHumanCustom processesStandard toolingSelf-serviceIntegrated services
AIRaw CLI, unstructured promptsMCP tools, structured schemasSelf-service composition, skill chainingMulti-agent orchestration
OperationsHumanBy requestCentrally trackedCentrally enabledManaged services
AIOne-off scriptsCentrally managed skillsAutomated curation, health checksAgents maintain own platform
MeasurementHumanAd hocConsistent collectionInsightsQuantitative and qualitative
AIManual log reviewSuccess/failure trackingAutomated pattern detectionAgents optimize agents
ContextAIEntire files in promptCurated system promptsDynamic context injectionAgents request what they need
MemoryAINo persistenceSession annotationsStructured long-term memoryAutomatic extraction, federated knowledge
AuthorityAINo boundariesStatic allowlistsDelegated authority with escalationDynamic trust based on track record

The key insight: most organizations are at different levels for human and AI consumers. A platform might be "Scalable" for developers — self-service portal, product mindset, dedicated team — but "Provisional" for agents — ad-hoc API keys, unstructured prompts, no memory.

Knowing where you stand on both axes tells you where to invest.

What Platform Teams Should Do

Audit your AI agent maturity. Walk through each dimension. Where are you Provisional? Most teams discover they have no structured context strategy and no memory infrastructure.

Invest in structured interfaces first. MCP tool schemas, typed APIs, and discoverable registries are the highest-ROI investment. They make your existing platform accessible to agents without rebuilding anything.

Treat context as a budget. Every token an agent spends reading irrelevant information is waste. Create CLAUDE.md files for your repositories. Write tool descriptions that include parameters and examples. Trim your system prompts.

Build memory infrastructure. Agents that start fresh every session repeat the same discovery work. Even simple persistence — session notes, project context files — eliminates massive redundancy.

Don't build a separate AI platform. The same APIs, the same CI/CD, the same infrastructure serves both audiences. What changes is the interface layer. Add tool schemas alongside your portal. Add structured context alongside your documentation.

The CNCF got the model right. The platform serves its consumers. Now it has two kinds.