Without OMA
- Traces pile up in Langfuse but never become PRs.
- Incident playbooks live in a wiki no on-call reads at 2am.
- Cost anomalies surface on the next month's invoice.
- Every operations decision is a human judgement call.
AIDLC automates design and construction. Operations — deploys, incidents, cost drift, regressions — still fall on the team. OMA is the plugin marketplace that closes the loop with AgenticOps: humans approve, agents execute everything between the checkpoints.
$ claude
> /plugin marketplace add aws-samples/sample-oh-my-aidlcops
> /plugin install ai-infra agenticops aidlc modernization
✔ 4 plugins enabled · 11 AWS MCP servers pinned
> /oma:autopilot "ship the anomaly detector end to end"
The gap
What changes
Spec → design → code → canary deploy → self-healing → cost attribution. /oma:autopilot drives the whole loop and pauses only at explicit approval checkpoints.
Langfuse traces feed /oma:self-improving. Failure patterns become draft PRs against the skills and prompts that produced them — regression tests run before the PR is opened.
Every Tier-0 workflow sandwiches agent-driven diagnosis, proposal, and execution between explicit human gates. The agent never silently mutates production.
Drop-in
Ship as a native Claude Code marketplace entry. Slash commands, keyword triggers, and the AWS hosted MCP layer work out of the box.
install/kiro.sh symlinks every skill into ~/.kiro/skills/ and wires kiro-agents profiles with pinned MCP server versions.
Tier-0 mode, project memory, and audit logs live in .omao/. Both harnesses read and write the same directory — switch without losing context.
Inception and Construction describe what will ship. Operations keeps it alive after it ships — and feeds learnings back to Construction without a human in the loop for routine corrections.
aidlcWorkspace detection, adaptive requirements, user stories, workflow plan. Output artifacts become the contract Construction must honor.
aidlcComponent design, code generation with human-approved gates, 12-category risk discovery, TDD for agentic systems, phase quality gates.
agenticopsAutopilot deploys, continuous eval, incident response, cost governance, and the self-improving loop that feeds learnings back into Construction.
Runtime (ai-infra) and brownfield entry (modernization) sit alongside the loop, not inside it.
AgenticOps capabilities
autopilot-deploy runs canary 1% → 10% → 50% → 100% with SLO-gated circuit breakers. Each stage waits for continuous-eval before promotion; regression trips auto-rollback.
incident-response classifies SEV1–4, pulls the matching runbook, issues diagnostic MCP queries, and drafts a remediation script for approval. SEV1 pages on-call; it never acts.
cost-governance attributes spend per agent, vetoes deploys that would breach the monthly ceiling, and drafts Opus → Sonnet → Haiku downgrade PRs. budget.yaml runs in a simpleeval sandbox — no Python eval, no RCE vector.
Every skill is reachable as a slash command in Claude Code or a direct skill call in Kiro. The full state lives under .omao/ and is portable between harnesses.
> /plugin marketplace add https://github.com/aws-samples/sample-oh-my-aidlcops> /plugin install ai-infra agenticops aidlc modernization> /oma:platform-bootstrap[1/5] Gather Context … ok[2/5] Pre-flight … ok
Nine Tier-0 workflows
Full AIDLC loop (Inception → Construction → Operations).
Single-feature Inception → Construction pass.
AIDLC Phase 1 only — spec, stories, workflow plan.
AIDLC Phase 2 only — design, codegen, agentic TDD.
Operations mode: continuous-eval + incident-response + cost-governance.
Langfuse traces → prompt / skill improvement PR.
5-checkpoint Agentic AI Platform bootstrap on EKS.
6-stage brownfield modernization (assessment → cutover).
Terminate the active Tier-0 mode.
Keyword triggers auto-suggest the right command when your prompt contains a match. See the trigger catalog.
Four plugins
AI runtime infrastructure on AWS. Ships EKS + vLLM + Inference Gateway + Langfuse + GPU + guardrails today; Bedrock / SageMaker runtime skills planned. MCP servers pinned to exact PyPI versions — no @latest.
agentic-eks-bootstrap · vllm-serving-setup · inference-gateway-routing · langfuse-observability · gpu-resource-management · ai-gateway-guardrails
AIDLC Phase 1 (Inception) + Phase 2 (Construction) opt-in extensions for awslabs/aidlc-workflows. Inception captures workspace, requirements, stories, and the workflow plan. Construction turns that plan into components, code, tests, and risk-discovered quality gates.
workspace-detection · requirements-analysis · user-stories · workflow-planning · component-design · code-generation · test-strategy · risk-discovery · quality-gates
Autonomous operations for production agentic workloads. Incident response, self-improving feedback loops, progressive rollouts with SLO circuit breakers, cost governance with a simpleeval sandbox, and verbatim audit trails.
self-improving-loop · autopilot-deploy · incident-response · continuous-eval · cost-governance · audit-trail
Brownfield legacy workload modernization using the AWS 6R strategy. Workload assessment with Five Lenses, 6R decision matrix, to-be architecture, containerization hardening, and production cutover planning with rollback triggers.
workload-assessment · modernization-strategy · to-be-architecture · containerization · cutover-planning
30-second install
claude
> /plugin marketplace add https://github.com/aws-samples/sample-oh-my-aidlcops> /plugin install ai-infra@oh-my-aidlcops
> /plugin install agenticops@oh-my-aidlcops
> /plugin install aidlc@oh-my-aidlcops
> /plugin install modernization@oh-my-aidlcops> /oma:autopilot "ship the anomaly detector end to end"Or start with a safer on-ramp: getting-started guide.
Secure by default
Every .mcp.json and agent profile references awslabs MCP servers by exact PyPI version. No @latest supply-chain surprises.
The Kiro agent profile does not enable --allow-write or --allow-sensitive-data-access by default; opt in explicitly.
langfuse-observability uses a bucket-scoped customer-managed policy. AmazonS3FullAccess is called out as a Bad Example.
cost-governance evaluates budget.yaml rules with simpleeval. Python eval() on user-editable config is a documented RCE vector.
.omao/state, .omao/plans, .omao/logs, audit-trail output, and project memory are gitignored. Verbatim prompts never leave the machine.
session-start.sh requires jq or python3 and refuses to emit shell-interpolated JSON, preventing state-file injection into context.
FAQ
profile.yaml v1 and the 8 ontology schemas are stable; CLI surfaces and the doctor report shape may still evolve before GA. Breaking changes land in CHANGELOG under an explicit "Breaking" heading. See the support policy for the full stability contract.
Install once. Approve at the checkpoints. Let agents carry the rest of the AIDLC loop.