feat: add Claude Code native implementation with multi-agent architecture#145
feat: add Claude Code native implementation with multi-agent architecture#145anhyobin wants to merge 1 commit intoawslabs:mainfrom
Conversation
…ture Add a Claude Code-specific implementation of AIDLC using native agents, skills, hooks, and rules. This provides the same AIDLC methodology with system-level role enforcement, on-demand context loading, and automated workflow management. - 14 skills (one per AIDLC stage, /aidlc-* slash commands) - 4 agents with tool isolation (analyst, architect, developer, reviewer) - 6 cross-cutting rules (content validation, error handling, terminology, workflow changes, 2 conditional extensions) - 2 hooks (SessionStart auto-detection, SubagentStop audit logging) - Benchmark: 100% pass rate across 76 assertions on 14 stages (vs 96.1% upstream rule file, 76.1% no guidance) Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
|
@anhyobin How do we maintain this? The AIDLC will evolve and your changes fully refactored the AIDLC into a group of skills. Is there any automation that can take the AIDLC flows and create this version of teh AIDLC? |
|
@harmjeff Great question. Let me explain the maintenance story. The mapping is 1:1 and modular.
There's no separate automation tool that generates these files, and by design it isn't needed. Cross-cutting logic (gate protocol, Q&A analysis, answer validation) already lives in the 4 agent files, not in skills. Skills delegate to agents and inherit shared protocols automatically. So when the AIDLC evolves:
These files are prompts, not compiled source. The content IS the product — each skill's core value is its domain-specific logic (question categories, artifact templates, analysis dimensions), which must be authored by hand regardless of any automation layer. Adding a generator on top would mean maintaining both the spec and the output, for changes that happen a few times per year at most. The structure follows what Anthropic calls harness design — each skill, agent, and rule is a self-contained file with a single responsibility. When the AIDLC evolves, you update the specific files that correspond to the changed stages. There's no monolithic prompt to reverse-engineer; the mapping from AIDLC stage to file is 1:1. In short: 24 focused markdown files in git, with shared behavior factored into agents (4) and rules (6). When bulk changes are needed, Claude Code itself is the multi-file edit tool. As the AIDLC methodology evolves, I can take ownership of maintaining the Claude Code harness to keep it in sync. |
|
I see what you are syaing, but my question still stands. Do we now need to maintain all changes to AIDLC rules in two places?
|
|
@harmjeff You're right — from a repo maintenance perspective, yes. ▎ - aidlc-rules/ is the platform-agnostic spec, usable by any AI coding tool. |
Summary
Add Claude Code-native AIDLC implementation using agents, skills, hooks, and rules. This provides the same AIDLC methodology with system-level role enforcement, on-demand context loading, and automated workflow management.
/aidlc-*slash commands)What Changed
New directory:
platforms/claude-code/(or location TBD by maintainers)Plus: README.md, LICENSE (MIT-0)
Why
Claude Code has native capabilities that the single rule-file approach cannot leverage:
disallowedTools: [Write, Edit](system)pathsfrontmatterTesting
Scenario: Serverless Order Management API (Cognito, Lambda, DynamoDB, SQS, EventBridge, Stripe integration. 4 units, TypeScript/CDK.)
Method: 42 agent evaluations across all 14 AIDLC stages, comparing 3 approaches:
Native outperforms upstream in:
functional: Technology-agnostic constraint enforcement (upstream leaked specific service names into business logic design)gate: 2-phase pipeline structure compliance (upstream merged code review and testing into a single undifferentiated pass)76 assertions cover: structural compliance (question format, [Answer]: tags, X) Other options, categorized questions), methodology adherence (4-dimension analysis, INVEST criteria, EXECUTE/SKIP decisions, GO/NO-GO + PASS/FAIL verdicts), and artifact completeness. The implementation is language-agnostic -- output language follows the user's language automatically.
Full benchmark data:
aidlc-skills-workspace/full-eval/benchmark.jsonUpstream Compatibility
aidlc-docs/)common/(content-validation, error-handling, terminology, workflow-changes)By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice.