Build production AI systems 3-5x faster with specialized agents that handle requirements, architecture, coding, and deployment automatically.
Status:
Version: See .repo-metadata.json for current version and agent counts.
15 minutes: Custom GPTs, Claude assistants, Cursor IDE modes
2 hours: Complete architecture + cost estimates
2-5 days: Working prototypes with tests
Complete lifecycle: Requirements → Architecture → Code → Deployment
# 1. Clone
git clone https://github.com/paulpham157/multi-agent-ai-development-framework
cd multi-agent-ai-development-framework
# 2. Deploy (Cursor → Settings → Chat → Custom Modes)
# Create "Supervisor Agent"
# Paste: supervisor_agent.system.prompt.md
# Enable: "All tools" → Save
# 3. Use
# Open AI Pane (Ctrl+Shift+L)
# Try: "Build a Streamlit chatbot with Claude"Done! Supervisor routes to specialized agents who generate code for you to review.
Other platforms: See docs/getting-started.md for Claude Projects and GitHub Copilot setup.
The Problem: Every AI project starts from scratch
- Requirements: 2 days → should be 2 hours
- Architecture: 2 weeks → should be 4 hours
- Prototypes: 2 months → should be 1 week
The Solution: Specialized agents automate 70% of repetitive work
| Task | Traditional | With Framework | Savings |
|---|---|---|---|
| Requirements | 2 days | 2 hours | 90% |
| Architecture | 2 weeks | 4 hours | 97% |
| Prototype | 2 months | 1 week | 87% |
Result: 3-5x faster delivery, 60% less rework, consistent quality
Main Supervisor → Routes your requests
Top-Level Domain Agents (5):
- Requirements: Discovers what you need (15-90 min workshops)
- Architecture: Designs system + estimates costs (AWS Well-Architected)
- Deployment: Creates platform-specific deployment guides
- Optimization: Analyzes and improves existing systems
- Prompt Engineering: Creates production-quality prompts
Engineering Supervisor (1): Coordinates 16 technology specialists
Engineering Specialists (16):
- Anthropic Claude (5): Code, Workspaces, SDK, MCP, Projects
- AWS Bedrock (2): AgentCore, Strands
- Other (9): Streamlit, LangChain, data, AWS infra/security, testing, GitHub, Cursor
See: docs/engineering-agents-guide.md for complete specialist reference
You: "Build a document summarization app with Streamlit and Claude"
Engineering Supervisor routes to specialists:
→ Streamlit UI Agent: Chat interface + file upload
→ Claude Code Agent: Claude SDK integration + streaming
→ Testing Agent: pytest suite with mocks
Output:
✅ streamlit_app.py, claude_client.py, tests/
✅ requirements.txt, README.md, .env.example
✅ 80% test coverage, production-ready
YOU: Review → Run locally → Deploy
Result: Working prototype same day
You: "Optimize my Claude chatbot at outputs/my-chatbot/"
Optimization Agent:
→ Discovers system structure
→ Assesses Well-Architected compliance (score: 6.2/10)
→ Proposes improvements (reduce tokens 30%, add caching)
→ Implements (with your approval)
→ Validates improvements
Result:
✅ Monthly cost: $120 → $78 (-35%)
✅ Response time: 3.2s → 2.4s (-25%)
✅ Well-Architected: 6.2 → 8.1 (+30%)
Result: Measurable improvements backed by data
More examples: See docs/workflow_guide.md and docs/examples/
Every architecture evaluated against 6 pillars + GenAI Lens:
- Operational Excellence • Security • Reliability
- Performance • Cost • Sustainability
- Model Selection • Prompt Engineering • RAG • Multi-Agent • Responsible AI
Test-Time Recursive Majority ensures quality:
- Generate → Validate → Improve → Re-validate
- Only present outputs meeting quality benchmarks
- Code coverage ≥80%, type hints ≥90%, 0 critical security issues
17 improvement prompts for continuous enhancement
- System-wide optimization
- Individual agent improvements
- 2-3 iterations per session (practical recursion prevention)
knowledge_base/system_config.json contains:
- 150+ technical documentation URLs
- AWS Well-Architected definitions
- Research papers (TRM, MetaGPT)
- Design patterns
- Quality benchmarks
✅ Junior Engineers: Learn from generated code, ship without years of experience
✅ Senior Engineers: Eliminate boilerplate, focus on complex decisions, 3-5x faster
✅ Consultants: Professional proposals in hours, accurate estimates
✅ Managers: Standardize processes, 5x faster onboarding
✅ Architects: Systematic Well-Architected designs, evidence-based recommendations
✅ CTOs: De-risk AI investments, scale without proportional hiring
Cursor IDE (Recommended - 5 min):
1. Settings → Chat → Custom Modes
2. Create "Supervisor Agent", paste supervisor_agent.system.prompt.md
3. Enable "All tools" → Save
Claude Projects (10 min):
1. Create project at https://claude.ai/projects (requires Anthropic account)
2. Upload knowledge_base/*.json
3. Custom Instructions: supervisor_agent.system.prompt.md
GitHub Copilot (15 min):
1. Create .github/copilot-instructions.md
2. Paste supervisor_agent.system.prompt.md
3. Use @workspace in VS Code
Full guide: docs/deployment-guide.md
Core: Python 3.12+ • Streamlit • Anthropic Claude • AWS Bedrock • MCP • LangChain
16 Engineering Specialists cover:
- 5 Anthropic Claude specialists
- 2 AWS Bedrock specialists
- 9 Other specialists (UI, orchestration, data, AWS infra/security, testing, platforms)
Centralized Docs: 150+ technical URLs in knowledge_base/system_config.json → technical_references
Two-layer supervisor-worker pattern:
Supervisor Agent
↓
┌───────────────┼───────────────┐
↓ ↓ ↓
Requirements Architecture Engineering Supervisor
↓ ↓ ↓
│ │ ┌──────┼──────┐
│ │ ↓ ↓ ↓
│ │ Streamlit Claude AWS
│ │ UI Code Bedrock
│ │ +14 more specialists
│ ↓
└───────→ Deployment + Optimization
Shared Knowledge Base:
├─ system_config.json (platform constraints, tech refs)
├─ user_requirements.json (business requirements)
└─ design_decisions.json (architecture, costs, plans)
See: ARCHITECTURE.md for complete details
Agents do (you review):
- ✅ Generate code/docs/configs
- ✅ Analyze systems
- ✅ Recommend improvements
- ✅ Validate quality
YOU do (always):
- ✅ Review all outputs
- ✅ Approve architectures
- ✅ Execute deployments
- ✅ Make critical decisions
Agents NEVER:
- ❌ Commit code automatically
- ❌ Deploy to production
- ❌ Make business decisions
- ❌ Spend money without approval
See: docs/human-ai-collaboration.md for complete guide
Essential (start here):
README.md- This file (overview + quick start)docs/getting-started.md- First project walkthrough (15 min)docs/deployment-guide.md- Platform deploymentdocs/human-ai-collaboration.md- Your role vs agent role
Reference:
docs/workflow_guide.md- Complete workflowsdocs/engineering-agents-guide.md- All 16 specialistsdocs/executive_overview.md- Business valueARCHITECTURE.md- System architectureknowledge_base/README.md- Knowledge base guidetemplates/- Requirements, architecture, checklistsprivate/README.md- Security guidelines for sensitive data
multi-agent-ai-development-framework/
├── .repo-metadata.json # Single source of truth (version, counts)
├── ai_agents/ # Agent system prompts
│ ├── supervisor_agent.system.prompt.md (main entry point)
│ ├── [5 top-level domain agents]
│ ├── engineering_supervisor_agent.system.prompt.md
│ └── [16 specialist agents]
├── knowledge_base/ # Shared state across agents
│ ├── system_config.json (150+ tech refs, Well-Architected defs, validation framework)
│ ├── user_requirements.json
│ ├── design_decisions.json
│ └── schemas/ (JSON schemas for validation)
├── user_prompts/ # Task-specific instructions
├── docs/ # Documentation
├── templates/ # Reusable templates
├── tests/ # Validation tests (auto-update metadata)
├── outputs/ # Generated systems go here
└── private/ # Sensitive data (NEVER committed to Git)
├── README.md (security guidelines)
└── sensitive-ai-agent-outputs/ (protected AI outputs)
See .repo-metadata.json for current agent/prompt counts
Note: Repository is under development locally, not yet published to GitHub.
What works:
- ✅ All agents functional
- ✅ Complete workflows (requirements → deployment)
- ✅ Code generation quality-assured (TRM validation)
- ✅ Well-Architected enforcement
- ✅ Self-improvement system
Known limitations:
⚠️ No production validation yet⚠️ Breaking changes expected before v1.0⚠️ Some edge cases untested⚠️ Documentation evolving (runuser_prompts/self_improvement/improve_all_documentation.user.prompt.mdfor latest)
Use at your own risk. Report issues on GitHub. Production-ready in v1.0.
Documentation: Start with docs/getting-started.md
Issues: GitHub Issues for bugs/features
Discussions: GitHub Discussions for questions
Contributing: See CONTRIBUTING.md
MIT License - Full commercial use permitted
- GitHub: Repository available locally (not yet published to GitHub)
- Getting Started:
docs/getting-started.md - Deployment Guide:
docs/deployment-guide.md - Engineering Specialists:
docs/engineering-agents-guide.md - Workflows:
docs/workflow_guide.md
Built with: Python • Streamlit • Anthropic Claude • AWS Bedrock • MCP • LangChain
🚀 Start building: Install Supervisor Agent in Cursor and say "Build a chatbot"