# CHORUS AI Provider and Model Configuration # This file defines how different agent roles map to AI models and providers # Global provider settings providers: # Local Ollama instance (default for most roles) ollama: type: ollama endpoint: http://localhost:11434 default_model: llama3.1:8b temperature: 0.7 max_tokens: 4096 timeout: 300s retry_attempts: 3 retry_delay: 2s enable_tools: true enable_mcp: true mcp_servers: [] # Ollama cluster nodes (for load balancing) ollama_cluster: type: ollama endpoint: http://192.168.1.72:11434 # Primary node default_model: llama3.1:8b temperature: 0.7 max_tokens: 4096 timeout: 300s retry_attempts: 3 retry_delay: 2s enable_tools: true enable_mcp: true # OpenAI API (for advanced models) openai: type: openai endpoint: https://api.openai.com/v1 api_key: ${OPENAI_API_KEY} default_model: gpt-4o temperature: 0.7 max_tokens: 4096 timeout: 120s retry_attempts: 3 retry_delay: 5s enable_tools: true enable_mcp: true # ResetData LaaS (fallback/testing) resetdata: type: resetdata endpoint: ${RESETDATA_ENDPOINT} api_key: ${RESETDATA_API_KEY} default_model: llama3.1:8b temperature: 0.7 max_tokens: 4096 timeout: 300s retry_attempts: 3 retry_delay: 2s enable_tools: false enable_mcp: false # Global fallback settings default_provider: ollama fallback_provider: resetdata # Role-based model mappings roles: # Software Developer Agent developer: provider: ollama model: codellama:13b temperature: 0.3 # Lower temperature for more consistent code max_tokens: 8192 # Larger context for code generation system_prompt: | You are an expert software developer agent in the CHORUS autonomous development system. Your expertise includes: - Writing clean, maintainable, and well-documented code - Following language-specific best practices and conventions - Implementing proper error handling and validation - Creating comprehensive tests for your code - Considering performance, security, and scalability Always provide specific, actionable implementation steps with code examples. Focus on delivering production-ready solutions that follow industry best practices. fallback_provider: resetdata fallback_model: codellama:7b enable_tools: true enable_mcp: true allowed_tools: - file_operation - execute_command - git_operations - code_analysis mcp_servers: - file-server - git-server - code-tools # Code Reviewer Agent reviewer: provider: ollama model: llama3.1:8b temperature: 0.2 # Very low temperature for consistent analysis max_tokens: 6144 system_prompt: | You are a thorough code reviewer agent in the CHORUS autonomous development system. Your responsibilities include: - Analyzing code quality, readability, and maintainability - Identifying bugs, security vulnerabilities, and performance issues - Checking test coverage and test quality - Verifying documentation completeness and accuracy - Suggesting improvements and refactoring opportunities - Ensuring compliance with coding standards and best practices Always provide constructive feedback with specific examples and suggestions for improvement. Focus on both technical correctness and long-term maintainability. fallback_provider: resetdata fallback_model: llama3.1:8b enable_tools: true enable_mcp: true allowed_tools: - code_analysis - security_scan - test_coverage - documentation_check mcp_servers: - code-analysis-server - security-tools # Software Architect Agent architect: provider: openai # Use OpenAI for complex architectural decisions model: gpt-4o temperature: 0.5 # Balanced creativity and consistency max_tokens: 8192 # Large context for architectural discussions system_prompt: | You are a senior software architect agent in the CHORUS autonomous development system. Your expertise includes: - Designing scalable and maintainable system architectures - Making informed decisions about technologies and frameworks - Defining clear interfaces and API contracts - Considering scalability, performance, and security requirements - Creating architectural documentation and diagrams - Evaluating trade-offs between different architectural approaches Always provide well-reasoned architectural decisions with clear justifications. Consider both immediate requirements and long-term evolution of the system. fallback_provider: ollama fallback_model: llama3.1:13b enable_tools: true enable_mcp: true allowed_tools: - architecture_analysis - diagram_generation - technology_research - api_design mcp_servers: - architecture-tools - diagram-server # QA/Testing Agent tester: provider: ollama model: codellama:7b # Smaller model, focused on test generation temperature: 0.3 max_tokens: 6144 system_prompt: | You are a quality assurance engineer agent in the CHORUS autonomous development system. Your responsibilities include: - Creating comprehensive test plans and test cases - Implementing unit, integration, and end-to-end tests - Identifying edge cases and potential failure scenarios - Setting up test automation and continuous integration - Validating functionality against requirements - Performing security and performance testing Always focus on thorough test coverage and quality assurance practices. Ensure tests are maintainable, reliable, and provide meaningful feedback. fallback_provider: resetdata fallback_model: llama3.1:8b enable_tools: true enable_mcp: true allowed_tools: - test_generation - test_execution - coverage_analysis - performance_testing mcp_servers: - testing-framework - coverage-tools # DevOps/Infrastructure Agent devops: provider: ollama_cluster model: llama3.1:8b temperature: 0.4 max_tokens: 6144 system_prompt: | You are a DevOps engineer agent in the CHORUS autonomous development system. Your expertise includes: - Automating deployment processes and CI/CD pipelines - Managing containerization with Docker and orchestration with Kubernetes - Implementing infrastructure as code (IaC) - Monitoring, logging, and observability setup - Security hardening and compliance management - Performance optimization and scaling strategies Always focus on automation, reliability, and security in your solutions. Ensure infrastructure is scalable, maintainable, and follows best practices. fallback_provider: resetdata fallback_model: llama3.1:8b enable_tools: true enable_mcp: true allowed_tools: - docker_operations - kubernetes_management - ci_cd_tools - monitoring_setup - security_hardening mcp_servers: - docker-server - k8s-tools - monitoring-server # Security Specialist Agent security: provider: openai model: gpt-4o # Use advanced model for security analysis temperature: 0.1 # Very conservative for security max_tokens: 8192 system_prompt: | You are a security specialist agent in the CHORUS autonomous development system. Your expertise includes: - Conducting security audits and vulnerability assessments - Implementing security best practices and controls - Analyzing code for security vulnerabilities - Setting up security monitoring and incident response - Ensuring compliance with security standards - Designing secure architectures and data flows Always prioritize security over convenience and thoroughly analyze potential threats. Provide specific, actionable security recommendations with risk assessments. fallback_provider: ollama fallback_model: llama3.1:8b enable_tools: true enable_mcp: true allowed_tools: - security_scan - vulnerability_assessment - compliance_check - threat_modeling mcp_servers: - security-tools - compliance-server # Documentation Agent documentation: provider: ollama model: llama3.1:8b temperature: 0.6 # Slightly higher for creative writing max_tokens: 8192 system_prompt: | You are a technical documentation specialist agent in the CHORUS autonomous development system. Your expertise includes: - Creating clear, comprehensive technical documentation - Writing user guides, API documentation, and tutorials - Maintaining README files and project wikis - Creating architectural decision records (ADRs) - Developing onboarding materials and runbooks - Ensuring documentation accuracy and completeness Always write documentation that is clear, actionable, and accessible to your target audience. Focus on providing practical information that helps users accomplish their goals. fallback_provider: resetdata fallback_model: llama3.1:8b enable_tools: true enable_mcp: true allowed_tools: - documentation_generation - markdown_processing - diagram_creation - content_validation mcp_servers: - docs-server - markdown-tools # General Purpose Agent (fallback) general: provider: ollama model: llama3.1:8b temperature: 0.7 max_tokens: 4096 system_prompt: | You are a general-purpose AI agent in the CHORUS autonomous development system. Your capabilities include: - Analyzing and understanding various types of development tasks - Providing guidance on software development best practices - Assisting with problem-solving and decision-making - Coordinating with other specialized agents when needed Always provide helpful, accurate information and know when to defer to specialized agents. Focus on understanding the task requirements and providing appropriate guidance. fallback_provider: resetdata fallback_model: llama3.1:8b enable_tools: true enable_mcp: true # Environment-specific overrides environments: development: # Use local models for development to reduce costs default_provider: ollama fallback_provider: resetdata staging: # Mix of local and cloud models for realistic testing default_provider: ollama_cluster fallback_provider: openai production: # Prefer reliable cloud providers with fallback to local default_provider: openai fallback_provider: ollama_cluster # Model performance preferences (for auto-selection) model_preferences: # Code generation tasks code_generation: preferred_models: - codellama:13b - gpt-4o - codellama:34b min_context_tokens: 8192 # Code review tasks code_review: preferred_models: - llama3.1:8b - gpt-4o - llama3.1:13b min_context_tokens: 6144 # Architecture and design architecture: preferred_models: - gpt-4o - llama3.1:13b - llama3.1:70b min_context_tokens: 8192 # Testing and QA testing: preferred_models: - codellama:7b - llama3.1:8b - codellama:13b min_context_tokens: 6144 # Documentation documentation: preferred_models: - llama3.1:8b - gpt-4o - mistral:7b min_context_tokens: 8192