Build a Software Development Team in 10 Minutes

🚀 Quick Start Guide

Experience the complete BDD-driven development lifecycle with SuperOptiX. See how evaluation, optimization, and multi-agent orchestration work together to create production-ready AI teams.

BDD Test Runner
DSPy Optimization
Multi-Criteria Evaluation
Multi-Agent Orchestra

🎯 What You'll Build in 10 Minutes

Software Development Team

Developer, QA Engineer, DevOps Engineer

BDD Test Runner

Professional specification execution

DSPy Optimization

Automatic performance improvement

Multi-Agent Orchestra

Production-ready deployment

🖥️ System Requirements

Ensure your system meets the requirements for optimal performance

Hardware Requirements

Minimum 16GB GPU RAM

Required for agent optimization (Step 5)

8GB+ system RAM

Recommended for smooth operation

Stable internet connection

For model downloads and API calls

Software Requirements

Python 3.11 or higher

Required for SuperOptiX compatibility

SuperOptiX Python package

Install with pip, conda, or uv

Ollama (for local LLMs)

Optional but recommended for local inference

AI Models (Required)

Install models before running agents

• Ollama: ollama pull llama3.2:1b
• SuperOptiX: super model install llama3.2:1b
• Auto-install: super model run llama3.2:1b "Hello"
🪟➡️On Windows, set PYTHONUTF8=1

⚠️ Important Note

Optimization (Step 5) is resource-intensive and can be costly. It requires significant GPU resources and may incur cloud costs. You can skip optimization if you don't meet the requirements or want to avoid costs - your agent will still work without optimization.

🛠️ Installation Options

Choose your preferred method to get started

# Install uv (if not already installed)
curl -LsSf https://astral.sh/uv/install.sh | sh
# Install SuperOptiX
uv pip install superoptix --prerelease=allow
# Verify installation
Note: First launch may take a few seconds
super

🔄 The Complete BDD-Driven Workflow

0
Install AI Models (Required)

Download models before proceeding - agents won't work without them

# Option 1: Using SuperOptiX CLI (Recommended)
super model install llama3.2:1b
# Option 2: Using Ollama directly
ollama pull llama3.2:1b
# Option 3: Auto-installation with direct execution
super model run llama3.2:1b "Hello world"

⚠️ Important Note

Models are required for agent execution. Without models, your agents won't be able to process requests. Choose the installation method that works best for your system.

Install AI models using any of the methods above. The SuperOptiX CLI method is recommended as it provides the most seamless experience.

1
Initialize Your SWE Project

super init swe
cd swe
ls -la

Creates the proper directory structure with `.super` configuration file for workspace management.

2
Pull a Developer Agent

💡 Model Requirement Check

Make sure you have installed AI models from Step 0 before pulling agents. The agent will use the models you installed.

super agent pull developer
🎉 AGENT ADDED SUCCESSFULLY! Pre-built Agent Ready
🤖 Name: Developer Assistant
🏢 Industry: Software | 🔮 Tier: Oracles
🔧 Tasks: 1 | 📁 Location: swe/agents/developer/playbook/developer_playbook.yaml

Downloads a production-ready agent with SuperSpec YAML configuration including BDD scenarios for testing.

3
Compile the Agent

super agent compile developer
🔨 Compiling agent 'developer'...
🎯 Agent: Developer Assistant
🏗️ Framework: DSPy (default) Junior Pipeline
🔧 Process: YAML playbook → Executable Python pipeline
📁 Output: swe/agents/developer/pipelines/developer_pipeline.py
✅ Compilation successful!

Converts the SuperSpec YAML playbook into an executable DSPy pipeline with BDD scenarios as evaluation tests.

4
BDD Test Runner - Establish Baseline

Critical: Always evaluate before optimizing to establish baseline performance!

super agent evaluate developer

Professional BDD Test Runner Output

🎭 Running BDD Test Suite for Agent: developer
============================================================
🎭 Executing BDD Scenarios...
📝 Running: basic_api_endpoint_creation
✅ PASSED
📝 Running: data_structure_design
✅ PASSED
📝 Running: algorithm_implementation
❌ FAILED: semantic meaning differs significantly
📝 Running: robust_error_handling
✅ PASSED
📝 Running: test_code_generation
✅ PASSED
📊 BDD Test Results Summary:
Total Scenarios: 5 | Passed: 4 ✅ | Failed: 1 ❌
Pass Rate: 80.0% | BDD Score: 0.800
🎉 BDD Test Suite: EXCELLENT (80.0%)
Multi-Criteria Evaluation
• Semantic Similarity: 50%
• Keyword Presence: 20%
• Structure Match: 20%
• Output Length: 10%
Quality Gates
🎉 ≥80%: EXCELLENT
⚠️ 60-79%: GOOD
❌ <60%: NEEDS WORK

The BDD Test Runner executes all scenarios defined in your SuperSpec playbook, providing detailed analysis and baseline metrics before optimization.

5
DSPy Optimization Engine

Transform BDD scenarios into training data for automatic agent improvement

⚠️ IMPORTANT WARNING: Optimization is Resource-Intensive

🚨 System Requirements:
  • • Minimum 16GB GPU RAM required for optimization
  • • 8GB+ system RAM recommended
  • • Stable internet connection for API calls
💡 Optimization Details:
  • • Makes multiple LLM API calls (can be costly)
  • • Uses DSPy BootstrapFewShot strategy
  • • May take several minutes to complete
  • • Use responsibly - monitor your API usage

🔄 Skip this step if you don't meet the requirements or want to avoid costs. Your agent will still work without optimization.

💡 Note: Optimization can also be triggered during evaluate or run stages depending on your workflow configuration.

super agent optimize developer

DSPy BootstrapFewShot Optimization

🚀 Optimizing agent 'developer'...
🎯 Agent: Developer
🔧 Strategy: DSPy BootstrapFewShot
📊 Data Source: BDD scenarios from playbook
💾 Output: swe/agents/developer/pipelines/developer_optimized.json
🚀 Starting optimization using 'bootstrap' strategy...
✅ Found 5 scenarios for optimization
🚀 Training with 5 examples...
💾 Optimized model saved
🎉 OPTIMIZATION SUCCESSFUL!
1. Learning from Examples

Uses BDD scenarios as training examples

2. Trial and Error

Tests different prompt variations

3. Weight Saving

Saves optimized configurations

The DSPy Optimization Engine uses your BDD scenarios as training data to automatically improve prompts, reasoning chains, and few-shot examples.

6
Measure Improvement

Re-run BDD tests to validate optimization effectiveness

super agent evaluate developer

Improved Performance Metrics

╭─────────────────── 🟢 Specification Results Summary ───────────────────╮
│ 📊 Total Specs: 5 🎯 Pass Rate: 80.0% │
│ ✅ Passed: 4 🤖 Model: llama3.2:1b │
│ ❌ Failed: 1 💪 Capability: 0.82 │
│ 🏆 Quality Gate: 🎉 EXCELLENT 🚀 Status: 🚀 Optimized │
╰─────────────────────────────────────────────────────────────────────╯
Before Optimization
Pass Rate: 60.0%
Capability Score: 0.65
Status: NEEDS WORK
After Optimization
Pass Rate: 80.0% ⬆️
Capability Score: 0.82 ⬆️
Status: EXCELLENT ⬆️

Quantitative improvement validated through the same BDD scenarios, showing measurable enhancement in agent performance.

7
Execute Production Goals

super agent run developer --goal "Build a snake game in Python"

Now your optimized agent can execute real production goals with improved performance validated by BDD scenarios.

8
Multi-Agent Orchestra

Coordinate multiple optimized agents for complex workflows

super agent pull qa_engineer
super agent pull devops_engineer
super agent compile --all
super orchestra create sdlc
super orchestra run sdlc --goal "Build a login feature"

Orchestra Execution

🎼 Running Orchestra: sdlc
📋 Executing 3 tasks sequentially...
🔄 Task 1/3: implement_feature
🤖 Running Developer Pipeline
✅ Task implement_feature completed in 2.95s
🔄 Task 2/3: configure_ci_pipeline
🤖 Running DevOps Engineer Pipeline
✅ Task configure_ci_pipeline completed in 2.29s
🔄 Task 3/3: create_test_plan
🤖 Running QA Engineer Pipeline
✅ Task create_test_plan completed in 2.13s
🎉 Orchestra completed successfully!

Kubernetes-style orchestration for AI agents - coordinate multiple optimized agents with automatic task decomposition and context passing.

🎯 Why This Workflow is Revolutionary

Evaluation-First Development

Test agents before optimizing, just like TDD for traditional software

BDD Scenarios as Training Data

Your tests become optimization training examples

Quality Gates

Deploy only when pass rates meet production standards

Automatic Optimization

DSPy automatically improves agent performance

Quantifiable Improvement

Measure exactly how much agents improved

Production Orchestration

Multi-agent coordination for complex workflows

🧠 What Just Happened?

Understanding the complete BDD-driven development lifecycle

🎯 The SuperOptiX Workflow

1. Evaluation-First: We started with BDD scenarios that define what success looks like, just like TDD in traditional software development.

2. Baseline Measurement: We established performance metrics before any optimization, ensuring we can measure improvement.

3. Systematic Optimization: DSPy used our BDD scenarios as training data to automatically improve prompts and reasoning chains.

4. Validation: We re-ran the same tests to validate that optimization actually improved performance.

5. Production Deployment: Our optimized agents can now execute real-world goals with confidence.

🔍 Why This Matters

Reproducible Results: Every optimization is based on your specific use cases, not generic examples.

Quality Assurance: BDD scenarios serve as both tests and training data, ensuring optimization improves real-world performance.

Measurable Improvement: You can quantify exactly how much your agents improved through the same evaluation metrics.

Production Ready: The evaluation-first approach ensures your agents meet quality gates before deployment.

📚 Learn More About Each Stage

For detailed explanations of what happens at each stage, comprehensive examples, and advanced techniques, visit our complete documentation:

Complete Quick Start Guide with Detailed Explanations

🎥 Watch the Complete Workflow

See the entire SuperOptiX workflow in action - from installation to running your first AI team

This demo shows the complete end-to-end workflow you just learned about:

Installation & Setup
BDD Evaluation & Optimization
Multi-Agent Orchestration