๐ Quick Start Guide: Two Paths to SuperOptiX Mastery
Choose Your Path
Part 1 โ Sentiment Analyzer Demo: A lightweight project that walks through evaluation and GEPA optimization in minutes.
Part 2 โ SWE Orchestration: A full multi-agent software engineering workflow that showcases the orchestration features.
Getting Started
You can complete Part 1 on its own, then move on to Part 2 when you're ready to build larger teams.
๐ Requirements
๐ฅ๏ธ Hardware
| Component | Requirement |
|---|---|
| GPU RAM | 16 GB recommended if you plan to run GEPA optimization |
| System RAM | 8 GB+ for smooth execution |
๐ Software
| Software | Version/Details |
|---|---|
| Python | 3.11 or higher |
| SuperOptiX package | Install via pip, conda, or uv |
| Ollama | For local LLMs (alternatives like MLX or Hugging Face also work) |
Install Ollama (if needed):
curl -fsSL https://ollama.com/install.sh | sh
Windows Users
Set PYTHONUTF8=1 in your environment before running the CLI.
๐ง Install Options
pip install superoptix
super --version
super
curl -LsSf https://astral.sh/uv/install.sh | sh
uv pip install superoptix
super --version
super
๐จ Part 1 โ Sentiment Analyzer Demo (Evaluation & Optimization)
Overview
This mini-project validates that your environment is ready. You'll initialize a project, pull a sample dataset, run the agent, evaluate it, and apply GEPA optimization.
Step 1. Initialize the Project
super init sentiment_analyzer
cd sentiment_analyzer
Step 2. Pull the Dataset
super dataset pull sentiment_reviews
Dataset Location
This stores sentiment_reviews.csv in your project's data/ directory.
Alternative: Create Dataset Manually
If you're on an older CLI without super dataset pull, create it manually:
mkdir -p data
cat <<'CSV' > data/sentiment_reviews.csv
text,label
I absolutely love this product! It exceeded all my expectations.,positive
This is terrible and I'm disappointed with the quality.,negative
The package arrived on time. It contains the items I ordered.,neutral
It's decent overall and I'm satisfied with the results.,positive
It works but I expected better for the price.,negative
Excellent service from start to finish.,positive
Poor quality and bad customer service.,negative
The restaurant is located downtown and serves Italian food.,neutral
CSV
Step 3. Pull & Compile the Agent
super agent pull sentiment_analyzer
super agent compile sentiment_analyzer
Step 4. Run the Agent
super agent run sentiment_analyzer \
--goal "Classify the sentiment of the review: 'I love this product but the shipping was slow.'"
Output
The agent responds with a sentiment label and a confidence score.
What Happened
- The pipeline (
agents/sentiment_analyzer/pipelines/sentiment_analyzer_pipeline.py) executed end-to-end with your goal. - DSPy configured the local Ollama model
llama3.1:8b(temperature 0.3, max 512 tokens). - The ReAct chain generated both the structured fields (
sentiment,confidence) and the reasoning trace. - Output is shown in the terminal and the pipeline remains inspectable under
agents/sentiment_analyzer/pipelines/.
Step 5. Evaluate the Agent
super agent evaluate sentiment_analyzer
What This Does
Runs the playbook scenarios plus the dataset samples.
What Happened
- Evaluation pulled every BDD scenario defined in
agents/sentiment_analyzer/playbook/sentiment_analyzer_playbook.yaml. - Each scenario is scored with the
answer_exact_matchmetric (threshold 0.7). - Examples from
data/sentiment_reviews.csvwere converted into DSPyExamples and included in the run. - A rich pass/fail summary (capability score, recommendations) was printed to the terminal.
- Improve results by:
- Expanding the
expected_outputkeywords so they better match the phrasing your model produces. - Trying a stronger base model (e.g.,
llama3.1:70b,gpt-4) via environment variables or CLI flags. - Extending
data/sentiment_reviews.csvwith domain-specific examples so evaluation reflects your requirements.
- Expanding the
Step 6. Optimize with GEPA & Re-evaluate
super agent optimize sentiment_analyzer --auto light
super agent evaluate sentiment_analyzer
GEPA Optimization
GEPA tunes prompts based on failed scenarios; the follow-up evaluation measures any change.
What Happened
- GEPA iteratively mutated the sentiment pipeline and scored each candidate against the same evaluation set.
- Optimized weights were saved to
agents/sentiment_analyzer/pipelines/sentiment_analyzer_optimized.json. - The second
evaluatecommand automatically loaded those weights before re-running the scenarios. - With the default local model, scores may still be lowโthis demo highlights the pipeline mechanics more than raw accuracy.
- To boost pass rates:
- Provide a more capable reflection model using
--reflection-lm(OpenAI, Anthropic, larger Ollama variants). - Increase
--autointensity ormax_full_evalsfor deeper GEPA exploration. - Update the playbook persona/reasoning to emphasize sentiment-specific cues you want.
- Add new dataset rows targeting failure cases, then rerun evaluation/optimization.
- Provide a more capable reflection model using
Part 1 Complete!
You've now completed the full evaluation-first loop! Continue exploring or move on to the multi-agent SWE workflow below.
๐๏ธ Part 2 โ SWE Multi-Agent Orchestration
Overview
In this section you'll build an end-to-end software development workflow with multiple cooperating agents.
Step 1. Initialize the SWE Project
cd .. # if you're still inside sentiment_analyzer
super init swe
cd swe
Project Structure
swe/
โโโ agents/
โโโ data/
โโโ orchestrations/
โโโ pipelines/
โโโ tools/
โโโ .super
Step 2. Pull & Compile the Developer Agent
super agent pull developer
super agent compile developer
Compilation Output
Compilation generates an explicit DSPy pipeline at agents/developer/pipelines/developer_pipeline.py. This is your starting point for customizationโopen the file to see imports, signature definitions, and the reasoning chain produced from the playbook.
Tip
Each compile also prints the next suggested command. Rerun super agent compile developer --verbose if you want extra scaffolding (test counts, customization advice, etc.).
Step 3. Run the Developer Agent
super agent run developer \
--goal "Create a Python function that validates email addresses using regex"
What to Expect
Watch the agent reason about the task and emit code along with explanations. The output file is stored in pipelines/ and the CLI displays the result inline.
Try Other Requests
super agent run developer --goal "Implement a REST API endpoint for user registration"
super agent run developer --goal "Fix a memory leak in a Python application"
super agent run developer --goal "Review and improve this sorting algorithm"
Need Automated Verification?
Run super agent evaluate developer --verbose to execute the BDD scenarios defined in the playbook. Keep in mind the developer playbook is intentionally minimalโtreat it as a baseline and extend it with domain-specific tests and knowledge.
Step 4. Add QA & DevOps Agents
super agent pull qa_engineer
super agent pull devops_engineer
super agent compile qa_engineer
super agent compile devops_engineer
Generated Pipelines
Each agent lives in its own subfolder under agents/. After compilation you'll see:
agents/qa_engineer/pipelines/qa_engineer_pipeline.pyagents/devops_engineer/pipelines/devops_engineer_pipeline.py
Tip
Open these pipelines to understand their personas, evaluation metrics, and prompts before orchestrating them.
Step 5. Create & Run the Orchestra
super orchestra create sdlc
super orchestra list
super orchestra run sdlc --goal "Build a task management web app with auth, CRUD, tests, and deployment config"
Orchestra Workflow
This generates orchestras/sdlc_orchestra.yaml and a compiled entry-point under pipelines/orchestras/. The sample goal walks through a three-phase SDLC:
- Developer: analyzes the goal, outlines the plan, and produces implementation artifacts.
- DevOps Engineer: translates the plan into CI/CD configuration and deployment notes.
- QA Engineer: derives comprehensive manual + automated test coverage from the preceding outputs.
Output Files
Orchestra results are saved to the project root (e.g., implement_feature_implementation.txt, configure_ci_pipeline_result.json, create_test_plan_test_plan.txt).
Customizing the Flow
Edit orchestras/sdlc_orchestra.yaml to insert additional agents, tweak execution order, or change prompts before re-running super orchestra run.
Step 6. Observe and Monitor
super observe traces developer
super observe dashboard
Observability Tools
- Traces: Step through each agent's reasoning, model calls, and artifacts
- Dashboard: Higher-level view for debugging orchestration runs or comparing pre/post optimization behavior
โ Summary
What You've Accomplished
Part 1: Demonstrated evaluation-first development using a sentiment analyzer, including GEPA optimization.
Part 2: Showed the full SWE orchestration flow with multiple agents collaborating on an SDLC task.
Next Steps
From here you can explore the marketplace (super market), design custom agents (super agent design), or build orchestras tailored to your workflows. Happy building! ๐