Skip to content

๐Ÿš€ Quick Start Guide: Two Paths to SuperOptiX Mastery

**Two hands-on experiences that highlight the SuperOptiX workflow**

Choose Your Path

Part 1 โ€“ Sentiment Analyzer Demo: A lightweight project that walks through evaluation and GEPA optimization in minutes.

Part 2 โ€“ SWE Orchestration: A full multi-agent software engineering workflow that showcases the orchestration features.

Getting Started

You can complete Part 1 on its own, then move on to Part 2 when you're ready to build larger teams.


๐Ÿ“‹ Requirements

๐Ÿ–ฅ๏ธ Hardware

Component Requirement
GPU RAM 16 GB recommended if you plan to run GEPA optimization
System RAM 8 GB+ for smooth execution

๐Ÿ Software

Software Version/Details
Python 3.11 or higher
SuperOptiX package Install via pip, conda, or uv
Ollama For local LLMs (alternatives like MLX or Hugging Face also work)

Install Ollama (if needed):

curl -fsSL https://ollama.com/install.sh | sh

Windows Users

Set PYTHONUTF8=1 in your environment before running the CLI.

๐Ÿ”ง Install Options

pip install superoptix
super --version
super
curl -LsSf https://astral.sh/uv/install.sh | sh
uv pip install superoptix
super --version
super

๐ŸŽจ Part 1 โ€” Sentiment Analyzer Demo (Evaluation & Optimization)

Overview

This mini-project validates that your environment is ready. You'll initialize a project, pull a sample dataset, run the agent, evaluate it, and apply GEPA optimization.

Step 1. Initialize the Project

super init sentiment_analyzer
cd sentiment_analyzer

Step 2. Pull the Dataset

super dataset pull sentiment_reviews

Dataset Location

This stores sentiment_reviews.csv in your project's data/ directory.

Alternative: Create Dataset Manually

If you're on an older CLI without super dataset pull, create it manually:

mkdir -p data
cat <<'CSV' > data/sentiment_reviews.csv
text,label
I absolutely love this product! It exceeded all my expectations.,positive
This is terrible and I'm disappointed with the quality.,negative
The package arrived on time. It contains the items I ordered.,neutral
It's decent overall and I'm satisfied with the results.,positive
It works but I expected better for the price.,negative
Excellent service from start to finish.,positive
Poor quality and bad customer service.,negative
The restaurant is located downtown and serves Italian food.,neutral
CSV

Step 3. Pull & Compile the Agent

super agent pull sentiment_analyzer
super agent compile sentiment_analyzer

Step 4. Run the Agent

super agent run sentiment_analyzer \
    --goal "Classify the sentiment of the review: 'I love this product but the shipping was slow.'"

Output

The agent responds with a sentiment label and a confidence score.

What Happened
  • The pipeline (agents/sentiment_analyzer/pipelines/sentiment_analyzer_pipeline.py) executed end-to-end with your goal.
  • DSPy configured the local Ollama model llama3.1:8b (temperature 0.3, max 512 tokens).
  • The ReAct chain generated both the structured fields (sentiment, confidence) and the reasoning trace.
  • Output is shown in the terminal and the pipeline remains inspectable under agents/sentiment_analyzer/pipelines/.

Step 5. Evaluate the Agent

super agent evaluate sentiment_analyzer

What This Does

Runs the playbook scenarios plus the dataset samples.

What Happened
  • Evaluation pulled every BDD scenario defined in agents/sentiment_analyzer/playbook/sentiment_analyzer_playbook.yaml.
  • Each scenario is scored with the answer_exact_match metric (threshold 0.7).
  • Examples from data/sentiment_reviews.csv were converted into DSPy Examples and included in the run.
  • A rich pass/fail summary (capability score, recommendations) was printed to the terminal.
  • Improve results by:
    • Expanding the expected_output keywords so they better match the phrasing your model produces.
    • Trying a stronger base model (e.g., llama3.1:70b, gpt-4) via environment variables or CLI flags.
    • Extending data/sentiment_reviews.csv with domain-specific examples so evaluation reflects your requirements.

Step 6. Optimize with GEPA & Re-evaluate

super agent optimize sentiment_analyzer --auto light
super agent evaluate sentiment_analyzer

GEPA Optimization

GEPA tunes prompts based on failed scenarios; the follow-up evaluation measures any change.

What Happened
  • GEPA iteratively mutated the sentiment pipeline and scored each candidate against the same evaluation set.
  • Optimized weights were saved to agents/sentiment_analyzer/pipelines/sentiment_analyzer_optimized.json.
  • The second evaluate command automatically loaded those weights before re-running the scenarios.
  • With the default local model, scores may still be lowโ€”this demo highlights the pipeline mechanics more than raw accuracy.
  • To boost pass rates:
    • Provide a more capable reflection model using --reflection-lm (OpenAI, Anthropic, larger Ollama variants).
    • Increase --auto intensity or max_full_evals for deeper GEPA exploration.
    • Update the playbook persona/reasoning to emphasize sentiment-specific cues you want.
    • Add new dataset rows targeting failure cases, then rerun evaluation/optimization.

Part 1 Complete!

You've now completed the full evaluation-first loop! Continue exploring or move on to the multi-agent SWE workflow below.


๐Ÿ—๏ธ Part 2 โ€” SWE Multi-Agent Orchestration

Overview

In this section you'll build an end-to-end software development workflow with multiple cooperating agents.

Step 1. Initialize the SWE Project

cd ..          # if you're still inside sentiment_analyzer
super init swe
cd swe

Project Structure

swe/
โ”œโ”€โ”€ agents/
โ”œโ”€โ”€ data/
โ”œโ”€โ”€ orchestrations/
โ”œโ”€โ”€ pipelines/
โ”œโ”€โ”€ tools/
โ””โ”€โ”€ .super

Step 2. Pull & Compile the Developer Agent

super agent pull developer
super agent compile developer

Compilation Output

Compilation generates an explicit DSPy pipeline at agents/developer/pipelines/developer_pipeline.py. This is your starting point for customizationโ€”open the file to see imports, signature definitions, and the reasoning chain produced from the playbook.

Tip

Each compile also prints the next suggested command. Rerun super agent compile developer --verbose if you want extra scaffolding (test counts, customization advice, etc.).


Step 3. Run the Developer Agent

super agent run developer \
    --goal "Create a Python function that validates email addresses using regex"

What to Expect

Watch the agent reason about the task and emit code along with explanations. The output file is stored in pipelines/ and the CLI displays the result inline.

Try Other Requests
super agent run developer --goal "Implement a REST API endpoint for user registration"
super agent run developer --goal "Fix a memory leak in a Python application"
super agent run developer --goal "Review and improve this sorting algorithm"

Need Automated Verification?

Run super agent evaluate developer --verbose to execute the BDD scenarios defined in the playbook. Keep in mind the developer playbook is intentionally minimalโ€”treat it as a baseline and extend it with domain-specific tests and knowledge.


Step 4. Add QA & DevOps Agents

super agent pull qa_engineer
super agent pull devops_engineer
super agent compile qa_engineer
super agent compile devops_engineer

Generated Pipelines

Each agent lives in its own subfolder under agents/. After compilation you'll see:

  • agents/qa_engineer/pipelines/qa_engineer_pipeline.py
  • agents/devops_engineer/pipelines/devops_engineer_pipeline.py

Tip

Open these pipelines to understand their personas, evaluation metrics, and prompts before orchestrating them.


Step 5. Create & Run the Orchestra

super orchestra create sdlc
super orchestra list
super orchestra run sdlc --goal "Build a task management web app with auth, CRUD, tests, and deployment config"

Orchestra Workflow

This generates orchestras/sdlc_orchestra.yaml and a compiled entry-point under pipelines/orchestras/. The sample goal walks through a three-phase SDLC:

  1. Developer: analyzes the goal, outlines the plan, and produces implementation artifacts.
  2. DevOps Engineer: translates the plan into CI/CD configuration and deployment notes.
  3. QA Engineer: derives comprehensive manual + automated test coverage from the preceding outputs.

Output Files

Orchestra results are saved to the project root (e.g., implement_feature_implementation.txt, configure_ci_pipeline_result.json, create_test_plan_test_plan.txt).

Customizing the Flow

Edit orchestras/sdlc_orchestra.yaml to insert additional agents, tweak execution order, or change prompts before re-running super orchestra run.


Step 6. Observe and Monitor

super observe traces developer
super observe dashboard

Observability Tools

  • Traces: Step through each agent's reasoning, model calls, and artifacts
  • Dashboard: Higher-level view for debugging orchestration runs or comparing pre/post optimization behavior

โœ… Summary

What You've Accomplished

Part 1: Demonstrated evaluation-first development using a sentiment analyzer, including GEPA optimization.

Part 2: Showed the full SWE orchestration flow with multiple agents collaborating on an SDLC task.

Next Steps

From here you can explore the marketplace (super market), design custom agents (super agent design), or build orchestras tailored to your workflows. Happy building! ๐ŸŽ‰