Skip to content

๐Ÿค– LLM Setup Guide

Welcome to SuperOptiX's LLM Setup Guide! This guide will help you configure and use local language models for your AI agents. We focus on local models for privacy, speed, and cost-effectiveness.

๐Ÿš€ Quick Start

New to local models? Start with Ollama - it's the easiest option for beginners!

๐ŸŽฏ Overview

SuperOptiX supports multiple local model backends, each optimized for different use cases:

Backend Best For Platform Ease of Use Performance
๐Ÿฆ™ Ollama Beginners, All platforms Cross-platform โญโญโญโญโญ โญโญโญโญ
๐Ÿค– GPT-OSS Advanced reasoning, Agentic tasks Cross-platform โญโญโญโญ โญโญโญโญโญ
๐ŸŽ MLX Apple Silicon users macOS only โญโญโญโญ โญโญโญโญโญ
๐ŸŽฎ LM Studio Windows users Windows/macOS โญโญโญ โญโญโญโญ
๐Ÿค— HuggingFace Advanced users All platforms โญโญ โญโญโญโญโญ

Production Inference Engines

vLLM, SGLang, and TGI are not included in the current version of SuperOptiX. These production-worthy inference engines are part of our enterprise offering.

Ollama is the easiest way to run local models on any platform. Perfect for beginners!

๐Ÿš€ Quick Setup

Bash
# Install Ollama
curl -fsSL https://ollama.ai/install.sh | sh

# Start Ollama (runs in background)
ollama serve
Bash
# Install Ollama
curl -fsSL https://ollama.ai/install.sh | sh

# Start Ollama
ollama serve
PowerShell
# Download from https://ollama.ai/download
# Or use winget
winget install Ollama.Ollama

# Start Ollama
ollama serve

๐Ÿ“ฆ Install Models with SuperOptiX

Bash
# Install recommended models by tier
super model install llama3.2:1b      # Oracles tier - Small tasks, fast responses
super model install llama3.2:8b      # Genies tier - Complex reasoning, tools, memory
super model install llama3.2:3b      # Alternative small model
super model install qwen2.5:7b       # Great all-rounder
Show Output
Text Only
๐Ÿš€ SuperOptiX Model Intelligence - Installing llama3.2:3b
๐Ÿฆ™ Pulling model llama3.2:3b from Ollama...
โณ This may take a few minutes depending on your internet connection and model size.

pulling manifest 
pulling dde5aa3fc5ff: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 2.0 GB                         
pulling 966de95ca8a6: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 1.4 KB                         
pulling fcc5a6bec9da: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 7.7 KB                         
pulling a70ff7e570d9: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 6.0 KB                         
pulling 56bb8bd477a5: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–   96 B                         
pulling 34bb5ab01051: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–  561 B                         
verifying sha256 digest 
writing manifest 
success 
โœ… Model pulled successfully!

๐Ÿ’ก You can now use it with SuperOptiX:
  super model dspy ollama/llama3.2:3b

๐Ÿ“Š Model details:
  โ€ข Size: small
  โ€ข Task: chat
  โ€ข Parameters: 3B

๐ŸŽ‰ Installation completed successfully!
๐Ÿฆ™ Ollama running on http://localhost:11434 ready to use with SuperOptiX!

๐Ÿ–ฅ๏ธ Server Management

๐Ÿ’ก Important: Ollama automatically starts its server when you run ollama serve or when you first use a model. You don't need to manually start the server unless you want custom configuration.

Bash
# Start Ollama server (runs on port 11434 by default)
ollama serve

# Or simply use a model - server starts automatically
ollama run llama3.2:1b

๐Ÿ”ง Custom Configuration: Only start the server manually if you need: - Different port: OLLAMA_HOST=0.0.0.0:8080 ollama serve - Custom model path: OLLAMA_MODELS=/custom/path ollama serve - GPU configuration: OLLAMA_GPU_LAYERS=35 ollama serve

โœ… Automatic Detection: SuperOptiX automatically detects and connects to Ollama running on the default port (11434). No additional configuration needed!

๐Ÿ“‹ Manage Ollama Models

Bash
# List installed models
super model list --backend ollama

Example Output:

Text Only
                ๐Ÿš€ SuperOptiX Model Intelligence - 3 models                 
โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“
โ”ƒ Model                   โ”ƒ  Backend  โ”ƒ    Status    โ”ƒ  Size   โ”ƒ   Task    โ”ƒ
โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ
โ”‚ llama3.1:8b             โ”‚ ๐Ÿฆ™ ollama โ”‚ โœ… installed โ”‚ medium  โ”‚   chat    โ”‚
โ”‚ llama3.2:1b             โ”‚ ๐Ÿฆ™ ollama โ”‚ โœ… installed โ”‚  tiny   โ”‚   chat    โ”‚
โ”‚ nomic-embed-text:latest โ”‚ ๐Ÿฆ™ ollama โ”‚ โœ… installed โ”‚ Unknown โ”‚ embedding โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Bash
# Get model information
super model info llama3.2:3b

# List all available models
super model list --all

๐Ÿค– GPT-OSS Models (OpenAI's Open Source)

GPT-OSS models are OpenAI's latest open-weight language models designed for powerful reasoning, agentic tasks, and versatile developer use cases. SuperOptiX now supports both GPT-OSS-20B and GPT-OSS-120B models with native Apple Silicon support!

๐ŸŽ Apple Silicon Support

MLX-LM v0.26.3 now provides native Apple Silicon support for GPT-OSS models, resolving the mixed precision issues that previously prevented these models from running on Apple Silicon.

Backend Model Status Performance Apple Silicon Recommendation
๐Ÿฆ™ Ollama gpt-oss:20b โœ… Works 19.7 t/s โœ… Optimized format โญ RECOMMENDED
๐ŸŽ MLX-LM openai_gpt-oss-20b โœ… Works 5.2 t/s โœ… Native support Apple Silicon only
๐Ÿค— HuggingFace openai/gpt-oss-20b โŒ Broken N/A โŒ Mixed precision errors โŒ Avoid on Apple Silicon

๐ŸŽฏ GPT-OSS Model Overview

Model Parameters Active Parameters Best For Hardware Requirements
GPT-OSS-20B 21B 3.6B Lower latency, local/specialized use cases 16GB+ RAM
GPT-OSS-120B 117B 5.1B Production, general purpose, high reasoning Single H100 GPU

๐Ÿš€ Recommended: Use Ollama for GPT-OSS Models

For the best performance and reliability with GPT-OSS models, we recommend using Ollama:

  • โœ… Best Performance: 19.7 t/s vs 5.2 t/s (MLX) vs N/A (HuggingFace)
  • โœ… Cross-Platform: Works on all platforms (Windows, macOS, Linux)
  • โœ… Easy Setup: Simple installation and model management
  • โœ… Optimized Format: GGUF format optimized for local inference
  • โœ… No Server Required: Direct model execution

Install and use GPT-OSS with Ollama:

Bash
# Install GPT-OSS models
super model install gpt-oss:20b
super model install gpt-oss:120b

# Use in playbooks
language_model:
  provider: ollama
  model: gpt-oss:20b
  api_base: http://localhost:11434

๐Ÿš€ Key Features

  • ๐Ÿ”“ Apache 2.0 License: Build freely without copyleft restrictions
  • โšก Native MXFP4 Quantization: Optimized for efficient inference
  • ๐ŸŽ Apple Silicon Native: No more mixed precision issues

๐Ÿ“ฆ Install GPT-OSS Models

Bash
# Install GPT-OSS models via Ollama (Best Performance)
super model install gpt-oss:20b
super model install gpt-oss:120b

# Or use direct Ollama commands
ollama pull gpt-oss:20b
ollama pull gpt-oss:120b

# Run with Ollama backend
super model run gpt-oss:20b "Your prompt" --backend ollama

Via MLX-LM (Apple Silicon - Native Support)

Bash
# Install GPT-OSS models via Ollama
super model install gpt-oss:20b
super model install gpt-oss:120b

# Or use direct Ollama commands
ollama pull gpt-oss:20b
ollama pull gpt-oss:120b

# Run with Ollama backend
super model run gpt-oss:20b "Your prompt" --backend ollama
Show Ollama Installation Output
Text Only
๐Ÿš€ SuperOptiX Model Intelligence - Installing gpt-oss:20b
๐Ÿฆ™ Pulling model gpt-oss:20b from Ollama...
โณ This may take a few minutes depending on your internet connection and model size.

pulling manifest 
pulling 8f7b3c2a1d4e: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 12.5 GB                         
pulling 9a2b4c6d8e0f: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 1.2 KB                         
verifying sha256 digest 
writing manifest 
success 
โœ… Model pulled successfully!

๐Ÿ’ก You can now use it with SuperOptiX:
  super model dspy ollama/gpt-oss:20b

๐Ÿ“Š Model details:
  โ€ข Size: large
  โ€ข Task: chat
  โ€ข Parameters: 21B (3.6B active)

๐ŸŽ‰ Installation completed successfully!
๐Ÿฆ™ Ollama running on http://localhost:11434 ready to use with SuperOptiX!

Via HuggingFace

Bash
# Install GPT-OSS models via HuggingFace
super model install openai/gpt-oss-20b --backend huggingface
super model install openai/gpt-oss-120b --backend huggingface

# Start HuggingFace server
super model server huggingface openai/gpt-oss-20b --port 8001
super model server huggingface openai/gpt-oss-120b --port 8002
Show HuggingFace Installation Output
Text Only
๐Ÿš€ SuperOptiX Model Intelligence - Installing openai/gpt-oss-20b
๐Ÿค— Downloading model from HuggingFace...
โณ This may take several minutes depending on your internet connection and model size.

Downloading model files...
  โ€ข config.json: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 2.1 KB
  โ€ข model.safetensors: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 12.5 GB
  โ€ข tokenizer.json: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 1.8 MB
  โ€ข tokenizer_config.json: 100% โ–•โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ– 1.2 KB

โœ… Model downloaded successfully!

๐Ÿ’ก You can now use it with SuperOptiX:
  super model server huggingface openai/gpt-oss-20b --port 8001

๐Ÿ“Š Model details:
  โ€ข Size: large
  โ€ข Task: chat
  โ€ข Parameters: 21B (3.6B active)
  โ€ข License: Apache 2.0

๐ŸŽ‰ Installation completed successfully!

๐ŸŽฏ Using GPT-OSS with SuperOptiX

1. Configure Playbook for GPT-OSS

YAML
# Example playbook configuration for GPT-OSS
language_model:
  provider: mlx  # or ollama or huggingface
  model: lmstudio-community/gpt-oss-20b-MLX-8bit  # for MLX-LM
  # model: gpt-oss:20b  # for Ollama
  # model: openai/gpt-oss-20b  # for HuggingFace
  api_base: http://localhost:11434  # for Ollama
  # api_base: http://localhost:8001  # for HuggingFace
  temperature: 0.7
  max_tokens: 2048

# GPT-OSS Language Model Configuration Examples

**๐Ÿฆ™ Ollama Backend (Cross-platform - RECOMMENDED):**
```yaml
language_model:
  provider: ollama
  model: gpt-oss:20b
  api_base: http://localhost:11434
  temperature: 0.7
  max_tokens: 4096

๐ŸŽ MLX Backend (Apple Silicon - Native Support):

YAML
language_model:
  provider: mlx
  model: lmstudio-community/gpt-oss-20b-MLX-8bit
  api_base: http://localhost:8000
  temperature: 0.7
  max_tokens: 4096

๐Ÿค— HuggingFace Backend (Limited on Apple Silicon):

YAML
language_model:
  provider: huggingface
  model: openai/gpt-oss-20b
  api_base: http://localhost:8001
  temperature: 0.7
  max_tokens: 4096

๐Ÿš€ Starting MLX Server for GPT-OSS

Before using GPT-OSS with MLX in your playbook, start the MLX server:

Bash
# Start MLX server for GPT-OSS model
super model server mlx lmstudio-community/gpt-oss-20b-MLX-8bit --port 8000

# Or start on a different port
super model server mlx lmstudio-community/gpt-oss-20b-MLX-8bit --port 9000

Server Output:

Text Only
๐ŸŽ MLX Local Server
Starting MLX server for lmstudio-community/gpt-oss-20b-MLX-8bit on port 8000...
๐Ÿš€ Starting MLX server...
python -m mlx_lm.server --model lmstudio-community/gpt-oss-20b-MLX-8bit --port 8000
โœ… MLX server is running on http://localhost:8000

Note: Keep the server running while using GPT-OSS models in your playbooks.

Text Only
#### 2. **Test GPT-OSS Models**

```bash
# Test with MLX-LM backend (Apple Silicon - Native)
super model run lmstudio-community/gpt-oss-20b-MLX-8bit "Explain quantum computing with detailed reasoning" --backend mlx

# Test with Ollama backend (Cross-platform - Best Performance)
super model run gpt-oss:20b "Explain quantum computing with detailed reasoning" --backend ollama

# Test with HuggingFace backend (Limited on Apple Silicon)
super model run openai/gpt-oss-20b "Write a Python function to solve the traveling salesman problem" --backend huggingface

3. Basic Usage Examples

Bash
# MLX-LM (Apple Silicon - Native support)
super model run lmstudio-community/gpt-oss-20b-MLX-8bit "What is 2+2?" --backend mlx
super model run lmstudio-community/gpt-oss-20b-MLX-8bit "Explain machine learning" --backend mlx
super model run lmstudio-community/gpt-oss-20b-MLX-8bit "Design a distributed system architecture" --backend mlx

# Ollama (Cross-platform - Best performance)
super model run gpt-oss:20b "What is 2+2?" --backend ollama
super model run gpt-oss:20b "Explain machine learning" --backend ollama
super model run gpt-oss:20b "Design a distributed system architecture" --backend ollama

๐Ÿ“‹ Manage GPT-OSS Models

Bash
# List installed GPT-OSS models
super model list | grep gpt-oss

# Get detailed information
super model info gpt-oss:20b
super model info openai/gpt-oss-120b

# Test model performance
super model test gpt-oss:20b "Hello, how are you?"

๐ŸŽฏ Performance Recommendations

Use Case Recommended Model Hardware
Quick responses GPT-OSS-20B 16GB+ RAM
Complex tasks GPT-OSS-120B H100 GPU
Local development GPT-OSS-20B 16GB+ RAM

๐Ÿ”ง Troubleshooting GPT-OSS

Error: error: 'mps.matmul' op detected operation with both F16 and BF16 operands which is not supported

Solution: ```bash

Use MLX-LM backend (native Apple Silicon support)

super model run lmstudio-community/gpt-oss-20b-MLX-8bit "prompt" --backend mlx

Text Only
# Or use Ollama backend (optimized format)
super model run gpt-oss:20b "prompt" --backend ollama
```

Error: Model not found or Model does not exist

Solution: ```bash

For MLX-LM (Apple Silicon)

super model install lmstudio-community/gpt-oss-20b-MLX-8bit --backend mlx

Text Only
# For Ollama
ollama pull gpt-oss:20b
ollama pull gpt-oss:120b

# For HuggingFace
super model install openai/gpt-oss-20b --backend huggingface
super model install openai/gpt-oss-120b --backend huggingface
```

Error: CUDA out of memory or Not enough memory

Solution:

Bash
# Use smaller model
super model install gpt-oss:20b  # Instead of 120b

# Use CPU inference
super model server huggingface openai/gpt-oss-20b --device cpu

Error: Connection refused or Cannot connect to server

Solution:

Bash
# Check Ollama server
ollama serve

# Check HuggingFace server
super model server huggingface openai/gpt-oss-20b --port 8001

๐Ÿ“š Resources

๐ŸŽ MLX (Apple Silicon)

MLX is Apple's native machine learning framework, offering blazing-fast inference on Apple Silicon Macs. MLX-LM v0.26.3 now provides native support for GPT-OSS models!

Apple Silicon Only

MLX only works on Apple Silicon Macs (M1, M2, M3). If you're on Intel Mac, use Ollama instead.

๐Ÿš€ Setup MLX

Bash
# Install MLX dependencies
pip install mlx-lm==0.26.3

# Or install with SuperOptiX
pip install "superoptix[mlx]"

๐Ÿ“ฆ Install MLX Models

Bash
# Install GPT-OSS models (native Apple Silicon support)
super model install openai/gpt-oss-20b --backend mlx
super model install openai/gpt-oss-120b --backend mlx

# Install popular MLX models
super model install -b mlx mlx-community/phi-2
super model install -b mlx mlx-community/Llama-3.2-3B-Instruct-4bit
super model install -b mlx mlx-community/Mistral-7B-Instruct-v0.2-4bit
super model install -b mlx lmstudio-community/gpt-oss-20b-MLX-8bit

๐Ÿ–ฅ๏ธ Start MLX Servers

Bash
# Start MLX server on specific port
super model server mlx phi-2 --port 8000
super model server mlx mlx-community/Llama-3.2-3B-Instruct-4bit --port 8000

Example Output:

Text Only
๐ŸŽ MLX Local Server
Starting MLX server for mlx-community_Llama-3.2-3B-Instruct-4bit on port 8000...
๐Ÿš€ Starting MLX server...
๐Ÿ“ก Server will be available at: http://localhost:8000
๐Ÿ’ก Use this URL in your playbook's api_base configuration
๐Ÿ”ง Manual server startup command:
   python -m mlx_lm.server --model mlx-community_Llama-3.2-3B-Instruct-4bit --port 8000
๐Ÿ“‹ Example playbook configuration:
   language_model:
     provider: mlx
     model: mlx-community_Llama-3.2-3B-Instruct-4bit
     api_base: http://localhost:8000
๐Ÿ”„ Executing: /path/to/python -m mlx_lm.server --model mlx-community_Llama-3.2-3B-Instruct-4bit --port 8000
โณ Server is starting... (Press Ctrl+C to stop)

๐Ÿ“‹ Manage MLX Models

Bash
# List MLX models
super model list --backend mlx

Example Output:

Text Only
                    ๐Ÿš€ SuperOptiX Model Intelligence - 1 models                     
โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”“
โ”ƒ Model                                    โ”ƒ Backend โ”ƒ    Status    โ”ƒ Size  โ”ƒ Task โ”ƒ
โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”ฉ
โ”‚ mlx-community_Llama-3.2-3B-Instruct-4bit โ”‚ ๐ŸŽ mlx  โ”‚ โœ… installed โ”‚ small โ”‚ chat โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Bash
# Get model information
super model info mlx-community/phi-2
super model info mlx-community_Llama-3.2-3B-Instruct-4bit

# Models are ready to use with SuperOptiX agents

๐ŸŽฎ LM Studio

LM Studio provides a user-friendly interface for running local models, especially popular on Windows.

๐Ÿš€ Setup LM Studio

  1. Download LM Studio from https://lmstudio.ai
  2. Install and launch LM Studio
  3. Download a model through the interface
  4. Start the server (default port: 1234)

๐Ÿ“ฆ Install Models with SuperOptiX

Bash
# Install models (use the name from LM Studio)
super model install -b lmstudio llama-3.2-1b-instruct
super model install -b lmstudio llama-3.2-3b
super model install -b lmstudio your-model-name

๐Ÿ–ฅ๏ธ Start LM Studio Servers

Bash
# Start server with specific model
super model server lmstudio llama-3.2-1b-instruct --port 1234
super model server lmstudio llama-3.2-3b --port 1234

Example Output:

Text Only
๐ŸŽฎ LM Studio Local Server
Starting LM Studio server for llama-3.2-1b-instruct on port 1234...
๐Ÿš€ Starting LM Studio server...
๐Ÿ“ก Server will be available at: http://localhost:1234
๐Ÿ’ก Use this URL in your playbook's api_base configuration
๐Ÿ”ง Manual server startup command:
   # Start server in LM Studio app first, then connect
๐Ÿ“‹ Example playbook configuration:
   language_model:
     provider: lmstudio
     model: llama-3.2-1b-instruct
     api_base: http://localhost:1234
โณ Server is starting... (Press Ctrl+C to stop)

๐Ÿ“‹ Manage LM Studio Models

Bash
# List LM Studio models
super model list --backend lmstudio

Example Output:

Text Only
                  ๐Ÿš€ SuperOptiX Model Intelligence - 3 models                  
โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”“
โ”ƒ Model                          โ”ƒ   Backend   โ”ƒ    Status    โ”ƒ  Size  โ”ƒ Task โ”ƒ
โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”ฉ
โ”‚ llama-3.2-1b-instruct          โ”‚ ๐ŸŽฎ lmstudio โ”‚ โœ… installed โ”‚ small  โ”‚ chat โ”‚
โ”‚ llama-3.3-70b-instruct         โ”‚ ๐ŸŽฎ lmstudio โ”‚ โœ… installed โ”‚ large  โ”‚ chat โ”‚
โ”‚ llama-4-scout-17b-16e-instruct โ”‚ ๐ŸŽฎ lmstudio โ”‚ โœ… installed โ”‚ medium โ”‚ chat โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Bash
# Get model information
super model info llama-3.2-1b-instruct

# Models are ready to use with SuperOptiX agents

๐Ÿค— HuggingFace

HuggingFace offers access to thousands of models, perfect for advanced users who want maximum flexibility.

๐Ÿš€ Setup HuggingFace

Bash
# Install HuggingFace dependencies
pip install transformers torch fastapi uvicorn

# Or install with SuperOptiX
pip install "superoptix[huggingface]"

๐Ÿ“ฆ Install HuggingFace Models

Bash
# Install popular models
super model install -b huggingface microsoft/Phi-4
super model install -b huggingface microsoft/DialoGPT-small
super model install -b huggingface microsoft/DialoGPT-medium
super model install -b huggingface meta-llama/Llama-2-7b-chat-hf

๐Ÿ–ฅ๏ธ Start HuggingFace Servers

Bash
# Start server with specific model
super model server huggingface microsoft/Phi-4 --port 8001
super model server huggingface microsoft/DialoGPT-small --port 8001
super model server huggingface microsoft/DialoGPT-medium --port 8001

Example Output:

Text Only
๐Ÿค— HuggingFace Local Server
Starting HuggingFace server for microsoft/DialoGPT-small on port 8002...
๐Ÿš€ Starting HuggingFace server...
๐Ÿ“ก Server will be available at: http://localhost:8002
๐Ÿ’ก Use this URL in your playbook's api_base configuration
๐Ÿ”ง Manual server startup command:
   python -m superoptix.models.backends.huggingface_server microsoft/DialoGPT-small --port 8002
๐Ÿ“‹ Example playbook configuration:
   language_model:
     provider: huggingface
     model: microsoft/DialoGPT-small
     api_base: http://localhost:8002
Device set to use cpu
INFO:     Started server process [4652]
INFO:     Waiting for application startup.
INFO:     Application startup complete.
INFO:     Uvicorn running on http://127.0.0.1:8002 (Press CTRL+C to quit)

๐Ÿ“‹ Manage HuggingFace Models

Bash
# List HuggingFace models
super model list --backend huggingface

Example Output:

Text Only
                ๐Ÿš€ SuperOptiX Model Intelligence - 2 models                
โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”“
โ”ƒ Model                    โ”ƒ    Backend     โ”ƒ    Status    โ”ƒ Size  โ”ƒ Task โ”ƒ
โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”ฉ
โ”‚ microsoft/DialoGPT-small โ”‚ ๐Ÿค— huggingface โ”‚ โœ… installed โ”‚ small โ”‚ chat โ”‚
โ”‚ microsoft/Phi-4          โ”‚ ๐Ÿค— huggingface โ”‚ โœ… installed โ”‚ small โ”‚ chat โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Bash
# Get model information
super model info microsoft/Phi-4
super model info microsoft/DialoGPT-small

# Models are ready to use with SuperOptiX agents

๐ŸŽฏ Model Management Commands

๐Ÿ–ฅ๏ธ Server Commands

Bash
# Get help for server commands
super model server --help

Example Output:

Text Only
usage: super model server [-h] [--port PORT] {mlx,huggingface,lmstudio} model_name

๐Ÿš€ Start local model servers for MLX, HuggingFace, or LM Studio. Examples: 
super model server mlx mlx-community/Llama-3.2-3B-Instruct-4bit 
super model server huggingface microsoft/DialoGPT-small --port 8001
super model server lmstudio llama-3.2-1b-instruct 

Backends: 
mlx Apple Silicon optimized (default: port 8000) 
huggingface Transformers models (default: port 8001) 
lmstudio Desktop app models (default: port 1234) 

Note: Ollama servers use 'ollama serve' command separately.

positional arguments:
  {mlx,huggingface,lmstudio}  Backend type
  model_name                   Model name to start server for

options:
  -h, --help                   show this help message and exit
  --port PORT, -p PORT         Port to run server on

๐Ÿ“‹ List and Explore Models

Bash
# List all installed models
super model list

Example Output:

Text Only
                           ๐Ÿš€ SuperOptiX Model Intelligence - 9 models                   
โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“
โ”ƒ Model                                    โ”ƒ    Backend     โ”ƒ    Status    โ”ƒ  Size   โ”ƒ   Task    โ”ƒ
โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ
โ”‚ llama-3.2-1b-instruct                    โ”‚  ๐ŸŽฎ lmstudio   โ”‚ โœ… installed โ”‚  small  โ”‚   chat    โ”‚
โ”‚ llama-3.3-70b-instruct                   โ”‚  ๐ŸŽฎ lmstudio   โ”‚ โœ… installed โ”‚  large  โ”‚   chat    โ”‚
โ”‚ llama-4-scout-17b-16e-instruct           โ”‚  ๐ŸŽฎ lmstudio   โ”‚ โœ… installed โ”‚ medium  โ”‚   chat    โ”‚
โ”‚ llama3.1:8b                              โ”‚   ๐Ÿฆ™ ollama    โ”‚ โœ… installed โ”‚ medium  โ”‚   chat    โ”‚
โ”‚ llama3.2:1b                              โ”‚   ๐Ÿฆ™ ollama    โ”‚ โœ… installed โ”‚  tiny   โ”‚   chat    โ”‚
โ”‚ microsoft/DialoGPT-small                 โ”‚ ๐Ÿค— huggingface โ”‚ โœ… installed โ”‚  small  โ”‚   chat    โ”‚
โ”‚ microsoft/Phi-4                          โ”‚ ๐Ÿค— huggingface โ”‚ โœ… installed โ”‚  small  โ”‚   chat    โ”‚
โ”‚ mlx-community_Llama-3.2-3B-Instruct-4bit โ”‚     ๐ŸŽ mlx     โ”‚ โœ… installed โ”‚  small  โ”‚   chat    โ”‚
โ”‚ nomic-embed-text:latest                  โ”‚   ๐Ÿฆ™ ollama    โ”‚ โœ… installed โ”‚ Unknown โ”‚ embedding โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

๐Ÿ” Discover more models: super model discover
๐Ÿ“ฅ Install a model: super model install <model_name>

Bash
# List all available models (including uninstalled)
super model list --all

# Filter by backend
super model list --backend ollama
super model list --backend mlx
super model list --backend lmstudio
super model list --backend huggingface

# Verbose information
super model list --verbose

๐Ÿ“Š Get Model Information

Bash
# Get detailed model info
super model info llama3.2:3b
super model info mlx-community/phi-2
super model info microsoft/Phi-4
super model info llama-3.2-1b-instruct

๐ŸŽฏ Choose Your Setup

Bash
# 1. Install Ollama
curl -fsSL https://ollama.ai/install.sh | sh

# 2. Install SuperOptiX
pip install superoptix

# 3. Install a model
super model install llama3.2:3b

# 4. Models are ready to use with SuperOptiX agents

๐ŸŽ Apple Silicon User

Bash
# 1. Install MLX dependencies
pip install mlx-lm

# 2. Install SuperOptiX
pip install superoptix

# 3. Install MLX model
super model install -b mlx mlx-community/phi-2

# 4. Start server
super model server mlx phi-2 --port 8000

# 5. Models are ready to use with SuperOptiX agents

๐ŸŽฎ Windows User

Bash
# 1. Install LM Studio from https://lmstudio.ai
# 2. Download a model in LM Studio
# 3. Start server in LM Studio

# 4. Install SuperOptiX
pip install superoptix

# 5. Connect to LM Studio
super model server lmstudio your-model-name --port 1234

# 6. Models are ready to use with SuperOptiX agents

๐Ÿค— Advanced User

Bash
# 1. Install HuggingFace dependencies
pip install transformers torch fastapi uvicorn

# 2. Install SuperOptiX
pip install superoptix

# 3. Install HuggingFace model
super model install -b huggingface microsoft/Phi-4

# 4. Start server
super model server huggingface microsoft/Phi-4 --port 8001

# 5. Models are ready to use with SuperOptiX agents

๐Ÿ”ง Advanced Configuration

๐ŸŒ Multiple Servers

Run multiple models simultaneously:

Bash
# Terminal 1: Ollama model
# Models are ready to use with SuperOptiX agents

# Terminal 2: MLX model (Apple Silicon)
super model server mlx phi-2 --port 8000
# Models are ready to use with SuperOptiX agents

# Terminal 3: HuggingFace model
super model server huggingface microsoft/Phi-4 --port 8001
# Models are ready to use with SuperOptiX agents

# Terminal 4: LM Studio model
super model server lmstudio llama-3.2-1b-instruct --port 1234
# Models are ready to use with SuperOptiX agents

๐Ÿšจ Troubleshooting

Common Issues

Error: Model not found or Model does not exist

Solution:

Bash
# Check available models
super model list --all

# Use correct model name
super model install llama3.2:3b  # โœ… Correct
super model install llama3.2     # โŒ Wrong

Error: Connection refused or Cannot connect to server

Solution:

Bash
# Check if server is running
# For Ollama: ollama serve
# For MLX: super model server mlx phi-2 --port 8000
# For LM Studio: Start in LM Studio app
# For HuggingFace: super model server huggingface model --port 8001

Error: Address already in use

Solution:

Bash
# Use different port
super model server mlx phi-2 --port 8001
super model server huggingface microsoft/Phi-4 --port 8002

Error: MLX requires Apple Silicon

Solution:

Bash
# Use Ollama instead
super model install llama3.2:3b
super model dspy ollama/llama3.2:3b

Error: ModuleNotFoundError: No module named 'mlx_lm' or ModuleNotFoundError: No module named 'transformers'

Solution:

Bash
# Install MLX dependencies
pip install mlx-lm

# Install HuggingFace dependencies
pip install transformers torch fastapi uvicorn

# Or install with SuperOptiX extras
pip install "superoptix[mlx]"
pip install "superoptix[huggingface]"

Error: Command 'ollama' not found or Command 'lms' not found

Solution:

Bash
# Install Ollama
curl -fsSL https://ollama.ai/install.sh | sh

# Install LM Studio
# Download from https://lmstudio.ai
# Or use winget on Windows:
winget install LMStudio.LMStudio

# Verify installation
ollama --version
lms --version

Error: 401 Unauthorized or Repository Not Found

Solution:

Bash
# For HuggingFace models, login:
huggingface-cli login

# For MLX models, ensure you have access:
# Some models require accepting terms on HuggingFace website

# Use public models instead:
super model install -b mlx mlx-community/phi-2
super model install -b huggingface microsoft/Phi-4

๐ŸŽ‰ Next Steps

Now that you have your local models set up:

  1. ๐Ÿš€ Quick Start Guide - Build your first agent with local models
  2. ๐Ÿค– Create Your First Genies Agent - Step-by-step tutorial
  3. ๐Ÿช Marketplace - Discover pre-built agents
  4. ๐Ÿ” Model Intelligence Guide - Advanced model management

๐Ÿ’ฌ Need Help?


๐Ÿค– Ready to Run Local Models?

๐Ÿค— HuggingFace

HuggingFace offers access to thousands of models, perfect for advanced users who want maximum flexibility.

๐Ÿš€ Setup HuggingFace

Bash
# Install HuggingFace dependencies
pip install transformers torch fastapi uvicorn

# Or install with SuperOptiX
pip install "superoptix[huggingface]"

๐Ÿ“ฆ Install HuggingFace Models

Bash
# Install popular models
super model install -b huggingface microsoft/Phi-4
super model install -b huggingface microsoft/DialoGPT-small
super model install -b huggingface microsoft/DialoGPT-medium
super model install -b huggingface meta-llama/Llama-2-7b-chat-hf

๐Ÿ–ฅ๏ธ Start HuggingFace Servers

Bash
# Start server with specific model
super model server huggingface microsoft/Phi-4 --port 8001
super model server huggingface microsoft/DialoGPT-small --port 8001
super model server huggingface microsoft/DialoGPT-medium --port 8001

Example Output:

Text Only
๐Ÿค— HuggingFace Local Server
Starting HuggingFace server for microsoft/DialoGPT-small on port 8002...
๐Ÿš€ Starting HuggingFace server...
๐Ÿ“ก Server will be available at: http://localhost:8002
๐Ÿ’ก Use this URL in your playbook's api_base configuration
๐Ÿ”ง Manual server startup command:
   python -m superoptix.models.backends.huggingface_server microsoft/DialoGPT-small --port 8002
๐Ÿ“‹ Example playbook configuration:
   language_model:
     provider: huggingface
     model: microsoft/DialoGPT-small
     api_base: http://localhost:8002
Device set to use cpu
INFO:     Started server process [4652]
INFO:     Waiting for application startup.
INFO:     Application startup complete.
INFO:     Uvicorn running on http://127.0.0.1:8002 (Press CTRL+C to quit)

๐Ÿ“‹ Manage HuggingFace Models

Bash
# List HuggingFace models
super model list --backend huggingface

Example Output:

Text Only
                ๐Ÿš€ SuperOptiX Model Intelligence - 2 models                
โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”“
โ”ƒ Model                    โ”ƒ    Backend     โ”ƒ    Status    โ”ƒ Size  โ”ƒ Task โ”ƒ
โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”ฉ
โ”‚ microsoft/DialoGPT-small โ”‚ ๐Ÿค— huggingface โ”‚ โœ… installed โ”‚ small โ”‚ chat โ”‚
โ”‚ microsoft/Phi-4          โ”‚ ๐Ÿค— huggingface โ”‚ โœ… installed โ”‚ small โ”‚ chat โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Bash
# Get model information
super model info microsoft/Phi-4
super model info microsoft/DialoGPT-small

# Models are ready to use with SuperOptiX agents