Model catalog
| Model | Provider | Aliases | Context | Cost (in/out per Mtok) | Speed |
|---|---|---|---|---|---|
claude-opus-4-6 | anthropic | opus, claude-opus | 1M | 75.00 | 25 tok/s |
claude-sonnet-4-6 | anthropic | sonnet, claude-sonnet | 200K | 15.00 | 50 tok/s |
claude-sonnet-4-5 | anthropic | 200K | 15.00 | 50 tok/s | |
claude-haiku-4-5 | anthropic | haiku, claude-haiku | 200K | 4.00 | 100 tok/s |
gpt-5.2 | openai | gpt5 | 1M | 14.00 | 65 tok/s |
gpt-5-mini | openai | gpt5-mini | 1M | 2.00 | 70 tok/s |
gpt-5.2-codex | openai | 1M | 14.00 | 100 tok/s | |
gpt-5.3-codex | openai | codex | 1M | 14.00 | 100 tok/s |
gpt-5.3-codex-spark | openai | codex-spark | 128K | n/a | 1000 tok/s |
gpt-5.4 | openai | gpt54 | 1M | 15.00 | 70 tok/s |
gpt-5.4-mini | openai | gpt54-mini | 400K | 4.50 | 140 tok/s |
gpt-5.4-pro | openai | gpt54-pro | 1M | 180.00 | 20 tok/s |
gemini-3.1-pro-preview | gemini | gemini-pro | 1M | 12.00 | 85 tok/s |
gemini-3.1-pro-preview-customtools | gemini | gemini-customtools | 1M | 12.00 | 85 tok/s |
gemini-3-flash-preview | gemini | gemini-flash | 1M | 3.00 | 150 tok/s |
gemini-3.1-flash-lite-preview | gemini | gemini-flash-lite | 1M | 1.50 | 200 tok/s |
kimi-k2.5 | kimi | kimi | 262K | 3.00 | 50 tok/s |
glm-4.7 | zai | glm, glm4 | 203K | 2.20 | 100 tok/s |
minimax-m2.5 | minimax | minimax | 197K | 1.20 | 45 tok/s |
mercury-2 | inception | mercury | 131K | 0.80 | 1000 tok/s |
ANTHROPIC_API_KEY, OPENAI_API_KEY, GEMINI_API_KEY). See the Quick Start for setup.
Default models
When no model or provider is specified, Fabro auto-detects the provider by checking which API keys are configured, using precedence order Anthropic > OpenAI > Gemini. If no keys are found, it falls back to Anthropic. Each provider has a default model:| Provider | Default model |
|---|---|
anthropic | claude-sonnet-4-6 |
openai | gpt-5.4 |
gemini | gemini-3.1-pro-preview |
kimi | kimi-k2.5 |
zai | glm-4.7 |
minimax | minimax-m2.5 |
inception | mercury |
Using models in workflows
Assign models to workflow nodes using model stylesheets, which use a CSS-like syntax:example.fabro
Overriding the default model
Model stylesheets set per-node models inside the workflow graph, but you can also override the default model for an entire run. This is useful for quick experimentation or when you want to swap models without editing the Graphviz file.CLI flags
Pass--model and optionally --provider to fabro run:
--provider for models not in the catalog or to force a specific provider.
Run config TOML
For repeatable runs, set the model in a run config file:run.toml
fallbacks array is optional. Each entry may be a bare provider token (like "gemini"), a bare model alias (like "gpt-5.4"), or a qualified "provider/model" reference. Fabro tries them in order when the primary provider is unavailable.
The precedence order is: node-level stylesheet > run config TOML > CLI flags > server defaults. More specific settings always win.