Model catalog
| Model | Provider | Aliases | Context | Cost (in/out per Mtok) | Speed |
|---|---|---|---|---|---|
claude-opus-4-6 | anthropic | opus, claude-opus | 1M | 75.00 | 25 tok/s |
claude-sonnet-4-5 | anthropic | sonnet, claude-sonnet | 200K | 15.00 | 50 tok/s |
claude-haiku-4-5 | anthropic | haiku, claude-haiku | 200K | 4.00 | 100 tok/s |
gpt-5.2 | openai | gpt5 | 1M | 14.00 | 65 tok/s |
gpt-5-mini | openai | gpt5-mini | 1M | 2.00 | 70 tok/s |
gpt-5.2-codex | openai | 1M | 14.00 | 100 tok/s | |
gpt-5.3-codex | openai | codex | 1M | 14.00 | 100 tok/s |
gpt-5.3-codex-spark | openai | codex-spark | 128K | n/a | 1000 tok/s |
gpt-5.4 | openai | gpt54 | 1M | 15.00 | 70 tok/s |
gpt-5.4-pro | openai | gpt54-pro | 1M | 180.00 | 20 tok/s |
gemini-3.1-pro-preview | gemini | gemini-pro | 1M | 12.00 | 85 tok/s |
gemini-3.1-pro-preview-customtools | gemini | gemini-customtools | 1M | 12.00 | 85 tok/s |
gemini-3-flash-preview | gemini | gemini-flash | 1M | 3.00 | 150 tok/s |
gemini-3.1-flash-lite-preview | gemini | gemini-flash-lite | 1M | 1.50 | 200 tok/s |
kimi-k2.5 | kimi | kimi | 262K | 3.00 | 50 tok/s |
glm-4.7 | zai | glm, glm4 | 203K | 2.20 | 100 tok/s |
minimax-m2.5 | minimax | minimax | 197K | 1.20 | 45 tok/s |
mercury-2 | inception | mercury | 131K | 0.80 | 1000 tok/s |
ANTHROPIC_API_KEY, OPENAI_API_KEY, GEMINI_API_KEY). See the Quick Start for setup.
Default models
When no model is specified, thefabro exec command uses a default model based on the provider:
| Provider | Default model |
|---|---|
anthropic | claude-opus-4-6 |
openai | gpt-5.2-codex |
gemini | gemini-3.1-pro-preview |
kimi | kimi-k2.5 |
zai | glm-4.7 |
minimax | minimax-m2.5 |
inception | mercury |
Using models in workflows
Assign models to workflow nodes using model stylesheets, which use a CSS-like syntax:example.fabro
Overriding the default model
Model stylesheets set per-node models inside the workflow graph, but you can also override the default model for an entire run. This is useful for quick experimentation or when you want to swap models without editing the DOT file.CLI flags
Pass--model and optionally --provider to fabro run:
--provider for models not in the catalog or to force a specific provider.
Run config TOML
For repeatable runs, set the model in a run config file:run.toml
[llm.fallbacks] table is optional. It maps each provider to an ordered list of fallback providers to try when the primary is unavailable.
The precedence order is: node-level stylesheet > run config TOML > CLI flags > server defaults. More specific settings always win.