Skip to content
8 changes: 5 additions & 3 deletions docs/cli/configuration/settings.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ Local overrides merge on top of the corresponding `settings.json` at the same le

| Setting | Options | Default | Description |
| ------- | ------- | ------- | ----------- |
| `model` | `opus`, `opus-4-6`, `opus-4-6-fast`, `sonnet`, `sonnet-4-6`, `gpt-5.4`, `gpt-5.2`, `gpt-5.2-codex`, `gpt-5.3-codex`, `haiku`, `gemini-3.1-pro`, `gemini-3-flash`, `droid-core`, `glm-5`, `kimi-k2.5`, `minimax-m2.5`, `custom-model` | `opus` | The default AI model used by droid |
| `model` | `opus`, `opus-4-7`, `opus-4-6`, `opus-4-6-fast`, `sonnet`, `sonnet-4-6`, `gpt-5.4`, `gpt-5.2`, `gpt-5.2-codex`, `gpt-5.3-codex`, `haiku`, `gemini-3.1-pro`, `gemini-3-flash`, `droid-core`, `glm-5`, `glm-5.1`, `kimi-k2.5`, `kimi-k2.6`, `minimax-m2.7`, `custom-model` | `opus` | The default AI model used by droid |
| `reasoningEffort` | `off`, `none`, `low`, `medium`, `high` (availability depends on the model) | Model-dependent default | Controls how much structured thinking the model performs. |
| `autonomyMode` | `normal`, `spec`, `auto-low`, `auto-medium`, `auto-high` | `normal` | Sets the default autonomy mode when starting droid. |
| `cloudSessionSync` | `true`, `false` | `true` | Mirror CLI sessions to Factory web. |
Expand Down Expand Up @@ -71,10 +71,12 @@ Choose the default AI model that powers your droid:
- **`haiku`** - Claude Haiku 4.5, fast and cost-effective
- **`gemini-3.1-pro`** - Gemini 3.1 Pro
- **`gemini-3-flash`** - Gemini 3 Flash, fast and cheap (0.2× multiplier)
- **`droid-core`** - GLM-4.7 open-source model
- **`droid-core`** - GLM-5.1 open-source model
- **`glm-5`** - GLM-5 open-source model
- **`glm-5.1`** - GLM-5.1 open-source model
- **`kimi-k2.5`** - Kimi K2.5 open-source model with image support
- **`minimax-m2.5`** - MiniMax M2.5 open-source model with reasoning support (0.12× multiplier)
- **`kimi-k2.6`** - Kimi K2.6 open-source model with image support and optional High reasoning
Comment thread
factory-zimu marked this conversation as resolved.
- **`minimax-m2.7`** - MiniMax M2.7 open-source model with reasoning support for cost-sensitive agentic workflows
- **`custom-model`** - Your own configured model via BYOK

[You can also add custom models and BYOK.](/cli/configuration/byok)
Expand Down
15 changes: 9 additions & 6 deletions docs/cli/droid-exec/overview.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -70,24 +70,27 @@ Options:

Supported models (examples):

- claude-opus-4-7
- claude-opus-4-6
- claude-opus-4-6-fast
- claude-opus-4-5-20251101
- claude-sonnet-4-6
- claude-sonnet-4-5-20250929
- claude-haiku-4-5-20251001
- gpt-5.1-codex
- gpt-5.1-codex-max
- gpt-5.1
- gpt-5.2
- gpt-5.2-codex
- gpt-5.3-codex
- gemini-3-pro-preview
- gpt-5.3-codex-fast
- gpt-5.4
- gpt-5.4-fast
- gpt-5.4-mini
- gemini-3.1-pro-preview
- gemini-3-flash-preview
- glm-4.7
- glm-5
- glm-5.1
- kimi-k2.5
- minimax-m2.5
- kimi-k2.6
Comment thread
factory-zimu marked this conversation as resolved.
- minimax-m2.7

<Note>
See the [model table](/pricing#available-models) for the full list of available models.
Expand Down
22 changes: 12 additions & 10 deletions docs/cli/user-guides/choosing-your-model.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -25,10 +25,11 @@ Model quality evolves quickly, and we tune the CLI defaults as the ecosystem shi
| 11 | **Claude Haiku 4.5** | Fast, cost-efficient for routine tasks and high-volume automation. |
| 12 | **Gemini 3.1 Pro** | Newer Gemini Pro generation with strong structured outputs and mixed reasoning controls for research-heavy tasks. |
| 13 | **Gemini 3 Flash** | Fast, cheap (0.2× multiplier) with full reasoning support; great for high-volume tasks where speed matters. |
| 14 | **Droid Core (MiniMax M2.5)** | Open-source, 0.12× multiplier with reasoning support (Low/Medium/High); cheapest model available. No image support. |
| 15 | **Droid Core (GLM-5)** | Open-source, 0.4× multiplier with updated GLM capabilities for bulk automation and air-gapped environments; no image support. |
| 16 | **Droid Core (GLM-4.7)** | Open-source, 0.25× multiplier, stable choice for bulk automation or air-gapped environments; note: no image support. |
| 17 | **Droid Core (Kimi K2.5)** | Open-source, 0.25× multiplier with image support; good for cost-sensitive work. |
| 14 | **Droid Core (MiniMax M2.7)** | Open-source, 0.12× multiplier with reasoning support (Low/Medium/High) and image support; cheapest model available. |
| 15 | **Droid Core (GLM-5.1)** | Open-source, 0.55× multiplier, newer GLM option for bulk automation and air-gapped environments; no image support. |
| 16 | **Droid Core (GLM-5)** | Open-source, 0.4× multiplier, stable choice for bulk automation and air-gapped environments; no image support. |
| 17 | **Droid Core (Kimi K2.6)** | Open-source, 0.4× multiplier with image support and optional High reasoning; good for cost-sensitive work when you still want a thinking toggle. |
| 18 | **Droid Core (Kimi K2.5)** | Open-source, 0.25× multiplier with image support; older Kimi option for cost-sensitive work. |

<Note>
We ship model updates regularly. When a new release overtakes the list above,
Expand All @@ -43,9 +44,9 @@ Model quality evolves quickly, and we tune the CLI defaults as the ecosystem shi
| ---------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------------------------------------ |
| **Deep planning, architecture reviews, ambiguous product specs** | Start with **Opus 4.7** for best depth and safety (1× promotional multiplier through April 30), or fall back to **Opus 4.6** / **Opus 4.6 Fast** for faster turnaround. Use **Sonnet 4.6** or **Sonnet 4.5** when you want balanced cost/quality, or **GPT-5.4** for large-context reasoning. |
| **Full-feature development, large refactors** | **Opus 4.7** or **Opus 4.6** for depth and safety. **GPT-5.4**, **GPT-5.3-Codex**, or **GPT-5.2-Codex** when you need speed plus **Extra High** reasoning; **Sonnet 4.6** or **Sonnet 4.5** for balanced loops. |
| **Repeatable edits, summarization, boilerplate generation** | **Haiku 4.5** or **Droid Core** (including **MiniMax M2.5** at 0.12×) for speed and cost. **GPT-5.2** when you need higher quality or structured outputs. |
| **Repeatable edits, summarization, boilerplate generation** | **Haiku 4.5** or **Droid Core** (including **MiniMax M2.7** at 0.12×) for speed and cost. **GPT-5.2** when you need higher quality or structured outputs. |
| **CI/CD or automation loops** | Favor **Haiku 4.5** or **Droid Core** for predictable, low-cost throughput. Use **GPT-5.3-Codex** or **GPT-5.4** when automation needs stronger reasoning. |
| **High-volume automation, frequent quick turns** | **Haiku 4.5** for speedy feedback. **Droid Core** (especially **MiniMax M2.5** at 0.12× with reasoning) when cost is critical or you need air-gapped deployment. |
| **High-volume automation, frequent quick turns** | **Haiku 4.5** for speedy feedback. **Droid Core** (especially **MiniMax M2.7** at 0.12× with reasoning) when cost is critical or you need air-gapped deployment. |

<Tip>
**Claude Opus 4.7** is the newest top-tier option for extremely complex architecture decisions or critical work where you need maximum reasoning capability—and it runs at a promotional 1× multiplier through April 30 (2× afterward). **Claude Opus 4.6** remains an excellent alternative, and **Opus 4.6 Fast** is tuned for faster responses at a higher cost. Most tasks don't require Opus-level power—start with Sonnet 4.6 or Sonnet 4.5 and escalate only if needed.
Expand Down Expand Up @@ -77,9 +78,10 @@ Tip: you can swap models mid-session with `/model` or by toggling in the setting
- **Gemini 3.1 Pro**: Low / Medium / High (default: High)
- **Gemini 3 Flash**: Minimal / Low / Medium / High (default: High)
- **Droid Core (GLM-5)**: None only (default: None; no image support)
- **Droid Core (GLM-4.7)**: None only (default: None; no image support)
- **Droid Core (GLM-5.1)**: None only (default: None; no image support)
- **Droid Core (Kimi K2.6)**: Off / High (default: High)
- **Droid Core (Kimi K2.5)**: None only (default: None)
- **Droid Core (MiniMax M2.5)**: Low / Medium / High (default: High)
- **Droid Core (MiniMax M2.7)**: Low / Medium / High (default: High)

Reasoning effort increases latency and cost—start low for simple work and escalate as needed. **Max** is available on Claude Opus 4.7, the Opus 4.6 family (Opus 4.6 and Opus 4.6 Fast), and Sonnet 4.6. **Extra High** is available on GPT-5.4, GPT-5.2, GPT-5.2-Codex, and GPT-5.3-Codex.

Expand All @@ -96,14 +98,14 @@ Factory ships with managed Anthropic and OpenAI access. If you prefer to run aga

### Open-source models

**Droid Core (GLM-5)**, **Droid Core (GLM-4.7)**, **Droid Core (Kimi K2.5)**, and **Droid Core (MiniMax M2.5)** are open-source alternatives available in the CLI. They're useful for:
**Droid Core (GLM-5)**, **Droid Core (GLM-5.1)**, **Droid Core (Kimi K2.6)**, **Droid Core (Kimi K2.5)**, and **Droid Core (MiniMax M2.7)** are open-source alternatives available in the CLI. They're useful for:

- **Air-gapped environments** where external API calls aren't allowed
- **Cost-sensitive projects** needing unlimited local inference
- **Privacy requirements** where code cannot leave your infrastructure
- **Experimentation** with open-source model capabilities

**Note:** GLM-5, GLM-4.7, and MiniMax M2.5 do not support image attachments. Kimi K2.5 does support images. MiniMax M2.5 is the cheapest model available (0.12× multiplier) and uniquely supports reasoning (Low/Medium/High) among Droid Core models. For image-based workflows, use Claude, GPT, or Kimi models.
**Note:** GLM-5 and GLM-5.1 do not support image attachments. Kimi K2.5, Kimi K2.6, and MiniMax M2.7 do support images. Kimi K2.6 adds an Off/High reasoning toggle, while MiniMax M2.7 (the cheapest model available, with 0.12× multiplier) supports Low/Medium/High reasoning. For image-based workflows, use Claude, GPT, Kimi, or MiniMax M2.7.

To use open-source models, you'll need to configure them via BYOK with a local inference server (like Ollama) or a hosted provider. See [BYOK documentation](/cli/configuration/byok) for setup instructions.

Expand Down
4 changes: 2 additions & 2 deletions docs/enterprise/hierarchical-settings-and-org-control.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -217,11 +217,11 @@ The following model identifiers can be used in `allowedModelIds` and `blockedMod

| Provider | Model IDs |
|---|---|
| **Anthropic** | `claude-sonnet-4-5-20250929`, `claude-sonnet-4-6`, `claude-opus-4-5-20251101`, `claude-opus-4-6`, `claude-opus-4-6-fast`, `claude-haiku-4-5-20251001` |
| **Anthropic** | `claude-sonnet-4-5-20250929`, `claude-sonnet-4-6`, `claude-opus-4-5-20251101`, `claude-opus-4-6`, `claude-opus-4-6-fast`, `claude-opus-4-7`, `claude-haiku-4-5-20251001` |
| **OpenAI** | `gpt-5.1`, `gpt-5.1-codex`, `gpt-5.1-codex-max`, `gpt-5.2`, `gpt-5.2-codex`, `gpt-5.3-codex` |
Comment thread
factory-zimu marked this conversation as resolved.
| **Google** | `gemini-3-pro-preview`, `gemini-3-flash-preview`, `gemini-3.1-pro-preview` |
| **XAI** | `grok-code-fast-1` |
| **Open source** | `glm-4.7`, `glm-5`, `kimi-k2.5`, `minimax-m2.5` |
| **Open source** | `glm-5`, `glm-5.1`, `kimi-k2.5`, `kimi-k2.6`, `minimax-m2.7` |

---

Expand Down
2 changes: 1 addition & 1 deletion docs/guides/power-user/prompt-crafting.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -376,7 +376,7 @@ Match the model to the task:
| **Feature implementation** | Sonnet 4.5 or GPT-5.1-Codex | Medium |
| **Quick edits, formatting** | Haiku 4.5 | Off/Low |
| **Code review** | GPT-5.1-Codex-Max | High |
| **Bulk automation** | GLM-4.7 (Droid Core) | None |
| **Bulk automation** | GLM-5 (Droid Core) | None |
| **Research/analysis** | Gemini 3 Pro | High |

---
Expand Down
12 changes: 6 additions & 6 deletions docs/guides/power-user/token-efficiency.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -134,13 +134,13 @@ Different models have different cost multipliers and capabilities. Match the mod

| Model | Multiplier | Best For |
|-------|------------|----------|
| Droid Core (MiniMax M2.5) | 0.12× | Cheapest option with reasoning support |
| Droid Core (MiniMax M2.7) | 0.12× | Cheapest option with reasoning support |
| Gemini 3 Flash | 0.2× | Fast, cheap for high-volume tasks |
| Droid Core (GLM-4.7) | 0.25× | Bulk automation, simple tasks |
| Droid Core (Kimi K2.5) | 0.25× | Cost-sensitive work, supports images |
| Claude Haiku 4.5 | 0.4× | Quick edits, routine work |
| Droid Core (GLM-5) | 0.4× | Newer open-source GLM model when you want stronger quality in Droid Core |
| GPT-5.1 / GPT-5.1-Codex | 0.5× | Implementation, debugging |
| Droid Core (Kimi K2.6) | 0.4× | Cost-sensitive work, supports images, optional High reasoning |
| Droid Core (GLM-5) | 0.4× | Bulk automation, simple tasks |
| Droid Core (GLM-5.1) | 0.55× | Newer open-source GLM model when you want stronger quality in Droid Core |
| GPT-5.2-Codex / GPT-5.3-Codex | 0.7× | Advanced coding with Extra High reasoning |
| Gemini 3 Pro | 0.8× | Research, analysis |
| Gemini 3.1 Pro | 0.8× | Research, analysis with newer Gemini generation |
Expand All @@ -154,10 +154,10 @@ Different models have different cost multipliers and capabilities. Match the mod

```
Simple edit, formatting → Haiku 4.5 (0.4×)
Implement feature from spec → GPT-5.1-Codex (0.5×)
Implement feature from spec → GPT-5.3-Codex (0.7×)
Debug complex issue → Sonnet 4.5 (1.2×)
Architecture planning → Opus 4.7 (2×, 1× until April 30) or Opus 4.6 (2×)
Bulk file processing → Droid Core (GLM-4.7 at 0.25× or GLM-5 at 0.4×)
Bulk file processing → Droid Core (GLM-5 at 0.4× or GLM-5.1 at 0.55×)
```

### Reasoning Effort Impact
Expand Down
8 changes: 5 additions & 3 deletions docs/jp/cli/configuration/settings.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ droidの設定を構成するには:

| 設定 | オプション | デフォルト | 説明 |
| ------- | ------- | ------- | ----------- |
| `model` | `opus`, `opus-4-6`, `opus-4-6-fast`, `sonnet`, `sonnet-4-6`, `gpt-5.4`, `gpt-5.2`, `gpt-5.2-codex`, `gpt-5.3-codex`, `haiku`, `gemini-3.1-pro`, `gemini-3-flash`, `droid-core`, `glm-5`, `kimi-k2.5`, `minimax-m2.5`, `custom-model` | `opus` | droidが使用するデフォルトのAIモデル |
| `model` | `opus`, `opus-4-7`, `opus-4-6`, `opus-4-6-fast`, `sonnet`, `sonnet-4-6`, `gpt-5.4`, `gpt-5.2`, `gpt-5.2-codex`, `gpt-5.3-codex`, `haiku`, `gemini-3.1-pro`, `gemini-3-flash`, `droid-core`, `glm-5`, `glm-5.1`, `kimi-k2.5`, `kimi-k2.6`, `minimax-m2.7`, `custom-model` | `opus` | droidが使用するデフォルトのAIモデル |
| `reasoningEffort` | `off`, `none`, `low`, `medium`, `high`(利用可能性はモデルに依存) | モデル依存のデフォルト | モデルが実行する構造化思考の量を制御します。 |
| `autonomyMode` | `normal`, `spec`, `auto-low`, `auto-medium`, `auto-high` | `normal` | droid開始時のデフォルト自律モードを設定します。 |
| `cloudSessionSync` | `true`, `false` | `true` | CLIセッションをFactory webにミラーします。 |
Expand Down Expand Up @@ -71,10 +71,12 @@ droidを動かすデフォルトのAIモデルを選択します:
- **`haiku`** - Claude Haiku 4.5、高速で費用対効果の高いモデル
- **`gemini-3.1-pro`** - Gemini 3.1 Pro
- **`gemini-3-flash`** - Gemini 3 Flash、高速で安価(0.2×倍率)
- **`droid-core`** - GLM-4.7 オープンソースモデル
- **`droid-core`** - GLM-5.1 オープンソースモデル
- **`glm-5`** - GLM-5 オープンソースモデル
- **`glm-5.1`** - GLM-5.1 オープンソースモデル
- **`kimi-k2.5`** - Kimi K2.5 画像サポート付きオープンソースモデル
- **`minimax-m2.5`** - MiniMax M2.5 推論サポート付きオープンソースモデル(0.12×倍率)
- **`kimi-k2.6`** - Kimi K2.6 画像サポートとHigh推論切り替え付きオープンソースモデル
- **`minimax-m2.7`** - MiniMax M2.7 コスト重視のエージェントワークフロー向け推論サポート付きオープンソースモデル
- **`custom-model`** - BYOK経由で設定されたカスタムモデル

[You can also add custom models and BYOK.](/cli/configuration/byok)
Expand Down
15 changes: 9 additions & 6 deletions docs/jp/cli/droid-exec/overview.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -70,24 +70,27 @@ Options:

サポートされているモデル(例):

- claude-opus-4-7
- claude-opus-4-6
- claude-opus-4-6-fast
- claude-opus-4-5-20251101
- claude-sonnet-4-6
- claude-sonnet-4-5-20250929
- claude-haiku-4-5-20251001
- gpt-5.1-codex
- gpt-5.1-codex-max
- gpt-5.1
- gpt-5.2
- gpt-5.2-codex
- gpt-5.3-codex
- gemini-3-pro-preview
- gpt-5.3-codex-fast
- gpt-5.4
- gpt-5.4-fast
- gpt-5.4-mini
- gemini-3.1-pro-preview
- gemini-3-flash-preview
- glm-4.7
- glm-5
- glm-5.1
- kimi-k2.5
- minimax-m2.5
- kimi-k2.6
- minimax-m2.7

<Note>
See the [model table](/pricing#available-models) for the full list of available models.
Expand Down
Loading