DeepMask is built around the principle that no single model is best for every task. You can switch models at any point — including mid-conversation — without losing your context. Use the tabs below to find the right model for what you’re working on right now.Documentation Index
Fetch the complete documentation index at: https://documentation.deepmask.io/llms.txt
Use this file to discover all available pages before exploring further.
- Coding & Engineering
- Research & Analysis
- Writing & Marketing
- Fast & Lightweight
- Reasoning
- EU-hosted only
These models perform well on software engineering tasks including code generation, debugging, architecture design, and long-horizon autonomous development.
| Model | Why it fits |
|---|---|
| GPT-5.2 / 5.3 / 5.4 | Most capable OpenAI models; strong reasoning and tool use across all coding tasks |
| Sonnet 4.5 / 4.6 | Gold standard for autonomous coding; handles 30+ hour engineering sessions with 1M token context |
| Gemini 2.5 Pro | Large context window suits large codebase analysis and multi-file refactors |
| Kimi K2 (DeepMask) | Agent Swarm Mode enables 100 parallel sub-agents for complex, multi-step builds |
| DeepSeek V3 | 671B MoE model with frontier-level coding and math; strong on STEM and security analysis |
| MiniMax M2 / M2.1 | Built specifically for elite multi-language coding and advanced agent workflows |
Model capability quick reference
What does 'extended thinking' mean?
What does 'extended thinking' mean?
Extended thinking (sometimes shown as “reasoning mode” or “thinking mode” in the UI) causes a model to work through a problem step-by-step before producing its final answer. The model generates an internal chain of reasoning that it uses to improve accuracy on complex tasks.Models in DeepMask that support extended or adaptive thinking include Sonnet 4.5 / 4.6, Haiku 4.5, Kimi K2 (DeepMask), GLM-4.7, and DeepSeek V3 (in its thinking mode). GPT-o3 Mini is also specifically optimized for reasoning tasks.Extended thinking increases response time but significantly improves results for graduate-level math, multi-step logic, planning, and any task where intermediate reasoning matters.
Which models support image analysis?
Which models support image analysis?
The following models in DeepMask can accept images as input and reason about their contents:
- OpenAI: GPT-4o, GPT-4.1, GPT-5.2, GPT-5.3, GPT-5.4
- Anthropic: Opus 4.5 / 4.6, Sonnet 4.5 / 4.6, Haiku 4.5
- Google: Gemini 2.5 Pro, Gemini 2.5 Flash, Gemma 3 27B (StackIT)
- MoonshotAI: Kimi K2 (DeepMask), Kimi K2.5 (via MoonViT multimodal)
- Alibaba: Qwen (DeepMask), Qwen3 (StackIT)
- Mistral: Mistral Large 3
Which models support tool use and MCP?
Which models support tool use and MCP?
Tool use (also called function calling) lets a model invoke external tools, APIs, or data connectors during a conversation. DeepMask exposes tool use through its MCP (Model Context Protocol) connector framework, which supports Google Drive, Gmail, SharePoint, Salesforce, and more.Models with strong tool use support include:
- Anthropic: Haiku 4.5 (95% success rate on complex JSON schemas), Sonnet 4.5 / 4.6, Opus 4.5 / 4.6
- OpenAI: GPT-4o, GPT-4.1, GPT-5.x, GPT-o3 Mini
- MoonshotAI: Kimi K2 (DeepMask) — maintains coherence across 300+ sequential tool calls
- Google: Gemini 2.5 Pro, Gemini 2.5 Flash
- Alibaba: Qwen (DeepMask), Qwen3 (StackIT)
- DeepSeek: DeepSeek V3, DeepSeek V3.1 (Infercom)
- MiniMax: M2, M2.1, M2.5 (Infercom)
- Z.ai: GLM-4.7
What is context window size?
What is context window size?
The context window is the maximum amount of text (measured in tokens, where 1 token ≈ 0.75 words) that a model can read and reason over in a single conversation. Larger context windows let you work with longer documents, more conversation history, and bigger codebases without losing earlier information.Context window sizes for key models in DeepMask:
For very long documents or multi-session projects, prefer Kimi K2, the Sonnet series, or the Gemini 2.5 models. Use DeepMask Projects to persist files and instructions across sessions regardless of the model you choose.
| Model | Context window |
|---|---|
| Kimi K2 (DeepMask) | 2,000,000 tokens |
| Sonnet 4.5 / 4.6 | 1,000,000 tokens |
| Gemini 2.5 Flash | 1,040,000 tokens |
| Gemini 2.5 Pro | ~1,000,000 tokens |
| Haiku 4.5 | 200,000 tokens |
| DeepSeek V3 / V3.1 | 128,000 – 164,000 tokens |
| MiniMax M2.5 (Infercom) | 164,000 tokens |
| GPT-4o | 128,000 tokens |
| GPT-4.1 | 128,000 tokens |