Skip to main content

Documentation Index

Fetch the complete documentation index at: https://documentation.deepmask.io/llms.txt

Use this file to discover all available pages before exploring further.

DeepMask is built around the principle that no single model is best for every task. You can switch models at any point — including mid-conversation — without losing your context. Use the tabs below to find the right model for what you’re working on right now.
These models perform well on software engineering tasks including code generation, debugging, architecture design, and long-horizon autonomous development.
ModelWhy it fits
GPT-5.2 / 5.3 / 5.4Most capable OpenAI models; strong reasoning and tool use across all coding tasks
Sonnet 4.5 / 4.6Gold standard for autonomous coding; handles 30+ hour engineering sessions with 1M token context
Gemini 2.5 ProLarge context window suits large codebase analysis and multi-file refactors
Kimi K2 (DeepMask)Agent Swarm Mode enables 100 parallel sub-agents for complex, multi-step builds
DeepSeek V3671B MoE model with frontier-level coding and math; strong on STEM and security analysis
MiniMax M2 / M2.1Built specifically for elite multi-language coding and advanced agent workflows
You can switch models at any point in a conversation. If a response isn’t working well, select a different model from the model picker and continue your conversation — DeepMask carries your context forward automatically.

Model capability quick reference

Extended thinking (sometimes shown as “reasoning mode” or “thinking mode” in the UI) causes a model to work through a problem step-by-step before producing its final answer. The model generates an internal chain of reasoning that it uses to improve accuracy on complex tasks.Models in DeepMask that support extended or adaptive thinking include Sonnet 4.5 / 4.6, Haiku 4.5, Kimi K2 (DeepMask), GLM-4.7, and DeepSeek V3 (in its thinking mode). GPT-o3 Mini is also specifically optimized for reasoning tasks.Extended thinking increases response time but significantly improves results for graduate-level math, multi-step logic, planning, and any task where intermediate reasoning matters.
The following models in DeepMask can accept images as input and reason about their contents:
  • OpenAI: GPT-4o, GPT-4.1, GPT-5.2, GPT-5.3, GPT-5.4
  • Anthropic: Opus 4.5 / 4.6, Sonnet 4.5 / 4.6, Haiku 4.5
  • Google: Gemini 2.5 Pro, Gemini 2.5 Flash, Gemma 3 27B (StackIT)
  • MoonshotAI: Kimi K2 (DeepMask), Kimi K2.5 (via MoonViT multimodal)
  • Alibaba: Qwen (DeepMask), Qwen3 (StackIT)
  • Mistral: Mistral Large 3
Models that do not support image input include GPT-OSS 120B (both StackIT and Infercom variants), DeepSeek V3 / V3.1, GPT-o3 Mini, and MiniMax M2.5 (Infercom).
Tool use (also called function calling) lets a model invoke external tools, APIs, or data connectors during a conversation. DeepMask exposes tool use through its MCP (Model Context Protocol) connector framework, which supports Google Drive, Gmail, SharePoint, Salesforce, and more.Models with strong tool use support include:
  • Anthropic: Haiku 4.5 (95% success rate on complex JSON schemas), Sonnet 4.5 / 4.6, Opus 4.5 / 4.6
  • OpenAI: GPT-4o, GPT-4.1, GPT-5.x, GPT-o3 Mini
  • MoonshotAI: Kimi K2 (DeepMask) — maintains coherence across 300+ sequential tool calls
  • Google: Gemini 2.5 Pro, Gemini 2.5 Flash
  • Alibaba: Qwen (DeepMask), Qwen3 (StackIT)
  • DeepSeek: DeepSeek V3, DeepSeek V3.1 (Infercom)
  • MiniMax: M2, M2.1, M2.5 (Infercom)
  • Z.ai: GLM-4.7
Gemma 3 27B (StackIT) does not support tool use in DeepMask.
The context window is the maximum amount of text (measured in tokens, where 1 token ≈ 0.75 words) that a model can read and reason over in a single conversation. Larger context windows let you work with longer documents, more conversation history, and bigger codebases without losing earlier information.Context window sizes for key models in DeepMask:
ModelContext window
Kimi K2 (DeepMask)2,000,000 tokens
Sonnet 4.5 / 4.61,000,000 tokens
Gemini 2.5 Flash1,040,000 tokens
Gemini 2.5 Pro~1,000,000 tokens
Haiku 4.5200,000 tokens
DeepSeek V3 / V3.1128,000 – 164,000 tokens
MiniMax M2.5 (Infercom)164,000 tokens
GPT-4o128,000 tokens
GPT-4.1128,000 tokens
For very long documents or multi-session projects, prefer Kimi K2, the Sonnet series, or the Gemini 2.5 models. Use DeepMask Projects to persist files and instructions across sessions regardless of the model you choose.