Alibaba’s Qwen3 models bring frontier-class reasoning and repository-scale coding to DeepMask in two deployment configurations. Qwen (DeepMask) is the 235B flagship with dual-mode inference, spatial-visual logic, and a 1M-token context window. Qwen3 (StackIT) is a StackIT-tuned variant co-developed for European cloud environments, with native infrastructure awareness and the same powerful reasoning core. Both models excel at multilingual tasks, document analysis, and agentic workflows.Documentation Index
Fetch the complete documentation index at: https://documentation.deepmask.io/llms.txt
Use this file to discover all available pages before exploring further.
- Qwen (DeepMask)
- Qwen3 (StackIT)
About
Qwen (DeepMask) is Alibaba’s 235B Qwen3 flagship model. It features Dual-Mode Inference, allowing you to toggle between “Instant” mode for fast chat and “Thinking” mode for deep, PhD-level problem solving. It leads on repository-scale coding — able to reason across tens of thousands of lines of code without context drift — and supports a 1M-token context window with efficient hardware use via a tiered KV cache.Qwen (DeepMask) is hosted on DeepMask infrastructure with a 1M-token context window. Your data remains within DeepMask’s EU-compliant environment.
Key Capabilities
Dual-Mode Reasoning
Toggles between fast chat and deep Thinking mode for complex, multi-step problem solving.
Spatial-Visual Logic
Excels at understanding complex diagrams, maps, technical blueprints, and spatial relationships.
1M Token Context
Handles up to 1 million tokens, enabling analysis of very large codebases and document sets.
Repository-Scale Coding
Understands the architectural intent behind a codebase, enabling whole-repo reasoning and refactoring.
Use Cases
- Enterprise software architecture — Plan and refactor multi-repository backend systems with full structural awareness.
- Global fintech analytics — Process large volumes of financial data for predictive market analysis.
- Creative design suite — Leverage native support for high-fidelity image understanding and natural speech tasks.
- Multilingual RAG — Build retrieval-augmented generation pipelines across multiple languages with strong reasoning.
Specifications
| Specification | Value |
|---|---|
| Model Provider | Alibaba |
| Main Use Cases | Agents Coding, Multilingual RAG |
| Reasoning Effort | High (Instant & Thinking) |
| GPQA Diamond | 89.3% |
| Max Context | 1M Tokens |
| Latency (TTFT) | 0.22s (Non-Thinking Mode) |
| Throughput | 145 Tokens/sec |