Google’s Gemini models bring two distinct capabilities to your DeepMask workspace: Gemini 2.5 Flash delivers high-throughput, low-cost multimodal processing at scale, while Gemini 2.5 Pro applies a native “Thinking” architecture to tackle complex reasoning, coding, and research tasks. Both models share a massive context window and full multimodal support for text, images, video, and audio.Documentation Index
Fetch the complete documentation index at: https://documentation.deepmask.io/llms.txt
Use this file to discover all available pages before exploring further.
- Gemini 2.5 Flash
- Gemini 2.5 Pro
About
Gemini 2.5 Flash is Google’s most efficient multimodal model, engineered for scale. It provides a massive 1-million-token context window at a fraction of the cost of Pro-tier models, and is specifically optimized for high-volume tasks such as real-time video summarization, large-scale document OCR, and high-speed data extraction. It is the most cost-effective way to process native audio and video inputs via API.Gemini 2.5 Flash is served via Google’s infrastructure. Your data is processed under DeepMask’s EU data-handling agreements.
Key Capabilities
Long-Context Retrieval
Maintains near-perfect accuracy (99%+) when finding specific data points across a million tokens.
Native Audio/Video Understanding
Processes video at 1 frame per second and audio at 16 kHz for high-fidelity temporal reasoning.
Context Caching
Store massive datasets — such as a 100-video training course — for rapid, cost-efficient recurring queries.
Real-Time Multimodal
Supports real-time, low-latency multimodal interactions for voice assistants and live monitoring pipelines.
Use Cases
- Real-time customer support — Power conversational bots that can understand user-uploaded screenshots or voice notes instantly.
- Large-scale document synthesis — Summarize hundreds of PDFs or hour-long meeting recordings in a single pass.
- Multimodal agents — Build assistants that can navigate your data across Gmail, Photos, and Workspace to perform complex cross-app tasks.
- High-speed data extraction — Process and reformat massive structured or semi-structured datasets with high throughput.
Specifications
| Specification | Value |
|---|---|
| Model Provider | |
| Main Use Cases | Data Extraction, Real-time Summarization, Large Codebase Search |
| Reasoning Effort | Adaptive (Balanced) |
| GPQA Diamond | 68.3% |
| Max Context | 1.04M Tokens |
| Latency (TTFT) | 0.15s |
| Throughput | 185 Tokens/sec |