Choosing the right Mira model

Mira offers a family of four models, each optimized for specific workloads. This guide helps you pick the right model based on your performance, context, and budget requirements.

Model comparison matrix

ModelContextMax outputSpeedCost
mira32K4KFast$
mira-pro64K8KMedium$$
mira-max128K16KSlower$$$

When to use each model

mira

The general-purpose workhorse for everyday tasks. Ideal for chatbots, question answering, text summarization, translation, and short code generation. The fastest and most cost-effective model in the family.

  • Best forsimple Q&A, summarization, translation, short code snippets
  • Context32K tokens (~50 pages of text)
  • Typical latency< 1 second to first token

mira-pro

A professional-grade model with an expanded context window and built-in thinking mode. Built for enterprise applications that demand large document processing, deep reasoning, and high accuracy.

  • Best forlong document analysis, enterprise applications, detailed code review
  • Context64K tokens (~100 pages of text)
  • Advantagebalances capability and cost for professional use

mira-max

The most powerful model with a maximum context window of 128K tokens. Designed for tasks requiring processing of entire codebases, lengthy legal documents, or complex multi-faceted analysis.

  • Best forentire codebases, legal documents, research papers, complex analysis
  • Context128K tokens (~200 pages of text)
  • Outputup to 16K tokens — the longest output of any model

Decision flowchart

Use this text-based flowchart to quickly identify the right model:

Model selection flowchart
1. Is your input > 64K tokens?
   ├─ Yes → mira-max (includes thinking mode)
   └─ No  → go to step 2

2. Is your input > 32K tokens OR do you need deep reasoning / max accuracy?
   ├─ Yes → mira-pro (includes thinking mode)
   └─ No  → mira (best cost-performance ratio)

Task-to-model matrix

Task typeRecommendationWhy
Chatbot / FAQmiraFast, cheap, accurate enough
Math problemsmira-proThinking mode improves accuracy
Code review (large PR)mira-pro64K context fits large diffs
Codebase analysismira-max128K context for entire projects
TranslationmiraGreat quality at low cost
Legal analysismira-maxLong context + high accuracy
Test generationmira-proUnderstands full project context

Cost vs performance trade-offs

Model costs scale with their capabilities. For most applications, we recommend starting with mira and upgrading only when output quality is insufficient.

Tip: use model routing — send simple queries to mira and complex ones to mira-pro or mira-max. This reduces average cost by 40-60%.

Migrating between models

All Mira models share the same API format. To switch between models, simply change the model parameter in your request. Prompts, tools, and system messages remain compatible.

Switching models
// Simply change the model parameter
const response = await fetch("https://api.vmira.ai/v1/chat/completions", {
  method: "POST",
  headers: {
    "Authorization": "Bearer sk-mira-YOUR_KEY",
    "Content-Type": "application/json",
  },
  body: JSON.stringify({
    model: "mira-pro",  // was "mira" — just change this line
    messages: [{ role: "user", content: "Analyze this document..." }],
  }),
});