LLM Comparison
DeepSeek V3.2 Exp vs MiniCPM
Side-by-side specs, pricing & capabilities · Updated May 2026
Add to comparison
2/6 modelsSame tier:
| Organization | ||
| OpenTools Score | 40 116 | |
| Family | DeepSeek | MiniCPM |
| Status | Current | Current |
| Release Date | Sep 2025 | — |
| Context Window | 164K tokens | 128K tokens |
| Input Price | $0.27/M tokens | Free |
| Output Price | $0.41/M tokens | Free |
| Pricing Notes | — | Open-weight GitHub and Hugging Face model family. There is no fixed vendor API price; runtime cost depends on the host, hardware, or inference provider. |
| Capabilities | textcode | textcodereasoninglocal-inference |
| Training Cutoff | — | Not publicly specified in queued source |
| Max Output | 66K tokens | 33K tokens |
| API Identifier | deepseek/deepseek-v3.2-exp | OpenBMB/MiniCPM |
| Benchmarks | ||
| MMLU-Pro | 85deepseek | — |
| GPQA Diamond | 79.9deepseek | — |
| AIME 2025 | 89deepseek | — |
| HMMT 2025 | 84deepseek | — |
| SWE-bench Verified | 67.8deepseek | — |
| Terminal-Bench | 37.7deepseek | — |
| BrowseComp | 40.1deepseek | — |
| MiniCPM-SALA standard benchmark average | — | 76.53official-github-readme |
| MiniCPM-SALA long-context average | — | 38.97official-github-readme |
| MiniCPM-SALA 2048K extrapolation score | — | 81.6official-github-readme |
| MiniCPM4.1 reasoning decoding speedup | — | 3official-github-readme |
| MiniCPM4 Jetson AGX Orin decoding speedup vs Qwen3-8B | — | 7official-github-readme |
| View DeepSeek V3.2 Exp | View MiniCPM | |
Cost Calculator
Enter your expected monthly token usage to compare costs.
| Model | Input | Output | Total / mo | vs Best |
|---|---|---|---|---|
| MiniCPMCheapest | $0.00 | $0.00 | $0.00 | — |
| DeepSeek V3.2 Exp | $0.27 | $0.21 | $0.48 | +0% |
DeepSeek
DeepSeek V3.2 Exp
DeepSeek V3.2 Exp is a large language model from DeepSeek. Supports up to 163,840 token context window. Achieves 87.1% on MMLU. Available from $0.27/M input tokens.
OpenBMB
MiniCPM
MiniCPM is OpenBMB’s ultra-efficient open language-model family for edge and end-device deployment. The MiniCPM4 and MiniCPM4.1 lines focus on fast local reasoning, while MiniCPM-SALA extends the family toward sparse/linear attention and million-token context research.
More Comparisons
Looking for more AI models?
Browse All LLMs