Original Reddit post

Significant improvements over Flash model. Xiaomi’s MiMo V2 Pro is very competitive in performance vs Claude Opus 4.6 and provides 8x in output cost efficiency. MiMo V2 Pro vs Opus 4.6: Input: $1/M vs $5/M (5x cost difference) Output: $3 vs $25/M (8.3x cost difference) Both coming in at a million context window. MiMo V2 Pro vs Leading Models General Agentic Capabilities:

PinchBench (avg.): 81.0%; Nearly ties Claude Opus 4.6 (81.5) ClawEval: 61.5%; Competitive with Claude models (66.3) GDPVal-AA: 1426; Strong complex tool-use DeepSearch QA-F1: 86.7%; Competitive with Sonnet 4.6 (89.2) and Opus 4.6 (91.3) t2-bench: 96.8%; Extremely close to the 98–99 leaders Coding Agentic Capabilities: SWE-bench Verified: 78.0%; Very competitive with Claude Opus 4.6 (80.8%), GPT-5.2 (80.0%), and Sonnet 4.6 (79.6%) SWE-bench Multilingual: 71.7% Terminal-Bench 2.0: 57.1%; Strong production coding performance Xiaomi has also significantly improved on hallucination, V2 Pro has 30% vs V2 Flash of 48% in AA Omniscience. This model sits between GLM5 & Kimi K2.5 on Artificial Analysis Intelligence Index. Could be a great general alternative from these numbers it seems. Would love to see this open sourced like they did with their Flash model in December. submitted by /u/hexxthegon

Originally posted by u/hexxthegon on r/ArtificialInteligence