Models

InclusionAI

Ling-flash-2.0

Ling-flash-2.0 is a model optimized for low-latency and rapid response, designed for applications requiring quick turnaround. It is part of the 'flash' series, emphasizing speed and efficiency.

Fast
Input / 1M tokens
$0.14
Output / 1M tokens
$0.57
Output tokens/s
86.79
First-token seconds
1.38s
Supported plans
1

Benchmark history

Evaluations

13

TAU2

Measured May 14, 2026Source

Score

0.21

Terminalbench Hard

Measured May 14, 2026Source

Score

0.11

Lcr

Measured May 14, 2026Source

Score

0.15

Ifbench

Measured May 14, 2026Source

Score

0.34

Aime 25

Measured May 14, 2026Source

Score

0.65

Scicode

Measured May 14, 2026Source

Score

0.29

Livecodebench

Measured May 14, 2026Source

Score

0.59

Hle

Measured May 14, 2026Source

Score

0.06

Gpqa

Measured May 14, 2026Source

Score

0.66

Mmlu Pro

Measured May 14, 2026Source

Score

0.78

Artificial Analysis Math Index

Measured May 14, 2026Source

Score

65.3

Artificial Analysis Coding Index

Measured May 14, 2026Source

Score

16.7

Artificial Analysis Intelligence Index

Measured May 14, 2026Source

Score

15.7

Plan availability

Products and plans that support this model

1
ZenMux Builder

ZenMux Builder

ZenMux is a unified API gateway that provides access to over 100 AI models from providers like OpenAI, Anthropic, Google, and xAI through a single account and API. It offers subscription plans (Builder Plan) for personal development, vibe coding, and testing, as well as pay-as-you-go for production use. The platform supports coding agents like Claude Code and OpenClaw, and provides automatic model routing, multi-provider failover, and AI quality insurance compensation.

Discussion

Thinking... Make sure you are connected to GitHub server