Models

DeepSeek R1 Distill Llama 8B

A distilled 8B parameter model from the DeepSeek R1 series, optimized for fast and efficient reasoning tasks. It inherits strong reasoning capabilities from larger R1 models while being lightweight and cost-effective.

ReasoningFastCheap
Input / 1M tokens
$0.00
Output / 1M tokens
$0.00
Supported plans
0

Benchmark history

Evaluations

15

Lcr

Measured May 14, 2026Source

Score

0

Ifbench

Measured May 14, 2026Source

Score

0.18

Aime 25

Measured May 14, 2026Source

Score

0.41

Aime

Measured May 14, 2026Source

Score

0.33

Math 500

Measured May 14, 2026Source

Score

0.85

Scicode

Measured May 14, 2026Source

Score

0.12

Livecodebench

Measured May 14, 2026Source

Score

0.23

Hle

Measured May 14, 2026Source

Score

0.04

Gpqa

Measured May 14, 2026Source

Score

0.3

Mmlu Pro

Measured May 14, 2026Source

Score

0.54

Artificial Analysis Math Index

Measured May 14, 2026Source

Score

41.3

Artificial Analysis Intelligence Index

Measured May 14, 2026Source

Score

12.1

TAU2

Measured May 14, 2026Source

Score

0.37

Terminalbench Hard

Measured May 14, 2026Source

Score

0.16

Artificial Analysis Coding Index

Measured May 14, 2026Source

Score

24

Plan availability

Products and plans that support this model

0
No products or plans have been linked to this model yet.

Discussion

Thinking... Make sure you are connected to GitHub server