Update README.md
Browse files
README.md
CHANGED
|
@@ -16,10 +16,10 @@ library_name: transformers
|
|
| 16 |
## Introduction
|
| 17 |
VibeThinker-1.5B is a 1.5-billion parameter dense language model. With a total training cost of only $7,800 USD, it achieves reasoning performance comparable to larger models like GPT OSS-20B Medium.
|
| 18 |
|
| 19 |
-
, underscoring its strong reasoning performance.
|
| 25 |
|
|
|
|
| 16 |
## Introduction
|
| 17 |
VibeThinker-1.5B is a 1.5-billion parameter dense language model. With a total training cost of only $7,800 USD, it achieves reasoning performance comparable to larger models like GPT OSS-20B Medium.
|
| 18 |
|
| 19 |
+

|
| 20 |
|
| 21 |
## Key Performance Data
|
| 22 |
+
💡 Mathematical Reasoning: On the three major math benchmarks AIME24, AIME25, and HMMT25, its scores (80.3, 74.4, and 50.4, respectively) all surpass those of the initial DeepSeek R1 model, which has over 400 times the parameters (scores of 79.8, 70.0, and 41.7, respectively).
|
| 23 |
|
| 24 |
🌱 Code Generation: It achieved scores of 55.9 on LiveCodeBench v5 and 51.1 on v6. Its v6 score slightly leads Magistral Medium (50.3), underscoring its strong reasoning performance.
|
| 25 |
|