YinZhiBin commited on
Commit
202e3a3
·
verified ·
1 Parent(s): d2a112d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -5
README.md CHANGED
@@ -13,19 +13,23 @@ library_name: transformers
13
  ---
14
  # VibeThinker-1.5B
15
  ## Introduction
16
- VibeThinker-1.5B is a 1.5-billion parameter dense language model. With a total training cost of only $7,800 USD, it achieves reasoning performance on several challenging benchmarks that matches or even surpasses that of significantly larger models. Its core innovation lies in the "Spectrum-to-Signal Principle" (SSP) training framework: it first explores solution diversity during the Supervised Fine-Tuning (SFT) stage, and then optimizes its policy to reinforce correct signals in the Reinforcement Learning (RL) stage.
17
 
18
- ![image](https://cdn-uploads.huggingface.co/production/uploads/64d1faaa1ed6649d70d1fa2f/N3mi0ApHWnGVv4f8kHhLU.png)
19
 
20
- ## Key Performance Data:
21
  Mathematical Reasoning: On the three major math benchmarks AIME24, AIME25, and HMMT25, its scores (80.9, 73.7, and 50.1, respectively) all surpass those of the initial DeepSeek R1 model, which has over 400 times the parameters (scores of 79.8, 70.0, and 41.7, respectively).
22
 
23
- Code Generation: It achieved a score of 51.0 on LiveCodeBench v6, slightly ahead of Magistral Medium (50.3) and significantly outperforming the base model (0.0).
24
 
25
- ![image](https://cdn-uploads.huggingface.co/production/uploads/64d1faaa1ed6649d70d1fa2f/6MgmRUkEubiEv4mT_wslp.png)
26
 
27
  ![image](https://cdn-uploads.huggingface.co/production/uploads/64d1faaa1ed6649d70d1fa2f/iIUFCoDjct1eygznATikH.png)
28
 
 
 
 
 
 
29
  ## Highlights Summary
30
  💡 Breakthrough Scale-Efficiency: With an extremely small parameter count of 1.5B, it achieves reasoning performance comparable to larger models like GPT OSS-20B Medium.
31
 
 
13
  ---
14
  # VibeThinker-1.5B
15
  ## Introduction
16
+ VibeThinker-1.5B is a 1.5-billion parameter dense language model. With a total training cost of only $7,800 USD, it achieves reasoning performance on several challenging benchmarks that matches or even surpasses that of significantly larger models.
17
 
18
+ ![image](https://cdn-uploads.huggingface.co/production/uploads/64d1faaa1ed6649d70d1fa2f/6MgmRUkEubiEv4mT_wslp.png)
19
 
 
20
  Mathematical Reasoning: On the three major math benchmarks AIME24, AIME25, and HMMT25, its scores (80.9, 73.7, and 50.1, respectively) all surpass those of the initial DeepSeek R1 model, which has over 400 times the parameters (scores of 79.8, 70.0, and 41.7, respectively).
21
 
22
+ Code Generation: It achieved a score of 55.3 on LiveCodeBench v5.
23
 
24
+ On the AIME 25 benchmark, VibeThinker-1.5B significantly extends the Pareto frontier of reasoning accuracy versus model scale, demonstrating that exceptional performance can be achieved with extreme parameter efficiency.
25
 
26
  ![image](https://cdn-uploads.huggingface.co/production/uploads/64d1faaa1ed6649d70d1fa2f/iIUFCoDjct1eygznATikH.png)
27
 
28
+
29
+ Its core innovation lies in the "Spectrum-to-Signal Principle" (SSP) training framework: it first explores solution diversity during the Supervised Fine-Tuning (SFT) stage, and then optimizes its policy to reinforce correct signals in the Reinforcement Learning (RL) stage.
30
+
31
+ ![image](https://cdn-uploads.huggingface.co/production/uploads/64d1faaa1ed6649d70d1fa2f/N3mi0ApHWnGVv4f8kHhLU.png)
32
+
33
  ## Highlights Summary
34
  💡 Breakthrough Scale-Efficiency: With an extremely small parameter count of 1.5B, it achieves reasoning performance comparable to larger models like GPT OSS-20B Medium.
35