Update README.md
Browse files
README.md
CHANGED
|
@@ -1,29 +1,13 @@
|
|
| 1 |
---
|
| 2 |
-
library_name: transformers
|
| 3 |
license: apache-2.0
|
| 4 |
datasets:
|
| 5 |
-
-
|
| 6 |
base_model:
|
| 7 |
- PrimeIntellect/Qwen3-0.6B
|
| 8 |
---
|
| 9 |
|
| 10 |
-
#
|
| 11 |
|
| 12 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 13 |
|
| 14 |
-
A
|
| 15 |
-
|
| 16 |
-
Created with this training command from [prime-rl](https://github.com/PrimeIntellect-ai/prime-rl) (commit hash: `8262560`)
|
| 17 |
-
|
| 18 |
-
```bash
|
| 19 |
-
uv run torchrun --nproc-per-node 8 src/prime_rl/trainer/sft/train.py \
|
| 20 |
-
--model.name PrimeIntellect/Qwen3-0.6B \
|
| 21 |
-
--data.name willcb/R1-reverse-wikipedia-paragraphs-v1-1000 \
|
| 22 |
-
--max-steps 100 \
|
| 23 |
-
--data.batch-size 16 \
|
| 24 |
-
--data.micro-batch-size 1 \
|
| 25 |
-
--data.seq-len 4096 \
|
| 26 |
-
--optim.lr 2e-5
|
| 27 |
-
```
|
| 28 |
-
|
| 29 |
-
Check the run out on [W&B](https://wandb.ai/primeintellect/mika/runs/odsfiekx?nw=nwusermikasenghaas_).
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
license: apache-2.0
|
| 3 |
datasets:
|
| 4 |
+
- willcb/V3-wordle
|
| 5 |
base_model:
|
| 6 |
- PrimeIntellect/Qwen3-0.6B
|
| 7 |
---
|
| 8 |
|
| 9 |
+
# Qwen-0.6B-Reverse-Text-SFT
|
| 10 |
|
| 11 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 12 |
|
| 13 |
+
A SFT fine-tune of `PrimeIntellect/Qwen-0.6B`. Details [here](https://github.com/PrimeIntellect-ai/prime-rl/tree/main/examples/reverse_text).
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|