AxionLab-official commited on
Commit
bc9bf49
·
verified ·
1 Parent(s): ec97241

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +109 -3
README.md CHANGED
@@ -1,3 +1,109 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - pt
5
+ library_name: transformers
6
+ ---
7
+
8
+ **Axion1.5-0.3B-Base**
9
+
10
+ 🧠 Axion1.5-0.3B-Base is a base language model with approximately 300 million parameters, trained purely for next-token prediction.
11
+
12
+ No instruction tuning.
13
+ No reinforcement learning.
14
+ No forced reasoning chains.
15
+
16
+ Just raw language modeling.
17
+
18
+ This model exists with a clear goal: to act as a clean, transparent baseline for future experiments focused on explicit reasoning and structured thinking.
19
+
20
+ What this model is
21
+
22
+ A foundation / base model
23
+
24
+ Trained only with next-token prediction
25
+
26
+ Not optimized for chat or instruction-following
27
+
28
+ Designed as a reference point for research and comparison
29
+
30
+ A functional “blank mind” before reasoning specialization
31
+
32
+ What this model is not
33
+
34
+ ❌ Not a chatbot
35
+
36
+ ❌ Not instruction-tuned
37
+
38
+ ❌ Not aligned for safety or helpfulness
39
+
40
+ ❌ Not optimized for long conversations
41
+
42
+ ❌ Not a reasoning model (yet)
43
+
44
+ If you are looking for a model that follows instructions or explains its thoughts, this is not it.
45
+
46
+ Why release a base model?
47
+
48
+ Releasing the base model publicly allows:
49
+
50
+ Transparent evaluation of raw language modeling quality
51
+
52
+ Fair comparison with future Axion reasoning variants
53
+
54
+ Reproducibility and honest benchmarking
55
+
56
+ A clear separation between language competence and reasoning behavior
57
+
58
+ Many projects hide their base models.
59
+ Axion does the opposite.
60
+
61
+ Intended use
62
+
63
+ Research and experimentation
64
+
65
+ Fine-tuning for instruction-following or reasoning tasks
66
+
67
+ Studying the effects of reasoning-oriented datasets
68
+
69
+ Serving as a backbone for Axion1.5-Reasoning variants
70
+
71
+ Limitations
72
+
73
+ Because this model is trained only for next-token prediction:
74
+
75
+ It may produce incoherent or incomplete responses
76
+
77
+ It does not reliably follow instructions
78
+
79
+ It does not reason step-by-step
80
+
81
+ It may hallucinate or contradict itself
82
+
83
+ These limitations are expected and acknowledged.
84
+
85
+ Future work
86
+
87
+ This release is part of a broader project:
88
+
89
+ Axion1.5-Reasoning – fine-tuned for structured reasoning
90
+
91
+ Axion-Critic – models focused on evaluation and self-critique
92
+
93
+ Experiments with short, verifiable reasoning traces
94
+
95
+ The base model will remain unchanged to preserve its value as a reference.
96
+
97
+ Philosophy
98
+
99
+ Scale is not intelligence.
100
+ Structure matters.
101
+
102
+ Axion explores whether smaller models, trained with the right constraints, can develop more meaningful reasoning behaviors.
103
+
104
+ This is an experiment.
105
+ And experiments are allowed to fail.
106
+
107
+ Acknowledgements
108
+
109
+ Created as an independent research project focused on understanding how reasoning emerges in language models.