lbourdois commited on
Commit
3c4a89f
·
verified ·
1 Parent(s): b9c8123

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +200 -195
README.md CHANGED
@@ -1,196 +1,201 @@
1
- ---
2
- language:
3
- - th
4
- - en
5
- library_name: transformers
6
- base_model:
7
- - Qwen/Qwen2.5-7B-Instruct
8
- - Qwen/Qwen2.5-7B
9
- pipeline_tag: text-generation
10
- model-index:
11
- - name: Tsunami-0.5-7B-Instruct
12
- results:
13
- - task:
14
- type: text-generation
15
- name: Text Generation
16
- dataset:
17
- name: IFEval (0-Shot)
18
- type: HuggingFaceH4/ifeval
19
- args:
20
- num_few_shot: 0
21
- metrics:
22
- - type: inst_level_strict_acc and prompt_level_strict_acc
23
- value: 74
24
- name: strict accuracy
25
- source:
26
- url: >-
27
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
28
- name: Open LLM Leaderboard
29
- - task:
30
- type: text-generation
31
- name: Text Generation
32
- dataset:
33
- name: BBH (3-Shot)
34
- type: BBH
35
- args:
36
- num_few_shot: 3
37
- metrics:
38
- - type: acc_norm
39
- value: 36.14
40
- name: normalized accuracy
41
- source:
42
- url: >-
43
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
44
- name: Open LLM Leaderboard
45
- - task:
46
- type: text-generation
47
- name: Text Generation
48
- dataset:
49
- name: MATH Lvl 5 (4-Shot)
50
- type: hendrycks/competition_math
51
- args:
52
- num_few_shot: 4
53
- metrics:
54
- - type: exact_match
55
- value: 0.15
56
- name: exact match
57
- source:
58
- url: >-
59
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
60
- name: Open LLM Leaderboard
61
- - task:
62
- type: text-generation
63
- name: Text Generation
64
- dataset:
65
- name: GPQA (0-shot)
66
- type: Idavidrein/gpqa
67
- args:
68
- num_few_shot: 0
69
- metrics:
70
- - type: acc_norm
71
- value: 7.83
72
- name: acc_norm
73
- source:
74
- url: >-
75
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
76
- name: Open LLM Leaderboard
77
- - task:
78
- type: text-generation
79
- name: Text Generation
80
- dataset:
81
- name: MuSR (0-shot)
82
- type: TAUR-Lab/MuSR
83
- args:
84
- num_few_shot: 0
85
- metrics:
86
- - type: acc_norm
87
- value: 12.21
88
- name: acc_norm
89
- source:
90
- url: >-
91
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
92
- name: Open LLM Leaderboard
93
- - task:
94
- type: text-generation
95
- name: Text Generation
96
- dataset:
97
- name: MMLU-PRO (5-shot)
98
- type: TIGER-Lab/MMLU-Pro
99
- config: main
100
- split: test
101
- args:
102
- num_few_shot: 5
103
- metrics:
104
- - type: acc
105
- value: 37.92
106
- name: accuracy
107
- source:
108
- url: >-
109
- https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
110
- name: Open LLM Leaderboard
111
- license: apache-2.0
112
- ---
113
-
114
- <img src="./Tsunami.webp" alt="Tsunami Model" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
115
-
116
- # Tsunami-0.5-7B-Instruct
117
- **TSUNAMI**: Transformative Semantic Understanding and Natural Augmentation Model for Intelligence.
118
-
119
- **TSUNAMI** full name was created by ChatGPT.
120
-
121
- ---
122
-
123
- ### infomation
124
- **Tsunami-0.5-7B-Instruct** is Thai Large Language Model that fine-tuned from **Qwen2.5-7B** around **60,000** rows in Thai-specific domain.
125
-
126
- ---
127
-
128
- ### Prompt Template
129
-
130
- This model uses `ChatML` prompt template:
131
-
132
- ```
133
- <|im_start|>system
134
- {System}<|im_end|>
135
- <|im_start|>user
136
- {User}<|im_end|>
137
- <|im_start|>assistant
138
- {Assistant}
139
- ````
140
-
141
- ### How to use
142
-
143
-
144
- ```python
145
-
146
- from transformers import AutoModelForCausalLM, AutoTokenizer
147
- import torch
148
-
149
- model_name = "Tsunami-th/Tsunami-0.5-7B-Instruct"
150
-
151
- model = AutoModelForCausalLM.from_pretrained(
152
- model_name,
153
- torch_dtype="auto",
154
- device_map="auto"
155
- )
156
- tokenizer = AutoTokenizer.from_pretrained(model_name)
157
-
158
- messages = [
159
- {"role": "system", "content": "You are a helpful assistant."},
160
- {"role": "user", "content": "สวัสดีครับ"}
161
- ]
162
- text = tokenizer.apply_chat_template(
163
- messages,
164
- tokenize=False,
165
- add_generation_prompt=True
166
- )
167
-
168
- inputs = tokenizer(text, return_tensors="pt")
169
- inputs = inputs.to(model.device)
170
- with torch.no_grad():
171
- output = model.generate(**inputs, max_new_tokens=512)
172
-
173
- response = tokenizer.decode(output[0, len(inputs['input_ids'][0]):], skip_special_tokens=True)
174
- ```
175
-
176
- ---
177
-
178
- ### Author
179
- - Pollakrit Lorprasertkul | game.pollakrit@gmail.com
180
-
181
- ---
182
-
183
- - **Tsunami-0.5-7B-Instruct** is the version 0.5 that did not train on the whole dataset.
184
- - **Tsunami-1.0-7B-Instruct** is coming soon.
185
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
186
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Tsunami-th__Tsunami-0.5-7B-Instruct)
187
-
188
- | Metric |Value|
189
- |-------------------|----:|
190
- |Avg. |28.04|
191
- |IFEval (0-Shot) |74.00|
192
- |BBH (3-Shot) |36.14|
193
- |MATH Lvl 5 (4-Shot)| 0.15|
194
- |GPQA (0-shot) | 7.83|
195
- |MuSR (0-shot) |12.21|
 
 
 
 
 
196
  |MMLU-PRO (5-shot) |37.92|
 
1
+ ---
2
+ language:
3
+ - zho
4
+ - eng
5
+ - fra
6
+ - spa
7
+ - por
8
+ - deu
9
+ - ita
10
+ - rus
11
+ - jpn
12
+ - kor
13
+ - vie
14
+ - tha
15
+ - ara
16
+ library_name: transformers
17
+ base_model:
18
+ - Qwen/Qwen2.5-7B-Instruct
19
+ - Qwen/Qwen2.5-7B
20
+ pipeline_tag: text-generation
21
+ license: apache-2.0
22
+ model-index:
23
+ - name: Tsunami-0.5-7B-Instruct
24
+ results:
25
+ - task:
26
+ type: text-generation
27
+ name: Text Generation
28
+ dataset:
29
+ name: IFEval (0-Shot)
30
+ type: HuggingFaceH4/ifeval
31
+ args:
32
+ num_few_shot: 0
33
+ metrics:
34
+ - type: inst_level_strict_acc and prompt_level_strict_acc
35
+ value: 74
36
+ name: strict accuracy
37
+ source:
38
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
39
+ name: Open LLM Leaderboard
40
+ - task:
41
+ type: text-generation
42
+ name: Text Generation
43
+ dataset:
44
+ name: BBH (3-Shot)
45
+ type: BBH
46
+ args:
47
+ num_few_shot: 3
48
+ metrics:
49
+ - type: acc_norm
50
+ value: 36.14
51
+ name: normalized accuracy
52
+ source:
53
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
54
+ name: Open LLM Leaderboard
55
+ - task:
56
+ type: text-generation
57
+ name: Text Generation
58
+ dataset:
59
+ name: MATH Lvl 5 (4-Shot)
60
+ type: hendrycks/competition_math
61
+ args:
62
+ num_few_shot: 4
63
+ metrics:
64
+ - type: exact_match
65
+ value: 0.15
66
+ name: exact match
67
+ source:
68
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
69
+ name: Open LLM Leaderboard
70
+ - task:
71
+ type: text-generation
72
+ name: Text Generation
73
+ dataset:
74
+ name: GPQA (0-shot)
75
+ type: Idavidrein/gpqa
76
+ args:
77
+ num_few_shot: 0
78
+ metrics:
79
+ - type: acc_norm
80
+ value: 7.83
81
+ name: acc_norm
82
+ source:
83
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
84
+ name: Open LLM Leaderboard
85
+ - task:
86
+ type: text-generation
87
+ name: Text Generation
88
+ dataset:
89
+ name: MuSR (0-shot)
90
+ type: TAUR-Lab/MuSR
91
+ args:
92
+ num_few_shot: 0
93
+ metrics:
94
+ - type: acc_norm
95
+ value: 12.21
96
+ name: acc_norm
97
+ source:
98
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
99
+ name: Open LLM Leaderboard
100
+ - task:
101
+ type: text-generation
102
+ name: Text Generation
103
+ dataset:
104
+ name: MMLU-PRO (5-shot)
105
+ type: TIGER-Lab/MMLU-Pro
106
+ config: main
107
+ split: test
108
+ args:
109
+ num_few_shot: 5
110
+ metrics:
111
+ - type: acc
112
+ value: 37.92
113
+ name: accuracy
114
+ source:
115
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5-7B-Instruct
116
+ name: Open LLM Leaderboard
117
+ ---
118
+
119
+ <img src="./Tsunami.webp" alt="Tsunami Model" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
120
+
121
+ # Tsunami-0.5-7B-Instruct
122
+ **TSUNAMI**: Transformative Semantic Understanding and Natural Augmentation Model for Intelligence.
123
+
124
+ **TSUNAMI** full name was created by ChatGPT.
125
+
126
+ ---
127
+
128
+ ### infomation
129
+ **Tsunami-0.5-7B-Instruct** is Thai Large Language Model that fine-tuned from **Qwen2.5-7B** around **60,000** rows in Thai-specific domain.
130
+
131
+ ---
132
+
133
+ ### Prompt Template
134
+
135
+ This model uses `ChatML` prompt template:
136
+
137
+ ```
138
+ <|im_start|>system
139
+ {System}<|im_end|>
140
+ <|im_start|>user
141
+ {User}<|im_end|>
142
+ <|im_start|>assistant
143
+ {Assistant}
144
+ ````
145
+
146
+ ### How to use
147
+
148
+
149
+ ```python
150
+
151
+ from transformers import AutoModelForCausalLM, AutoTokenizer
152
+ import torch
153
+
154
+ model_name = "Tsunami-th/Tsunami-0.5-7B-Instruct"
155
+
156
+ model = AutoModelForCausalLM.from_pretrained(
157
+ model_name,
158
+ torch_dtype="auto",
159
+ device_map="auto"
160
+ )
161
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
162
+
163
+ messages = [
164
+ {"role": "system", "content": "You are a helpful assistant."},
165
+ {"role": "user", "content": "สวัสดีครับ"}
166
+ ]
167
+ text = tokenizer.apply_chat_template(
168
+ messages,
169
+ tokenize=False,
170
+ add_generation_prompt=True
171
+ )
172
+
173
+ inputs = tokenizer(text, return_tensors="pt")
174
+ inputs = inputs.to(model.device)
175
+ with torch.no_grad():
176
+ output = model.generate(**inputs, max_new_tokens=512)
177
+
178
+ response = tokenizer.decode(output[0, len(inputs['input_ids'][0]):], skip_special_tokens=True)
179
+ ```
180
+
181
+ ---
182
+
183
+ ### Author
184
+ - Pollakrit Lorprasertkul | game.pollakrit@gmail.com
185
+
186
+ ---
187
+
188
+ - **Tsunami-0.5-7B-Instruct** is the version 0.5 that did not train on the whole dataset.
189
+ - **Tsunami-1.0-7B-Instruct** is coming soon.
190
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
191
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Tsunami-th__Tsunami-0.5-7B-Instruct)
192
+
193
+ | Metric |Value|
194
+ |-------------------|----:|
195
+ |Avg. |28.04|
196
+ |IFEval (0-Shot) |74.00|
197
+ |BBH (3-Shot) |36.14|
198
+ |MATH Lvl 5 (4-Shot)| 0.15|
199
+ |GPQA (0-shot) | 7.83|
200
+ |MuSR (0-shot) |12.21|
201
  |MMLU-PRO (5-shot) |37.92|