lbourdois commited on
Commit
c12db51
·
verified ·
1 Parent(s): f86045d

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +103 -90
README.md CHANGED
@@ -1,90 +1,103 @@
1
- ---
2
- base_model:
3
- - Bllossom/llama-3.2-Korean-Bllossom-3B
4
- - Qwen/Qwen2.5-3B
5
- - ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
6
- - meta-llama/Llama-3.2-3B
7
- - Qwen/Qwen2.5-3B-Instruct
8
- - chuanli11/Llama-3.2-3B-Instruct-uncensored
9
- - ibm-granite/granite-3b-code-base-2k
10
- - stabilityai/stable-code-3b
11
- - PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
12
- - meta-llama/Llama-3.2-3B-Instruct
13
- library_name: transformers
14
- tags:
15
- - mergekit
16
- - merge
17
-
18
- ---
19
- # merge
20
-
21
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
22
-
23
- ## Merge Details
24
- ### Merge Method
25
-
26
- This model was merged using the passthrough merge method using [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B) as a base.
27
-
28
- ### Models Merged
29
-
30
- The following models were included in the merge:
31
- * [Bllossom/llama-3.2-Korean-Bllossom-3B](https://huggingface.co/Bllossom/llama-3.2-Korean-Bllossom-3B)
32
- * [Qwen/Qwen2.5-3B](https://huggingface.co/Qwen/Qwen2.5-3B)
33
- * [ICEPVP8977/Uncensored_llama_3.2_3b_safetensors](https://huggingface.co/ICEPVP8977/Uncensored_llama_3.2_3b_safetensors)
34
- * [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct)
35
- * [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
36
- * [ibm-granite/granite-3b-code-base-2k](https://huggingface.co/ibm-granite/granite-3b-code-base-2k)
37
- * [stabilityai/stable-code-3b](https://huggingface.co/stabilityai/stable-code-3b)
38
- * [PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B](https://huggingface.co/PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B)
39
- * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
40
-
41
- ### Configuration
42
-
43
- The following YAML configuration was used to produce this model:
44
-
45
- ```yaml
46
- slices:
47
- - sources:
48
- - layer_range: [0, 1]
49
- model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
50
-
51
- - sources:
52
- - layer_range: [0, 1]
53
- model: meta-llama/Llama-3.2-3B-Instruct
54
-
55
- - sources:
56
- - layer_range: [0, 1]
57
- model: meta-llama/Llama-3.2-3B
58
-
59
- - sources:
60
- - layer_range: [0, 1]
61
- model: chuanli11/Llama-3.2-3B-Instruct-uncensored
62
-
63
- - sources:
64
- - layer_range: [0, 1]
65
- model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
66
-
67
- - sources:
68
- - layer_range: [0, 1]
69
- model: Bllossom/llama-3.2-Korean-Bllossom-3B
70
-
71
- - sources:
72
- - layer_range: [0, 1]
73
- model: Qwen/Qwen2.5-3B-Instruct
74
-
75
- - sources:
76
- - layer_range: [0, 1]
77
- model: Qwen/Qwen2.5-3B
78
-
79
- - sources:
80
- - layer_range: [0, 1]
81
- model: stabilityai/stable-code-3b
82
-
83
- - sources:
84
- - layer_range: [0, 1]
85
- model: ibm-granite/granite-3b-code-base-2k
86
-
87
- merge_method: passthrough
88
- dtype: float16
89
- base_model: meta-llama/Llama-3.2-3B
90
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Bllossom/llama-3.2-Korean-Bllossom-3B
4
+ - Qwen/Qwen2.5-3B
5
+ - ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
6
+ - meta-llama/Llama-3.2-3B
7
+ - Qwen/Qwen2.5-3B-Instruct
8
+ - chuanli11/Llama-3.2-3B-Instruct-uncensored
9
+ - ibm-granite/granite-3b-code-base-2k
10
+ - stabilityai/stable-code-3b
11
+ - PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
12
+ - meta-llama/Llama-3.2-3B-Instruct
13
+ library_name: transformers
14
+ tags:
15
+ - mergekit
16
+ - merge
17
+ language:
18
+ - zho
19
+ - eng
20
+ - fra
21
+ - spa
22
+ - por
23
+ - deu
24
+ - ita
25
+ - rus
26
+ - jpn
27
+ - kor
28
+ - vie
29
+ - tha
30
+ - ara
31
+ ---
32
+ # merge
33
+
34
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
35
+
36
+ ## Merge Details
37
+ ### Merge Method
38
+
39
+ This model was merged using the passthrough merge method using [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B) as a base.
40
+
41
+ ### Models Merged
42
+
43
+ The following models were included in the merge:
44
+ * [Bllossom/llama-3.2-Korean-Bllossom-3B](https://huggingface.co/Bllossom/llama-3.2-Korean-Bllossom-3B)
45
+ * [Qwen/Qwen2.5-3B](https://huggingface.co/Qwen/Qwen2.5-3B)
46
+ * [ICEPVP8977/Uncensored_llama_3.2_3b_safetensors](https://huggingface.co/ICEPVP8977/Uncensored_llama_3.2_3b_safetensors)
47
+ * [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct)
48
+ * [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
49
+ * [ibm-granite/granite-3b-code-base-2k](https://huggingface.co/ibm-granite/granite-3b-code-base-2k)
50
+ * [stabilityai/stable-code-3b](https://huggingface.co/stabilityai/stable-code-3b)
51
+ * [PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B](https://huggingface.co/PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B)
52
+ * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
53
+
54
+ ### Configuration
55
+
56
+ The following YAML configuration was used to produce this model:
57
+
58
+ ```yaml
59
+ slices:
60
+ - sources:
61
+ - layer_range: [0, 1]
62
+ model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
63
+
64
+ - sources:
65
+ - layer_range: [0, 1]
66
+ model: meta-llama/Llama-3.2-3B-Instruct
67
+
68
+ - sources:
69
+ - layer_range: [0, 1]
70
+ model: meta-llama/Llama-3.2-3B
71
+
72
+ - sources:
73
+ - layer_range: [0, 1]
74
+ model: chuanli11/Llama-3.2-3B-Instruct-uncensored
75
+
76
+ - sources:
77
+ - layer_range: [0, 1]
78
+ model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
79
+
80
+ - sources:
81
+ - layer_range: [0, 1]
82
+ model: Bllossom/llama-3.2-Korean-Bllossom-3B
83
+
84
+ - sources:
85
+ - layer_range: [0, 1]
86
+ model: Qwen/Qwen2.5-3B-Instruct
87
+
88
+ - sources:
89
+ - layer_range: [0, 1]
90
+ model: Qwen/Qwen2.5-3B
91
+
92
+ - sources:
93
+ - layer_range: [0, 1]
94
+ model: stabilityai/stable-code-3b
95
+
96
+ - sources:
97
+ - layer_range: [0, 1]
98
+ model: ibm-granite/granite-3b-code-base-2k
99
+
100
+ merge_method: passthrough
101
+ dtype: float16
102
+ base_model: meta-llama/Llama-3.2-3B
103
+ ```