Upload folder using huggingface_hub
Browse files- README.md +3 -2
- model.safetensors +2 -2
README.md
CHANGED
|
@@ -15,8 +15,8 @@ This tiny model is intended for debugging. It is randomly initialized using the
|
|
| 15 |
|
| 16 |
```python
|
| 17 |
import torch
|
| 18 |
-
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
| 19 |
import transformers
|
|
|
|
| 20 |
transformers.utils.import_utils.is_torch_fx_available = transformers.utils.import_utils.is_torch_available
|
| 21 |
|
| 22 |
model_id = "tiny-random/bailing-moe-v2.5"
|
|
@@ -41,6 +41,7 @@ from pathlib import Path
|
|
| 41 |
|
| 42 |
import accelerate
|
| 43 |
import torch
|
|
|
|
| 44 |
from huggingface_hub import file_exists, hf_hub_download
|
| 45 |
from transformers import (
|
| 46 |
AutoConfig,
|
|
@@ -49,7 +50,6 @@ from transformers import (
|
|
| 49 |
GenerationConfig,
|
| 50 |
set_seed,
|
| 51 |
)
|
| 52 |
-
import transformers
|
| 53 |
transformers.utils.import_utils.is_torch_fx_available = transformers.utils.import_utils.is_torch_available
|
| 54 |
source_model_id = "inclusionAI/Ring-2.5-1T"
|
| 55 |
save_folder = "/tmp/tiny-random/bailing-moe-v25"
|
|
@@ -98,6 +98,7 @@ with torch.no_grad():
|
|
| 98 |
for name, p in sorted(model.named_parameters()):
|
| 99 |
torch.nn.init.normal_(p, 0, 0.1)
|
| 100 |
print(name, p.shape)
|
|
|
|
| 101 |
model.save_pretrained(save_folder)
|
| 102 |
print(model)
|
| 103 |
with open(f"{save_folder}/config.json", "r", encoding='utf-8') as f:
|
|
|
|
| 15 |
|
| 16 |
```python
|
| 17 |
import torch
|
|
|
|
| 18 |
import transformers
|
| 19 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
| 20 |
transformers.utils.import_utils.is_torch_fx_available = transformers.utils.import_utils.is_torch_available
|
| 21 |
|
| 22 |
model_id = "tiny-random/bailing-moe-v2.5"
|
|
|
|
| 41 |
|
| 42 |
import accelerate
|
| 43 |
import torch
|
| 44 |
+
import transformers
|
| 45 |
from huggingface_hub import file_exists, hf_hub_download
|
| 46 |
from transformers import (
|
| 47 |
AutoConfig,
|
|
|
|
| 50 |
GenerationConfig,
|
| 51 |
set_seed,
|
| 52 |
)
|
|
|
|
| 53 |
transformers.utils.import_utils.is_torch_fx_available = transformers.utils.import_utils.is_torch_available
|
| 54 |
source_model_id = "inclusionAI/Ring-2.5-1T"
|
| 55 |
save_folder = "/tmp/tiny-random/bailing-moe-v25"
|
|
|
|
| 98 |
for name, p in sorted(model.named_parameters()):
|
| 99 |
torch.nn.init.normal_(p, 0, 0.1)
|
| 100 |
print(name, p.shape)
|
| 101 |
+
model.model.layers[1].mlp.gate.expert_bias = model.model.layers[1].mlp.gate.expert_bias.float()
|
| 102 |
model.save_pretrained(save_folder)
|
| 103 |
print(model)
|
| 104 |
with open(f"{save_folder}/config.json", "r", encoding='utf-8') as f:
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:19a618643f76a64b5b7b204ca792355fa0c8c161e7586981e55d0d81caf34eb3
|
| 3 |
+
size 6681112
|