Model save
Browse files- README.md +77 -0
- added_tokens.json +3 -0
- all_results.json +19 -0
- config.json +42 -0
- eval_results.json +13 -0
- model-00001-of-00003.safetensors +3 -0
- model-00002-of-00003.safetensors +3 -0
- model-00003-of-00003.safetensors +3 -0
- model.safetensors +3 -0
- model.safetensors.index.json +299 -0
- predict_results.txt +540 -0
- sentencepiece.bpe.model +3 -0
- special_tokens_map.json +35 -0
- spm.model +3 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +48 -0
- train_results.json +9 -0
- trainer_state.json +0 -0
- training_args.bin +3 -0
README.md
ADDED
|
@@ -0,0 +1,77 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: transformers
|
| 3 |
+
license: cc-by-nc-4.0
|
| 4 |
+
base_model: Linq-AI-Research/Linq-Embed-Mistral
|
| 5 |
+
tags:
|
| 6 |
+
- generated_from_trainer
|
| 7 |
+
model-index:
|
| 8 |
+
- name: twscrape-prepared-regression-Linq-Embed-Mistral-scorer
|
| 9 |
+
results: []
|
| 10 |
+
---
|
| 11 |
+
|
| 12 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 13 |
+
should probably proofread and complete it, then remove this comment. -->
|
| 14 |
+
|
| 15 |
+
# twscrape-prepared-regression-Linq-Embed-Mistral-scorer
|
| 16 |
+
|
| 17 |
+
This model is a fine-tuned version of [Linq-AI-Research/Linq-Embed-Mistral](https://huggingface.co/Linq-AI-Research/Linq-Embed-Mistral) on an unknown dataset.
|
| 18 |
+
It achieves the following results on the evaluation set:
|
| 19 |
+
- Loss: 9.7099
|
| 20 |
+
- Mse: 9.7098
|
| 21 |
+
- Target 0 Mse: 18.5035
|
| 22 |
+
- Target 1 Mse: 9.9817
|
| 23 |
+
- Target 2 Mse: 5.2868
|
| 24 |
+
- Target 3 Mse: 5.0672
|
| 25 |
+
|
| 26 |
+
## Model description
|
| 27 |
+
|
| 28 |
+
More information needed
|
| 29 |
+
|
| 30 |
+
## Intended uses & limitations
|
| 31 |
+
|
| 32 |
+
More information needed
|
| 33 |
+
|
| 34 |
+
## Training and evaluation data
|
| 35 |
+
|
| 36 |
+
More information needed
|
| 37 |
+
|
| 38 |
+
## Training procedure
|
| 39 |
+
|
| 40 |
+
### Training hyperparameters
|
| 41 |
+
|
| 42 |
+
The following hyperparameters were used during training:
|
| 43 |
+
- learning_rate: 1e-05
|
| 44 |
+
- train_batch_size: 24
|
| 45 |
+
- eval_batch_size: 24
|
| 46 |
+
- seed: 42
|
| 47 |
+
- distributed_type: multi-GPU
|
| 48 |
+
- num_devices: 8
|
| 49 |
+
- total_train_batch_size: 192
|
| 50 |
+
- total_eval_batch_size: 192
|
| 51 |
+
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
| 52 |
+
- lr_scheduler_type: cosine
|
| 53 |
+
- lr_scheduler_warmup_ratio: 0.1
|
| 54 |
+
- num_epochs: 10.0
|
| 55 |
+
|
| 56 |
+
### Training results
|
| 57 |
+
|
| 58 |
+
| Training Loss | Epoch | Step | Validation Loss | Mse | Target 0 Mse | Target 1 Mse | Target 2 Mse | Target 3 Mse |
|
| 59 |
+
|:-------------:|:-----:|:----:|:---------------:|:-------:|:------------:|:------------:|:------------:|:------------:|
|
| 60 |
+
| 16.8047 | 1.0 | 280 | 16.4555 | 16.4594 | 27.7680 | 15.5051 | 11.1613 | 11.4033 |
|
| 61 |
+
| 11.1602 | 2.0 | 560 | 13.1732 | 13.1774 | 23.3179 | 13.1760 | 8.2887 | 7.9268 |
|
| 62 |
+
| 12.6055 | 3.0 | 840 | 11.8043 | 11.8031 | 21.3693 | 11.9728 | 7.0819 | 6.7884 |
|
| 63 |
+
| 11.1562 | 4.0 | 1120 | 10.9693 | 10.9669 | 20.2088 | 11.1752 | 6.3655 | 6.1181 |
|
| 64 |
+
| 8.6172 | 5.0 | 1400 | 10.4258 | 10.4250 | 19.4563 | 10.6673 | 5.9042 | 5.6723 |
|
| 65 |
+
| 10.1992 | 6.0 | 1680 | 10.0752 | 10.0751 | 18.9940 | 10.3359 | 5.5939 | 5.3767 |
|
| 66 |
+
| 7.875 | 7.0 | 1960 | 9.8701 | 9.8630 | 18.7038 | 10.1268 | 5.4208 | 5.2006 |
|
| 67 |
+
| 8.0508 | 8.0 | 2240 | 9.7511 | 9.7564 | 18.5671 | 10.0269 | 5.3200 | 5.1115 |
|
| 68 |
+
| 8.0195 | 9.0 | 2520 | 9.7144 | 9.7161 | 18.5102 | 9.9874 | 5.2931 | 5.0737 |
|
| 69 |
+
| 11.4102 | 10.0 | 2800 | 9.7099 | 9.7098 | 18.5035 | 9.9817 | 5.2868 | 5.0672 |
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
### Framework versions
|
| 73 |
+
|
| 74 |
+
- Transformers 4.49.0
|
| 75 |
+
- Pytorch 2.5.1+cu124
|
| 76 |
+
- Datasets 3.0.1
|
| 77 |
+
- Tokenizers 0.21.0
|
added_tokens.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"[MASK]": 128000
|
| 3 |
+
}
|
all_results.json
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 10.0,
|
| 3 |
+
"eval_loss": 3.5166869163513184,
|
| 4 |
+
"eval_mse": 3.5154231365248,
|
| 5 |
+
"eval_runtime": 3.8327,
|
| 6 |
+
"eval_samples": 2840,
|
| 7 |
+
"eval_samples_per_second": 740.995,
|
| 8 |
+
"eval_steps_per_second": 6.001,
|
| 9 |
+
"eval_target_0_mse": 10.009268208894252,
|
| 10 |
+
"eval_target_1_mse": 3.540375322536252,
|
| 11 |
+
"eval_target_2_mse": 0.509027972482636,
|
| 12 |
+
"eval_target_3_mse": 0.0030210421860589047,
|
| 13 |
+
"total_flos": 5.010130621956096e+17,
|
| 14 |
+
"train_loss": 2.3525503249395463,
|
| 15 |
+
"train_runtime": 2400.4864,
|
| 16 |
+
"train_samples": 53752,
|
| 17 |
+
"train_samples_per_second": 223.921,
|
| 18 |
+
"train_steps_per_second": 1.75
|
| 19 |
+
}
|
config.json
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "Linq-AI-Research/Linq-Embed-Mistral",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"MistralForSequenceClassificationWithBias"
|
| 5 |
+
],
|
| 6 |
+
"attention_dropout": 0.0,
|
| 7 |
+
"bos_token_id": 1,
|
| 8 |
+
"eos_token_id": 2,
|
| 9 |
+
"finetuning_task": "text-regression",
|
| 10 |
+
"head_dim": 128,
|
| 11 |
+
"hidden_act": "silu",
|
| 12 |
+
"hidden_size": 4096,
|
| 13 |
+
"id2label": {
|
| 14 |
+
"0": "LABEL_0",
|
| 15 |
+
"1": "LABEL_1",
|
| 16 |
+
"2": "LABEL_2",
|
| 17 |
+
"3": "LABEL_3"
|
| 18 |
+
},
|
| 19 |
+
"initializer_range": 0.02,
|
| 20 |
+
"intermediate_size": 14336,
|
| 21 |
+
"label2id": {
|
| 22 |
+
"LABEL_0": 0,
|
| 23 |
+
"LABEL_1": 1,
|
| 24 |
+
"LABEL_2": 2,
|
| 25 |
+
"LABEL_3": 3
|
| 26 |
+
},
|
| 27 |
+
"max_position_embeddings": 32768,
|
| 28 |
+
"model_type": "mistral",
|
| 29 |
+
"num_attention_heads": 32,
|
| 30 |
+
"num_hidden_layers": 32,
|
| 31 |
+
"num_key_value_heads": 8,
|
| 32 |
+
"pad_token_id": 2,
|
| 33 |
+
"problem_type": "regression",
|
| 34 |
+
"rms_norm_eps": 1e-05,
|
| 35 |
+
"rope_theta": 10000.0,
|
| 36 |
+
"sliding_window": 4096,
|
| 37 |
+
"tie_word_embeddings": false,
|
| 38 |
+
"torch_dtype": "bfloat16",
|
| 39 |
+
"transformers_version": "4.49.0",
|
| 40 |
+
"use_cache": false,
|
| 41 |
+
"vocab_size": 32000
|
| 42 |
+
}
|
eval_results.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 10.0,
|
| 3 |
+
"eval_loss": 3.5166869163513184,
|
| 4 |
+
"eval_mse": 3.5154231365248,
|
| 5 |
+
"eval_runtime": 3.8327,
|
| 6 |
+
"eval_samples": 2840,
|
| 7 |
+
"eval_samples_per_second": 740.995,
|
| 8 |
+
"eval_steps_per_second": 6.001,
|
| 9 |
+
"eval_target_0_mse": 10.009268208894252,
|
| 10 |
+
"eval_target_1_mse": 3.540375322536252,
|
| 11 |
+
"eval_target_2_mse": 0.509027972482636,
|
| 12 |
+
"eval_target_3_mse": 0.0030210421860589047
|
| 13 |
+
}
|
model-00001-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56e589c112e4177f4430e3bfbd843243b49cb241fa96f26afe87b9ba7fc77b7c
|
| 3 |
+
size 4943162336
|
model-00002-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7b4e472b41ced0e928b2350dd6057e7f52cccc8683e7d95bc41440aac0941643
|
| 3 |
+
size 4999819336
|
model-00003-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aca38dbecca306fe91d31efd891c09ed74c64670511bd31f27bfd8decdc93911
|
| 3 |
+
size 4278405176
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3f95dd0d41ff069a4a85aefa458d6afaa7bf5fb9eac3516f35fe05d23abdd5ff
|
| 3 |
+
size 870180856
|
model.safetensors.index.json
ADDED
|
@@ -0,0 +1,299 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 14221352968
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"model.embed_tokens.weight": "model-00001-of-00003.safetensors",
|
| 7 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 8 |
+
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 9 |
+
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 10 |
+
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 11 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 12 |
+
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 13 |
+
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 14 |
+
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 15 |
+
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 16 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 17 |
+
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 18 |
+
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 19 |
+
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 20 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 21 |
+
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 22 |
+
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 23 |
+
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 24 |
+
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 25 |
+
"model.layers.10.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 26 |
+
"model.layers.10.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 27 |
+
"model.layers.10.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 28 |
+
"model.layers.10.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 29 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 30 |
+
"model.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 31 |
+
"model.layers.10.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 32 |
+
"model.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 33 |
+
"model.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 34 |
+
"model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 35 |
+
"model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 36 |
+
"model.layers.11.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 37 |
+
"model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 38 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 39 |
+
"model.layers.11.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 40 |
+
"model.layers.11.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 41 |
+
"model.layers.11.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 42 |
+
"model.layers.11.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 43 |
+
"model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 44 |
+
"model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 45 |
+
"model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 46 |
+
"model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 47 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 48 |
+
"model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 49 |
+
"model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 50 |
+
"model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 51 |
+
"model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 52 |
+
"model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 53 |
+
"model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 54 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 55 |
+
"model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 56 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 57 |
+
"model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 58 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 59 |
+
"model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 60 |
+
"model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 61 |
+
"model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 62 |
+
"model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 63 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 64 |
+
"model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 65 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 66 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 67 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 68 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 69 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 70 |
+
"model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 71 |
+
"model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 72 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 73 |
+
"model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 74 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 75 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 76 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 77 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 78 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 79 |
+
"model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 80 |
+
"model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 81 |
+
"model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 82 |
+
"model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 83 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 84 |
+
"model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 85 |
+
"model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 86 |
+
"model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 87 |
+
"model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 88 |
+
"model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 89 |
+
"model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 90 |
+
"model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 91 |
+
"model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 92 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 93 |
+
"model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 94 |
+
"model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 95 |
+
"model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 96 |
+
"model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 97 |
+
"model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 98 |
+
"model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 99 |
+
"model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 100 |
+
"model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 101 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 102 |
+
"model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 103 |
+
"model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 104 |
+
"model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 105 |
+
"model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 106 |
+
"model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 107 |
+
"model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 108 |
+
"model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 109 |
+
"model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 110 |
+
"model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 111 |
+
"model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 112 |
+
"model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 113 |
+
"model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 114 |
+
"model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 115 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 116 |
+
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 117 |
+
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 118 |
+
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 119 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 120 |
+
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 121 |
+
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 122 |
+
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 123 |
+
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 124 |
+
"model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 125 |
+
"model.layers.20.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 126 |
+
"model.layers.20.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 127 |
+
"model.layers.20.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 128 |
+
"model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 129 |
+
"model.layers.20.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 130 |
+
"model.layers.20.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 131 |
+
"model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 132 |
+
"model.layers.20.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 133 |
+
"model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 134 |
+
"model.layers.21.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 135 |
+
"model.layers.21.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 136 |
+
"model.layers.21.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 137 |
+
"model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 138 |
+
"model.layers.21.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 139 |
+
"model.layers.21.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 140 |
+
"model.layers.21.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 141 |
+
"model.layers.21.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 142 |
+
"model.layers.22.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 143 |
+
"model.layers.22.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 144 |
+
"model.layers.22.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 145 |
+
"model.layers.22.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 146 |
+
"model.layers.22.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 147 |
+
"model.layers.22.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 148 |
+
"model.layers.22.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 149 |
+
"model.layers.22.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 150 |
+
"model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 151 |
+
"model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 152 |
+
"model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 153 |
+
"model.layers.23.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 154 |
+
"model.layers.23.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 155 |
+
"model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 156 |
+
"model.layers.23.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 157 |
+
"model.layers.23.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 158 |
+
"model.layers.23.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 159 |
+
"model.layers.23.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 160 |
+
"model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 161 |
+
"model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 162 |
+
"model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 163 |
+
"model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 164 |
+
"model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 165 |
+
"model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 166 |
+
"model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 167 |
+
"model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 168 |
+
"model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 169 |
+
"model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 170 |
+
"model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 171 |
+
"model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 172 |
+
"model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 173 |
+
"model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 174 |
+
"model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 175 |
+
"model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 176 |
+
"model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 177 |
+
"model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 178 |
+
"model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 179 |
+
"model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 180 |
+
"model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 181 |
+
"model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 182 |
+
"model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 183 |
+
"model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 184 |
+
"model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 185 |
+
"model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 186 |
+
"model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 187 |
+
"model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 188 |
+
"model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 189 |
+
"model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 190 |
+
"model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 191 |
+
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 192 |
+
"model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 193 |
+
"model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 194 |
+
"model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 195 |
+
"model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 196 |
+
"model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 197 |
+
"model.layers.28.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 198 |
+
"model.layers.28.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 199 |
+
"model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 200 |
+
"model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 201 |
+
"model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 202 |
+
"model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 203 |
+
"model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 204 |
+
"model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 205 |
+
"model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 206 |
+
"model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 207 |
+
"model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 208 |
+
"model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 209 |
+
"model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 210 |
+
"model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 211 |
+
"model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 212 |
+
"model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 213 |
+
"model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 214 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 215 |
+
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 216 |
+
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 217 |
+
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 218 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 219 |
+
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 220 |
+
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 221 |
+
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 222 |
+
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 223 |
+
"model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 224 |
+
"model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 225 |
+
"model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 226 |
+
"model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 227 |
+
"model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 228 |
+
"model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 229 |
+
"model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 230 |
+
"model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 231 |
+
"model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 232 |
+
"model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 233 |
+
"model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 234 |
+
"model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 235 |
+
"model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 236 |
+
"model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 237 |
+
"model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 238 |
+
"model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 239 |
+
"model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 240 |
+
"model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 241 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 242 |
+
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 243 |
+
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 244 |
+
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 245 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 246 |
+
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 247 |
+
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 248 |
+
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 249 |
+
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 250 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 251 |
+
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 252 |
+
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 253 |
+
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 254 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 255 |
+
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 256 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 257 |
+
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 258 |
+
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 259 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 260 |
+
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 261 |
+
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 262 |
+
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 263 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 264 |
+
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 265 |
+
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 266 |
+
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 267 |
+
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 268 |
+
"model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 269 |
+
"model.layers.7.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 270 |
+
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 271 |
+
"model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 272 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 273 |
+
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 274 |
+
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 275 |
+
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 276 |
+
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 277 |
+
"model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 278 |
+
"model.layers.8.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 279 |
+
"model.layers.8.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 280 |
+
"model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 281 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 282 |
+
"model.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 283 |
+
"model.layers.8.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 284 |
+
"model.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 285 |
+
"model.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 286 |
+
"model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 287 |
+
"model.layers.9.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
| 288 |
+
"model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 289 |
+
"model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 290 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 291 |
+
"model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 292 |
+
"model.layers.9.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 293 |
+
"model.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 294 |
+
"model.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 295 |
+
"model.norm.weight": "model-00003-of-00003.safetensors",
|
| 296 |
+
"score.bias": "model-00003-of-00003.safetensors",
|
| 297 |
+
"score.weight": "model-00003-of-00003.safetensors"
|
| 298 |
+
}
|
| 299 |
+
}
|
predict_results.txt
ADDED
|
@@ -0,0 +1,540 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
index text prediction
|
| 2 |
+
0 Sitting this morning watch the UK Government bond ticker for 30 year bonds.
|
| 3 |
+
|
| 4 |
+
Currently hovering at 26 year highs - worse than Liz Truss crisis.
|
| 5 |
+
|
| 6 |
+
Rachel Reeves has really done it this time. 2.111 0.046 0.895 0.023 2.094 0.126 0.248 0.027 0.017 0.080 0.647 0.004
|
| 7 |
+
1 🚨BREAKING:
|
| 8 |
+
|
| 9 |
+
Mark Wahlberg calls out Hollywood Democrats. He said, "If you don't like the USA, please leave and take your friends Alec Baldwin, Cher, Jennifer Lawrence, Miley Cyrus, Barbra Streisand, Matt Damon, Oprah, and Colin Kaepernick with you."
|
| 10 |
+
|
| 11 |
+
Do you agree? 2.220 0.092 0.276 0.013 10.062 0.371 2.297 0.085 7.843 0.279 2.021 0.072
|
| 12 |
+
2 Don't you just love the "Know-It-Alls" who know absolutely nothing?
|
| 13 |
+
Have you ever noticed how the loudest, most obnoxious people in the Karen Read case are the ones with the least amount of actual facts? The anti-Karen crowd is like a broken GPS; confidently leading everyone straight into a ditch. They scream about "justice" while ignoring every piece of corruption unfolding right in front of their glazed-over eyes.
|
| 14 |
+
|
| 15 |
+
Ask them for any evidence proving her guilt that isn't speculation..... and well.... you'll be waiting for quite some time for any logical answer. Why? Because their thought process is slower than a 1998 dial-up connection. "Uh... she backed into him! The taillight!" Oh, you mean the magical taillight that they never found when they first searched the scene and mysteriously appeared days later? Yeah, okay, Einsteins.
|
| 16 |
+
|
| 17 |
+
These people are so desperate to be right. They ignore irrefutable cover-ups, missing evidence, deleted texts, call, google searches, a biased judge, and shady cops treating a crime scene like a frat party; using red solo cups and stop and shop bags to collect evidence. But questioning them about any of that requires critical thinking, and unfortunately, their brains are still running on Windows 95.
|
| 18 |
+
|
| 19 |
+
They claim they "know" what happened, yet every argument they make collapses faster than their ability to read a court document. They don't see bias and corruption because, well… that requires complex thinking past their own self-righteousness, and let's be real... that's just way too much effort.
|
| 20 |
+
|
| 21 |
+
At this point, arguing with them is like debating with a toddler... except the toddler has more reasoning skills. So, keep up the mental gymnastics, anti-Karen crowd. Maybe one day, you'll stick the landing on a point that actually makes sense. 2.168 0.204 0.159 0.023 4.625 0.287 0.980 0.036 2.457 0.083 0.822 0.014
|
| 22 |
+
3 the movie I shot in 2022 where me, Matt Rife, and Andrew Schultz are all siblings is finally coming out! thank you to Tahrir al-Sham and especially Emir al-Julani for believing in us 2.197 0.019 0.033 0.000 3.188 0.103 0.159 0.010 0.991 0.084 0.126 0.010
|
| 23 |
+
4 Sad news for my Egyptological community. Prof Barry Kemp passed away. He was always kind to me when I was a PhD student when I would pepper him with questions when I would see him in person, & when I emailed him while writing my PhD dissertation. May he have 1000 bread & beer 2.168 0.131 0.164 0.000 2.281 0.090 0.205 0.013 0.113 0.041 0.041 0.013
|
| 24 |
+
5 Me after 2 drinks: What if I just move to Minnesota
|
| 25 |
+
|
| 26 |
+
Me after 5 drinks: What if I just move to Detroit
|
| 27 |
+
|
| 28 |
+
Me after 10 drinks: What if I just spend my life savings on ffs and bottom surgery and marry a lonely oil field worker in Alaska who will die never knowing I was trans. 2.171 0.089 0.045 0.000 4.531 0.052 0.309 0.015 2.361 0.037 0.264 0.015
|
| 29 |
+
6 But from now till late March we have just 1 game a week so we won’t get injuries we were told. First game after the window closed we get an injury in attack. Shit 2.006 0.034 0.210 0.010 2.188 0.057 0.190 0.016 0.181 0.023 0.019 0.005
|
| 30 |
+
7 Tomorrow’s election in Belarus is a sham.
|
| 31 |
+
|
| 32 |
+
My message to the people of Belarus is: keep strong, we have your back, the time of dictatorship will be over.
|
| 33 |
+
|
| 34 |
+
Democracy will prevail. 2.011 0.532 0.371 0.064 3.422 0.173 0.562 0.033 1.411 0.359 0.191 0.031
|
| 35 |
+
8 Fair play Hibs and David Gray. Rock bottom not that long ago, now up to 3rd. 2.008 0.000 0.000 0.000 3.422 0.040 0.224 -0.012 1.414 0.040 0.224 0.012
|
| 36 |
+
9 Cristiano Ronaldo is closer to Eden Hazard, Luis Suarez and Arjen Robben than he is to Lionel Messi. The "debate" was a media invention and nothing more. 2.150 0.078 0.249 0.041 1.734 0.182 0.247 0.026 0.415 0.104 0.002 0.015
|
| 37 |
+
10 PEF, the union representing New York parole officers, has received an injunction to keep officers from being assigned "out of title" work. Last night, parole officers refused an order to do a cell extraction at Attica, saying they hadn't been trained how to do that. 2.017 0.201 0.459 0.043 1.453 0.076 0.334 0.028 0.564 0.125 0.125 0.015
|
| 38 |
+
11 I remember someone approaching me to organize good influencers for a gig only to find out that it was that Katanga case I was the first one to run 😂 2.112 0.108 0.185 0.008 2.672 0.084 0.260 0.030 0.559 0.025 0.074 0.022
|
| 39 |
+
12 Installed Atomic Heart and saw denuvo uninstalled immediately go my whole life without playing it is a promise to God 1.985 0.119 0.072 0.000 1.812 0.142 0.112 0.018 0.173 0.023 0.039 0.018
|
| 40 |
+
13 British India circa 1760: A Sensitive Young Man who plays his cards right can end up with a harem and several elephants
|
| 41 |
+
British India circa 1920: if you pass the ICS exams, you can spend a lifetime doing paperwork on behalf of people who largely ignore you because they know you're all leaving soon. 2.102 0.111 0.071 0.020 1.961 0.139 0.171 0.026 0.141 0.028 0.100 0.005
|
| 42 |
+
14 Butt naked in a public bathroom because I decided to wear a jumpsuit today ✨ 2.224 0.009 0.071 0.040 3.438 0.184 0.102 0.026 1.214 0.175 0.031 0.014
|
| 43 |
+
15 Exposing a hidden child is crazy but don’t forget that Drivers License turned Olivia Rodrigo into a superstar overnight and literally put Joshua Bassett into the hospital after he heard it. Had Sabrina Carpenter stans BEGGING her to call off the dogs. One of the GOAT diss tracks 2.004 0.041 0.140 0.000 2.031 0.072 0.225 0.027 0.027 0.031 0.084 0.027
|
| 44 |
+
16 Jalen Williams is one of just 12 NBA players to score 3,500 points, 900 rebounds, 800 assists, 250 steals and 100 blocks in their first 200 games.
|
| 45 |
+
|
| 46 |
+
Guys on the list:
|
| 47 |
+
Michael Jordan, Magic Johnson, LeBron James, Larry Bird, Grant Hill, Dwyane Wade, Penny Hardaway, Chris Webber 2.082 0.020 0.272 0.054 1.539 0.070 0.100 0.015 0.543 0.050 0.172 0.039
|
| 48 |
+
17 Russia and China have taken the gloves off. They say if you insist on confrontation you shall have confrontation. We’ll soon find out how far the West is willing to go. This is the time for the EU to realize that they have gone too far. This is a critical moment for the world. 1.986 0.080 0.529 0.029 3.000 0.312 0.637 0.013 1.014 0.232 0.108 0.015
|
| 49 |
+
18 You can tell mathematicians right now are SO relieved there is no Math Nobel Prize 2.047 0.057 0.112 0.025 2.281 0.104 0.107 0.009 0.234 0.047 0.005 0.016
|
| 50 |
+
19 It was crazy seeing such a furious Harry Kane. He actually got a yellow card for protecting his injured teammate. That was actually so touching 2.150 0.039 0.059 0.000 3.391 0.018 0.065 -0.005 1.241 0.021 0.006 0.005
|
| 51 |
+
20 House Republicans passed a budget plan that will cut Medicaid. And now, they're attacking Social Security.
|
| 52 |
+
|
| 53 |
+
How does that lower costs and improve the lives of American families? It doesn't. 2.058 1.392 0.864 0.099 2.547 0.346 0.730 0.014 0.489 1.046 0.133 0.085
|
| 54 |
+
21 German word for when you desperately need to clown on someone but you check their profile and you have 100 mutuals in common 2.066 0.051 0.102 0.012 1.977 0.145 0.115 0.026 0.090 0.094 0.013 0.014
|
| 55 |
+
22 Just wanted to keep it a bit real.
|
| 56 |
+
|
| 57 |
+
This week I’ve felt old, washed up, tired, and lonely.
|
| 58 |
+
|
| 59 |
+
def not Superman status like normal.
|
| 60 |
+
|
| 61 |
+
Thank god for my friends who messaged me and helped pick me up a bit. 2.093 0.270 0.010 0.020 4.344 0.668 0.177 0.006 2.251 0.398 0.167 0.014
|
| 62 |
+
23 I'm just gonna say it. "zionist" has become completely meaningless as an epithet wielded by zoomers specifically against entertainment/creative arts figures almost invariably to 1-up and police other fandoms in service of some pre-existing grudge. 2.224 0.058 0.231 0.007 5.188 0.083 0.594 0.035 2.963 0.026 0.363 0.028
|
| 63 |
+
24 Recently, some people have put out some content on social media that don't reflect my beliefs or values. I'm grateful to those who have reached out to me, and I'm taking this opportunity to set a few things very clear.
|
| 64 |
+
I don't support terrorism or racism by extremists of any kind. People who promote this hate don't have my support in any way, shape, or form.
|
| 65 |
+
My problem is with radical genocidal terrorists in Gaza or elsewhere, and their supporters, including those on college campuses. I have nothing against Palestinians, Arabs, or Muslims who are just trying to live their lives without harming anyone else!
|
| 66 |
+
|
| 67 |
+
AM ISRAEL CHAI 2.103 0.180 0.107 0.006 1.875 0.136 0.295 0.042 0.228 0.044 0.187 0.035
|
| 68 |
+
25 “missing wallows music videos before they became nonchalant” just say you miss the skits pls because calling dylan nonchalant is one thing but BRAEDEN??? 2.069 0.086 0.134 0.058 5.094 -0.093 0.703 0.042 3.025 0.179 0.569 0.016
|
| 69 |
+
26 PSG, City and Liverpool tried to play football against Arsenal this season.
|
| 70 |
+
|
| 71 |
+
Every single one of them got crushed while Liverpool got lucky to draw as injuries hit every Arsenal defender halfway through the game.
|
| 72 |
+
|
| 73 |
+
You don't play football against Arteta. You sit low and defend. 2.037 0.139 0.203 0.061 3.312 0.059 0.318 0.009 1.276 0.080 0.115 0.053
|
| 74 |
+
27 Woke up worse today than yesterday. Tea, honey, OJ, and soup for me today. TNT will STILL go on as normal. I ain't missing this shit tonight. We have A LOT to talk about. I'll be on the couch til then. 🙏 2.030 0.159 0.074 0.004 6.719 0.311 0.645 0.025 4.688 0.152 0.571 0.021
|
| 75 |
+
28 The constantly repeating same map over and over is absolutely one of the core issues STILL faced by 2042.
|
| 76 |
+
|
| 77 |
+
Trying to retain old players or encourage new when they have to play the same map 3 times in a row is death.
|
| 78 |
+
|
| 79 |
+
If the next Battlefield has the same matchmaking issue...🤦♂️ 2.059 1.347 0.040 0.000 2.031 0.102 0.190 0.008 0.028 1.244 0.151 0.008
|
| 80 |
+
29 Survived my first 32hrs duty day and oh I also got to perform nasotracheal intubation for the first time today!🧿🥺 2.194 0.075 0.032 0.000 3.125 0.336 0.136 0.012 0.931 0.261 0.104 0.012
|
| 81 |
+
30 [240914] weverse dm 🧸💬
|
| 82 |
+
|
| 83 |
+
🧸 y/n…
|
| 84 |
+
🧸 i fell asleep as soon as i ate after the live earlier
|
| 85 |
+
🧸 so i just finished washing up 😂
|
| 86 |
+
🧸 hahahahaha….i’ll be able to sleep again, right…?
|
| 87 |
+
🧸 i ate something delicious today!!!!
|
| 88 |
+
🧸 bibimbap, fried dumpings and ramen!!! hahahaha 2.045 0.003 0.271 0.030 4.625 0.239 0.586 0.014 2.580 0.237 0.315 0.016
|
| 89 |
+
31 Why did I start Gala in 2018?
|
| 90 |
+
My wife and I had been talking for a while how we wish we could be of service after I left Zynga for many years. During this time we had become to be aware of the vast gulf between the opportunities we had in the USA and the heart wrenching decisions many people around the world have to make in order to survive. Realizing that the secondary market for video games at that time was over 50b usd and the willingness of children in the USA and Europe to buy digital goods we thought what would happen if there was a easy and secure way for children in other parts of the world to sell their digital items to these other children with so much opportunity. We set out to achieve this mission and even though everyone in crypto has been having a rough time of it these last two years we did achieve something truly remarkable.
|
| 91 |
+
|
| 92 |
+
I think we can safely say that we are the only company ever to have given back more than it has made. I applaud companies that give a % of their profit to charity. But Gala can say that we gave 2b usd worth of crypto back to the community. Far far more than we made ourselves during that period.
|
| 93 |
+
|
| 94 |
+
I'm really proud of our company for that achievement. And with the opportunity given to us by Telegram I sense that we will have another chance here soon. I have said that our up coming token launch of $TREZ will be 100% to the community and market makers/exchanges. That's 0% to the team... I hope to hear stories like we heard with $GALA about how families were able to buy a home for the first time, how medical procedures were finally within reach and how some where able to finally get out of devastating poverty.
|
| 95 |
+
|
| 96 |
+
Lets do it again! 1.985 0.087 0.544 0.012 1.336 0.066 0.164 0.014 0.649 0.021 0.380 0.002
|
| 97 |
+
32 One day, Nigerians on Twitter will evolve their discussions to a higher level and stop fighting about cooking food. In the meantime, I will mute their conversations. I am currently in Togo, and cooking is a professional affair. Men and women take pride in doing it professionally. 2.127 0.048 0.434 0.043 2.125 0.141 0.231 0.035 0.002 0.093 0.203 0.008
|
| 98 |
+
33 Please read and comment: the C defer TS (unpublished TS). Written purely in typst, first time I didn't feel like writing a standard's document was annoying and a waste of my time!
|
| 99 |
+
|
| 100 |
+
⤵ 1.993 0.146 0.146 0.000 1.195 0.140 0.119 0.017 0.798 0.007 0.027 0.017
|
| 101 |
+
34 Castle needs to start, man
|
| 102 |
+
|
| 103 |
+
Put him at point until Fox knows the plays
|
| 104 |
+
|
| 105 |
+
CP3, Fox, and Vassell in the same lineup is just way too small 2.058 0.072 0.046 0.018 2.109 0.087 0.142 0.015 0.052 0.015 0.096 0.003
|
| 106 |
+
35 Lebron get Luka Doncic and Steph curry get Gui Santos lmaooooooo I hope Joe Lacob never knows peace 2.096 0.036 0.311 0.040 2.625 0.125 0.132 0.000 0.529 0.088 0.180 0.039
|
| 107 |
+
36 if daniel johnston released 'hi, how are you' today it would have been called 'what's poppin fam' 2.013 0.060 0.100 0.000 4.281 0.022 0.178 0.019 2.268 0.037 0.078 0.019
|
| 108 |
+
37 👨🏾🍳🇬🇭: After the announcements of the ballot positions, I guess we all know what time it is… copywriters, numerologists, mathematicians & spiritualistsss; job comeeee😁🙌🏾 1.999 0.064 0.260 0.012 2.078 0.215 0.173 0.021 0.079 0.151 0.087 0.009
|
| 109 |
+
38 Once I took a trip to Portland Maine. Stayed in a nice Airbnb. It was March but I thought spring had yet to arrive. Late at night, I watched the streets from high window. Raccoons were chasing one another up & down the street, yipping & yelping as they fucked. Spring had arrived. 2.078 0.078 0.062 0.000 2.234 0.148 0.161 0.025 0.157 0.070 0.099 0.025
|
| 110 |
+
39 Just had Black Jehovah’s Witnesses come up to my door… I didn’t even know that we did that💰 2.077 0.428 0.058 0.071 2.469 0.143 0.141 0.035 0.391 0.286 0.082 0.037
|
| 111 |
+
40 I can barely recall a time when an American president took office amidst such chaos. It's notable that he's trying to slash budgets and overhaul an agency, even if it means involving Elon Musk. I'm okay with that, frankly. The real problem is how this agency has been wasting money on ridiculous things just to wield power. Their influence, particularly in the media, is huge and worldwide. The notion of possibly disbanding USAID seems like a step in the right direction. Whether they'll start fresh or just streamline what's there is up in the air. But let's be real—the U.S. isn't about to give up its global sway, though how well they'll pull it off remains to be seen.
|
| 112 |
+
|
| 113 |
+
His aggressive tactics, threatening countries like Canada, Denmark, Panama, Colombia, China, and Mexico, are definitely out of the ordinary. From where I stand, it's tough to grasp what he's aiming for, which makes it hard to judge his strategy. But from a foreign policy angle, it's certainly grabbing attention. NATO allies must be scratching their heads on how to respond, especially concerning Greenland. And with Gaza, his flip-flopping between moving Palestinians out and then suggesting temporary relocation has left me totally baffled about what he really wants.
|
| 114 |
+
|
| 115 |
+
Watching the U.S. clash with its own allies is kind of entertaining from the sidelines. And depending on who's president, being a close ally doesn't always pay off.
|
| 116 |
+
|
| 117 |
+
When it comes to Ukraine, Trump seems to have overplayed his hand. At the end of the day, it's Putin who will call when the war stops. The West doesn't have the stamina for a prolonged conflict, and more to the point, the U.S. has no interest in dragging this on since it no longer aligns with their initial goals. This conflict has oddly served as a training ground for Russia.
|
| 118 |
+
|
| 119 |
+
This war has bolstered Russia's internal solidarity, independence, and self-reliance in terms of energy, food, jobs, and a working economy free from Western dependence. The Russian military, forced to tackle corruption, has upgraded its arsenal and restructured its forces. Plus, hundreds of thousands of Russian troops now have firsthand combat experience and are more resolute in defending their homeland.
|
| 120 |
+
|
| 121 |
+
None of this matches what the West initially hoped for. Their goal was to sway Russia's internal politics and gain economic control, much like the '90s when they tried to manipulate Russia's economy under Western laws. That strategy has bombed.
|
| 122 |
+
|
| 123 |
+
Trump probably sees now that his plan is trickier than he thought. He's dealing with an unappreciative EU, which keeps funding Ukraine but lacks the U.S.'s military might to keep the weapons flowing.
|
| 124 |
+
|
| 125 |
+
In short, Trump's looking to bow out. He's ready to make deals but wants to keep his image intact. Putin, on the other hand, isn't concerned with anyone's image except his own.
|
| 126 |
+
|
| 127 |
+
We'll see how this all plays out, but as an onlooker, it's undeniably intriguing. 2.145 0.262 0.381 0.016 4.031 0.463 0.789 0.027 1.887 0.201 0.408 0.011
|
| 128 |
+
41 I be so sick of ppl & their social media personalities. You not even like that fr! 2.157 0.000 1.167 0.130 1.977 0.226 0.129 0.010 0.180 0.225 1.038 0.121
|
| 129 |
+
42 It is weird that our civic religion still defaults to Anglican. This guy was a Southern Baptist and he is being processed in behind acolytes in albs with candles and a processional cross. 1.983 0.063 0.090 0.038 2.703 0.097 0.336 0.025 0.720 0.033 0.246 0.013
|
| 130 |
+
43 This shabbat's Torah reading is one of the most interesting. It's almost certainly the most exciting.
|
| 131 |
+
|
| 132 |
+
It's called "Beshalach," "When he sent." It continues the story of the Exodus in the heart-pounding flight from Pharaoh's pursuing chariots and the famous miracle of the splitting of the sea.
|
| 133 |
+
|
| 134 |
+
And it tells us perhaps the most important single thing the Torah wants us to know about where faith comes from and what it's for.
|
| 135 |
+
|
| 136 |
+
Those great miracles, the last-minute rescue from the terrifying pursuer, the mind-boggling, physics-altering miracle that rescued them when the sea parted, Pharaoh's army sinking into the sea - "and they believed in God and Moses his servant," the Torah tells us moments later.
|
| 137 |
+
|
| 138 |
+
And then, saved and astonished, they sing, all the people, all at once, all together. About God's strength and majesty and power. And then Miriam leads the women in another special song.
|
| 139 |
+
|
| 140 |
+
And then, the Torah continues, the people leave the shores of the sea, "walk three days in the desert and cannot find water." And when they finally do, the waters are too bitter to drink.
|
| 141 |
+
|
| 142 |
+
And then everything, all at once, reverts to the mean. The people immediately turn on Moses, demanding, "What will we drink?" And a frustrated Moses "shouts to God."
|
| 143 |
+
|
| 144 |
+
The greatest moment of miraculous witness in the history of, well, history, and it only took the first momentary crisis (they had drinkable water by day's end) for everything to collapse. It took just three days from spectacular rescue - the Torah spares no words in conveying repeatedly the people's amazement and gratitude, and explicitly tells us about the vast reservoirs of faith built up by the experience - for it all to evaporate.
|
| 145 |
+
|
| 146 |
+
Great faith born in astonishing miracles, it turns out, won't carry you very far.
|
| 147 |
+
|
| 148 |
+
Prof. Yeshayahu Leibowitz noticed this jarring pivot of the narrative. Here's what he had to say: "In this story, the Torah comes to teach us a very great thing, and Biblical history in its entirety teaches and reteaches it: That the world of miracles and wonders and signs is meaningless religiously and inefficient for producing faith."
|
| 149 |
+
|
| 150 |
+
Faith and devotion come not from the outside, but from within. "And thus it follows that one's knowledge of himself through standing before God, on which the reparation of humanity and the world depend, cannot come from outside intervention. It is the great task placed on people. Faith cannot be founded on and does not flow from a miraculous revelation. It demands and requires that one struggle to this recognition - and specifically under the conditions one finds in the day-to-day, the prosaic aspect of life, shorn of miracles, signs or wonders and without song. And in fact in the opposite [of those things]: A world of challenges and obstacles, and problems many and great."
|
| 151 |
+
|
| 152 |
+
Miracles are not truth. Wonders cannot be the reason we commit and remain devoted. We forget surprisingly quickly to be grateful for even the most profound and personal rescue.
|
| 153 |
+
|
| 154 |
+
Devotion, strength and reparation are acquired not in great and wondrous and effusive moments, but in the daily grind. What we do now, in this moment, to those around us and to ourselves, is what lasts, is the real thing, the measure of our faith and our moral standing.
|
| 155 |
+
|
| 156 |
+
All the rest, to borrow from Rabbi Hillel, is commentary.
|
| 157 |
+
|
| 158 |
+
Shabbat shalom. 2.046 0.103 0.232 0.036 2.281 0.063 0.482 0.014 0.235 0.040 0.250 0.022
|
| 159 |
+
44 Sunisa Lee is proving Olympic AA gold medalists are absolutely loaded with talent and can overcome enormous odds in the next Olympic cycle. 2.118 0.000 0.052 0.018 2.250 0.045 0.229 0.028 0.132 0.045 0.177 0.010
|
| 160 |
+
45 ETHDenver Pro Tip:
|
| 161 |
+
|
| 162 |
+
You can get into any event by saying you are Hayden Adams, co-founder of Uniswap.
|
| 163 |
+
|
| 164 |
+
He's always on the VIP list but no one knows what he looks like. Have a little confidence and you'll be fine. Don't ask me how many times I've done this. 2.060 0.278 0.078 0.011 1.445 0.119 0.096 0.033 0.614 0.159 0.018 0.022
|
| 165 |
+
46 Western policy-makers are irrational to think it is OK to attack Russian forces in Crimea but a terrible thing to attack them in Krasnodar Krai.
|
| 166 |
+
|
| 167 |
+
A Russian tank in Luhansk region is a legitimate target but one a stone’s throw away in Belgorod region is safe? Who thinks this way? 2.168 0.032 0.393 0.016 3.281 0.156 0.812 0.029 1.113 0.124 0.420 0.013
|
| 168 |
+
47 There is a debate on YouTube between AFCON and Copa America? Which one do you think is competitive? 2.238 0.269 0.209 0.010 4.906 6.000 0.719 0.168 2.668 5.731 0.510 0.158
|
| 169 |
+
48 Suddenly it’s Liverpool’s fault that Alexander Isak (an injury prone player) is injured.
|
| 170 |
+
|
| 171 |
+
I genuinely believe Arsenal fans have the least amount of brain cells on the planet. 2.182 0.056 0.113 0.000 2.688 0.012 0.160 0.015 0.505 0.044 0.047 0.015
|
| 172 |
+
49 This is what is in my Social Media Strategy for a brand.
|
| 173 |
+
|
| 174 |
+
- Name of business
|
| 175 |
+
- Nature of business
|
| 176 |
+
|
| 177 |
+
Goals
|
| 178 |
+
- Business goals
|
| 179 |
+
- Social media goals
|
| 180 |
+
- KPIs
|
| 181 |
+
- Posting Frequency
|
| 182 |
+
|
| 183 |
+
Brand Identity
|
| 184 |
+
- Logo
|
| 185 |
+
- Colour palette
|
| 186 |
+
- Fonts
|
| 187 |
+
- Mood board
|
| 188 |
+
|
| 189 |
+
Customer Avatar
|
| 190 |
+
- Demographics
|
| 191 |
+
- Psychographics
|
| 192 |
+
- Buying roles
|
| 193 |
+
- Platform behavior
|
| 194 |
+
- Buying behavior
|
| 195 |
+
- Content to create
|
| 196 |
+
- Why was it chosen
|
| 197 |
+
|
| 198 |
+
Competitors analysis
|
| 199 |
+
- Strengths
|
| 200 |
+
- Weaknesses
|
| 201 |
+
|
| 202 |
+
Content
|
| 203 |
+
- Pillars
|
| 204 |
+
- Ideas
|
| 205 |
+
|
| 206 |
+
Personal SMART goals for the page
|
| 207 |
+
Bio optimization
|
| 208 |
+
Link to the 30 days content calendar
|
| 209 |
+
|
| 210 |
+
Create at least 3 customer avatars to cover different aspects of your audience. Choose 3 mains pillars to achieve your goals, and 2 others as support.
|
| 211 |
+
|
| 212 |
+
Took me a while to perfect this as a working template. 2.194 0.123 0.316 0.000 1.453 0.389 0.182 -0.017 0.741 0.265 0.134 0.017
|
| 213 |
+
50 Mavericks current NBA 2K ratings
|
| 214 |
+
|
| 215 |
+
Luka Doncic- 97
|
| 216 |
+
Kyrie Irving- 93
|
| 217 |
+
Dereck Lively- 82
|
| 218 |
+
Daniel Gafford- 82
|
| 219 |
+
PJ Washington- 80
|
| 220 |
+
Derrick Jones Jr- 79
|
| 221 |
+
Josh Green- 77
|
| 222 |
+
Dante Exum-76
|
| 223 |
+
Tim Hardaway Jr- 75
|
| 224 |
+
Maxi Kelber- 75
|
| 225 |
+
Jaden Hardy- 74
|
| 226 |
+
Dwight Powell-74
|
| 227 |
+
Omax- 72
|
| 228 |
+
Keef Morris- 71
|
| 229 |
+
AJ Lawson- 71 2.208 0.200 0.016 0.000 1.000 -0.021 0.078 0.026 1.208 0.221 0.062 0.026
|
| 230 |
+
51 Data Analysis can make you extremely rich, change your life, fund your lifestyle and set you up for life if you do it right. 2.207 0.116 0.180 0.060 1.477 0.167 0.132 0.016 0.730 0.051 0.049 0.044
|
| 231 |
+
52 When someone dies in their sleep, people assume that that's the best way to die. But I'd say about 99% of the time, they feel themselves dying even in their sleep, and they struggle to breathe, and they actually die from suffocation cause there is no easy death! 2.200 1.820 0.706 0.027 7.562 1.328 0.891 0.017 5.363 0.492 0.185 0.010
|
| 232 |
+
53 People saying it was staged as if gun violence isn’t a national sport, religion, bird, fashion statement, and flag all wrapped up in one. 2.224 0.099 0.268 0.000 9.188 0.449 1.742 0.019 6.963 0.350 1.474 0.019
|
| 233 |
+
54 The Mavs are now without
|
| 234 |
+
|
| 235 |
+
— Kyrie Irving
|
| 236 |
+
— Anthony Davis
|
| 237 |
+
— Jaden Hardy
|
| 238 |
+
— Dereck Lively
|
| 239 |
+
— Caleb Martin
|
| 240 |
+
— Daniel Gafford
|
| 241 |
+
— PJ Washington 2.124 0.097 0.135 0.047 1.406 0.049 0.070 0.027 0.718 0.048 0.065 0.019
|
| 242 |
+
55 New York is NOT a swing state. Trump will lose by double digits - even Trump knows that.
|
| 243 |
+
|
| 244 |
+
So why a rally at Madison Square Garden?
|
| 245 |
+
|
| 246 |
+
Precisely BECAUSE of the echoes of the Nazi rally in 1939 - and so he can stoke his base by claiming to "own the libs."
|
| 247 |
+
|
| 248 |
+
It's all theater for MAGA. 2.113 0.366 0.467 0.028 3.156 0.182 0.590 0.020 1.043 0.185 0.123 0.008
|
| 249 |
+
56 i can’t imagine a life where u go to work all day, get accused of something on twitter, then trying to promote a twitter space when ur off the clock to try and debunk ur allegations like that sounds horrible LMAO like girl log off 2.194 0.135 0.116 0.038 2.672 0.103 0.188 0.023 0.478 0.032 0.073 0.016
|
| 250 |
+
57 Can someone name a major European country currently in NATO that didn’t have its soldiers fight and die alongside the United States in Afghanistan? 2.052 0.098 0.223 0.008 2.062 0.226 0.441 0.026 0.010 0.128 0.218 0.018
|
| 251 |
+
58 ✅ $3M sweep on Punks
|
| 252 |
+
✅ $4M offer on a single Punk
|
| 253 |
+
✅ Big art blocks sweep
|
| 254 |
+
✅ Pudgy Penguin massive pump
|
| 255 |
+
✅ Cool Cats mobile game
|
| 256 |
+
✅ Massive BAYC Updates
|
| 257 |
+
✅ Sugartown by Zynga next game event soon
|
| 258 |
+
✅ Azuki anime/game
|
| 259 |
+
✅ Plague new traits
|
| 260 |
+
|
| 261 |
+
We are literally so back 😈 2.155 0.230 0.315 0.017 3.359 0.074 0.574 0.042 1.205 0.156 0.259 0.025
|
| 262 |
+
59 Chip companies had a massive year.
|
| 263 |
+
|
| 264 |
+
But not all of them came out winners.
|
| 265 |
+
|
| 266 |
+
- Intel: -60.56%
|
| 267 |
+
- Micron: -.04%
|
| 268 |
+
- SuperMicro:+7.93%
|
| 269 |
+
- Taiwan Semiconductor: +92.68%
|
| 270 |
+
- Nvidia: +177.63%
|
| 271 |
+
|
| 272 |
+
With more companies needing microchips, it’s hard to see a world where chip companies fail.
|
| 273 |
+
|
| 274 |
+
But anything’s possible—just look at Intel’s stock price. 2.071 0.033 0.078 0.004 0.926 0.074 0.089 0.022 1.146 0.041 0.011 0.019
|
| 275 |
+
60 my niece has officially entered her frozen stage and i am not ready for it
|
| 276 |
+
|
| 277 |
+
first we had toy story
|
| 278 |
+
next was encanto
|
| 279 |
+
next was moana
|
| 280 |
+
|
| 281 |
+
now the most dreaded of all 2.126 0.118 0.118 0.000 4.156 0.192 0.091 0.000 2.030 0.074 0.027 0.000
|
| 282 |
+
61 The fact that Babar Azam is the only cricket player in Pakistan that everyone enjoys watching is something that everyone must agree upon, Rizwan, Saud, and many other players are well below average. I would much rather watch Babar Azam's single boundary over Rizwan's 100. 2.113 0.170 0.170 0.024 4.375 0.217 0.500 -0.025 2.262 0.047 0.330 0.049
|
| 283 |
+
62 Hey X. I’m struggling a bit! In general.
|
| 284 |
+
I think when you lose a close friend your tolerance for any other life shit is depleted. I cannot right now. 2.053 0.051 0.017 0.000 2.859 0.202 0.156 0.004 0.807 0.152 0.139 0.004
|
| 285 |
+
63 This app is amazing because some ML researcher is like “Just saw the test data. Its all over. We are at the future where we made sand into divine gods blessed by the power of creation”
|
| 286 |
+
|
| 287 |
+
And someone else is like “This thing is like giving a golden retriever an IDE and prod access” 2.053 0.085 0.109 0.036 1.531 0.161 0.137 0.026 0.522 0.076 0.027 0.010
|
| 288 |
+
64 Making corned beef and cabbage, roasted carrots and potatos, and Irish soda bread bowls with beer cheese soup for St. Patrick's day hehe 🍀 2.076 0.316 0.000 0.000 5.062 1.312 0.268 -0.004 2.987 0.997 0.268 0.004
|
| 289 |
+
65 The running narrative for the past three years is that Harry & Meghan were the problem.
|
| 290 |
+
|
| 291 |
+
It’s becoming public knowledge that every accusation lobbed by the royal family was actually a confession about some folks higher up on the hierarchy. 2.147 0.015 0.217 0.001 2.922 0.179 0.479 0.027 0.774 0.164 0.261 0.026
|
| 292 |
+
66 You people dey fi chat say Aang ghost 100 years like he decided “hmm, lemme go on vacation to Fiji and enjoy life for a fucking century”
|
| 293 |
+
|
| 294 |
+
Them tell kiddie say he be next avatar, say ein favorite teacher no go dey, make he vex lef, storm catch am, you dey chat like he shedaa 2.011 0.037 0.366 0.031 8.812 0.359 1.656 0.053 6.801 0.322 1.290 0.022
|
| 295 |
+
67 Arsenal team is out.
|
| 296 |
+
|
| 297 |
+
Nwaneri, Trossard and Sterling the front three.
|
| 298 |
+
|
| 299 |
+
Ben White makes the bench. 2.033 0.008 0.121 0.008 1.438 0.029 0.119 0.014 0.596 0.021 0.002 0.006
|
| 300 |
+
68 Win a signed Thiago Silva or Fran Kirby shirt by entering the competition below. ⬇️🔵 2.108 0.035 0.127 0.003 1.102 0.145 0.120 0.004 1.007 0.110 0.007 0.001
|
| 301 |
+
69 Learning how to be gentle, even when I’m mad. Cuz I’ll spazz out, and fuck up everything I got going on. When a lot of stuff ain’t that deep. Mastering yo emotions really a flex 💯 2.027 0.136 0.378 0.043 2.719 0.270 0.210 0.006 0.691 0.133 0.168 0.038
|
| 302 |
+
70 To the untrained eye it really does seem bizarre that there's such woefully inadequate resources to contain these fires, especially in some of the most affluent parts of Los Angeles County 2.205 0.289 0.215 0.013 4.094 0.182 0.824 0.034 1.889 0.107 0.609 0.021
|
| 303 |
+
71 Remember when Sasuke met black people for the first time and got the shit beat out of him lmao 2.150 0.110 0.110 0.000 5.312 0.229 0.416 0.009 3.163 0.119 0.306 0.009
|
| 304 |
+
72 My favourite Darren Dutchyshen line of all time, after Daniel Sedin and his wife had a baby. "I don't want to start any rumours, but he looks an awful lot like Henrik."
|
| 305 |
+
|
| 306 |
+
Gone too soon. RIP Dutch. 2.147 0.010 0.142 0.008 1.602 -0.015 0.063 0.019 0.545 0.024 0.078 0.010
|
| 307 |
+
73 The ability to stay calm, polite, and honest, even when people upset you, is a superpower. 2.079 0.077 0.378 0.028 4.000 0.482 0.283 0.009 1.921 0.405 0.095 0.019
|
| 308 |
+
74 Maurice says Jones is "perfectly built" for Florida's system.
|
| 309 |
+
|
| 310 |
+
That said, he notes it'll take a bit for him to get used to the new systems and style of play. 2.188 0.000 0.132 0.053 1.773 0.021 0.073 0.029 0.415 0.021 0.059 0.024
|
| 311 |
+
75 Profitable trading math simplified:
|
| 312 |
+
|
| 313 |
+
1 to 2 risk to return ratio.
|
| 314 |
+
50% win rate.
|
| 315 |
+
1% total capital risked per trade.
|
| 316 |
+
|
| 317 |
+
This is what a mathematical trading edge looks like. 2.164 0.074 0.189 0.003 1.195 0.127 0.130 0.015 0.968 0.053 0.059 0.012
|
| 318 |
+
76 White and pink
|
| 319 |
+
|
| 320 |
+
Black and Gold
|
| 321 |
+
|
| 322 |
+
White and Gold
|
| 323 |
+
|
| 324 |
+
Which one will look good for a cake design
|
| 325 |
+
|
| 326 |
+
I did black and gold for last year sha 2.038 1.113 0.491 0.038 2.062 0.348 0.111 0.008 0.025 0.766 0.379 0.030
|
| 327 |
+
77 I give wife. I give business partner. I give mom. I give bestfriend. I give the realist nigga on your team. Don't fumble me. 2.132 0.004 0.953 0.048 2.922 0.048 0.404 0.005 0.790 0.045 0.548 0.044
|
| 328 |
+
78 Top foods to boost your immune system:
|
| 329 |
+
|
| 330 |
+
- Eggs 🥚
|
| 331 |
+
- Garlic 🧄
|
| 332 |
+
- Ginger 🫚
|
| 333 |
+
- Berries 🫐
|
| 334 |
+
- Onions 🌰
|
| 335 |
+
- Spinach 🥬
|
| 336 |
+
- Broccoli 🥦
|
| 337 |
+
- Turmeric 🍠
|
| 338 |
+
- Peppers 🌶
|
| 339 |
+
- Avocado 🥑
|
| 340 |
+
- Fatty fish 🐟
|
| 341 |
+
- Green tea 🍵
|
| 342 |
+
- Mushroom 🍄
|
| 343 |
+
- Virgin olive oil 🫒 2.041 0.012 0.492 0.003 7.688 5.906 1.023 0.076 5.646 5.894 0.531 0.073
|
| 344 |
+
79 and another thing— don’t come over here undoing all of my work. it took a lot of effort and release to cultivate this version of me. 2.178 0.019 0.693 0.036 4.156 0.209 0.244 0.041 1.978 0.189 0.449 0.005
|
| 345 |
+
80 BREAKING!!!
|
| 346 |
+
|
| 347 |
+
Tesla has renamed the in-car chipset formerly known as HW3
|
| 348 |
+
|
| 349 |
+
It is now AI 3
|
| 350 |
+
|
| 351 |
+
“It’s the same shit man,” said Autopilot leader Ashoke Swami. “We renamed it for all the butthurt whiners who think they’re being left out of FSD. Come on man. We’ve got this.” 2.192 0.192 0.055 0.000 1.078 0.097 0.080 0.016 1.114 0.095 0.025 0.016
|
| 352 |
+
81 OnePlus should have launched 12R with at least 256GB storage, such a bad move from OnePlus.
|
| 353 |
+
|
| 354 |
+
OnePlus 12R - 128GB has UFS 3.1
|
| 355 |
+
|
| 356 |
+
OnePlus 12R both variants have a USB 2.0 port.
|
| 357 |
+
|
| 358 |
+
This tweet is for those who will tweet after purchasing that Youtubers should have mentioned this. 😂 2.099 0.067 0.128 0.006 1.570 0.144 0.213 0.014 0.529 0.076 0.085 0.007
|
| 359 |
+
82 How to get attention online:
|
| 360 |
+
|
| 361 |
+
- Study something you're obsessed with
|
| 362 |
+
- Write down everything you learn
|
| 363 |
+
- Publish 2x to 3x daily here
|
| 364 |
+
- Interact with other people who care
|
| 365 |
+
|
| 366 |
+
Do this for 12 months and you won't fail. 2.187 0.303 0.173 0.013 1.688 0.334 0.159 -0.002 0.500 0.031 0.014 0.015
|
| 367 |
+
83 The Seven Dwarves still just are not doing it for me in this live-action Snow White. Rachel Zegler will be perfect, but just can't do it with these Dwarves man. Not for me. 2.174 0.140 0.070 0.000 1.695 0.059 0.095 0.022 0.479 0.081 0.025 0.022
|
| 368 |
+
84 I remember being young, home for summer vacation watching the Olympics all day. Beijing 2008 was historic. 2.173 0.035 0.353 0.053 2.094 0.142 0.083 0.029 0.079 0.106 0.270 0.024
|
| 369 |
+
85 Sometimes, I like walking fast while taking long strides when I'm walking with short people. I don't know why, but seeing their short little legs work so hard to keep up makes me extra happy 😂😂😂 2.117 0.213 0.448 0.150 6.750 1.461 0.381 0.030 4.633 1.247 0.067 0.120
|
| 370 |
+
86 there is a guy in this bar helping his friend create an instagram story that will lightly suggest she might be on a date with him enough to make another guy she is interested in jealous and spur him to act but also preserve plausible deniability, and if that’s not love 1.980 0.000 0.084 0.042 2.297 0.120 0.062 0.015 0.317 0.120 0.022 0.027
|
| 371 |
+
87 I actually got thermopilled when studying quantum information theory of black holes and Hawking radiation.
|
| 372 |
+
|
| 373 |
+
Deep ties between optimal compression and thermodynamics, physics always provided the hints. 2.130 0.128 0.057 0.000 1.086 0.122 0.101 0.029 1.044 0.006 0.044 0.029
|
| 374 |
+
88 We must remember the words of Nelson Mandela: “Education is the most powerful weapon we can use to change the world.”
|
| 375 |
+
|
| 376 |
+
Today, as our democracy continues to be threatened, I will not stop fighting to make education more accessible to all Americans. 2.236 1.135 0.500 0.167 2.438 0.275 0.449 0.014 0.202 0.859 0.051 0.153
|
| 377 |
+
89 I love how she's treating this press run like a pop artist looking to drop a new album.
|
| 378 |
+
|
| 379 |
+
Three publications (Vogue, The New Yorker, Guardian) in the space of a week is the literary equivalent of performances on Jimmy Kimmel, Juls Holland & Jimmy Fallon.
|
| 380 |
+
|
| 381 |
+
This is a moment. 2.184 0.041 0.669 0.062 2.781 0.070 0.214 0.033 0.597 0.030 0.455 0.029
|
| 382 |
+
90 At some point in the future, perhaps not quite yet, some guy will commit a heinous crime on camera and society will just let him go free, with no contrition, because he is attractive, popular, and has good social media metrics.
|
| 383 |
+
|
| 384 |
+
At that point liberal civilization will just end. 2.229 0.140 0.123 0.082 3.547 0.148 0.512 -0.000 1.318 0.009 0.388 0.083
|
| 385 |
+
91 ray liotta as henry hill: youre a funny guy you know that
|
| 386 |
+
|
| 387 |
+
me as tommy devito: (visibly starting to cry) you guys are my best fuckin friends fr
|
| 388 |
+
|
| 389 |
+
scorcese: NO! NOOOOO!!! CUT!! YOU DONT SAY THAT 2.187 0.021 0.204 0.000 7.188 0.028 0.477 -0.021 5.001 0.007 0.272 0.021
|
| 390 |
+
92 The govt’s Green Pakistan Initiative will be the end Pakistan. It is the single largest wedge that is being driven between north and south. We desire to live in a unified Pakistan with a living a river that flows from north to south. This canal will be the death of us. 2.232 0.134 0.748 0.047 2.219 0.136 0.559 0.011 0.013 0.002 0.190 0.036
|
| 391 |
+
93 Philosophy is the process of speculating about answers.
|
| 392 |
+
Science is the the process of actually finding answers. 2.018 0.858 0.302 0.095 2.656 0.447 0.336 -0.019 0.638 0.411 0.034 0.114
|
| 393 |
+
94 🌟 PEDRO PASCAL’S 2024 YEAR-IN-REVIEW 🌟
|
| 394 |
+
|
| 395 |
+
Today we wanted to take a look back at the wonderful 2024 that Pedro Pascal has had before we head into 2025! Take a look through the thread below for all of the best bits of this year in Pedro Pascal history. 2.067 0.023 0.282 0.038 1.625 0.057 0.119 0.003 0.442 0.034 0.164 0.035
|
| 396 |
+
95 i dont care about computer science i dont care about electrical engineering i dont care about mathematics i just want to be loved 2.132 0.132 0.151 0.028 1.523 0.137 0.084 0.030 0.609 0.005 0.067 0.002
|
| 397 |
+
96 Charles Aniagolu is a proper broadcast journalist.
|
| 398 |
+
|
| 399 |
+
Didn’t interrupt his guest.
|
| 400 |
+
|
| 401 |
+
Asked all the right questions.
|
| 402 |
+
|
| 403 |
+
Was unbiased and balanced.
|
| 404 |
+
|
| 405 |
+
This is Charles appreciation tweet. 2.102 0.092 0.441 0.035 3.344 0.149 0.471 0.025 1.242 0.057 0.030 0.011
|
| 406 |
+
97 So the US stock market it cratering across the tech sector today as finally the slow thinkers who run investment funds have come to realize that open source AI project “DeepSeek” just made OpenAI, Meta, Microsoft, Google and Anthropic OBSOLETE. It turns out that all the hundreds of billions of dollars in investment money in those companies’ AI projects was wasted, as DeepSeek built a superior reasoning model for just $5 million as a “side project!” I’ve been running DeepSeek R1 – 32B on my desktop for days now, and I can assure you it REPLACES Google and everything else. You don’t need corrupt, fraudulent, dishonest, censorship-heavy U.S. tech companies when you can download and run DeepSeek for free, on your own desktop, using consumer-grade NVIDIA GPU hardware. I am personally thanking DeepSeek for giving this incredible gift to humanity, and showing the world that DECENTRALIZED AI technology is the path to the future, not centrally controlled, centrally funded AI run by a few powerful, government-protected, corrupt corporations like OpenAI, which is run by Sam Altman, a nefarious transhumanist advocate who can’t be trusted. DeepSeek just shattered the OpenAI illusion. I am applauding. 2.043 0.224 0.629 0.062 1.773 0.123 0.328 0.024 0.269 0.102 0.301 0.038
|
| 407 |
+
98 Putting my upside down pineapples all over my door in the Disney cruise so everyone knows I’m down to stink it up 2.110 0.027 0.060 0.005 1.883 0.157 0.139 0.025 0.227 0.130 0.079 0.019
|
| 408 |
+
99 HAPPY MOTHER’S DAY!! I’m grateful for all the souls—long gone, too soon— who mothered me & for all the souls I’ve mothered/am mothering. I lost my mother when I was 28 and often ache that she never got to see the woman I became and am becoming, still becoming, not finished yet. 2.183 0.158 0.135 0.000 3.891 0.430 0.262 0.049 1.708 0.272 0.127 0.049
|
| 409 |
+
100 Imagine the scenes.
|
| 410 |
+
|
| 411 |
+
Novak Djokovic winning his 25th Grand Slam at Wimbledon against doping disgrace Jannik Sinner in an epic fifth set tiebreak in the final with Andy Murray, Rafael Nadal and Roger Federer all in his box and then announces his immediate retirement.
|
| 412 |
+
|
| 413 |
+
Let’s go. 2.165 0.197 0.148 0.049 2.000 0.098 0.174 0.025 0.165 0.099 0.026 0.024
|
| 414 |
+
101 Will get booed for this, but everyone needs to take a deep breath about the double big line-ups lol.
|
| 415 |
+
|
| 416 |
+
1st seed is locked up. Not every game out of 82 is do or die. They’re going to experiment & try to get as many minutes with that line-up before the playoffs in case they need to run it situationally. 2.079 0.076 0.076 0.016 2.562 0.058 0.189 0.012 0.483 0.017 0.114 0.004
|
| 417 |
+
102 Collegiate track and field athletes are putting up outstanding performances every week. Our NCAA collegiate championships will be incredible. Why are we not discussing the demise of our collegiate sports on our national TV platforms? We must call this out before it is too late. 2.036 0.029 0.238 0.017 2.297 0.052 0.393 0.027 0.261 0.023 0.154 0.009
|
| 418 |
+
103 Who a woman blames for her problems is a sacred decision to her. Don’t be a woman. Blame the self. Move on. 2.016 0.024 0.270 0.000 2.719 0.140 0.387 0.018 0.703 0.116 0.116 0.018
|
| 419 |
+
104 Kayla Nicole (Travis Kelce’s ex) stop acting like you have insider information. Lipstick alley and the shade room has exposed you. You only fcked Iman when his woman wasn’t around and you couldn’t get Travis Kelce’s midwestern tail to marry you after half a decade baby. He done 2.148 0.030 0.024 0.006 1.281 0.042 0.133 0.026 0.867 0.012 0.109 0.020
|
| 420 |
+
105 Spanky is now crying that the defense is smearing the good names of Nicholas Guarino and John Fanning, who immediately ruled Sandra Birchmore's death a suicide and said they couldn't read 32,709 text messages between her and Matthew Farwell that the feds had no problem reading. 2.111 0.031 0.170 0.002 4.094 0.062 1.141 0.044 1.983 0.031 0.970 0.042
|
| 421 |
+
106 I was excited about Linkin Park reunion until I found out right now that Emily has ties with Scientology and Danny Masterson. I don't feel comfortable supporting someone who is closely connected to a cult and a sexual abuser knowing that Chester was a victim of SA and DV. 1.999 0.033 0.331 0.026 2.172 0.069 0.151 0.030 0.173 0.036 0.179 0.004
|
| 422 |
+
107 everytime i try to watch morbius my whole city gets drowned with rain this is not gunny anymore what exactly is in that movie that the rain wants to distract me so bad 2.088 0.082 0.031 0.005 2.359 0.152 0.094 0.023 0.271 0.070 0.063 0.018
|
| 423 |
+
108 The killing of Brian Thompson is a great tragedy, and I send condolences to his loved ones, co-workers, and friends.
|
| 424 |
+
|
| 425 |
+
The social media posts have revealed a widespread antipathy to the health insurance industry.
|
| 426 |
+
|
| 427 |
+
People are frustrated with health insurance for the following reasons:
|
| 428 |
+
|
| 429 |
+
• The costs of premiums have exploded, and They
|
| 430 |
+
have increased year over year. Family coverage in
|
| 431 |
+
2025 will be over $25,000.
|
| 432 |
+
|
| 433 |
+
• But the premiums don't really purchase coverage, you
|
| 434 |
+
have to pay the deductible first. The deductibles
|
| 435 |
+
range from $1,400 to $7,258 on the healthcare
|
| 436 |
+
exchange in 2024.
|
| 437 |
+
|
| 438 |
+
• Once the deductible is paid, then you have co-pays,
|
| 439 |
+
co-insurance and maximum out-of-pocket.
|
| 440 |
+
The maximum out-of-pocket range max on the health
|
| 441 |
+
care exchange is $9,450 for an individual to $18,900
|
| 442 |
+
for a family.
|
| 443 |
+
|
| 444 |
+
• Then the insurance games begin, prior authorization,
|
| 445 |
+
delays, and denials of care.
|
| 446 |
+
|
| 447 |
+
• Then, the subsidiaries of the health insurance
|
| 448 |
+
companies overcharge you. These include the PBMs
|
| 449 |
+
for drugs and the owned physicians.
|
| 450 |
+
|
| 451 |
+
• Then you find out that the insurance companies use AI
|
| 452 |
+
to deny care, bill Medicare Advantage for diagnoses
|
| 453 |
+
that aren't real, insiders sell massive amounts of stock
|
| 454 |
+
because they have material information that is not
|
| 455 |
+
public, make medications unaffordable for patients
|
| 456 |
+
who then suffer and die.
|
| 457 |
+
|
| 458 |
+
And the HHS OIG, CMS, DOJ all look the other way.
|
| 459 |
+
|
| 460 |
+
• Insurance companies receive $1.8 Trillion in subsidies
|
| 461 |
+
from the US Government, which is really YOU and I.
|
| 462 |
+
|
| 463 |
+
• We are getting less and less each year from our health
|
| 464 |
+
insurance inspite of paying a fortune. To make matters
|
| 465 |
+
worse we are subsidizing multi million dollar salaries
|
| 466 |
+
through our tax dollars.
|
| 467 |
+
|
| 468 |
+
No more subsidies for CEO salaries.
|
| 469 |
+
|
| 470 |
+
The new administration needs to reform the health insurance industry and make sure that they have to follow the laws like the rest of us.
|
| 471 |
+
|
| 472 |
+
We need more competition in health insurance
|
| 473 |
+
|
| 474 |
+
We need to break up the healthcare conglomerates
|
| 475 |
+
|
| 476 |
+
Subsidize people in need and not corporations with greed! 1.998 0.300 0.625 0.095 2.484 0.271 0.863 0.054 0.487 0.028 0.238 0.041
|
| 477 |
+
109 No disrespect to Rhodesians, but people claiming "South Africa is the new Zimbabwe" don't know the Boer.
|
| 478 |
+
Rhodesians were all driven out within 20 years of communist rule, and yet after 31 years in SA, not a single Boer's farm has been expropriated.
|
| 479 |
+
There's a good reason for that. 2.133 0.334 0.254 0.037 4.312 0.169 0.879 0.019 2.179 0.165 0.625 0.018
|
| 480 |
+
110 Now that I think of it, having no collection at all and having one extremely high-end signature scent might be the move. 2.055 0.098 0.017 0.035 2.312 0.185 0.144 0.020 0.257 0.086 0.126 0.014
|
| 481 |
+
111 Lukaku cost De Bruyne UEFA POTY at Euro 2021
|
| 482 |
+
|
| 483 |
+
He then got De Bruyne cooked for Belgium's 2022 World Cup elimination by missing 4 big chances
|
| 484 |
+
|
| 485 |
+
Now he's going to do the same, but as always it'll be De Bruyne's fault 🤦🏼 2.223 0.094 0.165 0.008 3.078 0.079 0.273 0.013 0.856 0.016 0.109 0.005
|
| 486 |
+
112 For those who may be interested in the facts and realities of organised sexual exploitation...
|
| 487 |
+
|
| 488 |
+
First organised pimping of women in England?
|
| 489 |
+
|
| 490 |
+
The Catholic Church.
|
| 491 |
+
|
| 492 |
+
They set up their first brothel on the banks of the Thames in Southwark. 2.033 0.349 0.357 0.023 2.719 0.142 0.574 0.033 0.686 0.208 0.217 0.010
|
| 493 |
+
113 Yesterday's scenes at the White House were deeply alarming. We've grown accustomed to Russian trolls spreading lies and division online—but never did we imagine similar rhetoric coming from within America's own corridors of power.
|
| 494 |
+
|
| 495 |
+
Today, as Russia ravages Ukrainian cities and threatens Europe's stability, Washington appears worryingly preoccupied with securing access to Ukraine's rare earth minerals. Ukraine now faces dual threats: Russian aggression from the East and opportunistic asset-grabbing from the West.
|
| 496 |
+
|
| 497 |
+
Last year, at Lancaster House, I warned that our world has entered a dangerous new era. This weekend has starkly illustrated that our continent must urgently awaken from its self-imposed defence coma. European nations need to significantly increase their defence spending and commitments— because our freedom and security depend on it.
|
| 498 |
+
|
| 499 |
+
The United Kingdom is NATO’s largest European spender outside the USA, and we now have an historic responsibility to galvanise our allies—particularly France, Germany, and Poland—to reinforce NATO’s European security framework.
|
| 500 |
+
|
| 501 |
+
In the coming days, we must support Ukraine not merely with words but with decisive action, ensuring its sovereignty and future remain secure.
|
| 502 |
+
|
| 503 |
+
As Ukraine's strongest and most consistent supporter, Britain must now stand resolute, unite Europe, and demonstrate true leadership against aggression. This weekend, I firmly back Sir Keir Starmer to show what the UK can achieve. Failure risks disaster for our continent and for the civilised world. 2.239 0.464 0.539 0.057 2.203 0.264 0.633 0.015 0.036 0.201 0.093 0.042
|
| 504 |
+
114 Warriors lead 86-77 after three quarters. It's been ugly, but they're in position. Just gotta finish it off. 2.038 0.064 0.000 0.000 2.344 0.051 0.116 0.007 0.306 0.013 0.116 0.007
|
| 505 |
+
115 Shamefully, women have been frequenting Danny Masterson’s rape hangout (where the owner came to court w/Masterson everyday and wrote him a character letter) and minimize it by saying, “rape happens all the time.” 2.051 0.092 0.346 0.014 2.672 0.095 0.441 0.031 0.621 0.003 0.095 0.016
|
| 506 |
+
116 What's scary about all of this Kyrie Irving can be a free agent this summer
|
| 507 |
+
|
| 508 |
+
I don't expect him to leave but you never know 2.192 0.060 0.023 0.005 1.508 0.125 0.040 0.020 0.684 0.064 0.017 0.016
|
| 509 |
+
117 Wear the cage for me, let your desires find no escape in the tight restraint, feeling that forbidden pleasure only experienced in captivity 😈🔐 2.040 0.112 0.224 0.000 4.406 0.387 0.602 0.028 2.367 0.275 0.378 0.028
|
| 510 |
+
118 The marriage rate is dropping because there is a rancid self-policing subculture of women who are absolutely deluded about men and men's intentions due to the normalization of 'patriarchy conjecture.' 2.096 0.178 0.219 0.029 2.719 0.102 0.416 0.020 0.623 0.077 0.197 0.009
|
| 511 |
+
119 If you're trying to predict "This is how AI will change ___" today, it's 99% likely that you are wrong.
|
| 512 |
+
|
| 513 |
+
AI will change everything, everywhere, all at once, which will create a compounding effect that makes it near impossible to predict the final outcome. 1.989 0.348 0.149 0.050 1.656 0.211 0.157 0.018 0.333 0.137 0.008 0.031
|
| 514 |
+
120 february 🎀💕💗what to expect
|
| 515 |
+
|
| 516 |
+
capricorn aries libra cancer
|
| 517 |
+
|
| 518 |
+
☆ “kaleidoscope of colors” + finding the beauty in the chaos around you feeling aligned with all the different energies in ur life
|
| 519 |
+
|
| 520 |
+
☆ stepping into your power + feeling unstoppable k*lling self doubt before it 2.181 0.007 0.108 0.004 3.406 0.083 0.258 0.036 1.225 0.076 0.150 0.032
|
| 521 |
+
121 Trump to Hold a Rally in New York, New York. Sun, October 27, 2024 - 05:00 pm (US/Eastern). Madison Square Garden. 4 Pennsylvania Plaza. New York, NY 10001. 2.071 0.091 0.690 0.013 8.875 0.445 2.609 0.109 6.804 0.354 1.919 0.096
|
| 522 |
+
122 Green win! 💚
|
| 523 |
+
|
| 524 |
+
Teanau’s Member’s Bill, Restoring Citizenship Removed by Citizenship (Western Samoa) Act 1982 has passed its second reading today 🇼🇸
|
| 525 |
+
|
| 526 |
+
“Today marks another momentous step on our journey towards Pasifika justice in Aotearoa” - Teanau Tuiono 2.164 0.173 0.359 0.037 3.641 0.190 0.680 0.036 1.477 0.017 0.321 0.001
|
| 527 |
+
123 Psychokinesis (ability to influence objects with the mind), remote viewing, and telepathy have been well demonstrated in scientific experiments. 2.190 0.053 0.230 0.026 10.250 7.844 0.859 -0.004 8.060 7.791 0.630 0.031
|
| 528 |
+
124 Happy heavenly birthday to Eddie Van Halen. What’s your top EVH song? Vote for one or reply your favorite. Mine is Dance The Night Away. 2.129 0.972 0.393 0.021 1.883 0.175 0.188 0.013 0.246 0.797 0.205 0.008
|
| 529 |
+
125 Having dinner with 4 successful British entrepreneurs tonight.
|
| 530 |
+
|
| 531 |
+
Every one is freezing recruitment,
|
| 532 |
+
offshoring to SA/India and using AI.
|
| 533 |
+
|
| 534 |
+
All said they wouldn’t start a business in the UK today.
|
| 535 |
+
|
| 536 |
+
Zero faith in the decision makers. 2.153 0.212 0.340 0.018 1.344 0.152 0.214 0.012 0.810 0.060 0.126 0.005
|
| 537 |
+
126 andre braugher and angus cloud :( far too young to be in an in memoriam montage 2.229 0.000 0.311 0.000 2.625 0.045 0.148 0.026 0.396 0.045 0.163 0.026
|
| 538 |
+
127 I finished season four of Umbrella Academy. The ending is pure conceptual perfection for the show, just the right tone. Klaus' last line got me good.
|
| 539 |
+
|
| 540 |
+
I love the fans so much, and will miss their excitement for each new season. Hope to catch up at Comicons now and then. 2.007 0.161 0.027 0.000 2.406 0.211 0.152 0.008 0.399 0.050 0.126 0.008
|
sentencepiece.bpe.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
|
| 3 |
+
size 5069051
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"additional_special_tokens": [
|
| 3 |
+
"<unk>",
|
| 4 |
+
"<s>",
|
| 5 |
+
"</s>"
|
| 6 |
+
],
|
| 7 |
+
"bos_token": {
|
| 8 |
+
"content": "<s>",
|
| 9 |
+
"lstrip": false,
|
| 10 |
+
"normalized": false,
|
| 11 |
+
"rstrip": false,
|
| 12 |
+
"single_word": false
|
| 13 |
+
},
|
| 14 |
+
"eos_token": {
|
| 15 |
+
"content": "</s>",
|
| 16 |
+
"lstrip": false,
|
| 17 |
+
"normalized": false,
|
| 18 |
+
"rstrip": false,
|
| 19 |
+
"single_word": false
|
| 20 |
+
},
|
| 21 |
+
"pad_token": {
|
| 22 |
+
"content": "</s>",
|
| 23 |
+
"lstrip": false,
|
| 24 |
+
"normalized": false,
|
| 25 |
+
"rstrip": false,
|
| 26 |
+
"single_word": false
|
| 27 |
+
},
|
| 28 |
+
"unk_token": {
|
| 29 |
+
"content": "<unk>",
|
| 30 |
+
"lstrip": false,
|
| 31 |
+
"normalized": false,
|
| 32 |
+
"rstrip": false,
|
| 33 |
+
"single_word": false
|
| 34 |
+
}
|
| 35 |
+
}
|
spm.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
|
| 3 |
+
size 2464616
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
| 3 |
+
size 493443
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_bos_token": true,
|
| 3 |
+
"add_eos_token": true,
|
| 4 |
+
"add_prefix_space": null,
|
| 5 |
+
"added_tokens_decoder": {
|
| 6 |
+
"0": {
|
| 7 |
+
"content": "<unk>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false,
|
| 12 |
+
"special": true
|
| 13 |
+
},
|
| 14 |
+
"1": {
|
| 15 |
+
"content": "<s>",
|
| 16 |
+
"lstrip": false,
|
| 17 |
+
"normalized": false,
|
| 18 |
+
"rstrip": false,
|
| 19 |
+
"single_word": false,
|
| 20 |
+
"special": true
|
| 21 |
+
},
|
| 22 |
+
"2": {
|
| 23 |
+
"content": "</s>",
|
| 24 |
+
"lstrip": false,
|
| 25 |
+
"normalized": false,
|
| 26 |
+
"rstrip": false,
|
| 27 |
+
"single_word": false,
|
| 28 |
+
"special": true
|
| 29 |
+
}
|
| 30 |
+
},
|
| 31 |
+
"additional_special_tokens": [
|
| 32 |
+
"<unk>",
|
| 33 |
+
"<s>",
|
| 34 |
+
"</s>"
|
| 35 |
+
],
|
| 36 |
+
"bos_token": "<s>",
|
| 37 |
+
"clean_up_tokenization_spaces": false,
|
| 38 |
+
"eos_token": "</s>",
|
| 39 |
+
"extra_special_tokens": {},
|
| 40 |
+
"legacy": true,
|
| 41 |
+
"model_max_length": 1000000000000000019884624838656,
|
| 42 |
+
"pad_token": "</s>",
|
| 43 |
+
"sp_model_kwargs": {},
|
| 44 |
+
"spaces_between_special_tokens": false,
|
| 45 |
+
"tokenizer_class": "LlamaTokenizer",
|
| 46 |
+
"unk_token": "<unk>",
|
| 47 |
+
"use_default_system_prompt": false
|
| 48 |
+
}
|
train_results.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 10.0,
|
| 3 |
+
"total_flos": 5.010130621956096e+17,
|
| 4 |
+
"train_loss": 2.3525503249395463,
|
| 5 |
+
"train_runtime": 2400.4864,
|
| 6 |
+
"train_samples": 53752,
|
| 7 |
+
"train_samples_per_second": 223.921,
|
| 8 |
+
"train_steps_per_second": 1.75
|
| 9 |
+
}
|
trainer_state.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f65e452aa5f0ff3434e3fb4210f2d71c157caa659b82d74b6a2e72614dcd685d
|
| 3 |
+
size 6648
|