mav23 commited on
Commit
9d40342
·
verified ·
1 Parent(s): 146de44

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ meta-llama-3-120b-instruct.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,113 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ tags:
4
+ - merge
5
+ - mergekit
6
+ - lazymergekit
7
+ base_model:
8
+ - meta-llama/Meta-Llama-3-70B-Instruct
9
+ - meta-llama/Meta-Llama-3-70B-Instruct
10
+ - meta-llama/Meta-Llama-3-70B-Instruct
11
+ - meta-llama/Meta-Llama-3-70B-Instruct
12
+ - meta-llama/Meta-Llama-3-70B-Instruct
13
+ - meta-llama/Meta-Llama-3-70B-Instruct
14
+ - meta-llama/Meta-Llama-3-70B-Instruct
15
+ ---
16
+
17
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/C-Xw_m97bhXaTA1TEpHB7.jpeg)
18
+
19
+ # Meta-Llama-3-120B-Instruct
20
+
21
+ Meta-Llama-3-120B-Instruct is a [meta-llama/Meta-Llama-3-70B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) self-merge made with [MergeKit](https://github.com/arcee-ai/mergekit/tree/main).
22
+
23
+ It was inspired by large merges like:
24
+
25
+ - [alpindale/goliath-120b](https://huggingface.co/alpindale/goliath-120b)
26
+ - [nsfwthrowitaway69/Venus-120b-v1.0](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.0)
27
+ - [cognitivecomputations/MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b)
28
+ - [wolfram/miquliz-120b-v2.0](https://huggingface.co/wolfram/miquliz-120b-v2.0).
29
+
30
+ Special thanks to [Eric Hartford](https://huggingface.co/ehartford) for both inspiring and evaluating this model and to [Charles Goddard](https://huggingface.co/chargoddard) for creating MergeKit.
31
+
32
+ ## 🔍 Applications
33
+
34
+ I recommend using this model for creative writing. It uses the Llama 3 chat template with a default context window of 8K (can be extended with rope theta).
35
+
36
+ Check the examples in the evaluation section to get an idea of its performance. The model is generally quite unhinged but has a good writing style. It sometimes outputs typos and is a big fan of uppercase.
37
+
38
+ ## ⚡ Quantized models
39
+
40
+ Thanks to [Bartowski](https://huggingface.co/ehartford), [elinas](https://huggingface.co/elinas), the [mlx-community](https://huggingface.co/mlx-community) and others for providing these models.
41
+
42
+ * **GGUF**: https://huggingface.co/lmstudio-community/Meta-Llama-3-120B-Instruct-GGUF
43
+ * **EXL2**: https://huggingface.co/elinas/Meta-Llama-3-120B-Instruct-4.0bpw-exl2
44
+ * **mlx**: https://huggingface.co/mlx-community/Meta-Llama-3-120B-Instruct-4bit
45
+
46
+ ## 🏆 Evaluation
47
+
48
+ This model is great for creative writing but struggles in other tasks. I'd say use it with caution and don't expect it to outperform GPT-4 outside of some very specific use cases.
49
+
50
+ * **X thread by Eric Hartford (creative writing)**: https://twitter.com/erhartford/status/1787050962114207886
51
+ * **X thread by Daniel Kaiser (creative writing)**: https://twitter.com/spectate_or/status/1787257261309518101
52
+ * **X thread by Simon (reasoning)**: https://twitter.com/NewDigitalEdu/status/1787403266894020893
53
+ * **r/LocalLLaMa**: https://www.reddit.com/r/LocalLLaMA/comments/1cl525q/goliath_lovers_where_is_the_feedback_about/
54
+
55
+ ### Creative Writing
56
+
57
+ Thanks to [Sam Paech](https://huggingface.co/sam-paech) for evaluating this model and sending me his outputs!
58
+
59
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/-LJ7ivCRIPR1ur-LJHk3m.png)
60
+
61
+ ## 🧩 Configuration
62
+
63
+ ```yaml
64
+ slices:
65
+ - sources:
66
+ - layer_range: [0, 20]
67
+ model: meta-llama/Meta-Llama-3-70B-Instruct
68
+ - sources:
69
+ - layer_range: [10, 30]
70
+ model: meta-llama/Meta-Llama-3-70B-Instruct
71
+ - sources:
72
+ - layer_range: [20, 40]
73
+ model: meta-llama/Meta-Llama-3-70B-Instruct
74
+ - sources:
75
+ - layer_range: [30, 50]
76
+ model: meta-llama/Meta-Llama-3-70B-Instruct
77
+ - sources:
78
+ - layer_range: [40, 60]
79
+ model: meta-llama/Meta-Llama-3-70B-Instruct
80
+ - sources:
81
+ - layer_range: [50, 70]
82
+ model: meta-llama/Meta-Llama-3-70B-Instruct
83
+ - sources:
84
+ - layer_range: [60, 80]
85
+ model: meta-llama/Meta-Llama-3-70B-Instruct
86
+ merge_method: passthrough
87
+ dtype: float16
88
+ ```
89
+
90
+ ## 💻 Usage
91
+
92
+ ```python
93
+ !pip install -qU transformers accelerate
94
+
95
+ from transformers import AutoTokenizer
96
+ import transformers
97
+ import torch
98
+
99
+ model = "mlabonne/Meta-Llama-3-120B-Instruct"
100
+ messages = [{"role": "user", "content": "What is a large language model?"}]
101
+
102
+ tokenizer = AutoTokenizer.from_pretrained(model)
103
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
104
+ pipeline = transformers.pipeline(
105
+ "text-generation",
106
+ model=model,
107
+ torch_dtype=torch.float16,
108
+ device_map="auto",
109
+ )
110
+
111
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
112
+ print(outputs[0]["generated_text"])
113
+ ```
meta-llama-3-120b-instruct.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:033d960a145208d9537ef1f6db64f560c0785760960747b6516a7de379e655cc
3
+ size 45098778016