test
#1
by
van-qa
- opened
- .gitattributes +0 -11
- README.md +5 -6
- llama-3.1-8b-instruct-q2_k.gguf +0 -3
- llama-3.1-8b-instruct-q3_k_l.gguf +0 -3
- llama-3.1-8b-instruct-q3_k_m.gguf +0 -3
- llama-3.1-8b-instruct-q3_k_s.gguf +0 -3
- llama-3.1-8b-instruct-q4_k_m.gguf +0 -3
- llama-3.1-8b-instruct-q4_k_s.gguf +0 -3
- llama-3.1-8b-instruct-q5_k_m.gguf +0 -3
- llama-3.1-8b-instruct-q5_k_s.gguf +0 -3
- llama-3.1-8b-instruct-q6_k.gguf +0 -3
- llama-3.1-8b-instruct-q8_0.gguf +0 -3
- metadata.yml +0 -5
- model.yml +0 -23
.gitattributes
CHANGED
|
@@ -33,14 +33,3 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
-
model.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
-
llama-3.1-8b-instruct-q2_k.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
-
llama-3.1-8b-instruct-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text
|
| 39 |
-
llama-3.1-8b-instruct-q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
| 40 |
-
llama-3.1-8b-instruct-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 41 |
-
llama-3.1-8b-instruct-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
| 42 |
-
llama-3.1-8b-instruct-q4_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 43 |
-
llama-3.1-8b-instruct-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
| 44 |
-
llama-3.1-8b-instruct-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 45 |
-
llama-3.1-8b-instruct-q6_k.gguf filter=lfs diff=lfs merge=lfs -text
|
| 46 |
-
llama-3.1-8b-instruct-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
README.md
CHANGED
|
@@ -1,8 +1,5 @@
|
|
| 1 |
---
|
| 2 |
license: llama3
|
| 3 |
-
pipeline_tag: text-generation
|
| 4 |
-
tags:
|
| 5 |
-
- cortex.cpp
|
| 6 |
---
|
| 7 |
|
| 8 |
## Overview
|
|
@@ -13,13 +10,15 @@ Meta developed and released the [Meta Llama 3](https://huggingface.co/meta-llama
|
|
| 13 |
|
| 14 |
| No | Variant | Cortex CLI command |
|
| 15 |
| --- | --- | --- |
|
| 16 |
-
| 1 | [
|
|
|
|
|
|
|
| 17 |
|
| 18 |
## Use it with Jan (UI)
|
| 19 |
|
| 20 |
1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart)
|
| 21 |
2. Use in Jan model Hub:
|
| 22 |
-
```
|
| 23 |
cortexso/llama3
|
| 24 |
```
|
| 25 |
|
|
@@ -27,7 +26,7 @@ Meta developed and released the [Meta Llama 3](https://huggingface.co/meta-llama
|
|
| 27 |
|
| 28 |
1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart)
|
| 29 |
2. Run the model with command:
|
| 30 |
-
```
|
| 31 |
cortex run llama3
|
| 32 |
```
|
| 33 |
|
|
|
|
| 1 |
---
|
| 2 |
license: llama3
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
|
| 5 |
## Overview
|
|
|
|
| 10 |
|
| 11 |
| No | Variant | Cortex CLI command |
|
| 12 |
| --- | --- | --- |
|
| 13 |
+
| 1 | [onnx](https://huggingface.co/cortexso/llama3/tree/onnx) | `cortex run llama3:onnx` |
|
| 14 |
+
| 2 | [gguf](https://huggingface.co/cortexso/llama3/tree/gguf) | `cortex run llama3:gguf` |
|
| 15 |
+
| 3 | [default](https://huggingface.co/cortexso/llama3/tree/default) | `cortex run llama3` |
|
| 16 |
|
| 17 |
## Use it with Jan (UI)
|
| 18 |
|
| 19 |
1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart)
|
| 20 |
2. Use in Jan model Hub:
|
| 21 |
+
```
|
| 22 |
cortexso/llama3
|
| 23 |
```
|
| 24 |
|
|
|
|
| 26 |
|
| 27 |
1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart)
|
| 28 |
2. Run the model with command:
|
| 29 |
+
```
|
| 30 |
cortex run llama3
|
| 31 |
```
|
| 32 |
|
llama-3.1-8b-instruct-q2_k.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:7ba07f0e6f20ed9a9a52c3a430dda0ab1c5eb71b5f61f194dbdb2691ad28bd94
|
| 3 |
-
size 3179136256
|
|
|
|
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q3_k_l.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:a0dec74f5fd617ad11172f763de849332254632248a430bd4b014971d12ae58a
|
| 3 |
-
size 4321961216
|
|
|
|
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q3_k_m.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:85aac3fae4c469e566bfe7371ff67f143f6b7a99f347147da83479c379daed8d
|
| 3 |
-
size 4018922752
|
|
|
|
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q3_k_s.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:622914d9363361a6a24225cc1ccc5f952fe664517f1d2054dc84ab8fd3838017
|
| 3 |
-
size 3664504064
|
|
|
|
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q4_k_m.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:c33f5a84eb2bbfbefcee9ec25cc8e639836eff6b1a54818774f254c18dbb4f5b
|
| 3 |
-
size 4920739072
|
|
|
|
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q4_k_s.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:24dd25091cde35e4c709dbcc4120430d0500fa2176dde27e67018967fb231f1f
|
| 3 |
-
size 4692673792
|
|
|
|
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q5_k_m.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:0bf7159d0d4313a818da7270399ddcbb2dbfdb25d502b82e1680e40b55a1a7a7
|
| 3 |
-
size 5732992256
|
|
|
|
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q5_k_s.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:6ba5070018af5e009c832fbc90f7a7ab224b578be78fa7ada80c9e61ebbd39ec
|
| 3 |
-
size 5599298816
|
|
|
|
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q6_k.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:e4537e7b02653bfa7fe109ac917ac374bedf69cb15b9c6fea678cda4cb8de3b2
|
| 3 |
-
size 6596011264
|
|
|
|
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q8_0.gguf
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:8349b563753acc025716792c2cad15de4f7e7ad5ea72a91f65c5df2bc226726f
|
| 3 |
-
size 8540775680
|
|
|
|
|
|
|
|
|
|
|
|
metadata.yml
DELETED
|
@@ -1,5 +0,0 @@
|
|
| 1 |
-
# metadata.yml
|
| 2 |
-
version: 1
|
| 3 |
-
name: llama3
|
| 4 |
-
default: 8b
|
| 5 |
-
author: "meta-llama"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
model.yml
DELETED
|
@@ -1,23 +0,0 @@
|
|
| 1 |
-
name: Llama 3
|
| 2 |
-
model: llama3:8B
|
| 3 |
-
version: 1
|
| 4 |
-
|
| 5 |
-
# Results Preferences
|
| 6 |
-
stop:
|
| 7 |
-
- <|end_of_text|>
|
| 8 |
-
- <|eot_id|>
|
| 9 |
-
top_p: 0.95
|
| 10 |
-
temperature: 0.7
|
| 11 |
-
frequency_penalty: 0
|
| 12 |
-
presence_penalty: 0
|
| 13 |
-
max_tokens: 8192 # Infer from base config.json -> max_position_embeddings
|
| 14 |
-
stream: true # true | false
|
| 15 |
-
|
| 16 |
-
# Engine / Model Settings
|
| 17 |
-
ngl: 33 # Infer from base config.json -> num_attention_heads
|
| 18 |
-
ctx_len: 8192 # Infer from base config.json -> max_position_embeddings
|
| 19 |
-
engine: llama-cpp
|
| 20 |
-
prompt_template: "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
|
| 21 |
-
# Prompt template: Can only be retrieved from instruct model
|
| 22 |
-
# - https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/tokenizer_config.json#L2053
|
| 23 |
-
# - Requires jinja format parser
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|