Add/update the quantized ONNX model files and README.md for Transformers.js v3 (#2)
Browse files- Add/update the quantized ONNX model files and README.md for Transformers.js v3 (a91aee49444b9673dff6a5035b81987810c49c55)
Co-authored-by: Yuichiro Tachibana <whitphx@users.noreply.huggingface.co>
- README.md +3 -3
- onnx/model_bnb4.onnx +3 -0
- onnx/model_int8.onnx +3 -0
- onnx/model_q4.onnx +3 -0
- onnx/model_q4f16.onnx +3 -0
- onnx/model_uint8.onnx +3 -0
README.md
CHANGED
|
@@ -9,15 +9,15 @@ https://huggingface.co/colbert-ir/colbertv2.0 with ONNX weights to be compatible
|
|
| 9 |
|
| 10 |
## Usage (Transformers.js)
|
| 11 |
|
| 12 |
-
If you haven't already, you can install the [Transformers.js](https://huggingface.co/docs/transformers.js) JavaScript library from [NPM](https://www.npmjs.com/package/@
|
| 13 |
```bash
|
| 14 |
-
npm i @
|
| 15 |
```
|
| 16 |
|
| 17 |
You can then use the model to compute embeddings like this:
|
| 18 |
|
| 19 |
```js
|
| 20 |
-
import { pipeline } from '@
|
| 21 |
|
| 22 |
// Create a feature-extraction pipeline
|
| 23 |
const extractor = await pipeline('feature-extraction', 'Xenova/colbertv2.0');
|
|
|
|
| 9 |
|
| 10 |
## Usage (Transformers.js)
|
| 11 |
|
| 12 |
+
If you haven't already, you can install the [Transformers.js](https://huggingface.co/docs/transformers.js) JavaScript library from [NPM](https://www.npmjs.com/package/@huggingface/transformers) using:
|
| 13 |
```bash
|
| 14 |
+
npm i @huggingface/transformers
|
| 15 |
```
|
| 16 |
|
| 17 |
You can then use the model to compute embeddings like this:
|
| 18 |
|
| 19 |
```js
|
| 20 |
+
import { pipeline } from '@huggingface/transformers';
|
| 21 |
|
| 22 |
// Create a feature-extraction pipeline
|
| 23 |
const extractor = await pipeline('feature-extraction', 'Xenova/colbertv2.0');
|
onnx/model_bnb4.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4aa580d95d00e89463edf07ff073a70be6f583b5b31058cfd9d7411a13148ac6
|
| 3 |
+
size 143893455
|
onnx/model_int8.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2af3646dd0e27cd8e0814e93091fe845ca906b61aee03f8bd0a5c19f5d656c52
|
| 3 |
+
size 109622387
|
onnx/model_q4.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:26b975434571f99a076361268c415342663665ff496083df1161f0f7d02d69a0
|
| 3 |
+
size 149201343
|
onnx/model_q4f16.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a81a450ce3e85beaa30bfffa16af59bd46542272af54f9545776b1f22135ca14
|
| 3 |
+
size 95979116
|
onnx/model_uint8.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5750b3ce904b82962e1006c2934c404696ea3f1da8c593e19a664f3a69506c81
|
| 3 |
+
size 109622429
|