Add/update the quantized ONNX model files and README.md for Transformers.js v3
#1
by
whitphx - opened
- README.md +18 -2
- onnx/model_bnb4.onnx +3 -0
- onnx/model_int8.onnx +3 -0
- onnx/model_q4.onnx +3 -0
- onnx/model_q4f16.onnx +3 -0
- onnx/model_uint8.onnx +3 -0
README.md
CHANGED
|
@@ -1,8 +1,24 @@
|
|
| 1 |
---
|
| 2 |
-
base_model:
|
| 3 |
library_name: transformers.js
|
| 4 |
---
|
| 5 |
|
| 6 |
-
https://huggingface.co/
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
|
| 8 |
Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [π€ Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model: Xenova/UMLSBert_ENG
|
| 3 |
library_name: transformers.js
|
| 4 |
---
|
| 5 |
|
| 6 |
+
https://huggingface.co/Xenova/UMLSBert_ENG with ONNX weights to be compatible with Transformers.js.
|
| 7 |
+
|
| 8 |
+
## Usage (Transformers.js)
|
| 9 |
+
|
| 10 |
+
If you haven't already, you can install the [Transformers.js](https://huggingface.co/docs/transformers.js) JavaScript library from [NPM](https://www.npmjs.com/package/@huggingface/transformers) using:
|
| 11 |
+
```bash
|
| 12 |
+
npm i @huggingface/transformers
|
| 13 |
+
```
|
| 14 |
+
|
| 15 |
+
**Example:** Run feature extraction.
|
| 16 |
+
|
| 17 |
+
```js
|
| 18 |
+
import { pipeline } from '@huggingface/transformers';
|
| 19 |
+
|
| 20 |
+
const extractor = await pipeline('feature-extraction', 'Xenova/UMLSBert_ENG');
|
| 21 |
+
const output = await extractor('This is a simple test.');
|
| 22 |
+
```
|
| 23 |
|
| 24 |
Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [π€ Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
|
onnx/model_bnb4.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5688e108fec9340473a00d2f86f055abc60da4af2b16d047ca91a199394aacdf
|
| 3 |
+
size 143893470
|
onnx/model_int8.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7fb8d6c3c6ff4ecde3f89d7191cba1682dc94bc8518d1193695364eb05d46089
|
| 3 |
+
size 109622399
|
onnx/model_q4.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c2b3a1351f560789e55328e51a74b2cb667f19e9431c875f6a8bd6e8dbc00118
|
| 3 |
+
size 149201358
|
onnx/model_q4f16.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:021b3dea9bdd1b8aae7f32f8621ab47747b2bd16168394c55c453e4a22cc617c
|
| 3 |
+
size 95979131
|
onnx/model_uint8.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ea06a3faf0223766daa88a412e910fb9e7052fc6fc78e3b498ea4a93baaac1a3
|
| 3 |
+
size 109622443
|