nteku1 commited on
Commit
81e05cd
·
verified ·
1 Parent(s): af7bc6d

Push model using huggingface_hub.

Browse files
Files changed (3) hide show
  1. README.md +3 -3
  2. pytorch_model.bin +1 -1
  3. tokenizer.json +2 -16
README.md CHANGED
@@ -25,7 +25,7 @@ You can then generate text as follows:
25
  ```python
26
  from transformers import pipeline
27
 
28
- generator = pipeline("text-generation", model="nteku1//tmp/tmpeczj8d97/nteku1/firstppomodel")
29
  outputs = generator("Hello, my llama is cute")
30
  ```
31
 
@@ -35,8 +35,8 @@ If you want to use the model for training or to obtain the outputs from the valu
35
  from transformers import AutoTokenizer
36
  from trl import AutoModelForCausalLMWithValueHead
37
 
38
- tokenizer = AutoTokenizer.from_pretrained("nteku1//tmp/tmpeczj8d97/nteku1/firstppomodel")
39
- model = AutoModelForCausalLMWithValueHead.from_pretrained("nteku1//tmp/tmpeczj8d97/nteku1/firstppomodel")
40
 
41
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
42
  outputs = model(**inputs, labels=inputs["input_ids"])
 
25
  ```python
26
  from transformers import pipeline
27
 
28
+ generator = pipeline("text-generation", model="nteku1//tmp/tmpwi2wt4ig/nteku1/firstppomodel")
29
  outputs = generator("Hello, my llama is cute")
30
  ```
31
 
 
35
  from transformers import AutoTokenizer
36
  from trl import AutoModelForCausalLMWithValueHead
37
 
38
+ tokenizer = AutoTokenizer.from_pretrained("nteku1//tmp/tmpwi2wt4ig/nteku1/firstppomodel")
39
+ model = AutoModelForCausalLMWithValueHead.from_pretrained("nteku1//tmp/tmpwi2wt4ig/nteku1/firstppomodel")
40
 
41
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
42
  outputs = model(**inputs, labels=inputs["input_ids"])
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5da979194b4e65f142613b5ed61d003101351fe01d80d89e0c9fae8bb69a96c8
3
  size 6652
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d46fe139897b8f7906e860e5e1f3eeeb3174d96c7e45580ca2728406371a3a2e
3
  size 6652
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 512
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 50256,
16
- "pad_type_id": 0,
17
- "pad_token": "<|endoftext|>"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 50256,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 50256,