Mithilhf01 commited on
Commit
69749de
·
verified ·
1 Parent(s): 8256ac3

Push model using huggingface_hub.

Browse files
Files changed (4) hide show
  1. README.md +3 -3
  2. config.json +5 -5
  3. model.safetensors +2 -2
  4. tokenizer.json +1 -6
README.md CHANGED
@@ -26,7 +26,7 @@ You can then generate text as follows:
26
  ```python
27
  from transformers import pipeline
28
 
29
- generator = pipeline("text-generation", model="Mithilhf01//tmp/tmpp90aqkcd/Mithilhf01/mistral-ppo")
30
  outputs = generator("Hello, my llama is cute")
31
  ```
32
 
@@ -36,8 +36,8 @@ If you want to use the model for training or to obtain the outputs from the valu
36
  from transformers import AutoTokenizer
37
  from trl import AutoModelForCausalLMWithValueHead
38
 
39
- tokenizer = AutoTokenizer.from_pretrained("Mithilhf01//tmp/tmpp90aqkcd/Mithilhf01/mistral-ppo")
40
- model = AutoModelForCausalLMWithValueHead.from_pretrained("Mithilhf01//tmp/tmpp90aqkcd/Mithilhf01/mistral-ppo")
41
 
42
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
43
  outputs = model(**inputs, labels=inputs["input_ids"])
 
26
  ```python
27
  from transformers import pipeline
28
 
29
+ generator = pipeline("text-generation", model="Mithilhf01//tmp/tmpxg2uubos/Mithilhf01/mistral-ppo")
30
  outputs = generator("Hello, my llama is cute")
31
  ```
32
 
 
36
  from transformers import AutoTokenizer
37
  from trl import AutoModelForCausalLMWithValueHead
38
 
39
+ tokenizer = AutoTokenizer.from_pretrained("Mithilhf01//tmp/tmpxg2uubos/Mithilhf01/mistral-ppo")
40
+ model = AutoModelForCausalLMWithValueHead.from_pretrained("Mithilhf01//tmp/tmpxg2uubos/Mithilhf01/mistral-ppo")
41
 
42
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
43
  outputs = model(**inputs, labels=inputs["input_ids"])
config.json CHANGED
@@ -6,19 +6,19 @@
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
- "head_dim": 32,
10
  "hidden_act": "silu",
11
- "hidden_size": 512,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 3072,
14
  "max_position_embeddings": 512,
15
  "model_type": "mistral",
16
- "num_attention_heads": 16,
17
- "num_hidden_layers": 4,
18
  "num_key_value_heads": 8,
19
  "rms_norm_eps": 1e-06,
20
  "rope_theta": 10000.0,
21
- "sliding_window": 512,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.47.1",
 
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
+ "head_dim": 24,
10
  "hidden_act": "silu",
11
+ "hidden_size": 768,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 3072,
14
  "max_position_embeddings": 512,
15
  "model_type": "mistral",
16
+ "num_attention_heads": 32,
17
+ "num_hidden_layers": 8,
18
  "num_key_value_heads": 8,
19
  "rms_norm_eps": 1e-06,
20
  "rope_theta": 10000.0,
21
+ "sliding_window": 768,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.47.1",
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:859e7582744a4c7cffc9125ace6d98fd4101d5402263c72e4206afe5afb2e3bb
3
- size 219177356
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc494b524ea8be5b41dba1d8f51b10a55372c79c495cbee259bd4760f0e17052
3
+ size 470350116
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 1024,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {