update new info
Browse files- README.md +19 -17
- config.json +1 -1
- tokenizer_config.json +6 -6
README.md
CHANGED
|
@@ -1,14 +1,12 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
-
datasets:
|
| 4 |
-
- togethercomputer/RedPajama-Data-1T
|
| 5 |
language:
|
| 6 |
- en
|
| 7 |
-
|
| 8 |
-
-
|
| 9 |
pipeline_tag: text-generation
|
| 10 |
---
|
| 11 |
-
#
|
| 12 |
|
| 13 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
|
|
@@ -22,20 +20,18 @@ This modelcard aims to be a base template for new models. It has been generated
|
|
| 22 |
|
| 23 |
|
| 24 |
|
| 25 |
-
- **Developed by:**
|
| 26 |
-
- **
|
| 27 |
-
- **
|
| 28 |
-
- **
|
| 29 |
-
- **
|
| 30 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
| 31 |
|
| 32 |
-
### Model Sources
|
| 33 |
|
| 34 |
<!-- Provide the basic links for the model. -->
|
| 35 |
|
| 36 |
-
- **Repository:**
|
| 37 |
-
- **
|
| 38 |
-
- **Demo [optional]:** [More Information Needed]
|
| 39 |
|
| 40 |
## Uses
|
| 41 |
|
|
@@ -75,7 +71,13 @@ Users (both direct and downstream) should be made aware of the risks, biases and
|
|
| 75 |
|
| 76 |
Use the code below to get started with the model.
|
| 77 |
|
| 78 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 79 |
|
| 80 |
## Training Details
|
| 81 |
|
|
@@ -83,7 +85,7 @@ Use the code below to get started with the model.
|
|
| 83 |
|
| 84 |
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 85 |
|
| 86 |
-
|
| 87 |
|
| 88 |
### Training Procedure
|
| 89 |
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
| 3 |
language:
|
| 4 |
- en
|
| 5 |
+
tags:
|
| 6 |
+
- hawk
|
| 7 |
pipeline_tag: text-generation
|
| 8 |
---
|
| 9 |
+
# Hawk-demo
|
| 10 |
|
| 11 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 12 |
|
|
|
|
| 20 |
|
| 21 |
|
| 22 |
|
| 23 |
+
- **Developed by:** Rexopia
|
| 24 |
+
- **Reach me:** ruiji.zhang@outlook.com
|
| 25 |
+
- **Language(s) (NLP):** English
|
| 26 |
+
- **License:** Apache license 2.0
|
| 27 |
+
- **Pretrained model [optional]:** True
|
|
|
|
| 28 |
|
| 29 |
+
### Model Sources
|
| 30 |
|
| 31 |
<!-- Provide the basic links for the model. -->
|
| 32 |
|
| 33 |
+
- **Github Repository:** Coming soon
|
| 34 |
+
- **Demo version:** True
|
|
|
|
| 35 |
|
| 36 |
## Uses
|
| 37 |
|
|
|
|
| 71 |
|
| 72 |
Use the code below to get started with the model.
|
| 73 |
|
| 74 |
+
```python
|
| 75 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 76 |
+
|
| 77 |
+
tokenizer = AutoTokenizer.from_pretrained("Rexopia/HawkLM-demo", trust_remote_code=True)
|
| 78 |
+
|
| 79 |
+
model = AutoModelForCausalLM.from_pretrained("Rexopia/HawkLM-demo", device_map="auto", trust_remote_code=True)
|
| 80 |
+
```
|
| 81 |
|
| 82 |
## Training Details
|
| 83 |
|
|
|
|
| 85 |
|
| 86 |
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 87 |
|
| 88 |
+
We sampled from Redpajama 1T datasets without any Arxiv and GitHub tags.
|
| 89 |
|
| 90 |
### Training Procedure
|
| 91 |
|
config.json
CHANGED
|
@@ -12,7 +12,7 @@
|
|
| 12 |
"embd_pdrop": 0.0,
|
| 13 |
"eos_token_id": 65535,
|
| 14 |
"initializer_range": 0.02,
|
| 15 |
-
"layer_norm_epsilon": 1e-
|
| 16 |
"model_type": "hawk",
|
| 17 |
"n_embd": 1024,
|
| 18 |
"n_head": 16,
|
|
|
|
| 12 |
"embd_pdrop": 0.0,
|
| 13 |
"eos_token_id": 65535,
|
| 14 |
"initializer_range": 0.02,
|
| 15 |
+
"layer_norm_epsilon": 1e-06,
|
| 16 |
"model_type": "hawk",
|
| 17 |
"n_embd": 1024,
|
| 18 |
"n_head": 16,
|
tokenizer_config.json
CHANGED
|
@@ -1,6 +1,12 @@
|
|
| 1 |
{
|
| 2 |
"add_bos_token": false,
|
| 3 |
"add_prefix_space": false,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
"bos_token": {
|
| 5 |
"__type": "AddedToken",
|
| 6 |
"content": "<s>",
|
|
@@ -36,11 +42,5 @@
|
|
| 36 |
"normalized": true,
|
| 37 |
"rstrip": false,
|
| 38 |
"single_word": false
|
| 39 |
-
},
|
| 40 |
-
"auto_map": {
|
| 41 |
-
"AutoTokenizer": [
|
| 42 |
-
"tokenization_hawk.HawkTokenizer",
|
| 43 |
-
null
|
| 44 |
-
]
|
| 45 |
}
|
| 46 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"add_bos_token": false,
|
| 3 |
"add_prefix_space": false,
|
| 4 |
+
"auto_map": {
|
| 5 |
+
"AutoTokenizer": [
|
| 6 |
+
"tokenization_hawk.HawkTokenizer",
|
| 7 |
+
null
|
| 8 |
+
]
|
| 9 |
+
},
|
| 10 |
"bos_token": {
|
| 11 |
"__type": "AddedToken",
|
| 12 |
"content": "<s>",
|
|
|
|
| 42 |
"normalized": true,
|
| 43 |
"rstrip": false,
|
| 44 |
"single_word": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 45 |
}
|
| 46 |
}
|