Update README.md
Browse files
README.md
CHANGED
|
@@ -35,7 +35,7 @@ from transformers import LlamaTokenizer, LlamaForCausalLM
|
|
| 35 |
tokenizer = LlamaTokenizer.from_pretrained("LLM360/AmberChat")
|
| 36 |
model = LlamaForCausalLM.from_pretrained("LLM360/AmberChat")
|
| 37 |
|
| 38 |
-
input_text = "
|
| 39 |
input_ids = tokenizer(input_text, return_tensors="pt").input_ids
|
| 40 |
|
| 41 |
outputs = model.generate(input_ids)
|
|
@@ -45,11 +45,11 @@ print(tokenizer.decode(outputs[0]))
|
|
| 45 |
# AmberChat Finetuning Details
|
| 46 |
|
| 47 |
## DataMix
|
| 48 |
-
| Subset | Number of rows |
|
| 49 |
-
| ----------- | ----------- |
|
| 50 |
-
| WizardLM/WizardLM_evol_instruct_V2_196k | 143k |
|
| 51 |
-
| icybee/share_gpt_90k_v1 | 90k |
|
| 52 |
-
| Total | 233k |
|
| 53 |
|
| 54 |
## Hyperparameters
|
| 55 |
| Hyperparameter | Value |
|
|
|
|
| 35 |
tokenizer = LlamaTokenizer.from_pretrained("LLM360/AmberChat")
|
| 36 |
model = LlamaForCausalLM.from_pretrained("LLM360/AmberChat")
|
| 37 |
|
| 38 |
+
input_text = "How old are you?"
|
| 39 |
input_ids = tokenizer(input_text, return_tensors="pt").input_ids
|
| 40 |
|
| 41 |
outputs = model.generate(input_ids)
|
|
|
|
| 45 |
# AmberChat Finetuning Details
|
| 46 |
|
| 47 |
## DataMix
|
| 48 |
+
| Subset | Number of rows | License |
|
| 49 |
+
| ----------- | ----------- | ----------- |
|
| 50 |
+
| WizardLM/WizardLM_evol_instruct_V2_196k | 143k | |
|
| 51 |
+
| icybee/share_gpt_90k_v1 | 90k | cc0-1.0 |
|
| 52 |
+
| Total | 233k | |
|
| 53 |
|
| 54 |
## Hyperparameters
|
| 55 |
| Hyperparameter | Value |
|