Update README.md
Browse files
README.md
CHANGED
|
@@ -9,6 +9,8 @@ model-index:
|
|
| 9 |
- name: llama3-conciser
|
| 10 |
results: []
|
| 11 |
pipeline_tag: text2text-generation
|
|
|
|
|
|
|
| 12 |
---
|
| 13 |
|
| 14 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
@@ -135,23 +137,26 @@ hub_model_id: chrislee973/llama3-conciser
|
|
| 135 |
|
| 136 |
# llama3-conciser
|
| 137 |
|
| 138 |
-
This model is a fine-tuned version of [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B) on
|
| 139 |
-
It achieves the following results on the evaluation set:
|
| 140 |
-
- Loss: 0.5174
|
| 141 |
-
|
| 142 |
-
## Model description
|
| 143 |
-
|
| 144 |
-
More information needed
|
| 145 |
-
|
| 146 |
-
## Intended uses & limitations
|
| 147 |
|
| 148 |
-
|
| 149 |
|
| 150 |
-
|
|
|
|
|
|
|
| 151 |
|
| 152 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 153 |
|
| 154 |
-
##
|
|
|
|
| 155 |
|
| 156 |
### Training hyperparameters
|
| 157 |
|
|
|
|
| 9 |
- name: llama3-conciser
|
| 10 |
results: []
|
| 11 |
pipeline_tag: text2text-generation
|
| 12 |
+
datasets:
|
| 13 |
+
- chrislee973/llama3-conciser-dataset
|
| 14 |
---
|
| 15 |
|
| 16 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
|
| 137 |
|
| 138 |
# llama3-conciser
|
| 139 |
|
| 140 |
+
This model is a fine-tuned version of [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B) on my [conciser dataset](https://huggingface.co/datasets/chrislee973/llama3-conciser-dataset).
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 141 |
|
| 142 |
+
## Uses
|
| 143 |
|
| 144 |
+
### Text Revision task
|
| 145 |
+
Given an input of a paragraph of text from a transcript, it lightly touches up and edits the sentences and phrases,
|
| 146 |
+
improving the flow and readability of the text while maintaining the speaker's original intention.
|
| 147 |
|
| 148 |
+
For example, given the following input text:
|
| 149 |
+
```
|
| 150 |
+
I think I sort of deep down believed in what we were doing, and I did some analysis. I was like, okay, well, what would I go do if I wasn't doing this? It's like, well, I really like building things, and I like helping people communicate, and I like understanding what's going on with people and the dynamics between people. So I think if I sold this company, I'd just go build another company like this. And I kind of like the one I have.
|
| 151 |
+
```
|
| 152 |
+
the revised output text is:
|
| 153 |
+
```
|
| 154 |
+
I believed deep down in what we were doing. I did some analysis. What would I go do if I wasn’t doing this? I really like building things, helping people communicate, understanding what’s going on with people and the dynamics between them. If I sold this company, I’d just go build another one like this. I kind of like the one I have.
|
| 155 |
+
```
|
| 156 |
+
There are still some rough edges around the model as a result of my dataset being so tiny (just 50 examples). I hope to smooth these imperfections out and close the quality gap by adding many more examples to the dataset.
|
| 157 |
|
| 158 |
+
## Usage
|
| 159 |
+
TODO: add sample inference code
|
| 160 |
|
| 161 |
### Training hyperparameters
|
| 162 |
|