Update README.md
Browse files
README.md
CHANGED
|
@@ -8,4 +8,26 @@ base_model:
|
|
| 8 |
- allenai/led-base-16384
|
| 9 |
pipeline_tag: summarization
|
| 10 |
library_name: transformers
|
| 11 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8 |
- allenai/led-base-16384
|
| 9 |
pipeline_tag: summarization
|
| 10 |
library_name: transformers
|
| 11 |
+
---
|
| 12 |
+
|
| 13 |
+
# Longformer fine-tuned to summarize Terms of Service
|
| 14 |
+
Terms of Service documents are lengthy, complex, and time-consuming to read. Due to its vague language, people often don’t understand what they are agreeing to. Hence, we have fine-tuned Longformer model to help summarize TOS and make it easy to read and understand.
|
| 15 |
+
|
| 16 |
+
This model is a fine-tuned version of [`allenai/led-base-16384`](https://huggingface.co/allenai/led-base-16384)
|
| 17 |
+
Dataset used: TL;DRLegal and TOS;DR website
|
| 18 |
+
It achieves the following results on the validation set:
|
| 19 |
+
ROUGE-1: 0.28
|
| 20 |
+
ROUGE-2: 0.13
|
| 21 |
+
ROUGE-L: 0.27
|
| 22 |
+
|
| 23 |
+
## How to Use
|
| 24 |
+
|
| 25 |
+
```python
|
| 26 |
+
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
| 27 |
+
|
| 28 |
+
tokenizer = AutoTokenizer.from_pretrained("aarushi-211/TOS-Longformer")
|
| 29 |
+
model = AutoModelForSequenceClassification.from_pretrained("aarushi-211/TOS-Longformer")
|
| 30 |
+
|
| 31 |
+
inputs = tokenizer("Your input text here", return_tensors="pt", truncation=True)
|
| 32 |
+
outputs = model(**inputs)
|
| 33 |
+
logits = outputs.logits
|