Update README.md
Browse files
README.md
CHANGED
|
@@ -1,20 +1,78 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
tags:
|
| 4 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
datasets:
|
| 6 |
-
-
|
| 7 |
-
|
| 8 |
-
-
|
| 9 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
---
|
| 11 |
|
| 12 |
-
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 13 |
-
should probably proofread and complete it, then remove this comment. -->
|
| 14 |
|
| 15 |
-
# bart-base-
|
| 16 |
|
| 17 |
-
This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the
|
| 18 |
|
| 19 |
## Model description
|
| 20 |
|
|
@@ -54,4 +112,4 @@ The following hyperparameters were used during training:
|
|
| 54 |
- Transformers 4.28.1
|
| 55 |
- Pytorch 2.0.1+cu117
|
| 56 |
- Datasets 2.12.0
|
| 57 |
-
- Tokenizers 0.13.3
|
|
|
|
| 1 |
---
|
| 2 |
+
languages:
|
| 3 |
+
- en
|
| 4 |
+
license:
|
| 5 |
+
- cc-by-nc-sa-4.0
|
| 6 |
+
- apache-2.0
|
| 7 |
tags:
|
| 8 |
+
- grammar
|
| 9 |
+
- spelling
|
| 10 |
+
- punctuation
|
| 11 |
+
- error-correction
|
| 12 |
+
- grammar synthesis
|
| 13 |
datasets:
|
| 14 |
+
- jfleg
|
| 15 |
+
widget:
|
| 16 |
+
- text: There car broke down so their hitching a ride to they're class.
|
| 17 |
+
example_title: compound-1
|
| 18 |
+
- text: i can has cheezburger
|
| 19 |
+
example_title: cheezburger
|
| 20 |
+
- text: >-
|
| 21 |
+
so em if we have an now so with fito ringina know how to estimate the tren
|
| 22 |
+
given the ereafte mylite trend we can also em an estimate is nod s i again
|
| 23 |
+
tort watfettering an we have estimated the trend an called wot to be called
|
| 24 |
+
sthat of exty right now we can and look at wy this should not hare a trend i
|
| 25 |
+
becan we just remove the trend an and we can we now estimate tesees ona
|
| 26 |
+
effect of them exty
|
| 27 |
+
example_title: Transcribed Audio Example 2
|
| 28 |
+
- text: >-
|
| 29 |
+
My coworker said he used a financial planner to help choose his stocks so he
|
| 30 |
+
wouldn't loose money.
|
| 31 |
+
example_title: incorrect word choice (context)
|
| 32 |
+
- text: >-
|
| 33 |
+
good so hve on an tadley i'm not able to make it to the exla session on
|
| 34 |
+
monday this week e which is why i am e recording pre recording an this
|
| 35 |
+
excelleision and so to day i want e to talk about two things and first of
|
| 36 |
+
all em i wont em wene give a summary er about ta ohow to remove trents in
|
| 37 |
+
these nalitives from time series
|
| 38 |
+
example_title: lowercased audio transcription output
|
| 39 |
+
- text: Frustrated, the chairs took me forever to set up.
|
| 40 |
+
example_title: dangling modifier
|
| 41 |
+
- text: I would like a peice of pie.
|
| 42 |
+
example_title: miss-spelling
|
| 43 |
+
- text: >-
|
| 44 |
+
Which part of Zurich was you going to go hiking in when we were there for
|
| 45 |
+
the first time together? ! ?
|
| 46 |
+
example_title: chatbot on Zurich
|
| 47 |
+
- text: >-
|
| 48 |
+
Most of the course is about semantic or content of language but there are
|
| 49 |
+
also interesting topics to be learned from the servicefeatures except
|
| 50 |
+
statistics in characters in documents. At this point, Elvthos introduces
|
| 51 |
+
himself as his native English speaker and goes on to say that if you
|
| 52 |
+
continue to work on social scnce,
|
| 53 |
+
example_title: social science ASR summary output
|
| 54 |
+
- text: >-
|
| 55 |
+
they are somewhat nearby right yes please i'm not sure how the innish is
|
| 56 |
+
tepen thut mayyouselect one that istatte lo variants in their property e ere
|
| 57 |
+
interested and anyone basical e may be applyind reaching the browing
|
| 58 |
+
approach were
|
| 59 |
+
- medical course audio transcription
|
| 60 |
+
parameters:
|
| 61 |
+
max_length: 128
|
| 62 |
+
min_length: 4
|
| 63 |
+
num_beams: 8
|
| 64 |
+
repetition_penalty: 1.21
|
| 65 |
+
length_penalty: 1
|
| 66 |
+
early_stopping: true
|
| 67 |
+
language:
|
| 68 |
+
- en
|
| 69 |
+
pipeline_tag: text2text-generation
|
| 70 |
---
|
| 71 |
|
|
|
|
|
|
|
| 72 |
|
| 73 |
+
# bart-base-grammar-synthesis
|
| 74 |
|
| 75 |
+
This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an expanded version of the JFLEG dataset.
|
| 76 |
|
| 77 |
## Model description
|
| 78 |
|
|
|
|
| 112 |
- Transformers 4.28.1
|
| 113 |
- Pytorch 2.0.1+cu117
|
| 114 |
- Datasets 2.12.0
|
| 115 |
+
- Tokenizers 0.13.3
|