Update README.md
Browse files
README.md
CHANGED
|
@@ -12,7 +12,9 @@ tags:
|
|
| 12 |
|
| 13 |
# CapDec - NoiseLevel: 0.015
|
| 14 |
|
| 15 |
-
|
|
|
|
|
|
|
| 16 |
|
| 17 |
Their method aims to train CLIP with only text samples. Therefore they are injecting zero-mean Gaussian Noise into the text embeddings before decoding.
|
| 18 |
|
|
@@ -28,7 +30,9 @@ The "Noise Level" of 0.015 is equivalent to the Noise Variance which is the squa
|
|
| 28 |
The reported metrics are results of a model with a Noise Variance of 0.016, which the authors unfortunately do not provide in their repository.
|
| 29 |
This model with a Noise Variance 0.015 is the closest available pre-trained model to their best model.
|
| 30 |
|
|
|
|
|
|
|
| 31 |
|
| 32 |
## Performance
|
| 33 |
The authors don't explicitly report the performance for this NoiseLevel but it can be estimated from the following figure from the original paper:
|
| 34 |
-
|
|
|
|
| 12 |
|
| 13 |
# CapDec - NoiseLevel: 0.015
|
| 14 |
|
| 15 |
+
## Model Description
|
| 16 |
+
|
| 17 |
+
These are model weights originally provided by the authors of the paper [Text-Only Training for Image Captioning using Noise-Injected CLIP](https://arxiv.org/pdf/2211.00575.pdf).
|
| 18 |
|
| 19 |
Their method aims to train CLIP with only text samples. Therefore they are injecting zero-mean Gaussian Noise into the text embeddings before decoding.
|
| 20 |
|
|
|
|
| 30 |
The reported metrics are results of a model with a Noise Variance of 0.016, which the authors unfortunately do not provide in their repository.
|
| 31 |
This model with a Noise Variance 0.015 is the closest available pre-trained model to their best model.
|
| 32 |
|
| 33 |
+
## Datasets
|
| 34 |
+
The authors trained the model on MS-COCO and Flickr30k datasets.
|
| 35 |
|
| 36 |
## Performance
|
| 37 |
The authors don't explicitly report the performance for this NoiseLevel but it can be estimated from the following figure from the original paper:
|
| 38 |
+

|