add todos
Browse files- README.md +1 -0
- model/t5_vae.py +1 -0
- train.py +1 -0
README.md
CHANGED
|
@@ -10,6 +10,7 @@ Builds on T5, using an autoencoder to convert it into a VAE.
|
|
| 10 |
|
| 11 |
## ToDo
|
| 12 |
|
|
|
|
| 13 |
- [ ] Convert `transformers/examples/flax/language-modeling/run_clm_flax.py` into a new training script for transformer-VAE's.
|
| 14 |
- Use an "empty VAE" a.k.a just sends the encoding to the decoder with no regularisation loss, use the T5 encoder & decoder.
|
| 15 |
- [ ] Make a `autoencoders.py` version of `autoencoders.py`.
|
|
|
|
| 10 |
|
| 11 |
## ToDo
|
| 12 |
|
| 13 |
+
- [ ] Save a wikipedia sentences dataset to Huggingface (see original https://github.com/ChunyuanLI/Optimus/blob/master/data/download_datasets.md)
|
| 14 |
- [ ] Convert `transformers/examples/flax/language-modeling/run_clm_flax.py` into a new training script for transformer-VAE's.
|
| 15 |
- Use an "empty VAE" a.k.a just sends the encoding to the decoder with no regularisation loss, use the T5 encoder & decoder.
|
| 16 |
- [ ] Make a `autoencoders.py` version of `autoencoders.py`.
|
model/t5_vae.py
CHANGED
|
@@ -153,4 +153,5 @@ class FlaxT5VAEForAutoencoding(FlaxPreTrainedModel):
|
|
| 153 |
params: dict = None,
|
| 154 |
dropout_rng: PRNGKey = None,
|
| 155 |
):
|
|
|
|
| 156 |
raise NotImplementedError()
|
|
|
|
| 153 |
params: dict = None,
|
| 154 |
dropout_rng: PRNGKey = None,
|
| 155 |
):
|
| 156 |
+
# TODO run `FlaxT5ForConditionalGeneration.decode` with above args
|
| 157 |
raise NotImplementedError()
|
train.py
CHANGED
|
@@ -2,6 +2,7 @@
|
|
| 2 |
Pre-training/Fine-tuning seq2seq models on autoencoding a dataset.
|
| 3 |
|
| 4 |
TODO:
|
|
|
|
| 5 |
- [ ] Add reg loss
|
| 6 |
- [ ] config
|
| 7 |
- [ ] calculate MMD loss
|
|
|
|
| 2 |
Pre-training/Fine-tuning seq2seq models on autoencoding a dataset.
|
| 3 |
|
| 4 |
TODO:
|
| 5 |
+
- [ ] Don't make decoder input ids.
|
| 6 |
- [ ] Add reg loss
|
| 7 |
- [ ] config
|
| 8 |
- [ ] calculate MMD loss
|