Fraser commited on
Commit
db1be99
·
1 Parent(s): 03c20bd

add todos

Browse files
Files changed (3) hide show
  1. README.md +1 -0
  2. model/t5_vae.py +1 -0
  3. train.py +1 -0
README.md CHANGED
@@ -10,6 +10,7 @@ Builds on T5, using an autoencoder to convert it into a VAE.
10
 
11
  ## ToDo
12
 
 
13
  - [ ] Convert `transformers/examples/flax/language-modeling/run_clm_flax.py` into a new training script for transformer-VAE's.
14
  - Use an "empty VAE" a.k.a just sends the encoding to the decoder with no regularisation loss, use the T5 encoder & decoder.
15
  - [ ] Make a `autoencoders.py` version of `autoencoders.py`.
 
10
 
11
  ## ToDo
12
 
13
+ - [ ] Save a wikipedia sentences dataset to Huggingface (see original https://github.com/ChunyuanLI/Optimus/blob/master/data/download_datasets.md)
14
  - [ ] Convert `transformers/examples/flax/language-modeling/run_clm_flax.py` into a new training script for transformer-VAE's.
15
  - Use an "empty VAE" a.k.a just sends the encoding to the decoder with no regularisation loss, use the T5 encoder & decoder.
16
  - [ ] Make a `autoencoders.py` version of `autoencoders.py`.
model/t5_vae.py CHANGED
@@ -153,4 +153,5 @@ class FlaxT5VAEForAutoencoding(FlaxPreTrainedModel):
153
  params: dict = None,
154
  dropout_rng: PRNGKey = None,
155
  ):
 
156
  raise NotImplementedError()
 
153
  params: dict = None,
154
  dropout_rng: PRNGKey = None,
155
  ):
156
+ # TODO run `FlaxT5ForConditionalGeneration.decode` with above args
157
  raise NotImplementedError()
train.py CHANGED
@@ -2,6 +2,7 @@
2
  Pre-training/Fine-tuning seq2seq models on autoencoding a dataset.
3
 
4
  TODO:
 
5
  - [ ] Add reg loss
6
  - [ ] config
7
  - [ ] calculate MMD loss
 
2
  Pre-training/Fine-tuning seq2seq models on autoencoding a dataset.
3
 
4
  TODO:
5
+ - [ ] Don't make decoder input ids.
6
  - [ ] Add reg loss
7
  - [ ] config
8
  - [ ] calculate MMD loss