LKyluk commited on
Commit
6fc6b8a
·
verified ·
1 Parent(s): 90270bd

Update Readme

Browse files
Files changed (1) hide show
  1. README.md +2 -5
README.md CHANGED
@@ -18,17 +18,14 @@ These models are designed for memory-efficient temporal sequence prediction, par
18
  This repository contains pre-trained weights for the following models, as described in the research article:
19
 
20
  * **TempVerseFormer (Rev-Transformer):** The core Reversible Temporal Transformer architecture, leveraging reversible blocks and time-agnostic backpropagation for memory efficiency.
21
- * Checkpoints available for different training configurations (e.g., with/without temporal patterns).
22
  * **TempFormer (Vanilla-Transformer):** A standard Vanilla Transformer architecture with temporal chaining, serving as a baseline to compare against TempVerseFormer.
23
- * Checkpoints available for different training configurations (e.g., with/without temporal patterns).
24
- * **Standard Transformer (Pipe-Transformer):** A standard Transformer model processing the entire context at once, used as a non-sequential baseline.
25
- * Checkpoints available for different training configurations (e.g., with/without temporal patterns).
26
  * **LSTM:** A Long Short-Term Memory network, representing a traditional recurrent sequence modeling approach.
27
- * Checkpoints available for different training configurations (e.g., with/without temporal patterns).
28
  * **VAE Models:** Variational Autoencoder (VAE) models used for encoding and decoding images to and from a latent space:
29
  * **Vanilla VAE:** Standard VAE architecture.
30
 
31
  Each model checkpoint is provided as a `.pt` file containing the `state_dict` of the trained model.
 
32
 
33
  ## Intended Use
34
 
 
18
  This repository contains pre-trained weights for the following models, as described in the research article:
19
 
20
  * **TempVerseFormer (Rev-Transformer):** The core Reversible Temporal Transformer architecture, leveraging reversible blocks and time-agnostic backpropagation for memory efficiency.
 
21
  * **TempFormer (Vanilla-Transformer):** A standard Vanilla Transformer architecture with temporal chaining, serving as a baseline to compare against TempVerseFormer.
22
+ * **Standard Transformer (Pipe-Transformer):** Standard Transformer (Pipe-Transformer): A standard Transformer model that predicts only one next element at once.
 
 
23
  * **LSTM:** A Long Short-Term Memory network, representing a traditional recurrent sequence modeling approach.
 
24
  * **VAE Models:** Variational Autoencoder (VAE) models used for encoding and decoding images to and from a latent space:
25
  * **Vanilla VAE:** Standard VAE architecture.
26
 
27
  Each model checkpoint is provided as a `.pt` file containing the `state_dict` of the trained model.
28
+ * For all of the models checkpoints available for different training configurations (e.g., with/without temporal patterns).*
29
 
30
  ## Intended Use
31