Update README.md
Browse files
README.md
CHANGED
|
@@ -21,3 +21,29 @@ how to recontruction noicing image to cleaning image.
|
|
| 21 |
LCVLM Train similiar Diffusion model when model training by self supervised learning by noiced image as input and clean image as output.
|
| 22 |
|
| 23 |
LCVLM Encoder train by Flickr8K datasets with MSE function as loss function, Adam as optimizers and here is model bechmark:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
LCVLM Train similiar Diffusion model when model training by self supervised learning by noiced image as input and clean image as output.
|
| 22 |
|
| 23 |
LCVLM Encoder train by Flickr8K datasets with MSE function as loss function, Adam as optimizers and here is model bechmark:
|
| 24 |
+
|
| 25 |
+

|
| 26 |
+
|
| 27 |
+
this bechmark train by spesific model:
|
| 28 |
+
embedding_dim and d_model: 256 \n
|
| 29 |
+
Global_FFN: (1024,256) \n
|
| 30 |
+
ConvTranspose2D: (128,64,64) \n
|
| 31 |
+
num_LCT_block: 3
|
| 32 |
+
|
| 33 |
+
model_size: 8.6 mb \n
|
| 34 |
+
|
| 35 |
+
how to use mode:
|
| 36 |
+
bash
|
| 37 |
+
```
|
| 38 |
+
from lctvlm import *
|
| 39 |
+
model = PretrainedVIT(
|
| 40 |
+
image_size=224,
|
| 41 |
+
patch_size=8,
|
| 42 |
+
embdding_dim=256,
|
| 43 |
+
n_block=3
|
| 44 |
+
)
|
| 45 |
+
checkpoint = torch.load(path_model)
|
| 46 |
+
model.load_state_dict(checkpoint)
|
| 47 |
+
```
|
| 48 |
+
|
| 49 |
+
note: LCTVLM is still not finish to create, QFormers and LMLCT will update soon
|