alistairfoggin commited on
Commit
d93d773
·
verified ·
1 Parent(s): f37e20c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -23
README.md CHANGED
@@ -8,14 +8,14 @@ tags:
8
  - relighting
9
  ---
10
 
11
- # CroCoDiLight: Repurposing Cross-View Completion Encoders for Relighting
12
  Disentangles illumination from scene content in CroCo (Cross-view Completion) latent representations. A learned lighting extractor separates each encoder embedding into a single lighting vector and lighting-invariant patch features, which can then be recombined with target lighting conditions. This enables shadow removal, albedo estimation, lighting transfer, and interpolation, trained on datasets two orders of magnitude smaller than the original CroCo pretraining.
13
 
14
- **Paper:** [OpenReview](https://openreview.net/forum?id=GKvb3HCyNk) (ICLR 2026)
15
- **Code:** [GitHub](https://github.com/alistairfoggin/CroCoDiLight)
 
16
 
17
  ## Pretrained Model Weights
18
-
19
  | File | Required for | Description |
20
  |---|---|---|
21
  | **Inference** | | |
@@ -25,38 +25,32 @@ Disentangles illumination from scene content in CroCo (Cross-view Completion) la
25
  | **Training** | | |
26
  | [`CroCoDiLight_decoder.pth`](https://huggingface.co/alistairfoggin/CroCoDiLight/resolve/main/CroCoDiLight_decoder.pth?download=true) | Training of `CroCoDiLight.pth` | The pretrained monocular decoder for the CroCo v2 encoder |
27
 
28
- `CroCoDiLight.pth` is the base model needed by every inference and evaluation script. The mapper weights are only needed for their respective tasks. Lighting transfer, freezing, and interpolation use the base model only.
29
-
30
- `CroCoDiLight_decoder.pth` is not necessary for inference as it is embedded into `CroCoDiLight.pth`, but can be used as a standalone decoder for the CroCo v2 ViTLarge encoder (which is embedded in the model weights too).
31
 
32
  ## Usage
33
-
34
  See the [GitHub repository](https://github.com/alistairfoggin/CroCoDiLight) for setup instructions, inference scripts, Gradio demos, training, and evaluation.
35
 
36
- ## License
 
 
 
 
 
 
 
 
 
37
 
 
38
  This project, including its source code and pretrained model weights, is
39
  licensed under [CC BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/). The
40
  pretrained weights are additionally subject to the license terms of the upstream training
41
  data documented in the [NOTICE](https://github.com/alistairfoggin/CroCoDiLight/blob/main/NOTICE) file.
42
 
43
  ## Acknowledgements
44
-
45
  CroCoDiLight builds on [CroCo](https://github.com/naver/croco) (Weinzaepfel et al.),
46
- licensed under CC BY-NC-SA 4.0 by Naver Corporation.
47
-
48
  Model training was performed on the Viking cluster, a high performance compute facility
49
  provided by the University of York. We are grateful for computational support from the
50
  University of York, IT Services and the Research IT team.
51
 
52
- ## Citation BibTeX
53
-
54
- ```bibtex
55
- @inproceedings{foggin2026crocodilight,
56
- title={{CroCoDiLight}: Repurposing Cross-View Completion Encoders for Relighting},
57
- author={Foggin, Alistair J and Smith, William A P},
58
- booktitle={The Fourteenth International Conference on Learning Representations},
59
- year={2026},
60
- url={https://openreview.net/forum?id=GKvb3HCyNk}
61
- }
62
- ```
 
8
  - relighting
9
  ---
10
 
11
+ # [ICLR 2026] CroCoDiLight: Repurposing Cross-View Completion Encoders for Relighting
12
  Disentangles illumination from scene content in CroCo (Cross-view Completion) latent representations. A learned lighting extractor separates each encoder embedding into a single lighting vector and lighting-invariant patch features, which can then be recombined with target lighting conditions. This enables shadow removal, albedo estimation, lighting transfer, and interpolation, trained on datasets two orders of magnitude smaller than the original CroCo pretraining.
13
 
14
+ **Paper:** [OpenReview](https://openreview.net/forum?id=GKvb3HCyNk) (ICLR 2026)
15
+ **Code:** [GitHub](https://github.com/alistairfoggin/CroCoDiLight)
16
+ **Project Page:** [alistairfoggin.com/projects/crocodilight](https://alistairfoggin.com/projects/crocodilight)
17
 
18
  ## Pretrained Model Weights
 
19
  | File | Required for | Description |
20
  |---|---|---|
21
  | **Inference** | | |
 
25
  | **Training** | | |
26
  | [`CroCoDiLight_decoder.pth`](https://huggingface.co/alistairfoggin/CroCoDiLight/resolve/main/CroCoDiLight_decoder.pth?download=true) | Training of `CroCoDiLight.pth` | The pretrained monocular decoder for the CroCo v2 encoder |
27
 
28
+ `CroCoDiLight.pth` is the base model needed by every inference and evaluation script. The mapper weights are only needed for their respective tasks. Lighting transfer, freezing, and interpolation use the base model only. `CroCoDiLight_decoder.pth` is not necessary for inference as it is embedded into `CroCoDiLight.pth`, but can be used as a standalone decoder for the CroCo v2 ViTLarge encoder (which is embedded in the model weights too).
 
 
29
 
30
  ## Usage
 
31
  See the [GitHub repository](https://github.com/alistairfoggin/CroCoDiLight) for setup instructions, inference scripts, Gradio demos, training, and evaluation.
32
 
33
+ ## Citation BibTeX
34
+ ```bibtex
35
+ @inproceedings{foggin2026crocodilight,
36
+ title={{CroCoDiLight}: Repurposing Cross-View Completion Encoders for Relighting},
37
+ author={Foggin, Alistair J and Smith, William A P},
38
+ booktitle={The Fourteenth International Conference on Learning Representations},
39
+ year={2026},
40
+ url={https://openreview.net/forum?id=GKvb3HCyNk}
41
+ }
42
+ ```
43
 
44
+ ## License
45
  This project, including its source code and pretrained model weights, is
46
  licensed under [CC BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/). The
47
  pretrained weights are additionally subject to the license terms of the upstream training
48
  data documented in the [NOTICE](https://github.com/alistairfoggin/CroCoDiLight/blob/main/NOTICE) file.
49
 
50
  ## Acknowledgements
 
51
  CroCoDiLight builds on [CroCo](https://github.com/naver/croco) (Weinzaepfel et al.),
52
+ licensed under CC BY-NC-SA 4.0 by Naver Corporation.
 
53
  Model training was performed on the Viking cluster, a high performance compute facility
54
  provided by the University of York. We are grateful for computational support from the
55
  University of York, IT Services and the Research IT team.
56