Depth LoRA Adapters (All)
Collection
All uploaded best LoRA adapters from depth-LoRA experiments (Booster, Metric3D, mixed datasets, MUSES, Taskonomy FOV, vKITTI2 conditions). โข 30 items โข Updated
Configuration Parsing Warning: In adapter_config.json: "peft.base_model_name_or_path" must be a string
Configuration Parsing Warning: In adapter_config.json: "peft.task_type" must be a string
Base model:
metric3d_vit_giant2loaded viatorch.hub.load("yvanyin/metric3d", "metric3d_vit_giant2")
LoRA fine-tuning of Metric3D v2
(metric3d_vit_giant2) on the Booster stereo depth dataset.
| Base model | metric3d_vit_giant2 (torch.hub yvanyin/metric3d) |
| Dataset | Booster stereo (prepared split, metric depth in metres) |
| Loss | Direct metric-depth L1 + gradient loss |
| LoRA rank / alpha | 16 / 32 |
| LoRA targets | qkv, proj |
| Best val AbsRel | 0.1094 |
| Epochs trained | 7 |
import torch
from peft import PeftModel
import torch.nn as nn
import torch.utils.checkpoint as torch_checkpoint
# Load base model
model = torch.hub.load("yvanyin/metric3d", "metric3d_vit_giant2",
pretrain=True, trust_repo=True)
# Apply the same gradient-checkpointing wrapper used during training
# (needed so PEFT key names match the saved adapter)
def enable_gradient_checkpointing(model):
try:
encoder = model.depth_model.encoder
except AttributeError:
encoder = model.base_model.model.depth_model.encoder
class _CheckpointedBlock(nn.Module):
def __init__(self, block):
super().__init__()
self.block = block
def forward(self, x):
return torch_checkpoint.checkpoint(self.block, x, use_reentrant=False)
for blk_group in encoder.blocks:
for key in list(blk_group._modules.keys()):
blk_group._modules[key] = _CheckpointedBlock(blk_group._modules[key])
enable_gradient_checkpointing(model)
model = PeftModel.from_pretrained(model, "igzi/metric3d-vit-giant2-lora-booster")
model.eval()
# Inference: input pixel_values shape (B, 3, 616, 1064), values normalised
# with mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375]
pred_canonical, _, _ = model({"input": pixel_values})
# De-canonicalise: pred_metric = pred_canonical * (fx_scaled / 1000)