|
|
--- |
|
|
license: cc-by-4.0 |
|
|
task_categories: |
|
|
- other |
|
|
tags: |
|
|
- 3d |
|
|
- computer-vision |
|
|
- orientation-estimation |
|
|
--- |
|
|
|
|
|
# Orient Anything V2 Dataset |
|
|
|
|
|
[**Project Page**](https://orient-anythingv2.github.io/) | [**Paper**](https://huggingface.co/papers/2601.05573) | [**GitHub**](https://github.com/SpatialVision/Orient-Anything-V2) |
|
|
|
|
|
**Orient Anything V2** is an enhanced foundation model for unified understanding of object 3D orientation and rotation from single or paired images. This repository contains the training data (final rendering data) used for the model. |
|
|
|
|
|
## Sample Usage |
|
|
|
|
|
Below is a snippet to run inference using the model and data logic, as found in the [official GitHub repository](https://github.com/SpatialVision/Orient-Anything-V2): |
|
|
|
|
|
```python |
|
|
import numpy as np |
|
|
from PIL import Image |
|
|
import torch |
|
|
import tempfile |
|
|
import os |
|
|
|
|
|
from paths import * |
|
|
from vision_tower import VGGT_OriAny_Ref |
|
|
from inference import * |
|
|
from app_utils import * |
|
|
|
|
|
mark_dtype = torch.bfloat16 if torch.cuda.get_device_capability()[0] >= 8 else torch.float16 |
|
|
# device = 'cuda:0' |
|
|
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu') |
|
|
|
|
|
if os.path.exists(LOCAL_CKPT_PATH): |
|
|
ckpt_path = LOCAL_CKPT_PATH |
|
|
else: |
|
|
from huggingface_hub import hf_hub_download |
|
|
ckpt_path = hf_hub_download(repo_id="Viglong/Orient-Anything-V2", filename=HF_CKPT_PATH, repo_type="model", cache_dir='./', resume_download=True) |
|
|
|
|
|
model = VGGT_OriAny_Ref(out_dim=900, dtype=mark_dtype, nopretrain=True) |
|
|
model.load_state_dict(torch.load(ckpt_path, map_location='cpu')) |
|
|
model.eval() |
|
|
model = model.to(device) |
|
|
print('Model loaded.') |
|
|
|
|
|
@torch.no_grad() |
|
|
def run_inference(pil_ref, pil_tgt=None, do_rm_bkg=True): |
|
|
if pil_tgt is not None: |
|
|
if do_rm_bkg: |
|
|
pil_ref = background_preprocess(pil_ref, True) |
|
|
pil_tgt = background_preprocess(pil_tgt, True) |
|
|
else: |
|
|
if do_rm_bkg: |
|
|
pil_ref = background_preprocess(pil_ref, True) |
|
|
|
|
|
try: |
|
|
ans_dict = inf_single_case(model, pil_ref, pil_tgt) |
|
|
except Exception as e: |
|
|
print("Inference error:", e) |
|
|
raise gr.Error(f"Inference failed: {str(e)}") |
|
|
|
|
|
def safe_float(val, default=0.0): |
|
|
try: |
|
|
return float(val) |
|
|
except: |
|
|
return float(default) |
|
|
|
|
|
az = safe_float(ans_dict.get('ref_az_pred', 0)) |
|
|
el = safe_float(ans_dict.get('ref_el_pred', 0)) |
|
|
ro = safe_float(ans_dict.get('ref_ro_pred', 0)) |
|
|
alpha = int(ans_dict.get('ref_alpha_pred', 1)) |
|
|
|
|
|
if pil_tgt is not None: |
|
|
rel_az = safe_float(ans_dict.get('rel_az_pred', 0)) |
|
|
rel_el = safe_float(ans_dict.get('rel_el_pred', 0)) |
|
|
rel_ro = safe_float(ans_dict.get('rel_ro_pred', 0)) |
|
|
|
|
|
print("Relative Pose: Azi",rel_az,"Ele",rel_el,"Rot",rel_ro) |
|
|
|
|
|
image_ref_path = 'assets/examples/F35-0.jpg' |
|
|
image_tgt_path = 'assets/examples/F35-1.jpg' # optional |
|
|
|
|
|
image_ref = Image.open(image_ref_path).convert('RGB') |
|
|
image_tgt = Image.open(image_tgt_path).convert('RGB') |
|
|
|
|
|
run_inference(image_ref, image_tgt, True) |
|
|
``` |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you find this project useful, please consider citing: |
|
|
|
|
|
```bibtex |
|
|
@inproceedings{wangorient, |
|
|
title={Orient Anything V2: Unifying Orientation and Rotation Understanding}, |
|
|
author={Wang, Zehan and Zhang, Ziang and Xu, Jiayang and Wang, Jialei and Pang, Tianyu and Du, Chao and Zhao, Hengshuang and Zhao, Zhou}, |
|
|
booktitle={The Thirty-ninth Annual Conference on Neural Information Processing Systems} |
|
|
} |
|
|
``` |