Update README.md
Browse files
README.md
CHANGED
|
@@ -20,8 +20,8 @@ language:
|
|
| 20 |
LAM3C is a self-supervised learning method trained on video-generated point clouds reconstructed from unlabeled indoor walkthrough videos. This repository provides pretrained Point Transformerv3 (PTv3) backbones for feature extraction and downstream 3D scene understanding.
|
| 21 |
|
| 22 |
> [!IMPORTANT]
|
| 23 |
-
> -
|
| 24 |
-
> - The expected per-point input is 9D
|
| 25 |
> - The backbone checkpoints are feature extractors. They do not include a task-specific segmentation head unless explicitly stated.
|
| 26 |
|
| 27 |
**arXiv:** [3D sans 3D Scans: Scalable Pre-training from Video-Generated Point Clouds (CVPR 2026)](https://arxiv.org/abs/2512.23042)
|
|
@@ -29,7 +29,7 @@ LAM3C is a self-supervised learning method trained on video-generated point clou
|
|
| 29 |
|
| 30 |
## What makes LAM3C different?
|
| 31 |
|
| 32 |
-
Most 3D self-supervised learning methods rely on real 3D scans, which are expensive to collect at scale. LAM3C instead learns from
|
| 33 |
|
| 34 |
The method combines:
|
| 35 |
- **RoomTours**, a scalable VGPC pre-training dataset
|
|
|
|
| 20 |
LAM3C is a self-supervised learning method trained on video-generated point clouds reconstructed from unlabeled indoor walkthrough videos. This repository provides pretrained Point Transformerv3 (PTv3) backbones for feature extraction and downstream 3D scene understanding.
|
| 21 |
|
| 22 |
> [!IMPORTANT]
|
| 23 |
+
> - LAM3C is not a raw-video model. The released checkpoints take point clouds as input, not videos.
|
| 24 |
+
> - The expected per-point input is 9D: XYZ + RGB + normals.
|
| 25 |
> - The backbone checkpoints are feature extractors. They do not include a task-specific segmentation head unless explicitly stated.
|
| 26 |
|
| 27 |
**arXiv:** [3D sans 3D Scans: Scalable Pre-training from Video-Generated Point Clouds (CVPR 2026)](https://arxiv.org/abs/2512.23042)
|
|
|
|
| 29 |
|
| 30 |
## What makes LAM3C different?
|
| 31 |
|
| 32 |
+
Most 3D self-supervised learning methods rely on real 3D scans, which are expensive to collect at scale. LAM3C instead learns from RoomTours, a large collection of point clouds reconstructed from unlabeled room-tour videos gathered from the web.
|
| 33 |
|
| 34 |
The method combines:
|
| 35 |
- **RoomTours**, a scalable VGPC pre-training dataset
|