yamada-aistar commited on
Commit
8fa80ce
·
verified ·
1 Parent(s): 59405f9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -20,8 +20,8 @@ language:
20
  LAM3C is a self-supervised learning method trained on video-generated point clouds reconstructed from unlabeled indoor walkthrough videos. This repository provides pretrained Point Transformerv3 (PTv3) backbones for feature extraction and downstream 3D scene understanding.
21
 
22
  > [!IMPORTANT]
23
- > - **LAM3C is not a raw-video model.** The released checkpoints take **point clouds** as input, not videos.
24
- > - The expected per-point input is 9D**: **XYZ + RGB + normals.
25
  > - The backbone checkpoints are feature extractors. They do not include a task-specific segmentation head unless explicitly stated.
26
 
27
  **arXiv:** [3D sans 3D Scans: Scalable Pre-training from Video-Generated Point Clouds (CVPR 2026)](https://arxiv.org/abs/2512.23042)
@@ -29,7 +29,7 @@ LAM3C is a self-supervised learning method trained on video-generated point clou
29
 
30
  ## What makes LAM3C different?
31
 
32
- Most 3D self-supervised learning methods rely on real 3D scans, which are expensive to collect at scale. LAM3C instead learns from **RoomTours**, a large collection of point clouds reconstructed from unlabeled room-tour videos gathered from the web.
33
 
34
  The method combines:
35
  - **RoomTours**, a scalable VGPC pre-training dataset
 
20
  LAM3C is a self-supervised learning method trained on video-generated point clouds reconstructed from unlabeled indoor walkthrough videos. This repository provides pretrained Point Transformerv3 (PTv3) backbones for feature extraction and downstream 3D scene understanding.
21
 
22
  > [!IMPORTANT]
23
+ > - LAM3C is not a raw-video model. The released checkpoints take point clouds as input, not videos.
24
+ > - The expected per-point input is 9D: XYZ + RGB + normals.
25
  > - The backbone checkpoints are feature extractors. They do not include a task-specific segmentation head unless explicitly stated.
26
 
27
  **arXiv:** [3D sans 3D Scans: Scalable Pre-training from Video-Generated Point Clouds (CVPR 2026)](https://arxiv.org/abs/2512.23042)
 
29
 
30
  ## What makes LAM3C different?
31
 
32
+ Most 3D self-supervised learning methods rely on real 3D scans, which are expensive to collect at scale. LAM3C instead learns from RoomTours, a large collection of point clouds reconstructed from unlabeled room-tour videos gathered from the web.
33
 
34
  The method combines:
35
  - **RoomTours**, a scalable VGPC pre-training dataset