gordon-0115 commited on
Commit
0084fcf
·
verified ·
1 Parent(s): da4c1e3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -12
README.md CHANGED
@@ -1,12 +1,56 @@
1
- ---
2
- license: mit
3
- pipeline_tag: image-to-3d
4
- ---
5
-
6
- # LGM Full
7
-
8
- This custom pipeline encapsulates the full [LGM](https://huggingface.co/ashawkey/LGM) pipeline, including [multi-view diffusion](https://huggingface.co/ashawkey/imagedream-ipmv-diffusers).
9
-
10
- It is provided as a resource for the [ML for 3D Course](https://huggingface.co/learn/ml-for-3d-course).
11
-
12
- Original LGM paper: [LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation](https://huggingface.co/papers/2402.05054).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ pipeline_tag: image-to-3d
4
+ ---
5
+
6
+ # LGM Full
7
+
8
+ # NLP to 3D Model Custom Pipeline
9
+
10
+ ## Overview
11
+
12
+ This project showcases an experimental pipeline that bridges **natural language prompts to 3D model generation** using a modified version of a pre-trained multi-view diffusion model.
13
+
14
+ It is part of a final year project for the *Comprehensive Creative Technologies Project* at UWE Bristol. The primary aim was to explore the potential of AI-assisted 3D content creation using natural language input.
15
+
16
+ ---
17
+
18
+ ## Model Source & Attribution
19
+
20
+ This project **relies on the pre-trained model from the [LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation](https://huggingface.co/papers/2402.05054)**, developed for the **ML for 3D Course** by researchers at Google Research.
21
+
22
+ 🔗 Original Model: *[https://huggingface.co/spaces/dylanebert/LGM-tiny]*
23
+ 📄 Paper: [arXiv:2402.05054](https://arxiv.org/abs/2402.05054)
24
+ 🔒 License: MIT
25
+
26
+ I do not claim authorship of the model architecture or training process. This space serves as a **custom wrapper** for experimentation with **text-to-3D workflows**.
27
+
28
+ ---
29
+
30
+ ## What This Model Does
31
+
32
+ - Allows input of a natural language description
33
+ - Internally maps the input to a representative image or multi-view description
34
+ - Generates a 3D model using the LGM pipeline
35
+
36
+ ---
37
+
38
+ ## Limitations
39
+
40
+ - This is a prototype for academic use only.
41
+ - The model’s ability to handle complex or abstract text is limited.
42
+ - Performance and quality depend entirely on the base pre-trained model.
43
+
44
+ ---
45
+
46
+ ## Acknowledgements
47
+
48
+ Thanks to Hugging Face and the authors of LGM for making their models publicly available.
49
+
50
+ ---
51
+
52
+ ## Author
53
+
54
+ **Gordon CHIN HO AU**
55
+ Final Year BSc Digital Media
56
+ University of the West of England, Bristol