manbeast3b commited on
Commit
90aaa65
·
verified ·
1 Parent(s): c2db1a0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -18
README.md CHANGED
@@ -1,19 +1,4 @@
1
- # flux-schnell-edge-inference
2
 
3
- This holds the baseline for the FLUX Schnel NVIDIA GeForce RTX 4090 contest, which can be forked freely and optimized
4
-
5
- Some recommendations are as follows:
6
- - Installing dependencies should be done in `pyproject.toml`, including git dependencies
7
- - HuggingFace models should be specified in the `models` array in the `pyproject.toml` file, and will be downloaded before benchmarking
8
- - The pipeline does **not** have internet access so all dependencies and models must be included in the `pyproject.toml`
9
- - Compiled models should be hosted on HuggingFace and included in the `models` array in the `pyproject.toml` (rather than compiling during loading). Loading time matters far more than file sizes
10
- - Avoid changing `src/main.py`, as that includes mostly protocol logic. Most changes should be in `models` and `src/pipeline.py`
11
- - Ensure the entire repository (excluding dependencies and HuggingFace models) is under 16MB
12
-
13
- For testing, you need a docker container with pytorch and ubuntu 22.04.
14
- You can download your listed dependencies with `uv`, installed with:
15
- ```bash
16
- pipx ensurepath
17
- pipx install uv
18
- ```
19
- You can then relock with `uv lock`, and then run with `uv run start_inference`
 
1
+ Transformer copy of slobers/Flux.1.Schnella
2
 
3
+ + +vae encoder pruned to 0.15%
4
+ + +vae decoder pruned to 2%