Update README.md
#18
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -59,6 +59,19 @@ BLIP2 has not been tested in real world applications. It should not be directly
|
|
| 59 |
|
| 60 |
For code examples, we refer to the [documentation](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2#transformers.Blip2ForConditionalGeneration.forward.example).
|
| 61 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 62 |
#### Running the model on CPU
|
| 63 |
|
| 64 |
<details>
|
|
|
|
| 59 |
|
| 60 |
For code examples, we refer to the [documentation](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2#transformers.Blip2ForConditionalGeneration.forward.example).
|
| 61 |
|
| 62 |
+
### Memory requirements
|
| 63 |
+
|
| 64 |
+
The memory requirements differ based on the precision one uses. One can use 4-bit inference using [Bitsandbytes](https://huggingface.co/blog/4bit-transformers-bitsandbytes), which greatly reduce the memory requirements.
|
| 65 |
+
|
| 66 |
+
Training requires 4 times the
|
| 67 |
+
|
| 68 |
+
| dtype | Largest Layer or Residual Group | Total Size | Training using Adam |
|
| 69 |
+
|-------------------|---------------------------------|------------|----------------------|
|
| 70 |
+
| float32 | 490.94 MB | 14.43 GB | 57.72 GB |
|
| 71 |
+
| float16/bfloat16 | 245.47 MB | 7.21 GB | 28.86 GB |
|
| 72 |
+
| int8 | 122.73 MB | 3.61 GB | 14.43 GB |
|
| 73 |
+
| int4 | 61.37 MB | 1.8 GB | 7.21 GB |
|
| 74 |
+
|
| 75 |
#### Running the model on CPU
|
| 76 |
|
| 77 |
<details>
|