ybelkada nielsr HF staff commited on
Commit
235c75e
1 Parent(s): 6e723d9

Update README.md (#18)

Browse files

- Update README.md (482f69651497a402d12e1883c3ed36a31c643c20)


Co-authored-by: Niels Rogge <[email protected]>

Files changed (1) hide show
  1. README.md +13 -0
README.md CHANGED
@@ -59,6 +59,19 @@ BLIP2 has not been tested in real world applications. It should not be directly
59
 
60
  For code examples, we refer to the [documentation](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2#transformers.Blip2ForConditionalGeneration.forward.example).
61
 
 
 
 
 
 
 
 
 
 
 
 
 
 
62
  #### Running the model on CPU
63
 
64
  <details>
 
59
 
60
  For code examples, we refer to the [documentation](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2#transformers.Blip2ForConditionalGeneration.forward.example).
61
 
62
+ ### Memory requirements
63
+
64
+ The memory requirements differ based on the precision one uses. One can use 4-bit inference using [Bitsandbytes](https://huggingface.co/blog/4bit-transformers-bitsandbytes), which greatly reduce the memory requirements.
65
+
66
+ Training requires 4 times the
67
+
68
+ | dtype | Largest Layer or Residual Group | Total Size | Training using Adam |
69
+ |-------------------|---------------------------------|------------|----------------------|
70
+ | float32 | 490.94 MB | 14.43 GB | 57.72 GB |
71
+ | float16/bfloat16 | 245.47 MB | 7.21 GB | 28.86 GB |
72
+ | int8 | 122.73 MB | 3.61 GB | 14.43 GB |
73
+ | int4 | 61.37 MB | 1.8 GB | 7.21 GB |
74
+
75
  #### Running the model on CPU
76
 
77
  <details>