Spaces:
Running
on
T4
Running
on
T4
title: PaliGemma Demo | |
emoji: 🤲 | |
colorFrom: green | |
colorTo: yellow | |
sdk: gradio | |
sdk_version: 4.22.0 | |
app_file: app.py | |
pinned: false | |
license: apache-2.0 | |
# PaliGemma Demo | |
See [Blogpost] and [`big_vision README.md`] for details about the model. | |
[Blogpost]: https://huggingface.co/blog/paligemma | |
[`big_vision README.md`]: https://github.com/google-research/big_vision/blob/main/big_vision/configs/proj/paligemma/README.md | |
## Development | |
Local testing (CPU, Python 3.12): | |
```bash | |
pip -m venv env | |
. env/bin/activate | |
pip install -qr requirements-cpu.txt | |
python app.py | |
``` | |
Environment variables: | |
- `MOCK_MODEL=yes`: For quick UI testing. | |
- `RAM_CACHE_GB=18`: Enables caching of 3 bf16 models in memory: a single bf16 | |
model is about 5860 MB. Use with care on spaces with little RAM. For example, | |
on a `A10G large` space you can cache five models in RAM, so you would set | |
`RAM_CACHE_GB=30`. | |
- `HOST_COLOCATION=4`: If host RAM/disk is shared between 4 processes (e.g. the | |
Huggingface `A10 large` Spaces). | |
Loading models: | |
- The set of models loaded is defined in `./models.py`. | |
- You must first acknowledge usage conditions to access models. | |
- When testing locally, you'll have to run `huggingface_cli login`. | |
- When running in a Huggingface Space, you'll have to set a `HF_TOKEN` secret. | |