Video-Text-to-Text
Transformers
Safetensors
English
llava
text-generation
multimodal
Eval Results
Inference Endpoints

Add more packages to "pip install"

#9
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -157,7 +157,7 @@ The model was trained on [LLaVA-Video-178K](https://huggingface.co/datasets/lmms
157
  We provide the simple generation process for using our model. For more details, you could refer to [Github](https://github.com/LLaVA-VL/LLaVA-NeXT).
158
 
159
  ```python
160
- # pip install git+https://github.com/LLaVA-VL/LLaVA-NeXT.git
161
  from llava.model.builder import load_pretrained_model
162
  from llava.mm_utils import get_model_name_from_path, process_images, tokenizer_image_token
163
  from llava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN, IGNORE_INDEX
 
157
  We provide the simple generation process for using our model. For more details, you could refer to [Github](https://github.com/LLaVA-VL/LLaVA-NeXT).
158
 
159
  ```python
160
+ # pip install transformers einops decord accelerate git+https://github.com/LLaVA-VL/LLaVA-NeXT.git
161
  from llava.model.builder import load_pretrained_model
162
  from llava.mm_utils import get_model_name_from_path, process_images, tokenizer_image_token
163
  from llava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN, IGNORE_INDEX