Where can I deploy this model for inference?

#7
by catworld1212 - opened

Hi, I'm impressed with the work on InternVL and I'm interested in deploying its inference as an endpoint. Unfortunately, vLLM and TGI don't support this. Could anyone offer guidance on how to achieve this? I'd appreciate any suggestions you may have.

See "Chat Web Demo" at https://github.com/OpenGVLab/InternVL/blob/main/README.md

I want to deploy it as an inference not run it as a demo, Can you tell do InternVL-Chat-V1-5 requires flash attention?

Hi @whai362 @czczup what's the proper way to few-shot prompting (also called in-context learning? How do I give the previous context? I'm using lmdeploy to serve the inference can you help me, please?

OpenGVLab org

Hi @whai362 @czczup what's the proper way to few-shot prompting (also called in-context learning? How do I give the previous context? I'm using lmdeploy to serve the inference can you help me, please?

Hi, you can perform few-shot prompting in the form of multi-turn dialogue, placing the few-shot examples in the conversation history. I think this shouldn't be different from how other VLMs perform few-shot prompting.

czczup changed discussion status to closed

Sign up or log in to comment