Use V1 tokenizer instead

#10
by Rocketknight1 HF staff - opened
No description provided.
Rocketknight1 changed pull request title from Upload tokenizer to Use V1 tokenizer instead

There was an issue with the last PR - we used the V3 tokenizer, but this base model actually uses the V1 tokenizer. This should fix the issue!

@Rocketknight1 does it affect the vocab size? Model and tokenizer sizes are not matching. So model is failing to load.

@lbathen can you give me some code to reproduce that issue? From here it looks like the tokenizer and the model both have a vocab size of 32000

@Rocketknight1 I confirmed that both show same vocab of 32K now. I had pulled the wrong revision :)

Is this going to be merged soon?

@Rocketknight1 Could you merge this in? It's working on my end and I'm thankful to have this model back.

This command should get it running for anyone who needs it:

python -m vllm.entrypoints.openai.api_server --model mistralai/Mixtral-8x22B-v0.1 --revision c356b81 --served-model-name mistralai/Mixtral-8x22B-v0.1 --max-logprobs 100 --gpu-memory-utilization=0.85 --disable-log-requests --disable-log-stats --port 5001 --tensor-parallel-size 8
Mistral AI_ org

My apologies! Merging this PR!

pandora-s changed pull request status to closed
pandora-s changed pull request status to open
pandora-s changed pull request status to merged

Sign up or log in to comment