Can't load the model

#2
by Ming369 - opened

In config.json file, there is:
"architectures": [
"LlamaEncoderModel"
],
"auto_map": {
"AutoModel": "McGill-NLP/LLM2Vec-Meta-Llama-31-8B-Instruct-mntp--modeling_llama_encoder.LlamaEncoderModel"
}

But in modeling_llama_encoder.py, the model name is "BidirectionalLlama", not "LlamaEncoderModel". So when I tried to load the model, such error occurred:
AttributeError: module 'transformers_modules.McGill-NLP.LLM2Vec-Meta-Llama-31-8B-Instruct-mntp.1d49bff4203a867109580085c67e3b3cc2984a89.modeling_llama_encoder' has no attribute 'LlamaEncoderModel'

McGill NLP Group org

Thanks for bringing this to my attention. I have fixed this. Can you check now?

Thank you for fixing this. I can load the model correctly now :)

Ming369 changed discussion status to closed

Sign up or log in to comment