from opencompass.models import Llama2 # Please follow the instruction in the Meta AI website https://github.com/facebookresearch/llama/tree/llama_v1 # and download the LLaMA model and tokenizer to the path './models/llama/'. # # The LLaMA requirement is also needed to be installed. # *Note* that the LLaMA-2 branch is fully compatible with LLAMA-1, and the LLaMA-2 branch is used here. # # git clone https://github.com/facebookresearch/llama.git # cd llama # pip install -e . models = [ dict( abbr="llama-30b", type=Llama2, path="./models/llama/30B/", tokenizer_path="./models/llama/tokenizer.model", max_out_len=100, max_seq_len=2048, batch_size=16, run_cfg=dict(num_gpus=4, num_procs=4), ), ]