from opencompass.models import Llama2Chat # Please follow the instruction in the Meta AI website https://github.com/facebookresearch/llama # and download the LLaMA-2-Chat model and tokenizer to the path './models/llama2/llama/'. # # The LLaMA requirement is also needed to be installed. # # git clone https://github.com/facebookresearch/llama.git # cd llama # pip install -e . api_meta_template = dict( round=[ dict(role="HUMAN", api_role="HUMAN"), dict(role="BOT", api_role="BOT", generate=True), ], ) models = [ dict( abbr="llama-2-70b-chat", type=Llama2Chat, path="./models/llama2/llama/llama-2-70b-chat/", tokenizer_path="./models/llama2/llama/tokenizer.model", meta_template=api_meta_template, max_out_len=100, max_seq_len=2048, batch_size=16, run_cfg=dict(num_gpus=8, num_procs=8), ), ]