from opencompass.models import HuggingFaceCausalLM models = [ dict( type=HuggingFaceCausalLM, abbr='vicuna-33b-v1.3-hf', path="lmsys/vicuna-33b-v1.3", tokenizer_path='lmsys/vicuna-33b-v1.3', tokenizer_kwargs=dict( padding_side='left', truncation_side='left', use_fast=False, ), max_out_len=100, max_seq_len=2048, batch_size=8, model_kwargs=dict(device_map='auto'), batch_padding=False, # if false, inference with for-loop without batch padding use_fastchat_template=True, run_cfg=dict(num_gpus=4, num_procs=1) ) ]