from mmengine.config import read_base | |
with read_base(): | |
from .datasets.lveval.lveval import LVEval_datasets as datasets | |
from .models.hf_llama.hf_llama2_7b_chat import models | |
from .summarizers.lveval import summarizer | |
models[0][ | |
"path" | |
] = "/path/to/your/huggingface_models/Llama-2-7b-chat-hf" | |
models[0][ | |
"tokenizer_path" | |
] = "/path/to/your/huggingface_models/Llama-2-7b-chat-hf" | |
models[0]["max_seq_len"] = 4096 | |
models[0]["generation_kwargs"] = dict(do_sample=False) | |
models[0]["mode"] = "mid" # truncate in the middle | |