# GPT4ALL or llama-cpp-python model_kwargs | |
# GPT4ALl GPT-J type, from model explorer choice, so downloads | |
model_name_gptj=ggml-gpt4all-j-v1.3-groovy.bin | |
# llama-cpp-python type, supporting version 3 quantization, here from locally built llama.cpp q4 v3 quantization | |
# below uses prompt_type=wizard2 | |
model_path_llama=WizardLM-7B-uncensored.ggmlv3.q8_0.bin | |
# below assumes max_new_tokens=256 | |
n_ctx=1792 | |
# uncomment below if using llama-cpp-pyton with cublas built in | |
# n_gpu_layers=20 | |
# GPT4ALl LLaMa type, supporting version 2 quantization, here from model explorer choice so downloads | |
model_name_gpt4all_llama=ggml-wizardLM-7B.q4_2.bin | |
# PDF_CLASS_NAME=UnstructuredPDFLoader | |