--- library_name: transformers license: apache-2.0 --- # WestLake-7B-v2-laser-truthy-dpo ![westlake-header](westlake-header.png) ## Process + Trained [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) on jondurbin/truthy-dpo-v0.1 + Completed 2 epochs + 2e-5 learning rate ## Code Example ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_id = "macadeliccc/WestLake-7B-v2-laser-truthy-dpo" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained(model_id) text = "Create an idea for a TV show and write a short pilot script" inputs = tokenizer(text, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=4096) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Evaluations ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6455cc8d679315e4ef16fbec/9CJeaPxf4XGJv7w114LKo.png)
----Benchmark Complete---- 2024-01-27 16:44:07 Time taken: 29.6 mins Prompt Format: Mistral Model: macadeliccc/WestLake-7B-v2-laser-truthy-dpo Score (v2): 73.39 Parseable: 169.0 --------------- Batch completed Time taken: 29.6 mins ---------------## GGUF GGUF versions are available [here](https://huggingface.co/macadeliccc/WestLake-7B-v2-laser-truthy-dpo-GGUF)