Neversleep-3B-Instruct-v0.1 / README copy.md
crimsonjoo's picture
update neversleeps
25b7d16
metadata
license: apache-2.0
tags:
  - generated_from_trainer
base_model: yanolja/EEVE-Korean-2.8B-v1.0

Built with Axolotl

"We must sleep, but AI Never Sleeps!"

 

Prompt Template

A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
Human: {prompt}
Assistant:

Simple-Usage

from transformers import AutoTokenizer
from transformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("yanolja/EEVE-Korean-Instruct-2.8B-v1.0", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("yanolja/EEVE-Korean-Instruct-2.8B-v1.0", trust_remote_code=True)

prompt_template = "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.\nHuman: {prompt}\nAssistant:\n"
text = 'λ‹€μ΄μ–΄νŠΈμ‹ 메뉴λ₯Ό μΆ”μ²œν•΄μ£Όμ„Έμš”.\n\n(A) μƒλŸ¬λ“œ\n(B) μΉ˜ν‚¨\n(C) ν”Όμž\n(D) νŒŒμŠ€νƒ€'
model_inputs = tokenizer(prompt_template.format(prompt=text), return_tensors='pt')

outputs = model.generate(**model_inputs, max_new_tokens=256)
output_text = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]
print(output_text)

Example Output

A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
Human: λ‹€μ΄μ–΄νŠΈμ‹ 메뉴λ₯Ό μΆ”μ²œν•΄μ£Όμ„Έμš”.

(A) μƒλŸ¬λ“œ
(B) μΉ˜ν‚¨
(C) ν”Όμž
(D) νŒŒμŠ€νƒ€
Assistant:
(A) μƒλŸ¬λ“œλ₯Ό μΆ”μ²œλ“œλ¦½λ‹ˆλ‹€. μƒλŸ¬λ“œλŠ” μ €μΉΌλ‘œλ¦¬μ΄λ©΄μ„œλ„ μ˜μ–‘μ†Œκ°€ 풍뢀해 λ‹€μ΄μ–΄νŠΈμ‹μœΌλ‘œ μ ν•©ν•©λ‹ˆλ‹€. λ‹€μ–‘ν•œ μ±„μ†Œμ™€ λ‹¨λ°±μ§ˆμ„ μΆ”κ°€ν•˜μ—¬ κ· ν˜• 작힌 식사λ₯Ό λ§Œλ“œμ‹€ 수 μžˆμŠ΅λ‹ˆλ‹€.

About the Model

First of all, Overwhelming gratitude to 'yanolja/EEVE' Model & Team! This model is a fine-tuned version of crimsonjoo/Neversleep-3B-v0.1, which is a Korean vocabulary-extended version of microsoft/phi-2. Specifically, we utilized Direct Preference Optimization (DPO) through the use of Axolotl.

For more details, please refer to our technical report: Efficient and Effective Vocabulary Expansion Towards Multilingual Large Language Models.

Training Data