davidkim205 commited on
Commit
8a2225d
β€’
1 Parent(s): 4e0a9f7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +73 -1
README.md CHANGED
@@ -34,7 +34,79 @@ korean multi-task instruction dataset
34
  - CUDA Version: 12.2
35
 
36
  ## Training
37
- Refer github
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
 
39
  ------------------------------------------------
40
  # Original model card: Meta's Llama 2 7B-chat
 
34
  - CUDA Version: 12.2
35
 
36
  ## Training
37
+ Refer https://github.com/davidkim205/komt
38
+
39
+ ## Usage
40
+ ```
41
+ from transformers import AutoTokenizer, AutoModelForCausalLM
42
+ from transformers import TextStreamer, GenerationConfig
43
+
44
+ model_name='davidkim205/komt-llama2-7b-v1'
45
+ model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
46
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
47
+ streamer = TextStreamer(tokenizer)
48
+
49
+ def gen(x):
50
+ generation_config = GenerationConfig(
51
+ temperature=0.8,
52
+ top_p=0.8,
53
+ top_k=100,
54
+ max_new_tokens=512,
55
+ early_stopping=True,
56
+ do_sample=True,
57
+ )
58
+ q = f"### instruction: {x}\n\n### Response: "
59
+ gened = model.generate(
60
+ **tokenizer(
61
+ q,
62
+ return_tensors='pt',
63
+ return_token_type_ids=False
64
+ ).to('cuda'),
65
+ generation_config=generation_config,
66
+ pad_token_id=tokenizer.eos_token_id,
67
+ eos_token_id=tokenizer.eos_token_id,
68
+ streamer=streamer,
69
+ )
70
+ result_str = tokenizer.decode(gened[0])
71
+
72
+ start_tag = f"\n\n### Response: "
73
+ start_index = result_str.find(start_tag)
74
+
75
+ if start_index != -1:
76
+ result_str = result_str[start_index + len(start_tag):].strip()
77
+ return result_str
78
+
79
+ print(gen('μ œμ£Όλ„λ₯Ό 1λ°•2일둜 혼자 μ—¬ν–‰ν•˜λ €κ³  ν•˜λŠ”λ° μ—¬ν–‰ μ½”μŠ€λ₯Ό λ§Œλ“€μ–΄μ€˜'))
80
+ ```
81
+ output
82
+ ```
83
+ ### Response: μ œμ£Όλ„λ₯Ό 1λ°•2일둜 혼자 μ—¬ν–‰ν•˜λ €λ©΄ λ‹€μŒκ³Ό 같은 μ—¬ν–‰ μ½”μŠ€λ₯Ό λ§Œλ“€μ–΄ κ³„νšν•  수 μžˆμŠ΅λ‹ˆλ‹€:
84
+
85
+ 1일차:
86
+ - μ•„μΉ¨: μ œμ£Όλ„μ˜ μ•„λ¦„λ‹€μš΄ 해변을 κ΅¬κ²½ν•˜κΈ° μœ„ν•΄ 해변에 λ„μ°©ν•˜μ„Έμš”. μΌμΆœμ„ κ°μƒν•˜λ©° μžμ—°μ˜ 아름닀움을 λ§Œλ½ν•˜μ„Έμš”.
87
+ - μ˜€ν›„: μ œμ£Όλ„μ˜ λŒ€ν‘œμ μΈ 관광지인 ν•œλΌμ‚°μ„ νƒν—˜ν•˜μ„Έμš”. λ“±μ‚°λ‘œλ₯Ό 따라 μ˜¬λΌκ°€λ©΄μ„œ 경치λ₯Ό 즐기고 μ„€λͺ…을 λ“£μœΌλ©° μ‰¬μš΄ 산책을 μ¦κΈ°μ„Έμš”.
88
+ - 저녁: μ œμ£Όλ„μ˜ λ§›μžˆλŠ” μŒμ‹μ μ—μ„œ 저녁을 λ³΄λ‚΄μ„Έμš”. μ‹ μ„ ν•œ ν•΄μ‚°λ¬Όκ³Ό ν–₯μ‹ λ£Œλ‘œ λ§Œλ“  μŒμ‹μ„ λ§›λ³΄λŠ” 것은 μ œμ£Όλ„ μ—¬ν–‰μ˜ μ™„λ²½ν•œ κ²½ν—˜μ΄ 될 κ²ƒμž…λ‹ˆλ‹€.
89
+
90
+ 2일차:
91
+ - μ•„μΉ¨: ν•œλΌμ‚° μΌλŒ€λ₯Ό νƒν—˜ν•˜κΈ° μœ„ν•΄ ν•œλΌμ‚° μΌ€μ΄ν”„λ‘œ μ΄λ™ν•˜μ„Έμš”. 이 μΌ€μ΄ν”„λŠ” 등산을 μ¦κΈ°λŠ” μ‚¬λžŒλ“€μ—κ²Œ 졜적의 μ„ νƒμž…λ‹ˆλ‹€.
92
+
93
+ ```
94
+ ## Evaluation
95
+
96
+ | model | score | average(0~5) | percentage |
97
+ | --------------------------------------- | ------- | ------------ | ---------- |
98
+ | gpt-3.5-turbo(close) | 147 | 3.97 | 79.45% |
99
+ | naver Cue(close) | 140 | 3.78 | 75.67% |
100
+ | clova X(close) | 136 | 3.67 | 73.51% |
101
+ | WizardLM-13B-V1.2(open) | 96 | 2.59 | 51.89% |
102
+ | Llama-2-7b-chat-hf(open) | 67 | 1.81 | 36.21% |
103
+ | Llama-2-13b-chat-hf(open) | 73 | 1.91 | 38.37% |
104
+ | nlpai-lab/kullm-polyglot-12.8b-v2(open) | 70 | 1.89 | 37.83% |
105
+ | kfkas/Llama-2-ko-7b-Chat(open) | 96 | 2.59 | 51.89% |
106
+ | beomi/KoAlpaca-Polyglot-12.8B(open) | 100 | 2.70 | 54.05% |
107
+ | **komt-llama2-7b-v1 (open)(ours)** | **117** | **3.16** | **63.24%** |
108
+ | **komt-llama2-13b-v1 (open)(ours)** | **129** | **3.48** | **69.72%** |
109
+
110
 
111
  ------------------------------------------------
112
  # Original model card: Meta's Llama 2 7B-chat