File size: 1,798 Bytes
a25977f
5294cb7
c96faea
 
 
 
 
 
 
 
 
a25977f
5294cb7
0f1544f
5294cb7
0f1544f
 
5294cb7
0f1544f
 
5294cb7
0f1544f
 
 
 
c96faea
0f1544f
c96faea
0f1544f
 
 
 
 
 
c96faea
0f1544f
c96faea
0f1544f
 
 
 
 
 
c96faea
0f1544f
 
c96faea
 
 
 
 
 
 
0f1544f
c96faea
 
 
0f1544f
2be038b
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
---
library_name: peft
tags:
- tiiuae
- code
- instruct
- databricks-dolly-15k
- falcon-40b
datasets:
- databricks/databricks-dolly-15k
base_model: tiiuae/falcon-40b
---

### Finetuning Overview:

**Model Used:** tiiuae/falcon-40b  
**Dataset:** Databricks-dolly-15k  

#### Dataset Insights:
The Databricks-dolly-15k dataset, comprising over 15,000 records, stands as a testament to the dedication of numerous Databricks professionals. Aimed at refining the interactive capabilities of systems like ChatGPT, the dataset offers:

- Prompt/response pairs across eight distinct instruction categories.
- A blend of the seven categories from the InstructGPT paper and an open-ended category.
- Original content, devoid of generative AI influence and primarily offline-sourced, with exceptions for Wikipedia references.
- Interactive sessions where contributors could address and rephrase peer questions.

Note: Some data categories incorporate Wikipedia references, evident from bracketed citation numbers, e.g., [42]. Exclusion is recommended for downstream applications.

#### Finetuning Details:

Leveraging [MonsterAPI](https://monsterapi.ai)'s no-code [LLM finetuner](https://docs.monsterapi.ai/fine-tune-a-large-language-model-llm), our finetuning emphasized:

- **Cost-Effectiveness:** A complete run at just `$11.8`.
- **Efficiency:** Using an A6000 48GB GPU, the session concluded in 5 hours and 40 minutes.

#### Hyperparameters & Additional Details:

- **Epochs:** 1
- **Learning Rate:** 0.0002
- **Data Split:** Training 90% / Validation 10%
- **Gradient Accumulation Steps:** 4

---

### Prompt Structure:
```
### INSTRUCTION:
[instruction]

[context]

### RESPONSE:
[response]
```

Loss metrics

Training loss:
![training loss](train-loss.png "Training loss")


---

license: apache-2.0