File size: 1,817 Bytes
a25977f
5294cb7
c96faea
 
 
 
 
 
 
 
 
1732c47
a25977f
5294cb7
0f1544f
5294cb7
0f1544f
 
5294cb7
0f1544f
 
5294cb7
0f1544f
 
 
 
c96faea
0f1544f
c96faea
0f1544f
 
 
 
 
 
c96faea
0f1544f
c96faea
0f1544f
 
 
 
 
 
c96faea
0f1544f
 
c96faea
 
 
 
 
 
 
0f1544f
c96faea
 
 
0f1544f
2be038b
 
 
 
 
1732c47
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
---
library_name: peft
tags:
- tiiuae
- code
- instruct
- databricks-dolly-15k
- falcon-40b
datasets:
- databricks/databricks-dolly-15k
base_model: tiiuae/falcon-40b
license: apache-2.0
---

### Finetuning Overview:

**Model Used:** tiiuae/falcon-40b  
**Dataset:** Databricks-dolly-15k  

#### Dataset Insights:
The Databricks-dolly-15k dataset, comprising over 15,000 records, stands as a testament to the dedication of numerous Databricks professionals. Aimed at refining the interactive capabilities of systems like ChatGPT, the dataset offers:

- Prompt/response pairs across eight distinct instruction categories.
- A blend of the seven categories from the InstructGPT paper and an open-ended category.
- Original content, devoid of generative AI influence and primarily offline-sourced, with exceptions for Wikipedia references.
- Interactive sessions where contributors could address and rephrase peer questions.

Note: Some data categories incorporate Wikipedia references, evident from bracketed citation numbers, e.g., [42]. Exclusion is recommended for downstream applications.

#### Finetuning Details:

Leveraging [MonsterAPI](https://monsterapi.ai)'s no-code [LLM finetuner](https://docs.monsterapi.ai/fine-tune-a-large-language-model-llm), our finetuning emphasized:

- **Cost-Effectiveness:** A complete run at just `$11.8`.
- **Efficiency:** Using an A6000 48GB GPU, the session concluded in 5 hours and 40 minutes.

#### Hyperparameters & Additional Details:

- **Epochs:** 1
- **Learning Rate:** 0.0002
- **Data Split:** Training 90% / Validation 10%
- **Gradient Accumulation Steps:** 4

---

### Prompt Structure:
```
### INSTRUCTION:
[instruction]

[context]

### RESPONSE:
[response]
```

Loss metrics

Training loss:
![training loss](train-loss.png "Training loss")


---

license: apache-2.0