eluzhnica commited on
Commit
67ce01e
1 Parent(s): 2f0d883

Add clarifications and disclaimer

Browse files
Files changed (1) hide show
  1. README.md +9 -1
README.md CHANGED
@@ -19,6 +19,14 @@ inference: false
19
 
20
  # MPT-30B-Instruct
21
 
 
 
 
 
 
 
 
 
22
  MPT-30B-Instruct is a model for short-form instruction following.
23
  It is built by finetuning [MPT-30B](https://huggingface.co/mosaicml/mpt-30b) on [Dolly HHRLHF](https://huggingface.co/datasets/mosaicml/dolly_hhrlhf) derived from the [Databricks Dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) and the [Anthropic Helpful and Harmless (HH-RLHF)](https://huggingface.co/datasets/Anthropic/hh-rlhf) datasets. It is also trained on [Competition Math](https://huggingface.co/datasets/competition_math), [Duorc](https://huggingface.co/datasets/duorc), [CoT GSM8k](https://huggingface.co/datasets/conceptofmind/cot_submix_original), [Qasper](https://huggingface.co/datasets/allenai/qasper), [Quality](https://huggingface.co/datasets/emozilla/quality), [Summ Screen FD](https://huggingface.co/datasets/tau/scrolls) and [Spider](https://huggingface.co/datasets/spider).
24
  * License: _CC-By-SA-3.0_
@@ -232,4 +240,4 @@ for open-source foundation models},
232
  note = {Accessed: 2023-06-22},
233
  urldate = {2023-06-22}
234
  }
235
- ```
 
19
 
20
  # MPT-30B-Instruct
21
 
22
+ This is the MPT-30B-Instruct but with added support to finetune using peft (tested with qlora). It is not finetuned further, the weights are the same as the original MPT-30B-Instruct.
23
+
24
+ I have not traced through the whole huggingface stack to see if this is working correctly but it does finetune with qlora and outputs are reasonable.
25
+ Inspired by implementations here https://huggingface.co/cekal/mpt-7b-peft-compatible/commits/main
26
+ https://huggingface.co/mosaicml/mpt-7b/discussions/42.
27
+
28
+ The original description for MosaicML team below:
29
+
30
  MPT-30B-Instruct is a model for short-form instruction following.
31
  It is built by finetuning [MPT-30B](https://huggingface.co/mosaicml/mpt-30b) on [Dolly HHRLHF](https://huggingface.co/datasets/mosaicml/dolly_hhrlhf) derived from the [Databricks Dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) and the [Anthropic Helpful and Harmless (HH-RLHF)](https://huggingface.co/datasets/Anthropic/hh-rlhf) datasets. It is also trained on [Competition Math](https://huggingface.co/datasets/competition_math), [Duorc](https://huggingface.co/datasets/duorc), [CoT GSM8k](https://huggingface.co/datasets/conceptofmind/cot_submix_original), [Qasper](https://huggingface.co/datasets/allenai/qasper), [Quality](https://huggingface.co/datasets/emozilla/quality), [Summ Screen FD](https://huggingface.co/datasets/tau/scrolls) and [Spider](https://huggingface.co/datasets/spider).
32
  * License: _CC-By-SA-3.0_
 
240
  note = {Accessed: 2023-06-22},
241
  urldate = {2023-06-22}
242
  }
243
+ ```