Update README.md
Browse files
README.md
CHANGED
@@ -13,7 +13,9 @@ There seems to have been some sort of problem with the training that I cannot id
|
|
13 |
|
14 |
Typically, the model would response with long responses when asked, be much more contextually intelligent, and answer in a thoughtful way. However, for whatever reason - likely something to do with not training with LLM-Foundry - the model does not like longer responses, and typical responds quite breifly.
|
15 |
|
16 |
-
I don't believe this is a base model issue, as I compared this fine tune with MPT-7B Instruct model, and it had no problem at all producing extremely long responses, etc. If anyone has the time to investigate, please follow up with me in the community tab or on Twitter, @Teknium1!
|
|
|
|
|
17 |
|
18 |
You should load the model and tokenizer like so:
|
19 |
|
|
|
13 |
|
14 |
Typically, the model would response with long responses when asked, be much more contextually intelligent, and answer in a thoughtful way. However, for whatever reason - likely something to do with not training with LLM-Foundry - the model does not like longer responses, and typical responds quite breifly.
|
15 |
|
16 |
+
I don't believe this is a base model issue, or at least, I believe it is a base model issue related to it and the trainer, as I compared this fine tune with MPT-7B Instruct model, and it had no problem at all producing extremely long responses, etc. If anyone has the time to investigate, please follow up with me in the community tab or on Twitter, @Teknium1!
|
17 |
+
|
18 |
+
I trained Replit 3b with the same trainer, same settings, and it's results were phenomenal. So I would love any hypothesis on what may have made this different.
|
19 |
|
20 |
You should load the model and tokenizer like so:
|
21 |
|