Update README.md
Browse filesadding todays update :
README.md
CHANGED
@@ -31,10 +31,25 @@ pipeline_tag: text-generation
|
|
31 |
|
32 |
|
33 |
This is based on the Quiet Star Reasoning Project : which was abandoned earlier in the year :)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
|
35 |
-
Current update : UNDER TEST ! Currently loading in unsloth but there still is the cos-cache issue so the model will load as mistral :
|
36 |
-
to load as talking heads you will need to copy modelling.py/configuration.py to the mistral folder of the transformers library and compile the latest then it loads fine but as remote code it has this issue !
|
37 |
-
i will FIX IT !!!~
|
38 |
# Introduction :
|
39 |
|
40 |
## STAR REASONERS !
|
|
|
31 |
|
32 |
|
33 |
This is based on the Quiet Star Reasoning Project : which was abandoned earlier in the year :)
|
34 |
+
Current Update :
|
35 |
+
This model is working , but actually untrained : to load the model it requires trust-remote=TRUE::
|
36 |
+
But also if it does not load then you need to clone the github:
|
37 |
+
|
38 |
+
|
39 |
+
```
|
40 |
+
! git clone https://github.com/huggingface/transformers.git
|
41 |
+
## copy modeling_mistral.py and configuartion.py to the Transformers foler / Src/models/mistral and overwrite the existing files first:
|
42 |
+
## THEN :
|
43 |
+
!cd transformers
|
44 |
+
!pip install ./transformers
|
45 |
+
|
46 |
+
```
|
47 |
+
|
48 |
+
then restaet the environment: the model can then load without trust-remote and WILL work FINE !
|
49 |
+
it can even be trained : hence the 4 bit optimised version ::
|
50 |
+
|
51 |
+
|
52 |
|
|
|
|
|
|
|
53 |
# Introduction :
|
54 |
|
55 |
## STAR REASONERS !
|