FantasiaFoundry commited on
Commit
853f72f
1 Parent(s): 46682f7

llama.cpp/#6920 warning

Browse files
Files changed (1) hide show
  1. README.md +9 -4
README.md CHANGED
@@ -9,13 +9,18 @@ tags:
9
  ---
10
 
11
  > [!TIP]
12
- > **Credits:** <br>
 
13
  > Made with love by [**@Lewdiculous**](https://huggingface.co/Lewdiculous). <br>
14
- > *If this proves useful for you, feel free to credit and share the repository and authors.*
15
 
16
  > [!WARNING]
17
- > **Warning:** <br>
18
- > For **Llama-3** models that don't follow the ChatML, Alpaca, Vicuna and other conventional formats, at the moment, you have to use `gguf-imat-llama-3.py` and replace the config files with the ones in the [**ChaoticNeutrals/Llama3-Corrections**](https://huggingface.co/ChaoticNeutrals/Llama3-Corrections/tree/main) repository to properly quant and generate the imatrix data.
 
 
 
 
19
 
20
  Pull Requests with your own features and improvements to this script are always welcome.
21
 
 
9
  ---
10
 
11
  > [!TIP]
12
+ > **Credits:**
13
+ >
14
  > Made with love by [**@Lewdiculous**](https://huggingface.co/Lewdiculous). <br>
15
+ > If this proves useful for you, feel free to credit and share the repository and authors.
16
 
17
  > [!WARNING]
18
+ > **[Important] Llama-3:**
19
+ >
20
+ > For those converting LLama-3 BPE models, you'll have to read [**llama.cpp/#6920**](https://github.com/ggerganov/llama.cpp/pull/6920#issue-2265280504) for more context. <br>
21
+ > Make sure you're in the latest llama.cpp repo commit, then run the new `convert-hf-to-gguf-update.py` script inside the repo, afterwards you need to manually copy the config files from `llama.cpp\models\tokenizers\llama-bpe` into your downloaded **model** folder, replacing the existing ones. <br>
22
+ > Try again and the conversion procress should work as expected.
23
+
24
 
25
  Pull Requests with your own features and improvements to this script are always welcome.
26