|
--- |
|
datasets: |
|
- Yukang/LongAlpaca-16k-length |
|
language: |
|
- en |
|
pipeline_tag: text-generation |
|
tags: |
|
- facebook |
|
- meta |
|
- llama |
|
- llama-3 |
|
- GGUF |
|
license: other |
|
license_name: llama3 |
|
license_link: LICENSE |
|
--- |
|
# GGUFs of Llama-3-8B-16K |
|
GGUF conversion and quantization of https://huggingface.co/mattshumer/Llama-3-8B-16K |
|
|
|
Done with Maxime Labonne's AutoGGUF |
|
|
|
## Orginal model card |
|
This is an extended (16K) context version of LLaMA 3 8B (base, not instruct). Trained for five hours on 8x A6000 GPUs, using the `Yukang/LongAlpaca-16k-length` dataset. |
|
|
|
`rope_theta` was set to `1000000.0`. Trained with Axolotl. |