---
base_model: Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit
datasets:
- mahiatlinux/Reflection-Dataset-v2
- Harshkmr/orca-math-word-reflection
language:
- en
license: llama3.1
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
- sft
- reflection
- llama-cpp
- gguf-my-repo
---
# Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF
This model was converted to GGUF format from [`Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit`](https://huggingface.co/Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
Refer to the [original model card](https://huggingface.co/Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit) for more details on the model.
**Not Officially Benchmarked Yet!** (Please submit any benchmarking or eval results via the Community tab.)
# Uploaded model
- **Developed by:** Solshine (Caleb DeLeeuw)
- **License:** LLama 3.1 License
- **Finetuned from model :** Solshine/reflection-llama-3.1-8B-Solshine-trainround2-16bit
Inspired by and featuring the Reflection Tuning technique pioneered by Matt Shumer (possibly earlier innovated by the team at Anthropic, and Mlabbone's Hermes.)
*To the authors' knowledge, this is V3 of the first "reflection tuned" Llama 3.1 8B LLM*
**As per the inspiring model "mattshumer/Reflection-Llama-3.1-70B" (this model was not used in the training process nor as a foundational model, but only served as inspiration) :**
```
During sampling, the model will start by outputting reasoning inside and tags, and then once it is satisfied with its reasoning, it will output the final answer inside tags. Each of these tags are special tokens, trained into the model.
This enables the model to separate its internal thoughts and reasoning from its final answer, improving the experience for the user.
Inside the section, the model may output one or more tags, which signals the model has caught an error in its reasoning and will attempt to correct it before providing a final answer.
System Prompt:
The system prompt used for training this model is:
You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside tags, and then provide your final response inside