|
--- |
|
base_model: opencsg/opencsg-starcoder2-15b-v0.1 |
|
pipeline_tag: text-generation |
|
inference: false |
|
license: bigcode-openrail-m |
|
model_creator: OpenCSG |
|
model_name: opencsg-starcoder2-15b-v0.1 |
|
model_type: starcoder2 |
|
tags: |
|
- code |
|
quantized_by: arzeth |
|
--- |
|
# Model Info |
|
|
|
- Model creator: [OpenCSG](https://huggingface.co/opencsg) |
|
- Original card (has more info): https://huggingface.co/opencsg/opencsg-starcoder2-15b-v0.1 |
|
|
|
| Layers | Context | Template | |
|
| --- | --- | --- | |
|
| <pre>40</pre> | <pre>16384</pre> | I think it's Alpaca. ChatML seems to work too but answers are probably worse? | |
|
|
|
Below is Alpaca template. I think there should be a new line (`\n`) after `### Response:`. Sysprompt which is the first line can be changed of course: |
|
``` |
|
You are an exceptionally intelligent coding assistant that consistently delivers accurate and reliable responses to user instructions. |
|
|
|
### Instruction: |
|
{instruction} |
|
|
|
### Response: |
|
|
|
``` |
|
|
|
# Quantization info |
|
Without imatrix. Quantized with llama.cpp b2333 (2024-03-04). |