File size: 1,011 Bytes
83c8692
73f9d8e
 
 
83c8692
73f9d8e
 
 
 
 
 
83c8692
73f9d8e
 
 
 
 
77d2728
73f9d8e
8239866
 
 
 
 
 
 
 
 
 
5103802
8239866
d389314
 
28a2b2b
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
---
base_model: opencsg/opencsg-starcoder2-15b-v0.1
pipeline_tag: text-generation
inference: false
license: bigcode-openrail-m
model_creator: OpenCSG
model_name: opencsg-starcoder2-15b-v0.1
model_type: starcoder2
tags:
- code
quantized_by: arzeth
---
# Model Info

- Model creator: [OpenCSG](https://huggingface.co/opencsg)
- Original card (has more info): https://huggingface.co/opencsg/opencsg-starcoder2-15b-v0.1

| Layers | Context | Template |
| --- | --- | --- |
| <pre>40</pre> | <pre>16384</pre> | I think it's Alpaca. ChatML seems to work too but answers are probably worse? |

Below is Alpaca template. I think there should be a new line (`\n`) after `### Response:`. Sysprompt which is the first line can be changed of course:
```
You are an exceptionally intelligent coding assistant that consistently delivers accurate and reliable responses to user instructions.

### Instruction:
{instruction}

### Response:

```

# Quantization info
Without imatrix. Quantized with llama.cpp b2333 (2024-03-04).