File size: 1,321 Bytes
c2c28c2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5ac2048
c2c28c2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
license: other
tags:
- llama
- pytorch
- chatbot
- storywriting
- generalist-model
---

# chronos-13b-v2

This is the FP16 PyTorch / HF version of **chronos-13b-v2** based on the **LLaMA v2** model.

Only use this version for further quantization or if you would like to run in full precision, as long as you have the VRAM required.

This model is primarily focused on chat, roleplay, storywriting, with good reasoning and logic.

Chronos can generate very long outputs with coherent text, largely due to the human inputs it was trained on, and it supports context length up to 4096 tokens.

This model uses Alpaca formatting, so for optimal model performance, use and either use a frontend like SillyTavern WITH instruction mode enabled, or continue your story with it:
```
### Instruction:
Your instruction or question here.
### Response:
```
Not using the format will make the model perform significantly worse than intended.

## Other Versions
[4bit GPTQ Quantized version](https://huggingface.co/elinas/chronos-13b-v2-GPTQ)

[GGML Versions provided by @TheBloke]()

<!--**Support My Development of New Models**
<a href='https://ko-fi.com/Q5Q6MB734' target='_blank'><img height='36' style='border:0px;height:36px;' 
src='https://storage.ko-fi.com/cdn/kofi1.png?v=3' border='0' alt='Support Development' /></a>-->