File size: 2,069 Bytes
5d38bbc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
---
library_name: transformers
license: apache-2.0
language:
- en
tags:
- MemGPT
- function
- function calling
---

This is an uncensored release of DPO version of [MemGPT](https://github.com/cpacker/MemGPT) Language Model.

# WARNING
This model is **UNCENSORED**. That means this model is highly compliant to any requests, even unethical and potentially dangerous ones. I do not take any responsibility whatsoever for any damage caused by the model in this repo.


# Model Description
This repository contains an uncensored model of [Mistral 7B Instruct](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2). This model is specifically designed for operating within function calling environment in MemGPT. It demonstrates comparable performances to GPT-4 when it comes to working with MemGPT.


# Key Features
* Function calling
* Dedicated to working with MemGPT
* Supports medium-length context, up to sequences of 8,192


# Prompt Format
This model uses **ChatML** prompt format:
```
<|im_start|>system
{system_instruction}<|im_end|>
<|im_start|>user
{user_message}<|im_end|>
<|im_start|>assistant
{assistant_response}<|im_end|>
```


# Usage
This model is designed to be ran on multiple backends, such as [oogabooga's textgen WebUI](https://github.com/oobabooga/text-generation-webui).
Simply install your preferred backend, and then load up this model.
Then, configure MemGPT using `memgpt configure`, and chat with MemGPT via `memgpt run` command!


# Model Details
* Developed by: @starsnatched
* Model type: This repo contains a language model based on the transformer decoder architecture.
* Language: English
* Contact: For any questions, concerns or comments about this model, please contact me at Discord, @starsnatched.


# Training Infrastructure
* Hardware: The model in this repo was trained on 2x A100 80GB GPUs.


# Intended Use
The model is designed to be used as the base model for MemGPT agents.


# Limitations and Risks
The model may exhibit unreliable, unsafe, or biased behaviours. Please double check the results this model may produce.