Edit model card

DE-LM-7B

DE-LM-7B is a 7.04 billion parameter decoder-only text generation model, released under the Apache 2.0 license. This is an instruction tuned model built on top of Deci/DeciLM-7B fine-tuned for data filtering and API generation.

Model Description

  • Language(s) (NLP): English
  • License: Apache 2.0

Model Architecture

Parameters Layers Heads Sequence Length GQA num_key_value_heads*
7.04 billion 32 32 8192 Variable

Uses

The model is intended for commercial and research use in English and can be fine-tuned for various tasks and languages.

How to Get Started with the Model

Use the code below to get started with the model.

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "dataequity/DE-LM-7B"
device = "cuda" # for GPU usage or "cpu" for CPU usage

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype="auto", trust_remote_code=True).to(device)

inputs = tokenizer.encode("List the top 10 financial APIs", return_tensors="pt").to(device)
outputs = model.generate(inputs, max_new_tokens=100, do_sample=True, top_p=0.95)
print(tokenizer.decode(outputs[0]))

# The model can also be used via the text-generation pipeline interface
from transformers import pipeline
generator = pipeline("text-generation", "dataequity/DE-LM-7B", torch_dtype="auto", trust_remote_code=True, device=device)
outputs = generator("List the top 10 financial APIs", max_new_tokens=100, do_sample=True, top_p=0.95)
print(outputs[0]["generated_text"])

Ethical Considerations and Limitations

DE-LM-7B is a new technology that comes with inherent risks associated with its use. The testing conducted so far has been primarily in English and does not encompass all possible scenarios. Like those of all large language models, DE-LM-7B's outputs are unpredictable, and the model may generate responses that are inaccurate, biased, or otherwise objectionable. Consequently, developers planning to use DE-LM-7B should undertake thorough safety testing and tuning designed explicitly for their intended applications of the model before deployment.

Citation

@misc{DeciFoundationModels,
title = {DeciLM-7B},
author = {DeciAI Research Team},
year = {2023}
url={https://huggingface.co/Deci/DeciLM-7B},
}
Downloads last month
3
Safetensors
Model size
7.04B params
Tensor type
F32
·
Inference Examples
Inference API (serverless) does not yet support model repos that contain custom code.