Edit model card

Making AI Models Faster, Smaller, and Cleaner! 🥝

Contact me if you have suggestions for models to compress next or if you want to compress your own models!

Reduction: We reduced the Aya's size from 11 GB to 7GB, as well as reducing inference cost and gpu usage.

How does the compression work?
The model is compressed using Quanto to 8 bits.

How does the model quality change?
Quality might vary compared to the base model.

How is the model efficiency evaluated?
Results are based on testing with specific hardware and configurations. Efficiency may vary with different settings.

What is the model format?
We use safetensors.

What calibration data was used?
Calibration data includes WikiText, if needed by the compression method.

How to compress your own models?
Request access for more compression methods and support: Request Access

About

Aya 23 is an open weights research release of an instruction fine-tuned model with highly advanced multilingual capabilities. Aya 23 focuses on pairing a highly performant pre-trained Command family of models with the recently released Aya Collection. The result is a powerful multilingual large language model serving 23 languages.

This model card corresponds to the 8-billion version of the Aya 23 model. We also released a 35-billion version which you can find here.

We cover 23 languages: Arabic, Chinese (simplified & traditional), Czech, Dutch, English, French, German, Greek, Hebrew, Hindi, Indonesian, Italian, Japanese, Korean, Persian, Polish, Portuguese, Romanian, Russian, Spanish, Turkish, Ukrainian, and Vietnamese

Developed by: Cohere For AI and Cohere

Usage

Please install transformers from the source repository that includes the necessary changes for this model

# pip install transformers==4.41.1
from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "CohereForAI/aya-23-8B"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

# Format message with the command-r-plus chat template
messages = [{"role": "user", "content": "Anneme onu ne kadar sevdiğimi anlatan bir mektup yaz"}]
input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
## <BOS_TOKEN><|START_OF_TURN_TOKEN|><|USER_TOKEN|>Anneme onu ne kadar sevdiğimi anlatan bir mektup yaz<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>

gen_tokens = model.generate(
    input_ids, 
    max_new_tokens=100, 
    do_sample=True, 
    temperature=0.3,
    )

gen_text = tokenizer.decode(gen_tokens[0])
print(gen_text)

Configurations

Configuration details are in config.json.

Credits & License

The license of the model follows the original model’s license. Check the license of the base model for details.

Want to suggest or compress other models?
Send me a message!

Downloads last month
35
Safetensors
Model size
4.65B params
Tensor type
F32
·
U8
·
Inference Examples
Inference API (serverless) has been turned off for this model.