Edit model card

Model Card for syedzaidi-kiwi/Llama-2-7b-chat-finetune

This model is a fine-tuned version of Meta's Llama 2 7B variant for enhanced chat functionalities.

This modelcard aims to be a base template for new models. It has been generated using this raw template.

Model Details

Model Description

  • Developed by: Syed Asad
  • Model type: Fine-tuned Llama 2 7B variant
  • Language(s) (NLP): English
  • License: Apache-2.0
  • Finetuned from model: NousResearch/Llama-2-7b-chat-hf

Model Sources

Uses

Direct Use

The model is intended for direct use in applications requiring conversational responses, such as chatbots or virtual assistants.

Out-of-Scope Use

The model is not designed for tasks outside of conversational AI, such as document summarization or translation.

Bias, Risks, and Limitations

Users should be aware of potential biases in the training data and limitations in the model's understanding of nuanced human language. Further evaluation is recommended for specific use cases.

How to Get Started with the Model

from transformers import AutoModelForCausalLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("syedzaidi-kiwi/Llama-2-7b-chat-finetune")
model = AutoModelForCausalLM.from_pretrained("syedzaidi-kiwi/Llama-2-7b-chat-finetune")

inputs = tokenizer("Hello, how are you?", return_tensors="pt")
response = model.generate(**inputs)
print(tokenizer.decode(response[0], skip_special_tokens=True))

Training Details

Training Data

The model was fine-tuned using the dataset mlabonne/guanaco-llama2-1k.

Link: https://huggingface.co/datasets/mlabonne/guanaco-llama2-1k

Training Procedure

Training Hyperparameters

  • Training regime:

The model was fine-tuned using a mix of precision training techniques to balance training speed and model performance effectively.

While the exact precision format (e.g., fp32, fp16, bf16) utilized depends on the compute capabilities available, an emphasis was placed on leveraging mixed precision (fp16) training to accelerate the training process on compatible hardware. This approach allowed for faster computation and reduced memory usage without significant loss in training quality.

Users are encouraged to adjust the precision settings based on their hardware specifications to optimize performance further.

Speeds, Sizes, Times

To be tested by the KiwiTech Team

Evaluation

Testing Data, Factors & Metrics

Testing Data

The model's performance was evaluated on a held-out test set from the mlabonne/guanaco-llama2-1k dataset.

This dataset comprises diverse conversational contexts to assess the model's generalization and robustness across various topics. [https://huggingface.co/datasets/mlabonne/guanaco-llama2-1k]

Factors

Evaluation focused on several key factors to ensure the model's versatility and reliability in conversational AI applications:

Context understanding: The model's ability to maintain context and coherence over long conversations. Diversity of responses: The variety in the model's responses to similar prompts, indicating its creative and dynamic conversational capabilities. Safety and bias: Monitoring for any unintended biases in responses or generation of inappropriate content.

Metrics

To comprehensively assess the model's performance, the following metrics were utilized:

Perplexity (PPL): Lower perplexity scores indicate better understanding and generation of the text. BLEU Score: For measuring the similarity between the model's generated responses and a set of reference responses, indicating the model's accuracy in reproducing human-like answers. F1 Score: Evaluating the balance between precision and recall in the model's responses, useful for assessing conversational relevance. Safety and Bias Evaluation: Custom metrics were developed to quantify the model's performance in generating safe, unbiased content.

Results

To be Evaulated, will be updated in this section.

Summary

The fine-tuned model demonstrates significant improvements in generating coherent, diverse, and contextually appropriate responses across various conversational settings.

It represents a step forward in developing conversational AI systems that are both efficient and effective.

Continuous evaluation and monitoring are advised to further enhance and maintain the model's performance standards.

Technical Specifications

Model Architecture and Objective

Transformers

Compute Infrastructure

T4 GPU

Hardware

Fine Tuned on Apple M3 Pro (Silicon Chip)

Software

Google Colab Notebook Used

Citation

OriginalLlama2Citation Title: Llama 2: Open Foundation and Fine-Tuned Chat Models}, Authors: Hugo Touvron∗ Louis Martin† Kevin Stone† Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale Dan Bikel Lukas Blecher Cristian Canton Ferrer Moya Chen Guillem Cucurull David Esiobu Jude Fernandes Jeremy Fu Wenyin Fu Brian Fuller Cynthia Gao Vedanuj Goswami Naman Goyal Anthony Hartshorn Saghar Hosseini Rui Hou Hakan Inan Marcin Kardas Viktor Kerkez Madian Khabsa Isabel Kloumann Artem Korenev Punit Singh Koura Marie-Anne Lachaux Thibaut Lavril Jenya Lee Diana Liskovich Yinghai Lu Yuning Mao Xavier Martinet Todor Mihaylov Pushkar Mishra Igor Molybog Yixin Nie Andrew Poulton Jeremy Reizenstein Rashi Rungta Kalyan Saladi Alan Schelten Ruan Silva Eric Michael Smith Ranjan Subramanian Xiaoqing Ellen Tan Binh Tang Ross Taylor Adina Williams Jian Xiang Kuan Puxin Xu Zheng Yan Iliyan Zarov Yuchen Zhang Angela Fan Melanie Kambadur Sharan Narang Aurelien Rodriguez Robert Stojnic Sergey Edunov Thomas Scialom

Journal: Gen AI, Meta Year: 2023

Link to Research Paper: https://ai.meta.com/research/publications/llama-2-open-foundation-and-fine-tuned-chat-models/

Model Card Authors

Syed Asad

Model Card Contact

Syed Asad ([email protected])

Downloads last month
22
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train syedzaidi-kiwi/Llama-2-7b-chat-finetune