Edit model card

monsterapi/gemma-2-2b-norobots

Base Model for Fine-tuning: google/gemma-2-2b-it
Service Used: MonsterAPI
License: Apache-2.0

Overview

monsterapi/gemma-2-2b-norobots is a fine-tuned language model designed to improve instruction-following capabilities. The model was trained using the "No Robots" dataset, a high-quality set of 10,000 instructions and demonstrations curated by expert human annotators. This fine-tuning process enhances the base model's performance in understanding and executing single-turn instructions, similar to the goals outlined in OpenAI's InstructGPT.

Dataset Details

Dataset Summary:
The "No Robots" dataset is a collection of 10,000 high-quality instructions and demonstrations created by skilled human annotators. The dataset is modeled after the instruction dataset described in OpenAI's InstructGPT paper. It mainly includes single-turn instructions across various categories, aiming to improve the instruction-following capabilities of language models during supervised fine-tuning (SFT).

Fine-tuning Details

Fine-tuned Model Name: monsterapi/gemma-2-2b-norobots
Training Time: 31 minutes
Cost: $1.10
Epochs: 1
Gradient Accumulation Steps: 32

The model was fine-tuned using MonsterAPI's finetuning service, optimizing the base model google/gemma-2-2b-it to perform better on instruction-following tasks.

Hyperparameters & Additional Details

  • Base Model: google/gemma-2-2b-it
  • Dataset: No Robots (10,000 instructions and demonstrations)
  • Training Duration: 31 minutes
  • Cost per Epoch: $1.10
  • Total Finetuning Cost: $1.10
  • Gradient Accumulation Steps: 32

Use Cases

This model is well-suited for tasks that require improved instruction-following capabilities, such as:

  • Chatbots and virtual assistants
  • Content creation tools
  • Automated customer support systems
  • Task automation in various industries

How to Use

You can load the model directly using the Hugging Face Transformers library:

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "monsterapi/gemma-2-2b-norobots"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Example usage
input_text = "Explain the concept of supervised fine-tuning in simple terms."
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Acknowledgements

The fine-tuning process was carried out using MonsterAPI's finetuning service, which offers a seamless experience for optimizing large language models.

Contact

For further details or queries, please contact MonsterAPI or visit the official documentation.

Downloads last month
6
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for monsterapi/gemma-2-2b-norobots

Base model

google/gemma-2-2b
Adapter
(137)
this model

Dataset used to train monsterapi/gemma-2-2b-norobots