Edit model card

Model Card

Fine tuned EleutherAI/pythia-410m using gokaygokay/prompt_description_stable_diffusion_3k dataset.

Direct Use

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "gokaygokay/phytia410m_desctoprompt"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Your description
test_description = """
View to a rustic terrace filled with pots with autumn flowers and a vine full of red leaves and bunches of grapes. 
in the foreground a wooden table with a copious breakfast, coffee, bowls, vases and plates with fruits, nuts, chestnuts, hazelnuts, breads and buns.
"""

prompt_template = """### Description:
{description}

### Prompt:
"""

text = prompt_template.format(description=test_description)

def inference(text, model, tokenizer, max_input_tokens=1000, max_output_tokens=200):
  # Tokenize
    input_ids = tokenizer.encode(
          text,
          return_tensors="pt",
          truncation=True,
          max_length=max_input_tokens
    )

    # Generate
    device = model.device
    generated_tokens_with_prompt = model.generate(
    input_ids=input_ids.to(device),
    max_length=max_output_tokens,
    )

    # Decode
    generated_text_with_prompt = tokenizer.batch_decode(generated_tokens_with_prompt, skip_special_tokens=True)

    # Strip the prompt
    generated_text_answer = generated_text_with_prompt[0][len(text):]

    return generated_text_answer


print("Description input (test):", text)

print("Finetuned model's prompt: ")
print(inference(text, model, tokenizer))
Downloads last month
16
Safetensors
Model size
405M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train gokaygokay/phytia410m_desctoprompt