tokenspace / openclip /basic.py
ppbrown's picture
Create openclip/basic.py
9428212 verified
raw
history blame
596 Bytes
#basic openclip usage
import torch
from PIL import Image
import open_clip
mtype='ViT-B-32'
mname='laion2b_s34b_b79k'
print("Loading",mtype,mname)
model, _, preprocess = open_clip.create_model_and_transforms(mtype,
pretrained=mname)
tokenizer = open_clip.get_tokenizer(mtype)
#image = preprocess(Image.open("CLIP.png")).unsqueeze(0)
text = tokenizer(["a diagram", "a dog", "a cat"])
text = tokenizer("cat")
with torch.no_grad(), torch.cuda.amp.autocast():
# image_features = model.encode_image(image)
text_features = model.encode_text(text)
embedding=text_features[0]