Spaces:
Running
Running
from PIL import Image | |
from transformers import BlipProcessor , BlipForConditionalGeneration | |
import torch | |
processor = BlipProcessor.from_pretrained("src/models/Caption") | |
model = BlipForConditionalGeneration.from_pretrained("src/models/Caption") | |
def generateCaption(image_path): | |
image = Image.open(image_path).convert("RGB") | |
inputs = processor(images = image , return_tensors="pt") | |
output = model.generate(**inputs) | |
caption = processor.decode(output[0], skip_special_tokens = True) | |
return caption |