Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -99,14 +99,16 @@ tokenizer.pad_token = tokenizer.unk_token
|
|
99 |
def predict_event(image):
|
100 |
|
101 |
|
102 |
-
img = Image.open(image).convert("RGB")
|
103 |
|
104 |
-
generated_caption = tokenizer.decode(model.generate(feature_extractor(
|
105 |
|
106 |
# caption_vitgpt = generate_caption(model, image)
|
107 |
#caption_vitgpt = generate_caption(feature_extractor, cap_model, image, tokenizer)
|
|
|
|
|
108 |
|
109 |
-
return '\033[96m' +generated_caption[:85]+ '\033[0m'
|
110 |
|
111 |
|
112 |
|
@@ -125,7 +127,7 @@ description = "final capstone"
|
|
125 |
# inputs=gr.inputs.Image(type="pil")
|
126 |
|
127 |
iface = gr.Interface(predict_event,
|
128 |
-
inputs=["
|
129 |
# gr.Image(type="pil"),
|
130 |
outputs=["text"] )
|
131 |
|
|
|
99 |
def predict_event(image):
|
100 |
|
101 |
|
102 |
+
# img = Image.open(image).convert("RGB")
|
103 |
|
104 |
+
generated_caption = tokenizer.decode(model.generate(feature_extractor(image, return_tensors="pt").pixel_values.to("cuda"))[0])
|
105 |
|
106 |
# caption_vitgpt = generate_caption(model, image)
|
107 |
#caption_vitgpt = generate_caption(feature_extractor, cap_model, image, tokenizer)
|
108 |
+
preds = [pred.strip() for pred in generated_caption]
|
109 |
+
return preds
|
110 |
|
111 |
+
#return '\033[96m' +generated_caption[:85]+ '\033[0m'
|
112 |
|
113 |
|
114 |
|
|
|
127 |
# inputs=gr.inputs.Image(type="pil")
|
128 |
|
129 |
iface = gr.Interface(predict_event,
|
130 |
+
inputs=["image"],
|
131 |
# gr.Image(type="pil"),
|
132 |
outputs=["text"] )
|
133 |
|