import gradio as gr | |
from deepseek_vl import DeepSeekVL | |
# Initialize DeepSeek-VL model (CPU for free Spaces) | |
model = DeepSeekVL(model_path="deepseek-ai/deepseek-vl-7b", device="cpu") | |
def qa(image, question): | |
# Run DeepSeek-VL inference: image + question -> answer | |
return model.generate(image, question) | |
demo = gr.Interface( | |
fn=qa, | |
inputs=[ | |
gr.Image(type="pil", label="Upload Image"), | |
gr.Textbox(label="Enter your question") | |
], | |
outputs="text", | |
title="DeepSeek-VL Multimodal QA Demo", | |
description="Upload an image and enter a question. Experience DeepSeek-VL's vision-language capabilities." | |
) | |
if __name__ == "__main__": | |
demo.launch() |