maxorange commited on
Commit
642655a
·
verified ·
1 Parent(s): 9f69fbd

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +19 -11
app.py CHANGED
@@ -57,19 +57,27 @@ def inference(raw_image, model_n, question, strategy):
57
  answer = model_vq(image_vq, question, train=False, inference='generate')
58
  return 'answer: '+answer[0]
59
 
60
- inputs = [
61
- gr.Image(type='pil', interactive=True),
62
- gr.inputs.Radio(choices=['Image Captioning',"Visual Question Answering"],
63
- type="value",
64
- default="Image Captioning",
65
- label="Task"
66
- ),gr.inputs.Textbox(lines=2, label="Question"),gr.inputs.Radio(choices=['Beam search','Nucleus sampling'], type="value", default="Nucleus sampling", label="Caption Decoding Strategy")]
67
- outputs = gr.outputs.Textbox(label="Output")
68
 
69
  title = "BLIP"
70
 
71
- description = "Gradio demo for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation (Salesforce Research). We have now disable image uploading as of March 23. 2023. Click one of the examples to load them. Read more at the links below."
72
 
73
- article = """<p style='text-align: center'><a href='https://arxiv.org/abs/2201.12086' target='_blank'>BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation</a> | <a href='https://github.com/salesforce/BLIP' target='_blank'>Github Repo</a></p>"""
74
 
75
- gr.Interface(inference, inputs, outputs, title=title, description=description, article=article, examples=[['starrynight.jpeg',"Image Captioning","None","Nucleus sampling"]]).launch(enable_queue=True)
 
 
 
 
 
 
 
 
 
 
 
 
57
  answer = model_vq(image_vq, question, train=False, inference='generate')
58
  return 'answer: '+answer[0]
59
 
60
+ inputs = [gr.Image(type='pil'),
61
+ gr.Radio(choices=['Image Captioning',"Visual Question Answering"], type="value", value="Image Captioning", label="Task"),
62
+ gr.Textbox(lines=2, label="Question"),
63
+ gr.Radio(choices=['Beam search','Nucleus sampling'], type="value", value="Nucleus sampling", label="Caption Decoding Strategy")]
64
+ outputs = gr.Textbox(label="Output")
 
 
 
65
 
66
  title = "BLIP"
67
 
68
+ description = "Gradio demo for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation (Salesforce Research). To use it, simply upload your image, or click one of the examples to load them. Read more at the links below."
69
 
70
+ article = "<p style='text-align: center'><a href='https://arxiv.org/abs/2201.12086' target='_blank'>BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation</a> | <a href='https://github.com/salesforce/BLIP' target='_blank'>Github Repo</a></p>"
71
 
72
+
73
+ demo = gr.Interface(inference,
74
+ inputs,
75
+ outputs,
76
+ title=title,
77
+ description=description,
78
+ article=article,
79
+ examples=[['starrynight.jpeg',"Image Captioning","None","Nucleus sampling"]],
80
+ allow_flagging='never',
81
+ cache_examples="lazy",
82
+ delete_cache=(4000, 4000))
83
+ demo.queue(default_concurrency_limit=1).launch(show_error=True, show_api=True, mcp_server=True)