BLIVA / bliva /configs /models /bliva_vicuna7b.yaml
gordonhubackup's picture
upload
e62d81d
raw
history blame contribute delete
807 Bytes
model:
arch: vicuna7b
load_finetuned: True
load_pretrained: False
pretrained: "https://storage.googleapis.com/sfr-vision-language-research/LAVIS/models/InstructBLIP/instruct_blip_vicuna7b_trimmed.pth"
finetuned: 'bliva_vicuna7b.pth'
# vit encoder
image_size: 224 #336
drop_path_rate: 0
use_grad_checkpoint: False
vit_precision: "fp16"
freeze_vit: True
# Q-Former
num_query_token: 32
# path to Vicuna checkpoint
llm_model: "hf_vicuna_7b"
# generation configs
prompt: ""
preprocess:
vis_processor:
train:
name: "blip2_image_train"
image_size: 224
eval:
name: "blip_image_eval"
image_size: 224
text_processor:
train:
name: "blip_caption"
eval:
name: "blip_caption"