weiyi01191 commited on
Commit
e2b0fd0
·
1 Parent(s): e8e4446

Update test_configs/llama2_test_config.yaml

Browse files
Files changed (1) hide show
  1. test_configs/llama2_test_config.yaml +55 -15
test_configs/llama2_test_config.yaml CHANGED
@@ -2,7 +2,7 @@ model:
2
  arch: mini_gpt4_llama_v2
3
  freeze_vit: True
4
  freeze_qformer: True
5
- max_txt_len: 384
6
  low_resource: True
7
  image_size: 224
8
  end_sym: "</s>"
@@ -10,47 +10,87 @@ model:
10
  ckpt: "checkpoints/video_llama_checkpoint_last.pth"
11
  use_grad_checkpoint: True
12
  chat_template: True
13
- lora_r: 96
14
- lora_alpha: 24
15
- length: 50
16
  use_grad_checkpoint_llm: True
17
- max_context_len: 4096
18
  architectures: [
19
  "MiniGPT4_Video"
20
  ]
21
  device: "cuda"
22
- drop_path_rate: 0
23
  img_size: 224
24
  model_type: "minigpt4_video"
25
- num_query_token: 48
26
  prompt: ""
27
  torch_dtype: "float16"
28
- transformers_version: "4.42.3"
29
  vit_precision: "fp16"
30
  vit_model: "eva_clip_g"
31
  token_pooling: true
32
- lora_target_modules : ["q_proj","v_proj","k_proj","o_proj","gate_proj","up_proj","down_proj"]
33
- lora_dropout: 0.08
34
  remove_template: false
35
  prompt_path: ""
36
  minigpt4_gpu_id: 0
37
  whisper_gpu_id: 0
38
  answer_module_gpu_id: 0
39
-
40
-
41
-
 
 
 
 
 
 
 
42
 
43
  datasets:
44
  video_chatgpt: #99378 row - 13224 video
45
- batch_size: 3
 
46
  vis_processor:
47
  train:
48
  name: "blip2_image_train"
49
  image_size: 224
 
 
 
 
 
50
  text_processor:
51
  train:
52
  name: "blip_caption"
53
- sample_ratio: 200
 
 
 
 
54
  run:
55
  seed: 42
56
  amp: true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  arch: mini_gpt4_llama_v2
3
  freeze_vit: True
4
  freeze_qformer: True
5
+ max_txt_len: 512
6
  low_resource: True
7
  image_size: 224
8
  end_sym: "</s>"
 
10
  ckpt: "checkpoints/video_llama_checkpoint_last.pth"
11
  use_grad_checkpoint: True
12
  chat_template: True
13
+ lora_r: 64
14
+ lora_alpha: 16
15
+ length: 45
16
  use_grad_checkpoint_llm: True
17
+ max_context_len: 3072
18
  architectures: [
19
  "MiniGPT4_Video"
20
  ]
21
  device: "cuda"
22
+ drop_path_rate: 0.1
23
  img_size: 224
24
  model_type: "minigpt4_video"
25
+ num_query_token: 32
26
  prompt: ""
27
  torch_dtype: "float16"
28
+ transformers_version: "4.37.2"
29
  vit_precision: "fp16"
30
  vit_model: "eva_clip_g"
31
  token_pooling: true
32
+ lora_target_modules: ["q_proj", "v_proj", "k_proj", "o_proj", "gate_proj", "up_proj"]
33
+ lora_dropout: 0.05
34
  remove_template: false
35
  prompt_path: ""
36
  minigpt4_gpu_id: 0
37
  whisper_gpu_id: 0
38
  answer_module_gpu_id: 0
39
+ gradient_accumulation_steps: 1
40
+ warmup_steps: 100
41
+ save_steps: 1000
42
+ logging_steps: 50
43
+ eval_steps: 500
44
+ max_new_tokens: 512
45
+ temperature: 0.7
46
+ top_p: 0.9
47
+ do_sample: true
48
+ num_beams: 1
49
 
50
  datasets:
51
  video_chatgpt: #99378 row - 13224 video
52
+ batch_size: 2
53
+ num_workers: 2
54
  vis_processor:
55
  train:
56
  name: "blip2_image_train"
57
  image_size: 224
58
+ mean: [0.48145466, 0.4578275, 0.40821073]
59
+ std: [0.26862954, 0.26130258, 0.27577711]
60
+ eval:
61
+ name: "blip2_image_eval"
62
+ image_size: 224
63
  text_processor:
64
  train:
65
  name: "blip_caption"
66
+ max_words: 512
67
+ eval:
68
+ name: "blip_caption"
69
+ max_words: 512
70
+ sample_ratio: 100
71
  run:
72
  seed: 42
73
  amp: true
74
+ distributed: false
75
+ gpu_id: 0
76
+ world_size: 1
77
+ rank: 0
78
+ dataloader_num_workers: 2
79
+ pin_memory: true
80
+ persistent_workers: true
81
+ prefetch_factor: 2
82
+ clip_grad_norm: 1.0
83
+ weight_decay: 0.01
84
+ adam_epsilon: 1e-8
85
+ adam_beta1: 0.9
86
+ adam_beta2: 0.999
87
+
88
+ inference:
89
+ batch_size: 1
90
+ max_frames: 45
91
+ frame_interval: 2
92
+ subtitle_max_len: 400
93
+ enable_subtitles: true
94
+ whisper_model: "base"
95
+ response_format: "detailed"
96
+ include_timestamps: false