svjack commited on
Commit
6dbab84
·
verified ·
1 Parent(s): bcb05c1

Upload stage2_512.yaml

Browse files
Files changed (1) hide show
  1. configs/unet/stage2_512.yaml +99 -0
configs/unet/stage2_512.yaml ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ syncnet_config_path: configs/syncnet/syncnet_16_pixel_attn.yaml
3
+ train_output_dir: debug/unet
4
+ train_fileslist: /mnt/bn/maliva-gen-ai-v2/chunyu.li/fileslist/data_v10_core.txt
5
+ train_data_dir: ""
6
+ audio_embeds_cache_dir: /mnt/bn/maliva-gen-ai-v2/chunyu.li/audio_cache/embeds
7
+ audio_mel_cache_dir: /mnt/bn/maliva-gen-ai-v2/chunyu.li/audio_cache/mel
8
+
9
+ val_video_path: assets/demo1_video.mp4
10
+ val_audio_path: assets/demo1_audio.wav
11
+ batch_size: 1 # 4
12
+ num_workers: 12 # 12
13
+ num_frames: 16
14
+ resolution: 512
15
+ mask_image_path: latentsync/utils/mask.png
16
+ audio_sample_rate: 16000
17
+ video_fps: 25
18
+ audio_feat_length: [2, 2]
19
+
20
+ ckpt:
21
+ resume_ckpt_path: checkpoints/latentsync_unet.pt
22
+ save_ckpt_steps: 10000
23
+
24
+ run:
25
+ pixel_space_supervise: true
26
+ use_syncnet: true
27
+ sync_loss_weight: 0.05
28
+ perceptual_loss_weight: 0.1 # 0.1
29
+ recon_loss_weight: 1 # 1
30
+ guidance_scale: 2.0 # [1.0 - 3.0]
31
+ trepa_loss_weight: 10
32
+ inference_steps: 20
33
+ trainable_modules:
34
+ - motion_modules.
35
+ - attentions.
36
+ seed: 1247
37
+ use_mixed_noise: true
38
+ mixed_noise_alpha: 1 # 1
39
+ mixed_precision_training: true
40
+ enable_gradient_checkpointing: true
41
+ max_train_steps: 10000000
42
+ max_train_epochs: -1
43
+
44
+ optimizer:
45
+ lr: 1e-5
46
+ scale_lr: false
47
+ max_grad_norm: 1.0
48
+ lr_scheduler: constant
49
+ lr_warmup_steps: 0
50
+
51
+ model:
52
+ act_fn: silu
53
+ add_audio_layer: true
54
+ attention_head_dim: 8
55
+ block_out_channels: [320, 640, 1280, 1280]
56
+ center_input_sample: false
57
+ cross_attention_dim: 384
58
+ down_block_types:
59
+ [
60
+ "CrossAttnDownBlock3D",
61
+ "CrossAttnDownBlock3D",
62
+ "CrossAttnDownBlock3D",
63
+ "DownBlock3D",
64
+ ]
65
+ mid_block_type: UNetMidBlock3DCrossAttn
66
+ up_block_types:
67
+ [
68
+ "UpBlock3D",
69
+ "CrossAttnUpBlock3D",
70
+ "CrossAttnUpBlock3D",
71
+ "CrossAttnUpBlock3D",
72
+ ]
73
+ downsample_padding: 1
74
+ flip_sin_to_cos: true
75
+ freq_shift: 0
76
+ in_channels: 13 # 49
77
+ layers_per_block: 2
78
+ mid_block_scale_factor: 1
79
+ norm_eps: 1e-5
80
+ norm_num_groups: 32
81
+ out_channels: 4 # 16
82
+ sample_size: 64
83
+ resnet_time_scale_shift: default # Choose between [default, scale_shift]
84
+
85
+ use_motion_module: true
86
+ motion_module_resolutions: [1, 2, 4, 8]
87
+ motion_module_mid_block: false
88
+ motion_module_decoder_only: false
89
+ motion_module_type: Vanilla
90
+ motion_module_kwargs:
91
+ num_attention_heads: 8
92
+ num_transformer_block: 1
93
+ attention_block_types:
94
+ - Temporal_Self
95
+ - Temporal_Self
96
+ temporal_position_encoding: true
97
+ temporal_position_encoding_max_len: 24
98
+ temporal_attention_dim_div: 1
99
+ zero_initialize: true