Spaces:
				
			
			
	
			
			
		Runtime error
		
	
	
	
			
			
	
	
	
	
		
		
		Runtime error
		
	
		unknown
		
	commited on
		
		
					Commit 
							
							·
						
						1197691
	
1
								Parent(s):
							
							91d0a5f
								
move to gpu in foley
Browse files
    	
        app.py
    CHANGED
    
    | @@ -120,11 +120,7 @@ class FoleyController: | |
| 120 |  | 
| 121 | 
             
                    self.pipeline.load_ip_adapter(fc_ckpt, subfolder='semantic', weight_name='semantic_adapter.bin', image_encoder_folder=None)
         | 
| 122 |  | 
| 123 | 
            -
                     | 
| 124 | 
            -
                    self.time_detector.to(self.device)
         | 
| 125 | 
            -
                    self.pipeline.to(self.device)
         | 
| 126 | 
            -
                    self.vocoder.to(self.device)
         | 
| 127 | 
            -
                    self.image_encoder.to(self.device)
         | 
| 128 |  | 
| 129 | 
             
                    gr.Info("Load Finish!")
         | 
| 130 | 
             
                    print("Load Finish!")
         | 
| @@ -145,14 +141,20 @@ class FoleyController: | |
| 145 | 
             
                    cfg_scale_slider,
         | 
| 146 | 
             
                    seed_textbox, 
         | 
| 147 | 
             
                ): 
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 148 | 
             
                    vision_transform_list = [
         | 
| 149 | 
             
                        torchvision.transforms.Resize((128, 128)),
         | 
| 150 | 
             
                        torchvision.transforms.CenterCrop((112, 112)),
         | 
| 151 | 
             
                        torchvision.transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
         | 
| 152 | 
             
                    ]
         | 
| 153 | 
             
                    video_transform = torchvision.transforms.Compose(vision_transform_list)
         | 
| 154 | 
            -
                     | 
| 155 | 
            -
             | 
| 156 | 
             
                    generator  = torch.Generator()
         | 
| 157 | 
             
                    if seed_textbox != "":
         | 
| 158 | 
             
                        torch.manual_seed(int(seed_textbox))
         | 
|  | |
| 120 |  | 
| 121 | 
             
                    self.pipeline.load_ip_adapter(fc_ckpt, subfolder='semantic', weight_name='semantic_adapter.bin', image_encoder_folder=None)
         | 
| 122 |  | 
| 123 | 
            +
                    self.move_to_device()
         | 
|  | |
|  | |
|  | |
|  | |
| 124 |  | 
| 125 | 
             
                    gr.Info("Load Finish!")
         | 
| 126 | 
             
                    print("Load Finish!")
         | 
|  | |
| 141 | 
             
                    cfg_scale_slider,
         | 
| 142 | 
             
                    seed_textbox, 
         | 
| 143 | 
             
                ): 
         | 
| 144 | 
            +
                    # move to gpu
         | 
| 145 | 
            +
                    self.time_detector.to(self.device)
         | 
| 146 | 
            +
                    self.pipeline.to(self.device)
         | 
| 147 | 
            +
                    self.vocoder.to(self.device)
         | 
| 148 | 
            +
                    self.image_encoder.to(self.device)
         | 
| 149 | 
            +
             | 
| 150 | 
             
                    vision_transform_list = [
         | 
| 151 | 
             
                        torchvision.transforms.Resize((128, 128)),
         | 
| 152 | 
             
                        torchvision.transforms.CenterCrop((112, 112)),
         | 
| 153 | 
             
                        torchvision.transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
         | 
| 154 | 
             
                    ]
         | 
| 155 | 
             
                    video_transform = torchvision.transforms.Compose(vision_transform_list)
         | 
| 156 | 
            +
                    if not self.loaded:
         | 
| 157 | 
            +
                        raise gr.Error("Error with loading model")
         | 
| 158 | 
             
                    generator  = torch.Generator()
         | 
| 159 | 
             
                    if seed_textbox != "":
         | 
| 160 | 
             
                        torch.manual_seed(int(seed_textbox))
         | 
