mirror of
				https://github.com/Wan-Video/Wan2.1.git
				synced 2025-11-04 06:15:17 +00:00 
			
		
		
		
	Update text2video.py to reduce GPU memory by emptying cache (#44)
* Update text2video.py to reduce GPU memory by emptying cache If offload_model is set, empty_cache() must be called after the model is moved to CPU to actually free the GPU. I verified on a RTX 4090 that without calling empty_cache the model remains in memory and the subsequent vae decoding never finishes. * Update text2video.py only one empty_cache needed before vae decode
This commit is contained in:
		
							parent
							
								
									9ab8f963c8
								
							
						
					
					
						commit
						fb6dbad54c
					
				@ -252,6 +252,7 @@ class WanT2V:
 | 
			
		||||
            x0 = latents
 | 
			
		||||
            if offload_model:
 | 
			
		||||
                self.model.cpu()
 | 
			
		||||
                torch.cuda.empty_cache()
 | 
			
		||||
            if self.rank == 0:
 | 
			
		||||
                videos = self.vae.decode(x0)
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
		Loading…
	
		Reference in New Issue
	
	Block a user