mirror of
				https://github.com/Wan-Video/Wan2.1.git
				synced 2025-11-04 06:15:17 +00:00 
			
		
		
		
	I need rest!
This commit is contained in:
		
							parent
							
								
									b5d26646c5
								
							
						
					
					
						commit
						34f9333fdc
					
				@ -7,23 +7,23 @@ import torch.nn.functional as F
 | 
			
		||||
major, minor = torch.cuda.get_device_capability(None)
 | 
			
		||||
bfloat16_supported =  major >= 8 
 | 
			
		||||
 | 
			
		||||
# try:
 | 
			
		||||
#     from xformers.ops import memory_efficient_attention
 | 
			
		||||
# except ImportError:
 | 
			
		||||
#     memory_efficient_attention = None
 | 
			
		||||
try:
 | 
			
		||||
    from xformers.ops import memory_efficient_attention
 | 
			
		||||
except ImportError:
 | 
			
		||||
    memory_efficient_attention = None
 | 
			
		||||
 | 
			
		||||
# try:
 | 
			
		||||
#     import flash_attn_interface
 | 
			
		||||
#     FLASH_ATTN_3_AVAILABLE = True
 | 
			
		||||
# except ModuleNotFoundError:
 | 
			
		||||
#     FLASH_ATTN_3_AVAILABLE = False
 | 
			
		||||
try:
 | 
			
		||||
    import flash_attn_interface
 | 
			
		||||
    FLASH_ATTN_3_AVAILABLE = True
 | 
			
		||||
except ModuleNotFoundError:
 | 
			
		||||
    FLASH_ATTN_3_AVAILABLE = False
 | 
			
		||||
 | 
			
		||||
# try:
 | 
			
		||||
#     import flash_attn
 | 
			
		||||
#     FLASH_ATTN_2_AVAILABLE = True
 | 
			
		||||
# except ModuleNotFoundError:
 | 
			
		||||
#     FLASH_ATTN_2_AVAILABLE = False
 | 
			
		||||
#     flash_attn = None
 | 
			
		||||
try:
 | 
			
		||||
    import flash_attn
 | 
			
		||||
    FLASH_ATTN_2_AVAILABLE = True
 | 
			
		||||
except ModuleNotFoundError:
 | 
			
		||||
    FLASH_ATTN_2_AVAILABLE = False
 | 
			
		||||
    flash_attn = None
 | 
			
		||||
 | 
			
		||||
try:
 | 
			
		||||
    from sageattention import sageattn_varlen
 | 
			
		||||
 | 
			
		||||
		Loading…
	
		Reference in New Issue
	
	Block a user