mirror of
https://github.com/Wan-Video/Wan2.1.git
synced 2025-11-04 14:16:57 +00:00
oops
This commit is contained in:
parent
b00354374c
commit
b5d26646c5
@ -21,11 +21,11 @@ WanGP supports the Wan (and derived models), Hunyuan Video and LTV Video models
|
|||||||
|
|
||||||
## 🔥 Latest Updates
|
## 🔥 Latest Updates
|
||||||
### July 8 2025: WanGP v6.6, WanGP offers you **Vace Multitalk Dual Voices Fusionix Infinite** :
|
### July 8 2025: WanGP v6.6, WanGP offers you **Vace Multitalk Dual Voices Fusionix Infinite** :
|
||||||
**Vace** our beloved super Control Net has been combined with **Multitalk** the new king in town that can animate two people speaking (**Dual Voices**). It is accelerated by the **Fusionix** model and thanks to *Sliding Windows* support and *Adaptive Projected Guidance* (much slower but should reduce the reddish effect with long videos) your two people will be able to talk for very a long time (which is an **Infinite** amount of time in the field of video generation).
|
**Vace** our beloved super Control Net has been combined with **Multitalk** the new king in town that can animate up to two people speaking (**Dual Voices**). It is accelerated by the **Fusionix** model and thanks to *Sliding Windows* support and *Adaptive Projected Guidance* (much slower but should reduce the reddish effect with long videos) your two people will be able to talk for very a long time (which is an **Infinite** amount of time in the field of video generation).
|
||||||
|
|
||||||
Of course you will get as well *Multitalk* vanilla and also *Multitalk 720p* as a bonus.
|
Of course you will get as well *Multitalk* vanilla and also *Multitalk 720p* as a bonus.
|
||||||
|
|
||||||
And since I am mister nice guy I had enclosed as an exclusivity an *Audio Separator* that will save you time to isolate each voice when using Multitalk with two people.
|
And since I am mister nice guy I have enclosed as an exclusivity an *Audio Separator* that will save you time to isolate each voice when using Multitalk with two people.
|
||||||
|
|
||||||
As I feel like a resting a bit I haven't produced a nice sample Video to illustrate all these new capabilities. But here is the thing, I ams sure you will publish in the *Share Your Best Video* channel your *Master Pieces*. The best one will be added to the *Announcements Channel* and will bring eternal fame to its author.
|
As I feel like a resting a bit I haven't produced a nice sample Video to illustrate all these new capabilities. But here is the thing, I ams sure you will publish in the *Share Your Best Video* channel your *Master Pieces*. The best one will be added to the *Announcements Channel* and will bring eternal fame to its author.
|
||||||
|
|
||||||
|
|||||||
@ -7,23 +7,23 @@ import torch.nn.functional as F
|
|||||||
major, minor = torch.cuda.get_device_capability(None)
|
major, minor = torch.cuda.get_device_capability(None)
|
||||||
bfloat16_supported = major >= 8
|
bfloat16_supported = major >= 8
|
||||||
|
|
||||||
try:
|
# try:
|
||||||
from xformers.ops import memory_efficient_attention
|
# from xformers.ops import memory_efficient_attention
|
||||||
except ImportError:
|
# except ImportError:
|
||||||
memory_efficient_attention = None
|
# memory_efficient_attention = None
|
||||||
|
|
||||||
try:
|
# try:
|
||||||
import flash_attn_interface
|
# import flash_attn_interface
|
||||||
FLASH_ATTN_3_AVAILABLE = True
|
# FLASH_ATTN_3_AVAILABLE = True
|
||||||
except ModuleNotFoundError:
|
# except ModuleNotFoundError:
|
||||||
FLASH_ATTN_3_AVAILABLE = False
|
# FLASH_ATTN_3_AVAILABLE = False
|
||||||
|
|
||||||
try:
|
# try:
|
||||||
import flash_attn
|
# import flash_attn
|
||||||
FLASH_ATTN_2_AVAILABLE = True
|
# FLASH_ATTN_2_AVAILABLE = True
|
||||||
except ModuleNotFoundError:
|
# except ModuleNotFoundError:
|
||||||
FLASH_ATTN_2_AVAILABLE = False
|
# FLASH_ATTN_2_AVAILABLE = False
|
||||||
flash_attn = None
|
# flash_attn = None
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from sageattention import sageattn_varlen
|
from sageattention import sageattn_varlen
|
||||||
|
|||||||
@ -5,19 +5,19 @@ from einops import rearrange, repeat
|
|||||||
from .multitalk_utils import RotaryPositionalEmbedding1D, normalize_and_scale, split_token_counts_and_frame_ids
|
from .multitalk_utils import RotaryPositionalEmbedding1D, normalize_and_scale, split_token_counts_and_frame_ids
|
||||||
from wan.modules.attention import pay_attention
|
from wan.modules.attention import pay_attention
|
||||||
|
|
||||||
import xformers.ops
|
# import xformers.ops
|
||||||
|
|
||||||
try:
|
# try:
|
||||||
import flash_attn_interface
|
# import flash_attn_interface
|
||||||
FLASH_ATTN_3_AVAILABLE = True
|
# FLASH_ATTN_3_AVAILABLE = True
|
||||||
except ModuleNotFoundError:
|
# except ModuleNotFoundError:
|
||||||
FLASH_ATTN_3_AVAILABLE = False
|
# FLASH_ATTN_3_AVAILABLE = False
|
||||||
|
|
||||||
try:
|
# try:
|
||||||
import flash_attn
|
# import flash_attn
|
||||||
FLASH_ATTN_2_AVAILABLE = True
|
# FLASH_ATTN_2_AVAILABLE = True
|
||||||
except ModuleNotFoundError:
|
# except ModuleNotFoundError:
|
||||||
FLASH_ATTN_2_AVAILABLE = False
|
# FLASH_ATTN_2_AVAILABLE = False
|
||||||
|
|
||||||
import warnings
|
import warnings
|
||||||
|
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user