mirror of
https://github.com/Wan-Video/Wan2.1.git
synced 2025-11-04 14:16:57 +00:00
fixed docs and requirements
This commit is contained in:
parent
c495c41ad7
commit
f4537e0cf2
@ -233,7 +233,7 @@ git clone https://github.com/deepbeepmeep/Wan2GP.git
|
|||||||
cd Wan2GP
|
cd Wan2GP
|
||||||
conda create -n wan2gp python=3.10.9
|
conda create -n wan2gp python=3.10.9
|
||||||
conda activate wan2gp
|
conda activate wan2gp
|
||||||
pip install torch==2.7.0 torchvision==0.22.0 torchaudio==2.7.0 --extra-index-url https://download.pytorch.org/whl/cu124
|
pip install torch==2.7.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/test/cu128
|
||||||
pip install -r requirements.txt
|
pip install -r requirements.txt
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|||||||
@ -8,7 +8,7 @@ This guide covers installation for different GPU generations and operating syste
|
|||||||
- Conda or Python venv
|
- Conda or Python venv
|
||||||
- Compatible GPU (RTX 10XX or newer recommended)
|
- Compatible GPU (RTX 10XX or newer recommended)
|
||||||
|
|
||||||
## Installation for RTX 10XX to RTX 40XX (Stable)
|
## Installation for RTX 10XX to RTX 50XX (Stable)
|
||||||
|
|
||||||
This installation uses PyTorch 2.7.0 which is well-tested and stable.
|
This installation uses PyTorch 2.7.0 which is well-tested and stable.
|
||||||
|
|
||||||
@ -28,7 +28,7 @@ conda activate wan2gp
|
|||||||
|
|
||||||
```shell
|
```shell
|
||||||
# Install PyTorch 2.7.0 with CUDA 12.4
|
# Install PyTorch 2.7.0 with CUDA 12.4
|
||||||
pip install torch==2.7.0 torchvision==0.22.0 torchaudio==2.7.0 --extra-index-url https://download.pytorch.org/whl/cu124
|
pip install torch==2.7.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/test/cu128
|
||||||
```
|
```
|
||||||
|
|
||||||
### Step 3: Install Dependencies
|
### Step 3: Install Dependencies
|
||||||
@ -40,7 +40,7 @@ pip install -r requirements.txt
|
|||||||
|
|
||||||
### Step 4: Optional Performance Optimizations
|
### Step 4: Optional Performance Optimizations
|
||||||
|
|
||||||
#### Sage Attention (30% faster)
|
#### Sage Attention (30% faster), don't install with RTX 50xx as it is not compatible
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
# Windows only: Install Triton
|
# Windows only: Install Triton
|
||||||
@ -71,62 +71,7 @@ pip install -e .
|
|||||||
pip install flash-attn==2.7.2.post1
|
pip install flash-attn==2.7.2.post1
|
||||||
```
|
```
|
||||||
|
|
||||||
## Installation for RTX 50XX (Beta)
|
|
||||||
|
|
||||||
RTX 50XX GPUs require PyTorch 2.7.0 (beta). This version may be less stable.
|
|
||||||
|
|
||||||
⚠️ **Important:** Use Python 3.10 for compatibility with pip wheels.
|
|
||||||
|
|
||||||
### Step 1: Setup Environment
|
|
||||||
|
|
||||||
```shell
|
|
||||||
# Clone and setup (same as above)
|
|
||||||
python -m pip install "setuptools<=75.8.2" --force-reinstall
|
|
||||||
git clone https://github.com/deepbeepmeep/Wan2GP.git
|
|
||||||
cd Wan2GP
|
|
||||||
conda create -n wan2gp python=3.10.9
|
|
||||||
conda activate wan2gp
|
|
||||||
```
|
|
||||||
|
|
||||||
### Step 2: Install PyTorch Beta
|
|
||||||
|
|
||||||
```shell
|
|
||||||
# Install PyTorch 2.7.0 with CUDA 12.8
|
|
||||||
pip install torch==2.7.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/test/cu128
|
|
||||||
```
|
|
||||||
|
|
||||||
### Step 3: Install Dependencies
|
|
||||||
|
|
||||||
```shell
|
|
||||||
pip install -r requirements.txt
|
|
||||||
```
|
|
||||||
|
|
||||||
### Step 4: Optional Optimizations for RTX 50XX
|
|
||||||
|
|
||||||
#### Sage Attention
|
|
||||||
|
|
||||||
```shell
|
|
||||||
# Windows
|
|
||||||
pip install triton-windows
|
|
||||||
pip install sageattention==1.0.6
|
|
||||||
|
|
||||||
# Linux
|
|
||||||
pip install sageattention==1.0.6
|
|
||||||
```
|
|
||||||
|
|
||||||
#### Sage 2 Attention
|
|
||||||
|
|
||||||
```shell
|
|
||||||
# Windows
|
|
||||||
pip install triton-windows
|
|
||||||
pip install https://github.com/woct0rdho/SageAttention/releases/download/v2.1.1-windows/sageattention-2.1.1+cu128torch2.7.0-cp310-cp310-win_amd64.whl
|
|
||||||
|
|
||||||
# Linux (manual compilation)
|
|
||||||
git clone https://github.com/thu-ml/SageAttention
|
|
||||||
cd SageAttention
|
|
||||||
pip install -e .
|
|
||||||
```
|
|
||||||
|
|
||||||
## Attention Modes
|
## Attention Modes
|
||||||
|
|
||||||
WanGP supports several attention implementations:
|
WanGP supports several attention implementations:
|
||||||
@ -136,6 +81,12 @@ WanGP supports several attention implementations:
|
|||||||
- **Sage2**: 40% speed boost
|
- **Sage2**: 40% speed boost
|
||||||
- **Flash**: Good performance, may be complex to install on Windows
|
- **Flash**: Good performance, may be complex to install on Windows
|
||||||
|
|
||||||
|
### Attention GPU Compatibility
|
||||||
|
|
||||||
|
- RTX 10XX, 20XX: SDPA
|
||||||
|
- RTX 30XX, 40XX: SDPA, Flash Attention, Xformers, Sage, Sage2
|
||||||
|
- RTX 50XX: SDPA, SDPA, Flash Attention, Xformers, Sage2
|
||||||
|
|
||||||
## Performance Profiles
|
## Performance Profiles
|
||||||
|
|
||||||
Choose a profile based on your hardware:
|
Choose a profile based on your hardware:
|
||||||
@ -163,10 +114,5 @@ If Sage attention doesn't work:
|
|||||||
- Use Profile 4 for lower VRAM usage
|
- Use Profile 4 for lower VRAM usage
|
||||||
- Consider using 1.3B models instead of 14B models
|
- Consider using 1.3B models instead of 14B models
|
||||||
|
|
||||||
### GPU Compatibility
|
|
||||||
|
|
||||||
- RTX 10XX, 20XX: Supported with SDPA attention
|
|
||||||
- RTX 30XX, 40XX: Full feature support
|
|
||||||
- RTX 50XX: Beta support with PyTorch 2.7.0
|
|
||||||
|
|
||||||
For more troubleshooting, see [TROUBLESHOOTING.md](TROUBLESHOOTING.md)
|
For more troubleshooting, see [TROUBLESHOOTING.md](TROUBLESHOOTING.md)
|
||||||
|
|||||||
@ -11,7 +11,7 @@ sentencepiece
|
|||||||
open_clip_torch>=2.29.0
|
open_clip_torch>=2.29.0
|
||||||
|
|
||||||
# Video & media
|
# Video & media
|
||||||
moviepy @ git+https://github.com/Zulko/moviepy.git
|
moviepy==1.0.3
|
||||||
av
|
av
|
||||||
ffmpeg-python
|
ffmpeg-python
|
||||||
pygame>=2.1.0
|
pygame>=2.1.0
|
||||||
@ -45,6 +45,7 @@ torchdiffeq>=0.2.5
|
|||||||
tensordict>=0.6.1
|
tensordict>=0.6.1
|
||||||
mmgp==3.5.10
|
mmgp==3.5.10
|
||||||
peft==0.15.0
|
peft==0.15.0
|
||||||
|
matplotlib
|
||||||
|
|
||||||
# Utilities
|
# Utilities
|
||||||
ftfy
|
ftfy
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user