mirror of
https://github.com/Wan-Video/Wan2.1.git
synced 2026-01-10 00:03:32 +00:00
Update README.md
This commit is contained in:
parent
d16cc19945
commit
79df3aae64
27
README.md
27
README.md
@ -122,33 +122,6 @@ See full changelog: **[Changelog](docs/CHANGELOG.md)**
|
|||||||
|
|
||||||
## 🚀 Quick Start
|
## 🚀 Quick Start
|
||||||
|
|
||||||
### 🐳 Docker:
|
|
||||||
|
|
||||||
**For Debian-based systems (Ubuntu, Debian, etc.):**
|
|
||||||
|
|
||||||
```bash
|
|
||||||
./run-docker-cuda-deb.sh
|
|
||||||
```
|
|
||||||
|
|
||||||
This automated script will:
|
|
||||||
|
|
||||||
- Detect your GPU model and VRAM automatically
|
|
||||||
- Select optimal CUDA architecture for your GPU
|
|
||||||
- Install NVIDIA Docker runtime if needed
|
|
||||||
- Build a Docker image with all dependencies
|
|
||||||
- Run WanGP with optimal settings for your hardware
|
|
||||||
|
|
||||||
**Docker environment includes:**
|
|
||||||
|
|
||||||
- NVIDIA CUDA 12.4.1 with cuDNN support
|
|
||||||
- PyTorch 2.6.0 with CUDA 12.4 support
|
|
||||||
- SageAttention compiled for your specific GPU architecture
|
|
||||||
- Optimized environment variables for performance (TF32, threading, etc.)
|
|
||||||
- Automatic cache directory mounting for faster subsequent runs
|
|
||||||
- Current directory mounted in container - all downloaded models, loras, generated videos and files are saved locally
|
|
||||||
|
|
||||||
**Supported GPUs:** RTX 50XX, RTX 40XX, RTX 30XX, RTX 20XX, GTX 16XX, GTX 10XX, Tesla V100, A100, H100, and more.
|
|
||||||
|
|
||||||
**One-click installation:** Get started instantly with [Pinokio App](https://pinokio.computer/)
|
**One-click installation:** Get started instantly with [Pinokio App](https://pinokio.computer/)
|
||||||
|
|
||||||
**Manual installation:**
|
**Manual installation:**
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user