Files
ViGent/models/MuseTalk/DEPLOY.md
2026-01-15 15:50:23 +08:00

156 lines
4.2 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# MuseTalk 部署指南
## 硬件要求
| 配置 | 最低要求 | 推荐配置 |
|------|----------|----------|
| GPU | 8GB VRAM (如 RTX 3060) | 24GB VRAM (如 RTX 3090) |
| 内存 | 32GB | 64GB |
| CUDA | 11.7+ | 12.0+ |
---
## 📦 安装步骤
### 1. 克隆 MuseTalk 仓库
```bash
# 进入 ViGent 项目的 models 目录
cd /home/rongye/ProgramFiles/ViGent/models
# 克隆 MuseTalk 仓库
git clone https://github.com/TMElyralab/MuseTalk.git MuseTalk_repo
# 保留我们的自定义文件 (如果有)
# cp MuseTalk/DEPLOY.md MuseTalk_repo/
# cp MuseTalk/musetalk_api.py MuseTalk_repo/
# 替换目录
rm -rf MuseTalk
mv MuseTalk_repo MuseTalk
```
### 2. 创建虚拟环境
```bash
cd /home/rongye/ProgramFiles/ViGent/models/MuseTalk
conda create -n musetalk python=3.10 -y
conda activate musetalk
```
### 3. 安装 PyTorch (稳定版 2.0.1)
⚠️ **注意**:必须使用 PyTorch 2.0.1 配合 CUDA 11.8,因为这是 mmcv 预编译包支持的版本。
```bash
pip install torch==2.0.1 torchvision==0.15.2 torchaudio==2.0.2 --index-url https://download.pytorch.org/whl/cu118
```
### 4. 安装 MuseTalk 依赖 (MMLab)
严格按照以下顺序和版本安装:
```bash
pip install -r requirements.txt
# 安装 mmlab 系列
pip install --no-cache-dir -U openmim
mim install mmengine
mim install "mmcv==2.0.1"
mim install "mmdet==3.1.0"
pip install chumpy --no-build-isolation
pip install "mmpose==1.1.0" --no-deps
```
### 5. 下载模型权重 ⬇️
> **注意**:模型目录结构极其重要,必须严格按照以下步骤操作。
```bash
cd /home/rongye/ProgramFiles/ViGent/models/MuseTalk/models
# 1. 准备目录
mkdir -p musetalk musetalkV15 dwpose syncnet face-parse-bisent
# 2. 从 HuggingFace 下载基础模型
pip install huggingface_hub
# MuseTalk v1
huggingface-cli download TMElyralab/MuseTalk --local-dir ./musetalk_tmp --include "*.pth" "*.json" "*.bin"
mv musetalk_tmp/* musetalk/ && rm -rf musetalk_tmp
# MuseTalk v1.5
huggingface-cli download TMElyralab/MuseTalk --local-dir ./musetalkV15_tmp --include "unet.pth"
mv musetalkV15_tmp/* musetalkV15/ && rm -rf musetalkV15_tmp
# SD-VAE
huggingface-cli download stabilityai/sd-vae-ft-mse --local-dir ./sd-vae-ft-mse
# Whisper
huggingface-cli download openai/whisper-tiny --local-dir ./whisper
# DWPose
huggingface-cli download yzd-v/DWPose dw-ll_ucoco_384.pth --local-dir ./dwpose
# SyncNet
huggingface-cli download ByteDance/LatentSync latentsync_syncnet.pt --local-dir ./syncnet
# 3. 下载 Face Parse 模型 (需从 Google Drive 或 PyTorch 官网)
cd face-parse-bisent
wget https://download.pytorch.org/models/resnet18-5c106cde.pth -O resnet18-5c106cde.pth
# 79999_iter.pth主要从Google Drive下载或手动上传
# pip install gdown && gdown 154JgKpzCPW82qINcVieuPH3fZ2e0P812 -O 79999_iter.pth
cd ..
# 4. === 关键修复步骤 ===
# 创建必要的符号链接以匹配代码路径
ln -s sd-vae-ft-mse sd-vae
cd musetalk && ln -s musetalk.json config.json && cd ..
```
---
## 📂 最终目录结构验证
确保 `models/MuseTalk/models` 目录如下所示:
```
models/
├── musetalk/
│ ├── config.json -> musetalk.json # ⚠️ 必须有此软链
│ ├── musetalk.json
│ └── pytorch_model.bin
├── musetalkV15/
│ ├── musetalk.json
│ └── unet.pth
├── sd-vae -> sd-vae-ft-mse # ⚠️ 必须有此软链
├── sd-vae-ft-mse/
│ └── diffusion_pytorch_model.bin
├── whisper/
│ └── pytorch_model.bin ...
├── dwpose/
│ └── dw-ll_ucoco_384.pth
├── syncnet/
│ └── latentsync_syncnet.pt
└── face-parse-bisent/
├── 79999_iter.pth
└── resnet18-5c106cde.pth
```
---
## 🔧 验证安装
```bash
cd /home/rongye/ProgramFiles/ViGent/models/MuseTalk
conda activate musetalk
# 运行测试脚本 (需先准备测试素材)
# 确保 inference_config.yaml 格式正确
CUDA_VISIBLE_DEVICES=1 python -m scripts.inference \
--version v15 \
--inference_config configs/inference/test.yaml \
--result_dir ./results \
--use_float16
```