4.2 KiB
4.2 KiB
MuseTalk 部署指南
硬件要求
| 配置 | 最低要求 | 推荐配置 |
|---|---|---|
| GPU | 8GB VRAM (如 RTX 3060) | 24GB VRAM (如 RTX 3090) |
| 内存 | 32GB | 64GB |
| CUDA | 11.7+ | 12.0+ |
📦 安装步骤
1. 克隆 MuseTalk 仓库
# 进入 ViGent 项目的 models 目录
cd /home/rongye/ProgramFiles/ViGent/models
# 克隆 MuseTalk 仓库
git clone https://github.com/TMElyralab/MuseTalk.git MuseTalk_repo
# 保留我们的自定义文件 (如果有)
# cp MuseTalk/DEPLOY.md MuseTalk_repo/
# cp MuseTalk/musetalk_api.py MuseTalk_repo/
# 替换目录
rm -rf MuseTalk
mv MuseTalk_repo MuseTalk
2. 创建虚拟环境
cd /home/rongye/ProgramFiles/ViGent/models/MuseTalk
conda create -n musetalk python=3.10 -y
conda activate musetalk
3. 安装 PyTorch (稳定版 2.0.1)
⚠️ 注意:必须使用 PyTorch 2.0.1 配合 CUDA 11.8,因为这是 mmcv 预编译包支持的版本。
pip install torch==2.0.1 torchvision==0.15.2 torchaudio==2.0.2 --index-url https://download.pytorch.org/whl/cu118
4. 安装 MuseTalk 依赖 (MMLab)
严格按照以下顺序和版本安装:
pip install -r requirements.txt
# 安装 mmlab 系列
pip install --no-cache-dir -U openmim
mim install mmengine
mim install "mmcv==2.0.1"
mim install "mmdet==3.1.0"
pip install chumpy --no-build-isolation
pip install "mmpose==1.1.0" --no-deps
5. 下载模型权重 ⬇️
注意:模型目录结构极其重要,必须严格按照以下步骤操作。
cd /home/rongye/ProgramFiles/ViGent/models/MuseTalk/models
# 1. 准备目录
mkdir -p musetalk musetalkV15 dwpose syncnet face-parse-bisent
# 2. 从 HuggingFace 下载基础模型
pip install huggingface_hub
# MuseTalk v1
huggingface-cli download TMElyralab/MuseTalk --local-dir ./musetalk_tmp --include "*.pth" "*.json" "*.bin"
mv musetalk_tmp/* musetalk/ && rm -rf musetalk_tmp
# MuseTalk v1.5
huggingface-cli download TMElyralab/MuseTalk --local-dir ./musetalkV15_tmp --include "unet.pth"
mv musetalkV15_tmp/* musetalkV15/ && rm -rf musetalkV15_tmp
# SD-VAE
huggingface-cli download stabilityai/sd-vae-ft-mse --local-dir ./sd-vae-ft-mse
# Whisper
huggingface-cli download openai/whisper-tiny --local-dir ./whisper
# DWPose
huggingface-cli download yzd-v/DWPose dw-ll_ucoco_384.pth --local-dir ./dwpose
# SyncNet
huggingface-cli download ByteDance/LatentSync latentsync_syncnet.pt --local-dir ./syncnet
# 3. 下载 Face Parse 模型 (需从 Google Drive 或 PyTorch 官网)
cd face-parse-bisent
wget https://download.pytorch.org/models/resnet18-5c106cde.pth -O resnet18-5c106cde.pth
# 79999_iter.pth主要从Google Drive下载,或手动上传
# pip install gdown && gdown 154JgKpzCPW82qINcVieuPH3fZ2e0P812 -O 79999_iter.pth
cd ..
# 4. === 关键修复步骤 ===
# 创建必要的符号链接以匹配代码路径
ln -s sd-vae-ft-mse sd-vae
cd musetalk && ln -s musetalk.json config.json && cd ..
📂 最终目录结构验证
确保 models/MuseTalk/models 目录如下所示:
models/
├── musetalk/
│ ├── config.json -> musetalk.json # ⚠️ 必须有此软链
│ ├── musetalk.json
│ └── pytorch_model.bin
├── musetalkV15/
│ ├── musetalk.json
│ └── unet.pth
├── sd-vae -> sd-vae-ft-mse # ⚠️ 必须有此软链
├── sd-vae-ft-mse/
│ └── diffusion_pytorch_model.bin
├── whisper/
│ └── pytorch_model.bin ...
├── dwpose/
│ └── dw-ll_ucoco_384.pth
├── syncnet/
│ └── latentsync_syncnet.pt
└── face-parse-bisent/
├── 79999_iter.pth
└── resnet18-5c106cde.pth
🔧 验证安装
cd /home/rongye/ProgramFiles/ViGent/models/MuseTalk
conda activate musetalk
# 运行测试脚本 (需先准备测试素材)
# 确保 inference_config.yaml 格式正确
CUDA_VISIBLE_DEVICES=1 python -m scripts.inference \
--version v15 \
--inference_config configs/inference/test.yaml \
--result_dir ./results \
--use_float16