DaVinci-MagiHuman:Speed by Simplicity: A Single-Stream Architecture for Fast Audio-Video Generative Foundation Model
- add layer offload num to fit high vram 大显存 把offload开到你跑得动为止,小显存则从1开始测试,MagiCompiler库不用,但是改了麻烦,还是加了回来
In the ./ComfyUI/custom_nodes directory, run the following:
git clone https://github.com/smthemex/ComfyUI_MagiHuman
pip install -r requirements.txt
#python312以下,注释掉MagiCompiler目录下的pyproject.toml的13行的requires-python = ">=3.12" 再安装
git clone https://github.com/SandAI-org/MagiCompiler.git
cd MagiCompiler
pip install -r requirements.txt
pip install .
├── ComfyUI/models/
| ├── diffusion_models/
| ├──distill-merger_bf16.safetensors #28G
| ├──540p_sr_merge_bf16.safetensors #28g For SR ,放大用开源不下
| ├── vae/
| ├──sd_audio.safetensors #4.7GM
| ├──Wan2.2_VAE.pth # 2.7G
| ├── gguf
| ├──t5gemma-9b-9b-ul2-Q6_K.gguf # 11G
We thank the open-source community, and in particular Wan2.2 and Turbo-VAED, for their valuable contributions.
This project is released under the Apache License 2.0.
@misc{davinci-magihuman-2026,
title = {Speed by Simplicity: A Single-Stream Architecture for Fast Audio-Video Generative Foundation Model},
author = {SII-GAIR and Sand.ai},
year = {2026},
url = {https://github.com/GAIR-NLP/daVinci-MagiHuman}
}

