目录

MagicAvatar: Multimodal Avatar Generation and Animation

Jianfeng Zhang* · Hanshu Yan* · Zhongcong Xu* · Jiashi Feng · Jun Hao Liew†
ByteDance Inc.

Paper PDF Project Page Project Page

Introducing MagicAvatar, a multi-modal framework capable of converting various input modalities — text, video, and audio — into motion signals that subsequently generate/ animate an avatar.

For more general video editing applications, please also check our latest work MagicEdit!

Citing

If you find our work useful, please consider citing:

@inproceedings{zhang2023magicavatar,
    author    = {Zhang, Jianfeng and Yan, Hanshu and Xu, Zhongcong and Feng, Jiashi and Liew, Jun Hao},
    title     = {MagicAvatar: Multi-modal Avatar Generation and Animation},
    booktitle = {arXiv},
    year      = {2023}
}

@inproceedings{liew2023magicedit,
    author    = {Liew, Jun Hao and Yan, Hanshu and Zhang, Jianfeng and Xu, Zhongcong and Feng, Jiashi},
    title     = {MagicEdit: High-Fidelity and Temporally Coherent Video Editing},
    booktitle = {arXiv},
    year      = {2023}
}
关于
30.0 KB
邀请码
    Gitlink(确实开源)
  • 加入我们
  • 官网邮箱:gitlink@ccf.org.cn
  • QQ群
  • QQ群
  • 公众号
  • 公众号

版权所有:中国计算机学会技术支持:开源发展技术委员会
京ICP备13000930号-9 京公网安备 11010802032778号