{"id":18000370,"url":"https://github.com/mingyuan-zhang/motiondiffuse","last_synced_at":"2025-04-12T18:52:16.124Z","repository":{"id":58552671,"uuid":"532145731","full_name":"mingyuan-zhang/MotionDiffuse","owner":"mingyuan-zhang","description":"MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model","archived":false,"fork":false,"pushed_at":"2024-07-19T21:25:48.000Z","size":31221,"stargazers_count":913,"open_issues_count":26,"forks_count":76,"subscribers_count":28,"default_branch":"main","last_synced_at":"2025-04-03T21:13:46.120Z","etag":null,"topics":["3d-generation","diffusion-model","motion-generation","text-driven"],"latest_commit_sha":null,"homepage":"https://mingyuan-zhang.github.io/projects/MotionDiffuse.html","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"other","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/mingyuan-zhang.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2022-09-03T03:36:54.000Z","updated_at":"2025-04-02T12:29:47.000Z","dependencies_parsed_at":"2024-10-29T23:42:59.082Z","dependency_job_id":null,"html_url":"https://github.com/mingyuan-zhang/MotionDiffuse","commit_stats":null,"previous_names":[],"tags_count":0,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/mingyuan-zhang%2FMotionDiffuse","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/mingyuan-zhang%2FMotionDiffuse/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/mingyuan-zhang%2FMotionDiffuse/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/mingyuan-zhang%2FMotionDiffuse/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/mingyuan-zhang","download_url":"https://codeload.github.com/mingyuan-zhang/MotionDiffuse/tar.gz/refs/heads/main","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":248618229,"owners_count":21134200,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["3d-generation","diffusion-model","motion-generation","text-driven"],"created_at":"2024-10-29T23:11:35.039Z","updated_at":"2025-04-12T18:52:16.098Z","avatar_url":"https://github.com/mingyuan-zhang.png","language":"Python","readme":"\u003cdiv align=\"center\"\u003e\n\n\u003ch1\u003eMotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model\u003c/h1\u003e\n\n\u003cdiv\u003e\n    \u003ca href='https://scholar.google.com/citations?user=2QLD4fAAAAAJ\u0026hl=en' target='_blank'\u003eMingyuan Zhang\u003c/a\u003e\u003csup\u003e1\u003c/sup\u003e*\u0026emsp;\n    \u003ca href='https://caizhongang.github.io/' target='_blank'\u003eZhongang Cai\u003c/a\u003e\u003csup\u003e1,2\u003c/sup\u003e*\u0026emsp;\n    \u003ca href='https://scholar.google.com/citations?user=lSDISOcAAAAJ\u0026hl=zh-CN' target='_blank'\u003eLiang Pan\u003c/a\u003e\u003csup\u003e1\u003c/sup\u003e\u0026emsp;\n    \u003ca href='https://hongfz16.github.io/' target='_blank'\u003eFangzhou Hong\u003c/a\u003e\u003csup\u003e1\u003c/sup\u003e\u0026emsp;\n    \u003ca href='https://gxyes.github.io/' target='_blank'\u003eXinying Guo\u003c/a\u003e\u003csup\u003e1\u003c/sup\u003e\u0026emsp;\n    \u003ca href='https://yanglei.me/' target='_blank'\u003eLei Yang\u003c/a\u003e\u003csup\u003e2\u003c/sup\u003e\u0026emsp;\n    \u003ca href='https://liuziwei7.github.io/' target='_blank'\u003eZiwei Liu\u003c/a\u003e\u003csup\u003e1+\u003c/sup\u003e\n\u003c/div\u003e\n\u003cdiv\u003e\n    \u003csup\u003e1\u003c/sup\u003eS-Lab, Nanyang Technological University\u0026emsp;\n    \u003csup\u003e2\u003c/sup\u003eSenseTime Research\u0026emsp;\n\u003c/div\u003e\n\u003cdiv\u003e\n    *equal contribution\u0026emsp;\n    \u003csup\u003e+\u003c/sup\u003ecorresponding author\n\u003c/div\u003e\n\n\u003ctable\u003e\n\u003ctr\u003e\n    \u003ctd align='center' width='24%'\u003eplay the guitar\u003c/td\u003e\n    \u003ctd align='center' width='24%'\u003ewalk sadly\u003c/td\u003e\n    \u003ctd align='center' width='24%'\u003ewalk happily\u003c/td\u003e\n    \u003ctd align='center' width='24%'\u003echeck time\u003c/td\u003e\n\u003c/tr\u003e\n\u003ctr\u003e\n    \u003ctd\u003e\u003cimg src=\"figures/gallery/gen_00.gif\" width=\"100%\"/\u003e\u003c/td\u003e\n    \u003ctd\u003e\u003cimg src=\"figures/gallery/gen_03.gif\" width=\"100%\"/\u003e\u003c/td\u003e\n    \u003ctd\u003e\u003cimg src=\"figures/gallery/gen_05.gif\" width=\"100%\"/\u003e\u003c/td\u003e\n    \u003ctd\u003e\u003cimg src=\"figures/gallery/gen_06.gif\" width=\"100%\"/\u003e\u003c/td\u003e\n\u003c/tr\u003e\n\u003c/table\u003e\n\nThis repository contains the official implementation of _MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model_.\n\n---\n\n\u003ch4 align=\"center\"\u003e\n  \u003ca href=\"https://mingyuan-zhang.github.io/projects/MotionDiffuse.html\" target='_blank'\u003e[Project Page]\u003c/a\u003e •\n  \u003ca href=\"https://arxiv.org/abs/2208.15001\" target='_blank'\u003e[arXiv]\u003c/a\u003e •\n  \u003ca href=\"https://youtu.be/U5PTnw490SA\" target='_blank'\u003e[Video]\u003c/a\u003e •\n  \u003ca href=\"https://colab.research.google.com/drive/1Dp6VsZp2ozKuu9ccMmsDjyij_vXfCYb3?usp=sharing\" target='_blank'\u003e[Colab Demo]\u003c/a\u003e •\n  \u003ca href=\"https://huggingface.co/spaces/mingyuan/MotionDiffuse\" target='_blank'\u003e[Hugging Face Demo]\u003c/a\u003e\n\u003c/h4\u003e\n\n\u003c/div\u003e\n\n\n## Updates\n\n[10/2022] Add a [🤗Hugging Face Demo](https://huggingface.co/spaces/mingyuan/MotionDiffuse) for text-driven motion generation!\n\n[10/2022] Add a [Colab Demo](https://colab.research.google.com/drive/1Dp6VsZp2ozKuu9ccMmsDjyij_vXfCYb3?usp=sharing) for text-driven motion generation! [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1Dp6VsZp2ozKuu9ccMmsDjyij_vXfCYb3?usp=sharing)\n\n[10/2022] Code release for text-driven motion generation!\n\n[8/2022] [Paper](https://arxiv.org/abs/2208.15001) uploaded to arXiv. [![arXiv](https://img.shields.io/badge/arXiv-2208.15001-b31b1b.svg)](https://arxiv.org/abs/2208.15001)\n\n## Text-driven Motion Generation\n\nYou may refer to [this file](text2motion/README.md) for detailed introduction.\n\n## Citation\n\nIf you find our work useful for your research, please consider citing the paper:\n\n```\n@article{zhang2022motiondiffuse,\n  title={MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model},\n  author={Zhang, Mingyuan and Cai, Zhongang and Pan, Liang and Hong, Fangzhou and Guo, Xinying and Yang, Lei and Liu, Ziwei},\n  journal={arXiv preprint arXiv:2208.15001},\n  year={2022}\n}\n```\n\n## Acknowledgements\n\nThis study is supported by NTU NAP, MOE AcRF Tier 2 (T2EP20221-0033), and under the RIE2020 Industry Alignment Fund – Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry partner(s).\n","funding_links":[],"categories":[],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fmingyuan-zhang%2Fmotiondiffuse","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fmingyuan-zhang%2Fmotiondiffuse","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fmingyuan-zhang%2Fmotiondiffuse/lists"}