{"id":13441207,"url":"https://github.com/myshell-ai/OpenVoice","last_synced_at":"2025-03-20T11:37:38.558Z","repository":{"id":209854679,"uuid":"725089041","full_name":"myshell-ai/OpenVoice","owner":"myshell-ai","description":"Instant voice cloning by MIT and MyShell.","archived":false,"fork":false,"pushed_at":"2024-08-21T18:04:49.000Z","size":3877,"stargazers_count":28007,"open_issues_count":184,"forks_count":2740,"subscribers_count":212,"default_branch":"main","last_synced_at":"2024-08-21T20:24:14.301Z","etag":null,"topics":["text-to-speech","tts","voice-clone","zero-shot-tts"],"latest_commit_sha":null,"homepage":"https://research.myshell.ai/open-voice","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/myshell-ai.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2023-11-29T12:17:01.000Z","updated_at":"2024-08-21T18:05:38.000Z","dependencies_parsed_at":"2023-11-29T14:29:33.913Z","dependency_job_id":"e2eeee9c-6d25-4357-890d-37ee6d3ad4ed","html_url":"https://github.com/myshell-ai/OpenVoice","commit_stats":null,"previous_names":["myshell-ai/openvic","myshell-ai/openvoice"],"tags_count":0,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/myshell-ai%2FOpenVoice","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/myshell-ai%2FOpenVoice/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/myshell-ai%2FOpenVoice/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/myshell-ai%2FOpenVoice/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/myshell-ai","download_url":"https://codeload.github.com/myshell-ai/OpenVoice/tar.gz/refs/heads/main","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":221759943,"owners_count":16876322,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["text-to-speech","tts","voice-clone","zero-shot-tts"],"created_at":"2024-07-31T03:01:31.143Z","updated_at":"2025-03-20T11:37:38.549Z","avatar_url":"https://github.com/myshell-ai.png","language":"Python","readme":"\u003cdiv align=\"center\"\u003e\n  \u003cdiv\u003e\u0026nbsp;\u003c/div\u003e\n  \u003cimg src=\"resources/openvoicelogo.jpg\" width=\"400\"/\u003e \n\n[Paper](https://arxiv.org/abs/2312.01479) |\n[Website](https://research.myshell.ai/open-voice) \u003cbr\u003e \u003cbr\u003e\n\u003ca href=\"https://trendshift.io/repositories/6161\" target=\"_blank\"\u003e\u003cimg src=\"https://trendshift.io/api/badge/repositories/6161\" alt=\"myshell-ai%2FOpenVoice | Trendshift\" style=\"width: 250px; height: 55px;\" width=\"250\" height=\"55\"/\u003e\u003c/a\u003e\n\u003c/div\u003e\n\n## Introduction\n\n### OpenVoice V1\n\nAs we detailed in our [paper](https://arxiv.org/abs/2312.01479) and [website](https://research.myshell.ai/open-voice), the advantages of OpenVoice are three-fold:\n\n**1. Accurate Tone Color Cloning.**\nOpenVoice can accurately clone the reference tone color and generate speech in multiple languages and accents.\n\n**2. Flexible Voice Style Control.**\nOpenVoice enables granular control over voice styles, such as emotion and accent, as well as other style parameters including rhythm, pauses, and intonation. \n\n**3. Zero-shot Cross-lingual Voice Cloning.**\nNeither of the language of the generated speech nor the language of the reference speech needs to be presented in the massive-speaker multi-lingual training dataset.\n\n### OpenVoice V2\n\nIn April 2024, we released OpenVoice V2, which includes all features in V1 and has:\n\n**1. Better Audio Quality.**\nOpenVoice V2 adopts a different training strategy that delivers better audio quality.\n\n**2. Native Multi-lingual Support.**\nEnglish, Spanish, French, Chinese, Japanese and Korean are natively supported in OpenVoice V2.\n\n**3. Free Commercial Use.**\nStarting from April 2024, both V2 and V1 are released under MIT License. Free for commercial use.\n\n[Video](https://github.com/myshell-ai/OpenVoice/assets/40556743/3cba936f-82bf-476c-9e52-09f0f417bb2f)\n\nOpenVoice has been powering the instant voice cloning capability of [myshell.ai](https://app.myshell.ai/explore) since May 2023. Until Nov 2023, the voice cloning model has been used tens of millions of times by users worldwide, and witnessed the explosive user growth on the platform.\n\n## Main Contributors\n\n- [Zengyi Qin](https://www.qinzy.tech) at MIT\n- [Wenliang Zhao](https://wl-zhao.github.io) at Tsinghua University\n- [Xumin Yu](https://yuxumin.github.io) at Tsinghua University\n- [Ethan Sun](https://twitter.com/ethan_myshell) at MyShell\n\n## How to Use\nPlease see [usage](docs/USAGE.md) for detailed instructions.\n\n## Common Issues\n\nPlease see [QA](docs/QA.md) for common questions and answers. We will regularly update the question and answer list.\n\n## Citation\n```\n@article{qin2023openvoice,\n  title={OpenVoice: Versatile Instant Voice Cloning},\n  author={Qin, Zengyi and Zhao, Wenliang and Yu, Xumin and Sun, Xin},\n  journal={arXiv preprint arXiv:2312.01479},\n  year={2023}\n}\n```\n\n## License\nOpenVoice V1 and V2 are MIT Licensed. Free for both commercial and research use.\n\n## Acknowledgements\nThis implementation is based on several excellent projects, [TTS](https://github.com/coqui-ai/TTS), [VITS](https://github.com/jaywalnut310/vits), and [VITS2](https://github.com/daniilrobnikov/vits2). Thanks for their awesome work!\n","funding_links":[],"categories":["Python","Azure Cognitive Search \u0026 OpenAI","\u003cspan id=\"speech\"\u003eSpeech\u003c/span\u003e","Projects","App","音频","HarmonyOS","语音合成","Tools \u0026 Frameworks","Colab Notebooks","TTS Models","Repos","Voice \u0026 Multimodal (local) (16)"],"sub_categories":["\u003cspan id=\"tool\"\u003eLLM (LLM \u0026 Tool)\u003c/span\u003e","🎥 Video","Windows Manager","网络服务_其他","Open-source projects","OpenVoice V2 (Voice Conversion)"],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fmyshell-ai%2FOpenVoice","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fmyshell-ai%2FOpenVoice","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fmyshell-ai%2FOpenVoice/lists"}