Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/aifsh/comfyui-whisperx
a comfyui cuatom node for audio subtitling based on whisperX and translators
https://github.com/aifsh/comfyui-whisperx
srt-subtitles sutitles translation whisper
Last synced: about 2 months ago
JSON representation
a comfyui cuatom node for audio subtitling based on whisperX and translators
- Host: GitHub
- URL: https://github.com/aifsh/comfyui-whisperx
- Owner: AIFSH
- License: other
- Created: 2024-04-22T08:18:49.000Z (8 months ago)
- Default Branch: main
- Last Pushed: 2024-10-23T22:19:56.000Z (2 months ago)
- Last Synced: 2024-10-24T11:24:49.125Z (about 2 months ago)
- Topics: srt-subtitles, sutitles, translation, whisper
- Language: Python
- Homepage:
- Size: 1.12 MB
- Stars: 29
- Watchers: 1
- Forks: 6
- Open Issues: 7
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# ComfyUI-WhisperX
a comfyui cuatom node for audio subtitling based on [whisperX](https://github.com/m-bain/whisperX.git) and [translators](https://github.com/UlionTse/translators)
## Features
- export `srt` file for subtitle was supported
- translate was supported by [translators](https://github.com/UlionTse/translators) with huge number engine
- mutiple speaker diarization was supported by [pyannote-audio](https://github.com/pyannote/pyannote-audio)
- huge comfyui custom nodes can merge in whisperx## How to use
make sure `ffmpeg` is worked in your commandline
for Linux
```
apt update
apt install ffmpeg
```
for Windows,you can install `ffmpeg` by [WingetUI](https://github.com/marticliment/WingetUI) automaticallythen!
```
git clone https://github.com/AIFSH/ComfyUI-WhisperX.git
cd ComfyUI-WhisperX
pip install -r requirements.txt
```
`weights` will be downloaded from huggingface automatically! if you in china,make sure your internet attach the huggingface
or if you still struggle with huggingface, you may try follow [hf-mirror](https://hf-mirror.com/) to config your env.to speaker diarization, you need!
1. Accept [`pyannote/segmentation-3.0`](https://hf.co/pyannote/segmentation-3.0) user conditions
2. Accept [`pyannote/speaker-diarization-3.1`](https://hf.co/pyannote/speaker-diarization-3.1) user conditions
3. Create access token at [`hf.co/settings/tokens`](https://hf.co/settings/tokens).## Tutorial
[Demo](https://www.bilibili.com/video/BV19i421y7jb/)## WeChat Group && Donate
## Thanks
- [whisperX](https://github.com/m-bain/whisperX.git)
- [translators](https://github.com/UlionTse/translators)