English | 中文
- Video to Video Translation
- Voice Clone
- Lip Sync
- Add Subtitles
- Supports Chinese and English
Please check res.mp4
for the demo.
- ASR: fast-whisper
- Text Translation:
- facebook/m2m
- translators[default]
- Voice Clone:
- Lip Sync: videotalking
- Face Restore: gfpgan
- Video Merge: pyvideotrans
- Install ffmpeg and add it to the system environment variable, or simply put the executable file in the
ffmpeg
directory, and the directory structure should be as follows:
ffmpeg
|- ffmpeg.exe
|- ffprobe.exe
...
- Install dependencies
git clone [email protected]:halfzm/v2vt.git
conda create -n v2vt_clone python=3.11.0
conda activate v2vt_clone
cd v2vt_clone
pip install -r requirements.txt
pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118
- Put the
openvoice pretrained models
in theopenvoice_checkpoints
directory, and the directory structure should be as follows:[optional]
openvoice_checkpoints
|- base_speakers
|- ...
|- converter
|- ...
...
- Put the
coqui xtts pretrained models
in thetts_models
directory,and the directory structure should be as follows:
tts_models
|- tts_models--multilingual--multi-dataset--xtts_v2
|- config.json
|- model.pth
|- speakers_xtts.pth
|- vocal.json
|- ...
...
- Put the
video retalking checkpoints
in thevideo-retalking/checkpoints
directory,and the directory structure should be as follows:
video-retalking
|- checkpoints
|- ...
...
- Quick start
python app.py
-
Input file
The input video should not be too short, otherwise an error will occur when the voice is being cloned (preferably >=5 seconds) -
Output file
By default, you can find the outfile in current directory, namedoutpu.mp4
, which can also be downloaded directly from the webui.
About licence and code_of_conduct, we follow the video-retalking project.
You can see the details at LICENSE
and CODE_OF_CONDUCT
.