π¬ Video Translation with Synchronized Audio π
SonyTranslate is a powerful and user-friendly web application that allows you to easily translate videos into different languages. This repository hosts the code for the SonyTranslate web UI, which is built with the Gradio library to provide a seamless and interactive user experience.
Description | Link |
---|---|
π Colab Notebook | |
π Repository | |
π Online DEMO |
Language Code | Language |
---|---|
en | English |
fr | French |
de | German |
es | Spanish |
it | Italian |
ja | Japanese |
zh | Chinese |
nl | Dutch |
uk | Ukrainian |
pt | Portuguese |
ar | Arabic |
cs | Czech |
da | Danish |
fi | Finnish |
el | Greek |
he | Hebrew |
hu | Hungarian |
ko | Korean |
fa | Persian |
pl | Polish |
ru | Russian |
tr | Turkish |
ur | Urdu |
hi | Hindi |
vi | Vietnamese |
id | Indonesian |
bn | Bengali |
te | Telugu |
mr | Marathi |
ta | Tamil |
jw (or jv) | Javanese |
ca | Catalan |
ne | Nepali |
th | Thai |
sv | Swedish |
am | Amharic |
cy | Welsh |
hr | Croatian |
is | Icelandic |
ka | Georgian |
km | Khmer |
sk | Slovak |
sq | Albanian |
sr | Serbian |
az | Azerbaijani |
bg | Bulgarian |
gl | Galician |
gu | Gujarati |
kk | Kazakh |
kn | Kannada |
lt | Lithuanian |
lv | Latvian |
ml | Malayalam |
ro | Romanian |
si | Sinhala |
su | Sundanese |
Video_t.mp4
video_dub.mp4
To run SoniTranslate using Colab Runtime:
Before you start installing and using SoniTranslate, there are a few things you need to do:
- Install the NVIDIA drivers for CUDA 11.8.0, NVIDIA CUDA is a parallel computing platform and programming model that enables developers to use the power of NVIDIA graphics processing units (GPUs) to speed up compute-intensive tasks. You can find the drivers here. Follow the instructions on the website to download and install the drivers.
- Accept the license agreement for using Pyannote. You need to have an account on Hugging Face and
accept the license to use the models
: https://huggingface.co/pyannote/speaker-diarization and https://huggingface.co/pyannote/segmentation - Create a huggingface token. Hugging Face is a natural language processing platform that provides access to state-of-the-art models and tools. You will need to create a token in order to use some of the automatic model download features in SoniTranslate. Follow the instructions on the Hugging Face website to create a token.
- Install Anaconda. Anaconda is a free and open-source distribution of Python and R. It includes a package manager called conda that makes it easy to install and manage Python environments and packages. Follow the instructions on the Anaconda website to download and install Anaconda on your system.
- Install Git for your system. Git is a version control system that helps you track changes to your code and collaborate with other developers. You can install Git with Anaconda by running
conda install -c anaconda git -y
in your terminal (Do this after step 1 in the following section.). If you have trouble installing Git via Anaconda, you can use the following link instead:
Once you have completed these steps, you will be ready to install SoniTranslate.
To install SoniTranslate, follow these steps:
- Create a suitable anaconda environment for SoniTranslate and activate it:
conda create -n sonitr python=3.10 -y
conda activate sonitr
conda install pytorch torchvision torchaudio pytorch-cuda=11.8 -c pytorch -c nvidia
- Clone this github repository and navigate to it:
git clone https://github.com/r3gm/SoniTranslate.git
cd SoniTranslate
- Install required packages:
pip install -r requirements_base.txt -v
pip install -r requirements_extra.txt -v
pip install onnxruntime-gpu
-
Install ffmpeg. FFmpeg is a free software project that produces libraries and programs for handling multimedia data. You will need it to process audio and video files. You can install ffmpeg with Anaconda by running
conda install -y ffmpeg
in your terminal. If you have trouble installing ffmpeg via Anaconda, you can use the following link instead: (https://ffmpeg.org/ffmpeg.html). Once it is installed, make sure it is in your PATH by runningffmpeg -h
in your terminal. If you don't get an error message, you're good to go. -
Optional install:
After installing FFmpeg, you can install these optional packages.
Piper TTS is a fast, local neural text to speech system that sounds great and is optimized for the Raspberry Pi 4. Piper is used in a variety of projects. Voices are trained with VITS and exported to the onnxruntime.
pip install -q piper-tts==1.2.0
Coqui XTTS is a text-to-speech (TTS) model that lets you generate realistic voices in different languages. It can clone voices with just a short audio clip, even speak in a different language! It's like having a personal voice mimic for any text you need spoken.
pip install -q -r requirements_xtts.txt
pip install -q TTS==0.21.1 --no-deps
To run SoniTranslate locally, make sure the sonitr
conda environment is active:
conda activate sonitr
Setting your Hugging Face token as an environment variable in Linux:
export YOUR_HF_TOKEN="YOUR_HUGGING_FACE_TOKEN"
Then navigate to the SoniTranslate
folder and run either the app_rvc.py
python app_rvc.py
When the local URL
http://127.0.0.1:7860
is displayed in the terminal, simply open this URL in your web browser to access the SoniTranslate interface.
In most environments, you can stop the execution by pressing Ctrl+C in the terminal where you launched the script app_rvc.py
. This will interrupt the program and stop the Gradio app.
To deactivate the Conda environment, you can use the following command:
conda deactivate
This will deactivate the currently active Conda environment sonitr, and you'll return to the base environment or the global Python environment.
If you need to start over from scratch, you can delete the SoniTranslate
folder and remove the sonitr
conda environment with the following set of commands:
conda deactivate
conda env remove -n sonitr
With the sonitr
environment removed, you can start over with a fresh installation.
The app_rvc.py script supports command-line arguments to customize its behavior. Here's a brief guide on how to use them:
Argument command | Default | Value | Description |
---|---|---|---|
--theme | Taithrah/Minimal | String | Sets the theme for the interface. Themes can be found in the Theme Gallery. |
--language | english | String | Selects the interface language. Available options: arabic, azerbaijani, chinese_zh_cn, english, french, german, hindi, indonesian, italian, japanese, korean, marathi, polish, portuguese, russian, spanish, swedish, turkish, ukrainian, vietnamese. |
--verbosity_level | info | String | Sets the verbosity level of the logger: debug, info, warning, error, or critical. |
--public_url | Boolean | Enables a public link. | |
--logs_in_gui | Boolean | Shows the operations performed in Logs (obsolete). |
Example usage:
python app_rvc.py --theme aliabid94/new-theme --language french
This command sets the theme to a custom theme and selects French as the interface language. Feel free to customize these arguments according to your preferences and requirements.
π₯ 2024/03/02: Preserve file names in output. Multiple archives can now be submitted simultaneously by specifying their paths, directories or URLs separated by commas. Processing of a full YouTube playlist. About supported sites URL, please be aware that not all sites may work optimally. Added option for disabling diarization. Implemented soft subtitles. Format output (MP3, MP4, MKV, WAV, and OGG), and resolved issues related to file reading and diarization.
π₯ 2024/02/22: Added freevc for voice imitation, fixed voiceless track, divide segments. New languages support (Swedish, Amharic, Welsh, Croatian, Icelandic, Georgian, Khmer, Slovak, Albanian, Serbian, Azerbaijani, Bulgarian, Galician, Gujarati, Kazakh, Kannada, Lithuanian, Latvian, Malayalam, Romanian, Sinhala and Sundanese). New translations of the GUI (Spanish, French, German, Italian, Japanese, Chinese Simplified, Ukrainian, Arabic, Russian, Turkish, Indonesian, Portuguese, Hindi, Vietnamese, Polish, Swedish, Korean, Marathi and Azerbaijani). With subtitle file, no align and the media file is not needed to process the SRT file. Burn subtitles to video. Queue can accept multiple tasks simultaneously. Sound alert notification. Continue process from last checkpoint. Acceleration rate regulation.
π₯ 2024/01/16: Expanded language support (Thai, Nepali, Catalan, Javanese, Tamil, Marathi, Telugu, Bengali and Indonesian), the introduction of whisper large v3, configurable GUI options, integration of BARK, Facebook-mms, Coqui XTTS, and Piper-TTS. Additional features included audio separation utilities, XTTS WAV creation, use an SRT file as a base for translation, document translation, manual speaker editing, and flexible output options (video, audio, subtitles).
π₯ 2023/10/29: Edit the translated subtitle, download it, adjust volume and speed options.
π₯ 2023/08/03: Changed default options and added directory view of downloads.
π₯ 2023/08/02: Added support for Arabic, Czech, Danish, Finnish, Greek, Hebrew, Hungarian, Korean, Persian, Polish, Russian, Turkish, Urdu, Hindi, and Vietnamese languages. π
π₯ 2023/08/01: Add options for use RVC models.
π₯ 2023/07/27: Fix some bug processing the video and audio.
π₯ 2023/07/26: New UI and add mix options.
Welcome to contributions from the community! If you have any ideas, bug reports, or feature requests, please open an issue or submit a pull request. For more information, please refer to the contribution guidelines.
Although the code is licensed under Apache 2, the models or weights may have commercial restrictions, as seen with pyannote diarization.