Crates.io | gpt_sovits_rs |
lib.rs | gpt_sovits_rs |
version | |
source | src |
created_at | 2024-10-25 09:52:06.413602 |
updated_at | 2024-12-06 14:33:35.458055 |
description | A GPT_Sovits inference lib in Rust base on pytorch |
homepage | |
repository | https://github.com/second-state/gpt_sovits_rs |
max_upload_size | |
id | 1422287 |
Cargo.toml error: | TOML parse error at line 18, column 1 | 18 | autolib = false | ^^^^^^^ unknown field `autolib`, expected one of `name`, `version`, `edition`, `authors`, `description`, `readme`, `license`, `repository`, `homepage`, `documentation`, `build`, `resolver`, `links`, `default-run`, `default_dash_run`, `rust-version`, `rust_dash_version`, `rust_version`, `license-file`, `license_dash_file`, `license_file`, `licenseFile`, `license_capital_file`, `forced-target`, `forced_dash_target`, `autobins`, `autotests`, `autoexamples`, `autobenches`, `publish`, `metadata`, `keywords`, `categories`, `exclude`, `include` |
size | 0 |
This Rust project provides a library for integrating GPT-SoVITS inference in Rust applications. GPT-Sovits is a powerful tool for speech synthesis and voice conversion. By using this library, developers can easily leverage the capabilities of GPT-Sovits within their Rust projects.
Before you can use the GPT-SoVits-rs Inference Library, you must ensure that libtorch 2.4.0 is installed on your system. libtorch is the C++ frontend for PyTorch, which is required for running the GPT-Sovits models.
You can download and install libtorch 2.4.0 from the official PyTorch website:
Go to the PyTorch website. Select the appropriate options for your system (OS, Package Manager, Python version, etc.), making sure to choose the "LibTorch" option. Scroll down to the "Install" section and run the provided command to download and install libtorch 2.4.0.
For example, on a Linux system with CUDA support, the command might look like this:
wget https://download.pytorch.org/libtorch/cu124/libtorch-cxx11-abi-shared-with-deps-2.4.0%2Bcu124.zip
unzip libtorch-cxx11-abi-shared-with-deps-2.4.0+cu124.zip
After downloading and extracting the library, you may need to set environment variables to include the libtorch library path. For example:
export LIBTORCH=/path/to/libtorch
export LD_LIBRARY_PATH=$LIBTORCH/lib:$LD_LIBRARY_PATH
Replace /path/to/libtorch with the actual path where you extracted libtorch.
To use this library, add the following dependency to your Cargo.toml
file:
[dependencies]
gpt_sovits_rs = "0.1.0"
Replace "0.1.0" with the latest version of the library.
use gpt_sovits_rs::GPTSovitsConfig;
fn main() {
env_logger::init();
let gpt_config = GPTSovitsConfig::new(
"path/to/ssl_model.pt".to_string(),
).with_chinese(
"path/to/g2pw.pt".to_string(),
"path/to/bert_model.pt".to_string(),
"path/to/tokenizer.json".to_string());
// If you don't need to Chinese, you can not call `with_chinese`
let device = gpt_sovits_rs::Device::cuda_if_available();
log::info!("device: {:?}", device);
let mut gpt_sovits = gpt_config.build(device).unwrap();
let ref_text = "Speaker1 Reference Text";
let ref_path = "path/to/speaker1/reference_voice.wav";
let file = std::fs::File::open(ref_path).unwrap();
let (head, ref_audio_samples) = wav_io::read_from_file(file).unwrap();
log::info!("load ref_1 done");
gpt_sovits
.create_speaker(
"speaker1",
"path/to/speaker1/gpt_sovits_model.pt",
&ref_audio_samples,
head.sample_rate as usize,
ref_text,
)
.unwrap();
log::info!("init speaker1 done");
let ref_text = "Speaker2 Reference Text";
let ref_path = "path/to/speaker2/reference_voice.wav";
let file = std::fs::File::open(ref_path).unwrap();
let (head, ref_audio_samples) = wav_io::read_from_file(file).unwrap();
log::info!("load ref_2 done");
gpt_sovits
.create_speaker(
"speaker2",
"path/to/speaker2/gpt_sovits_model.pt",
&ref_audio_samples,
head.sample_rate as usize,
ref_text,
)
.unwrap();
log::info!("init speaker2 done");
let text1 = "What you want speaker1 to say";
let text2 = "What you want speaker2 to say";
let audio1 = gpt_sovits.infer("speaker1",text1).unwrap();
let audio2 = gpt_sovits.infer("speaker2",text2).unwrap();
log::info!("start write file");
let output1 = "speaker1.wav";
let output2 = "speaker2.wav";
let audio1_size = audio1.size1().unwrap() as usize;
let audio2_size = audio1.size1().unwrap() as usize;
// save speaker1.wav
let mut samples1 = vec![0f32; audio1_size];
audio1.f_copy_data(&mut samples1, audio1_size).unwrap();
let mut file_out = std::fs::File::create(output1).unwrap();
let header = wav_io::new_header(32000, 16, false, true);
wav_io::write_to_file(&mut file_out, &header, &samples1).unwrap();
// save speaker2.wav
let mut samples2 = vec![0f32; audio2_size];
audio1.f_copy_data(&mut samples2, audio2_size).unwrap();
let mut file_out = std::fs::File::create(output2).unwrap();
let header = wav_io::new_header(32000, 16, false, true);
wav_io::write_to_file(&mut file_out, &header, &samples2).unwrap();
}
After completing the training of a GPT-Sovits model, you might need to export the training results to a .pt (PyTorch) file for use in other environments. Below are the detailed steps to export the trained model:
Ensure that your GPT-Sovits model has finished training and that you have a trained model file available.
Use the following command to run the export script and export the training results to gpt_sovits_model.pt:
python GPT_SoVITS/export_torch_script.py --gpt_model GPT_weights_v2/xxx-e15.ckpt --sovits_model SoVITS_weights_v2/xxx_e8_s248.pth --ref_audio ref.wav --ref_text 'Reference Text' --output_path output --export_common_model
Now you can find gpt_sovits_model.pt
, ssl_model.pt
, and bert_model.pt
in the output
directory.
The ssl_model.pt
and bert_model.pt
are common model files, determined by the option --export_common_model
for whether to export, and they are not related to the trained model. Therefore, they do not need to be exported every time.
If you do not wish to export, you can go to Hugging Face to download the resource.zip that I have already exported. And remove the --export_common_model
when export model.
You can download g2pw.pt
from my Hugging Face repo