mirror of
https://github.com/neodyland/sbv2-api.git
synced 2025-12-23 07:59:56 +00:00
Compare commits
174 Commits
commit-8a0
...
commit-82c
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
82c9f13ce4 | ||
|
|
dd3ee380e5 | ||
|
|
7034c1aad6 | ||
|
|
f00ce04032 | ||
|
|
a7e6fadf6d | ||
|
|
951551822a | ||
|
|
205cc2ae51 | ||
|
|
3519664fcd | ||
|
|
7d32db19b9 | ||
|
|
2cb43aaf6f | ||
|
|
d5a0f16bb0 | ||
|
|
6aac94e731 | ||
|
|
36f031bfcc | ||
|
|
f587a2c735 | ||
|
|
ca5887e864 | ||
|
|
9fbe75796c | ||
|
|
c6349d18bf | ||
|
|
87bec1b8c2 | ||
|
|
d928962b07 | ||
|
|
e3c9c5b19b | ||
|
|
47f2af1413 | ||
|
|
d6bc99caa3 | ||
|
|
b9e2c0fec1 | ||
|
|
8d97d4228a | ||
|
|
0651b49226 | ||
|
|
56e663c395 | ||
|
|
40cfda707e | ||
|
|
be7c48a12a | ||
|
|
e76b963dfb | ||
|
|
26dfe6dcdb | ||
|
|
8e031387ba | ||
|
|
51c0bf8521 | ||
|
|
33d6f1ccf0 | ||
|
|
404fa0f733 | ||
|
|
0e0e60e04b | ||
|
|
ac2f5a2b42 | ||
|
|
1b78865246 | ||
|
|
3054ab94e0 | ||
|
|
f98c6b8226 | ||
|
|
043f380c75 | ||
|
|
703035cc47 | ||
|
|
fcfff9395b | ||
|
|
0cbb6783cf | ||
|
|
37477c1e30 | ||
|
|
93c121da95 | ||
|
|
69d2857d11 | ||
|
|
5cb4480ac7 | ||
|
|
3553d01a7f | ||
|
|
91ac2fad78 | ||
|
|
ebcb6f9081 | ||
|
|
e47f57ade5 | ||
|
|
605ea12450 | ||
|
|
ef8b2d829c | ||
|
|
2aa43e0e3d | ||
|
|
ef071f2b55 | ||
|
|
dc5fa88432 | ||
|
|
a5f45cd2ef | ||
|
|
84e9118d99 | ||
|
|
3050cc1e99 | ||
|
|
d5fcacd799 | ||
|
|
25ca89e341 | ||
|
|
0c2a397775 | ||
|
|
470a0348fe | ||
|
|
9a99b88b00 | ||
|
|
29f39f0795 | ||
|
|
9f22694df0 | ||
|
|
62ba2c802f | ||
|
|
4f5b936f6f | ||
|
|
3c8efc716c | ||
|
|
e9ced32b70 | ||
|
|
e7a1575cbc | ||
|
|
873bbb77b6 | ||
|
|
1725863fca | ||
|
|
55f05580e4 | ||
|
|
320664eae2 | ||
|
|
87903827fa | ||
|
|
9b8e9dc39d | ||
|
|
bbc38081b6 | ||
|
|
0b822f704a | ||
|
|
132eb6386d | ||
|
|
ee56e9591d | ||
|
|
3194e599b2 | ||
|
|
00f4787f6e | ||
|
|
4b6c72aa51 | ||
|
|
7db6bb67a4 | ||
|
|
b3c75f973e | ||
|
|
e9529be559 | ||
|
|
a6694b5d81 | ||
|
|
096859de66 | ||
|
|
dabdc6712f | ||
|
|
45c3255a91 | ||
|
|
bf39890b3d | ||
|
|
120bc608d7 | ||
|
|
2fc547e38b | ||
|
|
98ddaa3c58 | ||
|
|
656e405cd7 | ||
|
|
9d6aa46fdf | ||
|
|
2fe90c6ede | ||
|
|
7faba2447b | ||
|
|
02ac0885e0 | ||
|
|
1f96b09f3b | ||
|
|
d583c1ca1c | ||
|
|
c135aac852 | ||
|
|
f31fa1d4f9 | ||
|
|
efec7cce14 | ||
|
|
61914129dc | ||
|
|
97c63a2e23 | ||
|
|
3475f47305 | ||
|
|
5493b91a84 | ||
|
|
bca6d04e7b | ||
|
|
d44ebe873e | ||
|
|
96b53d42cd | ||
|
|
9765ef51d2 | ||
|
|
655be55605 | ||
|
|
e68f58d698 | ||
|
|
2124fe4650 | ||
|
|
0217c0a4d5 | ||
|
|
1de09597f5 | ||
|
|
38d86c9249 | ||
|
|
ddc132b27b | ||
|
|
558cd24677 | ||
|
|
6657b06786 | ||
|
|
2a8c9bafde | ||
|
|
d7065ac6eb | ||
|
|
0b1dbe4991 | ||
|
|
1ad588bfcf | ||
|
|
9733ba95fa | ||
|
|
843c16995c | ||
|
|
f0821ea957 | ||
|
|
abc9cec7c7 | ||
|
|
19e6b7f0e6 | ||
|
|
451f4497b6 | ||
|
|
e5e92f6211 | ||
|
|
b835577325 | ||
|
|
3caf93441a | ||
|
|
4deefc596b | ||
|
|
9174aa9b11 | ||
|
|
6bccf0468b | ||
|
|
bbb3f0003b | ||
|
|
46de7a9d3f | ||
|
|
252b27de48 | ||
|
|
1dd3e02562 | ||
|
|
4990261ecd | ||
|
|
e873892223 | ||
|
|
f081b2ed22 | ||
|
|
103eb51ca8 | ||
|
|
01541ff381 | ||
|
|
70c2341afd | ||
|
|
a5d783bd65 | ||
|
|
633dfc305e | ||
|
|
53d7daf11a | ||
|
|
5abfe732e4 | ||
|
|
48aef6cef4 | ||
|
|
64fc74eee6 | ||
|
|
6e01103c5d | ||
|
|
00e95cd77c | ||
|
|
01f2aaa406 | ||
|
|
3785faf81e | ||
|
|
70e16f95ad | ||
|
|
a67df43fc7 | ||
|
|
472d1c600f | ||
|
|
acf94a1283 | ||
|
|
dd5c536f39 | ||
|
|
07637f587d | ||
|
|
e8dbf956e1 | ||
|
|
2687af1a9b | ||
|
|
e915e2bc84 | ||
|
|
22ed557395 | ||
|
|
b8f0477318 | ||
|
|
f4de3e15ae | ||
|
|
fc944b9d33 | ||
|
|
4255e15748 | ||
|
|
8bf3906105 | ||
|
|
1d80eda325 |
15
.github/dependabot.yml
vendored
Normal file
15
.github/dependabot.yml
vendored
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
# To get started with Dependabot version updates, you'll need to specify which
|
||||||
|
# package ecosystems to update and where the package manifests are located.
|
||||||
|
# Please see the documentation for all configuration options:
|
||||||
|
# https://docs.github.com/code-security/dependabot/dependabot-version-updates/configuration-options-for-the-dependabot.yml-file
|
||||||
|
|
||||||
|
version: 2
|
||||||
|
updates:
|
||||||
|
- package-ecosystem: "cargo" # See documentation for possible values
|
||||||
|
directory: "/" # Location of package manifests
|
||||||
|
schedule:
|
||||||
|
interval: "weekly"
|
||||||
|
- package-ecosystem: "npm" # See documentation for possible values
|
||||||
|
directory: "/" # Location of package manifests
|
||||||
|
schedule:
|
||||||
|
interval: "weekly"
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
name: CI
|
name: Build
|
||||||
|
|
||||||
on:
|
on:
|
||||||
push:
|
push:
|
||||||
@@ -6,7 +6,6 @@ on:
|
|||||||
- main
|
- main
|
||||||
tags:
|
tags:
|
||||||
- '*'
|
- '*'
|
||||||
pull_request:
|
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
|
|
||||||
permissions:
|
permissions:
|
||||||
@@ -29,7 +28,7 @@ jobs:
|
|||||||
- uses: actions/setup-python@v5
|
- uses: actions/setup-python@v5
|
||||||
with:
|
with:
|
||||||
python-version: 3.x
|
python-version: 3.x
|
||||||
- run: docker build . -f .github/workflows/CI.Dockerfile --tag ci
|
- run: docker build . -f .github/workflows/build.Dockerfile --tag ci
|
||||||
- name: Build wheels
|
- name: Build wheels
|
||||||
uses: PyO3/maturin-action@v1
|
uses: PyO3/maturin-action@v1
|
||||||
with:
|
with:
|
||||||
@@ -111,6 +110,7 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
name: wheels-sdist
|
name: wheels-sdist
|
||||||
path: ./crates/sbv2_bindings/dist
|
path: ./crates/sbv2_bindings/dist
|
||||||
|
|
||||||
python-wheel:
|
python-wheel:
|
||||||
name: Wheel Upload
|
name: Wheel Upload
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
26
.github/workflows/lint.yml
vendored
Normal file
26
.github/workflows/lint.yml
vendored
Normal file
@@ -0,0 +1,26 @@
|
|||||||
|
name: Lint
|
||||||
|
|
||||||
|
on:
|
||||||
|
pull_request:
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
check:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
matrix:
|
||||||
|
components:
|
||||||
|
- rustfmt
|
||||||
|
- clippy
|
||||||
|
steps:
|
||||||
|
- name: Setup
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
- uses: actions-rust-lang/setup-rust-toolchain@v1
|
||||||
|
with:
|
||||||
|
components: ${{ matrix.components }}
|
||||||
|
- name: Format
|
||||||
|
if: ${{ matrix.components == 'rustfmt' }}
|
||||||
|
run: cargo fmt --all -- --check
|
||||||
|
- name: Lint
|
||||||
|
if: ${{ matrix.components == 'clippy' }}
|
||||||
|
run: cargo clippy --all-targets --all-features -- -D warnings
|
||||||
1057
Cargo.lock
generated
1057
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
@@ -1,18 +1,18 @@
|
|||||||
[workspace]
|
[workspace]
|
||||||
resolver = "2"
|
resolver = "3"
|
||||||
members = ["./crates/sbv2_api", "./crates/sbv2_core", "./crates/sbv2_bindings", "./crates/sbv2_wasm"]
|
members = ["./crates/sbv2_api", "./crates/sbv2_core", "./crates/sbv2_bindings", "./crates/sbv2_wasm"]
|
||||||
|
|
||||||
[workspace.package]
|
[workspace.package]
|
||||||
version = "0.2.0-alpha6"
|
version = "0.2.0"
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
description = "Style-Bert-VITSの推論ライブラリ"
|
description = "Style-Bert-VITSの推論ライブラリ"
|
||||||
license = "MIT"
|
license = "MIT"
|
||||||
readme = "./README.md"
|
readme = "./README.md"
|
||||||
repository = "https://github.com/tuna2134/sbv2-api"
|
repository = "https://github.com/neodyland/sbv2-api"
|
||||||
documentation = "https://docs.rs/sbv2_core"
|
documentation = "https://docs.rs/sbv2_core"
|
||||||
|
|
||||||
[workspace.dependencies]
|
[workspace.dependencies]
|
||||||
anyhow = "1.0.96"
|
anyhow = "1.0.100"
|
||||||
dotenvy = "0.15.7"
|
dotenvy = "0.15.7"
|
||||||
env_logger = "0.11.6"
|
env_logger = "0.11.6"
|
||||||
ndarray = "0.16.1"
|
ndarray = "0.16.1"
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
MIT License
|
MIT License
|
||||||
|
|
||||||
Copyright (c) 2024 tuna2134
|
Copyright (c) 2024 tuna2134
|
||||||
|
Copyright (c) 2025- neodyland
|
||||||
|
|
||||||
Permission is hereby granted, free of charge, to any person obtaining a copy
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||||
of this software and associated documentation files (the "Software"), to deal
|
of this software and associated documentation files (the "Software"), to deal
|
||||||
27
README.md
27
README.md
@@ -1,14 +1,23 @@
|
|||||||
# SBV2-API
|
# SBV2-API
|
||||||
|
|
||||||
## 注意:本バージョンはアルファ版です。
|
> [!CAUTION]
|
||||||
安定版を利用したい場合は[こちら](https://github.com/tuna2134/sbv2-api/tree/v0.1.x)をご覧ください。
|
> 本バージョンはアルファ版です。
|
||||||
|
>
|
||||||
|
> 安定版を利用したい場合は[こちら](https://github.com/neodyland/sbv2-api/tree/v0.1.x)をご覧ください。
|
||||||
|
|
||||||
## 注意: オプションの辞書はLGPLです。
|
> [!CAUTION]
|
||||||
オプションの辞書を使用する場合、バイナリの内部の辞書部分について、LGPLが適用されます。
|
> オプションの辞書はLGPLです。
|
||||||
|
>
|
||||||
|
> オプションの辞書を使用する場合、バイナリの内部の辞書部分について、LGPLが適用されます。
|
||||||
|
|
||||||
|
> [!NOTE]
|
||||||
|
> このレポジトリはメンテナンスの都合上、[tuna2134](https:://github.com/tuna2134)氏の所属する[Neodyland](https://neody.land/)へとリポジトリ所在地を移動しました。
|
||||||
|
>
|
||||||
|
> 引き続きtuna2134氏がメインメンテナとして管理しています。
|
||||||
|
|
||||||
## プログラミングに詳しくない方向け
|
## プログラミングに詳しくない方向け
|
||||||
|
|
||||||
[こちら](https://github.com/tuna2134/sbv2-gui?tab=readme-ov-file)を参照してください。
|
[こちら](https://github.com/tuna2134/sbv2-gui)を参照してください。
|
||||||
|
|
||||||
コマンドやpythonの知識なしで簡単に使えるバージョンです。(できることはほぼ同じ)
|
コマンドやpythonの知識なしで簡単に使えるバージョンです。(できることはほぼ同じ)
|
||||||
|
|
||||||
@@ -20,7 +29,7 @@ JP-Extra しか対応していません。(基本的に対応する予定もあ
|
|||||||
|
|
||||||
## 変換方法
|
## 変換方法
|
||||||
|
|
||||||
[こちら](https://github.com/tuna2134/sbv2-api/tree/main/scripts/convert)を参照してください。
|
[こちら](https://github.com/neodyland/sbv2-api/tree/main/scripts/convert)を参照してください。
|
||||||
|
|
||||||
## Todo
|
## Todo
|
||||||
|
|
||||||
@@ -66,7 +75,7 @@ CPUの場合は
|
|||||||
```sh
|
```sh
|
||||||
docker run -it --rm -p 3000:3000 --name sbv2 \
|
docker run -it --rm -p 3000:3000 --name sbv2 \
|
||||||
-v ./models:/work/models --env-file .env \
|
-v ./models:/work/models --env-file .env \
|
||||||
ghcr.io/tuna2134/sbv2-api:cpu
|
ghcr.io/neodyland/sbv2-api:cpu
|
||||||
```
|
```
|
||||||
|
|
||||||
<details>
|
<details>
|
||||||
@@ -81,7 +90,7 @@ CPUの場合は
|
|||||||
```bash
|
```bash
|
||||||
docker run --platform linux/amd64 -it --rm -p 3000:3000 --name sbv2 \
|
docker run --platform linux/amd64 -it --rm -p 3000:3000 --name sbv2 \
|
||||||
-v ./models:/work/models --env-file .env \
|
-v ./models:/work/models --env-file .env \
|
||||||
ghcr.io/tuna2134/sbv2-api:cpu
|
ghcr.io/neodyland/sbv2-api:cpu
|
||||||
```
|
```
|
||||||
</details>
|
</details>
|
||||||
|
|
||||||
@@ -90,7 +99,7 @@ CUDAの場合は
|
|||||||
docker run -it --rm -p 3000:3000 --name sbv2 \
|
docker run -it --rm -p 3000:3000 --name sbv2 \
|
||||||
-v ./models:/work/models --env-file .env \
|
-v ./models:/work/models --env-file .env \
|
||||||
--gpus all \
|
--gpus all \
|
||||||
ghcr.io/tuna2134/sbv2-api:cuda
|
ghcr.io/neodyland/sbv2-api:cuda
|
||||||
```
|
```
|
||||||
|
|
||||||
### 起動確認
|
### 起動確認
|
||||||
|
|||||||
@@ -10,14 +10,14 @@ license.workspace = true
|
|||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
anyhow.workspace = true
|
anyhow.workspace = true
|
||||||
axum = "0.8.0"
|
axum = "0.8.6"
|
||||||
dotenvy.workspace = true
|
dotenvy.workspace = true
|
||||||
env_logger.workspace = true
|
env_logger.workspace = true
|
||||||
log = "0.4.22"
|
log = "0.4.28"
|
||||||
sbv2_core = { version = "0.2.0-alpha6", path = "../sbv2_core", features = ["aivmx"] }
|
sbv2_core = { version = "0.2.0-alpha6", path = "../sbv2_core", features = ["aivmx"] }
|
||||||
serde = { version = "1.0.210", features = ["derive"] }
|
serde = { version = "1.0.228", features = ["derive"] }
|
||||||
tokio = { version = "1.40.0", features = ["full"] }
|
tokio = { version = "1.47.1", features = ["full"] }
|
||||||
utoipa = { version = "5.0.0", features = ["axum_extras"] }
|
utoipa = { version = "5.4.0", features = ["axum_extras"] }
|
||||||
utoipa-scalar = { version = "0.3.0", features = ["axum"] }
|
utoipa-scalar = { version = "0.3.0", features = ["axum"] }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
|
|||||||
@@ -53,12 +53,16 @@ struct SynthesizeRequest {
|
|||||||
text: String,
|
text: String,
|
||||||
ident: String,
|
ident: String,
|
||||||
#[serde(default = "sdp_default")]
|
#[serde(default = "sdp_default")]
|
||||||
|
#[schema(example = 0.0_f32)]
|
||||||
sdp_ratio: f32,
|
sdp_ratio: f32,
|
||||||
#[serde(default = "length_default")]
|
#[serde(default = "length_default")]
|
||||||
|
#[schema(example = 1.0_f32)]
|
||||||
length_scale: f32,
|
length_scale: f32,
|
||||||
#[serde(default = "style_id_default")]
|
#[serde(default = "style_id_default")]
|
||||||
|
#[schema(example = 0_i32)]
|
||||||
style_id: i32,
|
style_id: i32,
|
||||||
#[serde(default = "speaker_id_default")]
|
#[serde(default = "speaker_id_default")]
|
||||||
|
#[schema(example = 0_i64)]
|
||||||
speaker_id: i64,
|
speaker_id: i64,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ crate-type = ["cdylib"]
|
|||||||
[dependencies]
|
[dependencies]
|
||||||
anyhow.workspace = true
|
anyhow.workspace = true
|
||||||
ndarray.workspace = true
|
ndarray.workspace = true
|
||||||
pyo3 = { version = "0.23.0", features = ["anyhow"] }
|
pyo3 = { version = "0.26.0", features = ["anyhow"] }
|
||||||
sbv2_core = { path = "../sbv2_core", features = ["std"], default-features = false }
|
sbv2_core = { path = "../sbv2_core", features = ["std"], default-features = false }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
|
|||||||
@@ -136,6 +136,7 @@ impl TTSModel {
|
|||||||
/// -------
|
/// -------
|
||||||
/// voice_data : bytes
|
/// voice_data : bytes
|
||||||
/// 音声データ
|
/// 音声データ
|
||||||
|
#[allow(clippy::too_many_arguments)]
|
||||||
fn synthesize<'p>(
|
fn synthesize<'p>(
|
||||||
&'p mut self,
|
&'p mut self,
|
||||||
py: Python<'p>,
|
py: Python<'p>,
|
||||||
|
|||||||
@@ -16,16 +16,16 @@ env_logger.workspace = true
|
|||||||
hound = "3.5.1"
|
hound = "3.5.1"
|
||||||
jpreprocess = { version = "0.12.0", features = ["naist-jdic"] }
|
jpreprocess = { version = "0.12.0", features = ["naist-jdic"] }
|
||||||
ndarray.workspace = true
|
ndarray.workspace = true
|
||||||
npyz = { version = "0.8.3", optional = true }
|
npyz = { version = "0.8.4", optional = true }
|
||||||
num_cpus = "1.16.0"
|
num_cpus = "1.17.0"
|
||||||
once_cell.workspace = true
|
once_cell.workspace = true
|
||||||
ort = { git = "https://github.com/pykeio/ort.git", version = "2.0.0-rc.9", optional = true }
|
ort = { git = "https://github.com/pykeio/ort.git", version = "2.0.0-rc.9", optional = true }
|
||||||
regex = "1.10.6"
|
regex = "1.12.1"
|
||||||
serde = { version = "1.0.210", features = ["derive"] }
|
serde = { version = "1.0.228", features = ["derive"] }
|
||||||
serde_json = "1.0.128"
|
serde_json = "1.0.145"
|
||||||
tar = "0.4.41"
|
tar = "0.4.41"
|
||||||
thiserror = "2.0.11"
|
thiserror = "2.0.17"
|
||||||
tokenizers = { version = "0.21.0", default-features = false }
|
tokenizers = { version = "0.22.1", default-features = false }
|
||||||
zstd = "0.13.2"
|
zstd = "0.13.2"
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
@@ -44,4 +44,4 @@ base64 = ["dep:base64"]
|
|||||||
|
|
||||||
[build-dependencies]
|
[build-dependencies]
|
||||||
dirs = "6.0.0"
|
dirs = "6.0.0"
|
||||||
ureq = "3.0.6"
|
ureq = "3.1.2"
|
||||||
|
|||||||
@@ -5,21 +5,27 @@ use std::io::copy;
|
|||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
fn main() -> Result<(), Box<dyn std::error::Error>> {
|
fn main() -> Result<(), Box<dyn std::error::Error>> {
|
||||||
let static_path = home_dir().unwrap().join(".cache/sbv2/all.bin");
|
let static_dir = home_dir().unwrap().join(".cache/sbv2");
|
||||||
|
let static_path = static_dir.join("all.bin");
|
||||||
let out_path = PathBuf::from(&env::var("OUT_DIR").unwrap()).join("all.bin");
|
let out_path = PathBuf::from(&env::var("OUT_DIR").unwrap()).join("all.bin");
|
||||||
println!("cargo:rerun-if-changed=build.rs");
|
println!("cargo:rerun-if-changed=build.rs");
|
||||||
if static_path.exists() {
|
if static_path.exists() {
|
||||||
if fs::hard_link(&static_path, &out_path).is_err() {
|
println!("cargo:info=Dictionary file already exists, skipping download.");
|
||||||
fs::copy(static_path, out_path).unwrap();
|
|
||||||
};
|
|
||||||
} else {
|
} else {
|
||||||
println!("cargo:warning=Downloading dictionary file...");
|
println!("cargo:warning=Downloading dictionary file...");
|
||||||
let mut response =
|
let mut response =
|
||||||
ureq::get("https://huggingface.co/neody/sbv2-api-assets/resolve/main/dic/all.bin")
|
ureq::get("https://huggingface.co/neody/sbv2-api-assets/resolve/main/dic/all.bin")
|
||||||
.call()?;
|
.call()?;
|
||||||
let mut response = response.body_mut().as_reader();
|
let mut response = response.body_mut().as_reader();
|
||||||
let mut file = fs::File::create(&out_path)?;
|
if !static_dir.exists() {
|
||||||
|
fs::create_dir_all(static_dir)?;
|
||||||
|
}
|
||||||
|
let mut file = fs::File::create(&static_path)?;
|
||||||
copy(&mut response, &mut file)?;
|
copy(&mut response, &mut file)?;
|
||||||
}
|
}
|
||||||
|
if !out_path.exists() && fs::hard_link(&static_path, &out_path).is_err() {
|
||||||
|
println!("cargo:warning=Failed to create hard link, copying instead.");
|
||||||
|
fs::copy(static_path, out_path)?;
|
||||||
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -14,11 +14,9 @@ pub fn predict(
|
|||||||
"attention_mask" => TensorRef::from_array_view((vec![1, attention_masks.len() as i64], attention_masks.as_slice()))?,
|
"attention_mask" => TensorRef::from_array_view((vec![1, attention_masks.len() as i64], attention_masks.as_slice()))?,
|
||||||
}
|
}
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
let output = outputs["output"]
|
let output = outputs["output"]
|
||||||
.try_extract_tensor::<f32>()?
|
.try_extract_array::<f32>()?
|
||||||
.into_dimensionality::<Ix2>()?
|
.into_dimensionality::<Ix2>()?
|
||||||
.to_owned();
|
.to_owned();
|
||||||
|
|
||||||
Ok(output)
|
Ok(output)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -28,6 +28,8 @@ pub enum Error {
|
|||||||
Base64Error(#[from] base64::DecodeError),
|
Base64Error(#[from] base64::DecodeError),
|
||||||
#[error("other")]
|
#[error("other")]
|
||||||
OtherError(String),
|
OtherError(String),
|
||||||
|
#[error("Style error: {0}")]
|
||||||
|
StyleError(String),
|
||||||
}
|
}
|
||||||
|
|
||||||
pub type Result<T> = std::result::Result<T, Error>;
|
pub type Result<T> = std::result::Result<T, Error>;
|
||||||
|
|||||||
@@ -1,5 +1,177 @@
|
|||||||
|
/*
|
||||||
|
このファイルのコードは
|
||||||
|
https://github.com/litagin02/Style-Bert-VITS2/blob/master/style_bert_vits2/nlp/japanese/g2p.py
|
||||||
|
を参考にRustに書き換えています。
|
||||||
|
|
||||||
|
以下はライセンスです。
|
||||||
|
GNU LESSER GENERAL PUBLIC LICENSE
|
||||||
|
Version 3, 29 June 2007
|
||||||
|
|
||||||
|
Copyright (C) 2007 Free Software Foundation, Inc. <https://fsf.org/>
|
||||||
|
Everyone is permitted to copy and distribute verbatim copies
|
||||||
|
of this license document, but changing it is not allowed.
|
||||||
|
|
||||||
|
|
||||||
|
This version of the GNU Lesser General Public License incorporates
|
||||||
|
the terms and conditions of version 3 of the GNU General Public
|
||||||
|
License, supplemented by the additional permissions listed below.
|
||||||
|
|
||||||
|
0. Additional Definitions.
|
||||||
|
|
||||||
|
As used herein, "this License" refers to version 3 of the GNU Lesser
|
||||||
|
General Public License, and the "GNU GPL" refers to version 3 of the GNU
|
||||||
|
General Public License.
|
||||||
|
|
||||||
|
"The Library" refers to a covered work governed by this License,
|
||||||
|
other than an Application or a Combined Work as defined below.
|
||||||
|
|
||||||
|
An "Application" is any work that makes use of an interface provided
|
||||||
|
by the Library, but which is not otherwise based on the Library.
|
||||||
|
Defining a subclass of a class defined by the Library is deemed a mode
|
||||||
|
of using an interface provided by the Library.
|
||||||
|
|
||||||
|
A "Combined Work" is a work produced by combining or linking an
|
||||||
|
Application with the Library. The particular version of the Library
|
||||||
|
with which the Combined Work was made is also called the "Linked
|
||||||
|
Version".
|
||||||
|
|
||||||
|
The "Minimal Corresponding Source" for a Combined Work means the
|
||||||
|
Corresponding Source for the Combined Work, excluding any source code
|
||||||
|
for portions of the Combined Work that, considered in isolation, are
|
||||||
|
based on the Application, and not on the Linked Version.
|
||||||
|
|
||||||
|
The "Corresponding Application Code" for a Combined Work means the
|
||||||
|
object code and/or source code for the Application, including any data
|
||||||
|
and utility programs needed for reproducing the Combined Work from the
|
||||||
|
Application, but excluding the System Libraries of the Combined Work.
|
||||||
|
|
||||||
|
1. Exception to Section 3 of the GNU GPL.
|
||||||
|
|
||||||
|
You may convey a covered work under sections 3 and 4 of this License
|
||||||
|
without being bound by section 3 of the GNU GPL.
|
||||||
|
|
||||||
|
2. Conveying Modified Versions.
|
||||||
|
|
||||||
|
If you modify a copy of the Library, and, in your modifications, a
|
||||||
|
facility refers to a function or data to be supplied by an Application
|
||||||
|
that uses the facility (other than as an argument passed when the
|
||||||
|
facility is invoked), then you may convey a copy of the modified
|
||||||
|
version:
|
||||||
|
|
||||||
|
a) under this License, provided that you make a good faith effort to
|
||||||
|
ensure that, in the event an Application does not supply the
|
||||||
|
function or data, the facility still operates, and performs
|
||||||
|
whatever part of its purpose remains meaningful, or
|
||||||
|
|
||||||
|
b) under the GNU GPL, with none of the additional permissions of
|
||||||
|
this License applicable to that copy.
|
||||||
|
|
||||||
|
3. Object Code Incorporating Material from Library Header Files.
|
||||||
|
|
||||||
|
The object code form of an Application may incorporate material from
|
||||||
|
a header file that is part of the Library. You may convey such object
|
||||||
|
code under terms of your choice, provided that, if the incorporated
|
||||||
|
material is not limited to numerical parameters, data structure
|
||||||
|
layouts and accessors, or small macros, inline functions and templates
|
||||||
|
(ten or fewer lines in length), you do both of the following:
|
||||||
|
|
||||||
|
a) Give prominent notice with each copy of the object code that the
|
||||||
|
Library is used in it and that the Library and its use are
|
||||||
|
covered by this License.
|
||||||
|
|
||||||
|
b) Accompany the object code with a copy of the GNU GPL and this license
|
||||||
|
document.
|
||||||
|
|
||||||
|
4. Combined Works.
|
||||||
|
|
||||||
|
You may convey a Combined Work under terms of your choice that,
|
||||||
|
taken together, effectively do not restrict modification of the
|
||||||
|
portions of the Library contained in the Combined Work and reverse
|
||||||
|
engineering for debugging such modifications, if you also do each of
|
||||||
|
the following:
|
||||||
|
|
||||||
|
a) Give prominent notice with each copy of the Combined Work that
|
||||||
|
the Library is used in it and that the Library and its use are
|
||||||
|
covered by this License.
|
||||||
|
|
||||||
|
b) Accompany the Combined Work with a copy of the GNU GPL and this license
|
||||||
|
document.
|
||||||
|
|
||||||
|
c) For a Combined Work that displays copyright notices during
|
||||||
|
execution, include the copyright notice for the Library among
|
||||||
|
these notices, as well as a reference directing the user to the
|
||||||
|
copies of the GNU GPL and this license document.
|
||||||
|
|
||||||
|
d) Do one of the following:
|
||||||
|
|
||||||
|
0) Convey the Minimal Corresponding Source under the terms of this
|
||||||
|
License, and the Corresponding Application Code in a form
|
||||||
|
suitable for, and under terms that permit, the user to
|
||||||
|
recombine or relink the Application with a modified version of
|
||||||
|
the Linked Version to produce a modified Combined Work, in the
|
||||||
|
manner specified by section 6 of the GNU GPL for conveying
|
||||||
|
Corresponding Source.
|
||||||
|
|
||||||
|
1) Use a suitable shared library mechanism for linking with the
|
||||||
|
Library. A suitable mechanism is one that (a) uses at run time
|
||||||
|
a copy of the Library already present on the user's computer
|
||||||
|
system, and (b) will operate properly with a modified version
|
||||||
|
of the Library that is interface-compatible with the Linked
|
||||||
|
Version.
|
||||||
|
|
||||||
|
e) Provide Installation Information, but only if you would otherwise
|
||||||
|
be required to provide such information under section 6 of the
|
||||||
|
GNU GPL, and only to the extent that such information is
|
||||||
|
necessary to install and execute a modified version of the
|
||||||
|
Combined Work produced by recombining or relinking the
|
||||||
|
Application with a modified version of the Linked Version. (If
|
||||||
|
you use option 4d0, the Installation Information must accompany
|
||||||
|
the Minimal Corresponding Source and Corresponding Application
|
||||||
|
Code. If you use option 4d1, you must provide the Installation
|
||||||
|
Information in the manner specified by section 6 of the GNU GPL
|
||||||
|
for conveying Corresponding Source.)
|
||||||
|
|
||||||
|
5. Combined Libraries.
|
||||||
|
|
||||||
|
You may place library facilities that are a work based on the
|
||||||
|
Library side by side in a single library together with other library
|
||||||
|
facilities that are not Applications and are not covered by this
|
||||||
|
License, and convey such a combined library under terms of your
|
||||||
|
choice, if you do both of the following:
|
||||||
|
|
||||||
|
a) Accompany the combined library with a copy of the same work based
|
||||||
|
on the Library, uncombined with any other library facilities,
|
||||||
|
conveyed under the terms of this License.
|
||||||
|
|
||||||
|
b) Give prominent notice with the combined library that part of it
|
||||||
|
is a work based on the Library, and explaining where to find the
|
||||||
|
accompanying uncombined form of the same work.
|
||||||
|
|
||||||
|
6. Revised Versions of the GNU Lesser General Public License.
|
||||||
|
|
||||||
|
The Free Software Foundation may publish revised and/or new versions
|
||||||
|
of the GNU Lesser General Public License from time to time. Such new
|
||||||
|
versions will be similar in spirit to the present version, but may
|
||||||
|
differ in detail to address new problems or concerns.
|
||||||
|
|
||||||
|
Each version is given a distinguishing version number. If the
|
||||||
|
Library as you received it specifies that a certain numbered version
|
||||||
|
of the GNU Lesser General Public License "or any later version"
|
||||||
|
applies to it, you have the option of following the terms and
|
||||||
|
conditions either of that published version or of any later version
|
||||||
|
published by the Free Software Foundation. If the Library as you
|
||||||
|
received it does not specify a version number of the GNU Lesser
|
||||||
|
General Public License, you may choose any version of the GNU Lesser
|
||||||
|
General Public License ever published by the Free Software Foundation.
|
||||||
|
|
||||||
|
If the Library as you received it specifies that a proxy can decide
|
||||||
|
whether future versions of the GNU Lesser General Public License shall
|
||||||
|
apply, that proxy's public statement of acceptance of any version is
|
||||||
|
permanent authorization for you to choose that version for the
|
||||||
|
Library.
|
||||||
|
*/
|
||||||
use crate::error::{Error, Result};
|
use crate::error::{Error, Result};
|
||||||
use crate::mora::{MORA_KATA_TO_MORA_PHONEMES, VOWELS};
|
use crate::mora::{CONSONANTS, MORA_KATA_TO_MORA_PHONEMES, MORA_PHONEMES_TO_MORA_KATA, VOWELS};
|
||||||
use crate::norm::{replace_punctuation, PUNCTUATIONS};
|
use crate::norm::{replace_punctuation, PUNCTUATIONS};
|
||||||
use jpreprocess::{kind, DefaultTokenizer, JPreprocess, SystemDictionaryConfig, UserDictionary};
|
use jpreprocess::{kind, DefaultTokenizer, JPreprocess, SystemDictionaryConfig, UserDictionary};
|
||||||
use once_cell::sync::Lazy;
|
use once_cell::sync::Lazy;
|
||||||
@@ -76,6 +248,34 @@ static MORA_PATTERN: Lazy<Vec<String>> = Lazy::new(|| {
|
|||||||
});
|
});
|
||||||
static LONG_PATTERN: Lazy<Regex> = Lazy::new(|| Regex::new(r"(\w)(ー*)").unwrap());
|
static LONG_PATTERN: Lazy<Regex> = Lazy::new(|| Regex::new(r"(\w)(ー*)").unwrap());
|
||||||
|
|
||||||
|
fn phone_tone_to_kana(phones: Vec<String>, tones: Vec<i32>) -> Vec<(String, i32)> {
|
||||||
|
let phones = &phones[1..];
|
||||||
|
let tones = &tones[1..];
|
||||||
|
let mut results = Vec::new();
|
||||||
|
let mut current_mora = String::new();
|
||||||
|
for ((phone, _next_phone), (&tone, &next_tone)) in phones
|
||||||
|
.iter()
|
||||||
|
.zip(phones.iter().skip(1))
|
||||||
|
.zip(tones.iter().zip(tones.iter().skip(1)))
|
||||||
|
{
|
||||||
|
if PUNCTUATIONS.contains(&phone.clone().as_str()) {
|
||||||
|
results.push((phone.to_string(), tone));
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
if CONSONANTS.contains(&phone.clone()) {
|
||||||
|
assert_eq!(current_mora, "");
|
||||||
|
assert_eq!(tone, next_tone);
|
||||||
|
current_mora = phone.to_string()
|
||||||
|
} else {
|
||||||
|
current_mora += phone;
|
||||||
|
let kana = MORA_PHONEMES_TO_MORA_KATA.get(¤t_mora).unwrap();
|
||||||
|
results.push((kana.to_string(), tone));
|
||||||
|
current_mora = String::new();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
results
|
||||||
|
}
|
||||||
|
|
||||||
pub struct JTalkProcess {
|
pub struct JTalkProcess {
|
||||||
jpreprocess: Arc<JPreprocessType>,
|
jpreprocess: Arc<JPreprocessType>,
|
||||||
parsed: Vec<String>,
|
parsed: Vec<String>,
|
||||||
@@ -95,24 +295,24 @@ impl JTalkProcess {
|
|||||||
.map(|(_letter, tone)| *tone)
|
.map(|(_letter, tone)| *tone)
|
||||||
.collect();
|
.collect();
|
||||||
if tone_values.len() == 1 {
|
if tone_values.len() == 1 {
|
||||||
assert!(tone_values == hash_set![0], "{:?}", tone_values);
|
assert!(tone_values == hash_set![0], "{tone_values:?}");
|
||||||
Ok(phone_tone_list)
|
Ok(phone_tone_list)
|
||||||
} else if tone_values.len() == 2 {
|
} else if tone_values.len() == 2 {
|
||||||
if tone_values == hash_set![0, 1] {
|
if tone_values == hash_set![0, 1] {
|
||||||
return Ok(phone_tone_list);
|
Ok(phone_tone_list)
|
||||||
} else if tone_values == hash_set![-1, 0] {
|
} else if tone_values == hash_set![-1, 0] {
|
||||||
return Ok(phone_tone_list
|
Ok(phone_tone_list
|
||||||
.iter()
|
.iter()
|
||||||
.map(|x| {
|
.map(|x| {
|
||||||
let new_tone = if x.1 == -1 { 0 } else { 1 };
|
let new_tone = if x.1 == -1 { 0 } else { 1 };
|
||||||
(x.0.clone(), new_tone)
|
(x.0.clone(), new_tone)
|
||||||
})
|
})
|
||||||
.collect());
|
.collect())
|
||||||
} else {
|
} else {
|
||||||
return Err(Error::ValueError("Invalid tone values 0".to_string()));
|
Err(Error::ValueError("Invalid tone values 0".to_string()))
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
return Err(Error::ValueError("Invalid tone values 1".to_string()));
|
Err(Error::ValueError("Invalid tone values 1".to_string()))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -165,6 +365,11 @@ impl JTalkProcess {
|
|||||||
Ok((phones, tones, new_word2ph))
|
Ok((phones, tones, new_word2ph))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn g2kana_tone(&self) -> Result<Vec<(String, i32)>> {
|
||||||
|
let (phones, tones, _) = self.g2p()?;
|
||||||
|
Ok(phone_tone_to_kana(phones, tones))
|
||||||
|
}
|
||||||
|
|
||||||
fn distribute_phone(n_phone: i32, n_word: i32) -> Vec<i32> {
|
fn distribute_phone(n_phone: i32, n_word: i32) -> Vec<i32> {
|
||||||
let mut phones_per_word = vec![0; n_word as usize];
|
let mut phones_per_word = vec![0; n_word as usize];
|
||||||
for _ in 0..n_phone {
|
for _ in 0..n_phone {
|
||||||
@@ -193,12 +398,12 @@ impl JTalkProcess {
|
|||||||
} else if PUNCTUATIONS.contains(&phone.as_str()) {
|
} else if PUNCTUATIONS.contains(&phone.as_str()) {
|
||||||
result.push((phone, 0));
|
result.push((phone, 0));
|
||||||
} else {
|
} else {
|
||||||
println!("phones {:?}", phone_with_punct);
|
println!("phones {phone_with_punct:?}");
|
||||||
println!("phone_tone_list: {:?}", phone_tone_list);
|
println!("phone_tone_list: {phone_tone_list:?}");
|
||||||
println!("result: {:?}", result);
|
println!("result: {result:?}");
|
||||||
println!("tone_index: {:?}", tone_index);
|
println!("tone_index: {tone_index:?}");
|
||||||
println!("phone: {:?}", phone);
|
println!("phone: {phone:?}");
|
||||||
return Err(Error::ValueError(format!("Mismatched phoneme: {}", phone)));
|
return Err(Error::ValueError(format!("Mismatched phoneme: {phone}")));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -243,8 +448,7 @@ impl JTalkProcess {
|
|||||||
}
|
}
|
||||||
if !KATAKANA_PATTERN.is_match(&text) {
|
if !KATAKANA_PATTERN.is_match(&text) {
|
||||||
return Err(Error::ValueError(format!(
|
return Err(Error::ValueError(format!(
|
||||||
"Input must be katakana only: {}",
|
"Input must be katakana only: {text}"
|
||||||
text
|
|
||||||
)));
|
)));
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -252,7 +456,7 @@ impl JTalkProcess {
|
|||||||
let mora = mora.to_string();
|
let mora = mora.to_string();
|
||||||
let (consonant, vowel) = MORA_KATA_TO_MORA_PHONEMES.get(&mora).unwrap();
|
let (consonant, vowel) = MORA_KATA_TO_MORA_PHONEMES.get(&mora).unwrap();
|
||||||
if consonant.is_none() {
|
if consonant.is_none() {
|
||||||
text = text.replace(&mora, &format!(" {}", vowel));
|
text = text.replace(&mora, &format!(" {vowel}"));
|
||||||
} else {
|
} else {
|
||||||
text = text.replace(
|
text = text.replace(
|
||||||
&mora,
|
&mora,
|
||||||
@@ -286,7 +490,7 @@ impl JTalkProcess {
|
|||||||
let (string, pron) = self.parse_to_string_and_pron(parts.clone());
|
let (string, pron) = self.parse_to_string_and_pron(parts.clone());
|
||||||
let mut yomi = pron.replace('’', "");
|
let mut yomi = pron.replace('’', "");
|
||||||
let word = replace_punctuation(string);
|
let word = replace_punctuation(string);
|
||||||
assert!(!yomi.is_empty(), "Empty yomi: {}", word);
|
assert!(!yomi.is_empty(), "Empty yomi: {word}");
|
||||||
if yomi == "、" {
|
if yomi == "、" {
|
||||||
if !word
|
if !word
|
||||||
.chars()
|
.chars()
|
||||||
@@ -297,7 +501,7 @@ impl JTalkProcess {
|
|||||||
yomi = word.clone();
|
yomi = word.clone();
|
||||||
}
|
}
|
||||||
} else if yomi == "?" {
|
} else if yomi == "?" {
|
||||||
assert!(word == "?", "yomi `?` comes from: {}", word);
|
assert!(word == "?", "yomi `?` comes from: {word}");
|
||||||
yomi = "?".to_string();
|
yomi = "?".to_string();
|
||||||
}
|
}
|
||||||
seq_text.push(word);
|
seq_text.push(word);
|
||||||
|
|||||||
@@ -30,8 +30,7 @@ fn main_inner() -> anyhow::Result<()> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let audio =
|
let audio = tts_holder.easy_synthesize(ident, text, 0, 0, tts::SynthesizeOptions::default())?;
|
||||||
tts_holder.easy_synthesize(ident, &text, 0, 0, tts::SynthesizeOptions::default())?;
|
|
||||||
fs::write("output.wav", audio)?;
|
fs::write("output.wav", audio)?;
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
|
|||||||
@@ -21,10 +21,7 @@ pub fn load_model<P: AsRef<[u8]>>(model_file: P, bert: bool) -> Result<Session>
|
|||||||
#[cfg(feature = "cuda")]
|
#[cfg(feature = "cuda")]
|
||||||
{
|
{
|
||||||
#[allow(unused_mut)]
|
#[allow(unused_mut)]
|
||||||
let mut cuda = ort::execution_providers::CUDAExecutionProvider::default()
|
let mut cuda = ort::execution_providers::CUDAExecutionProvider::default();
|
||||||
.with_conv_algorithm_search(
|
|
||||||
ort::execution_providers::cuda::CUDAExecutionProviderCuDNNConvAlgoSearch::Default,
|
|
||||||
);
|
|
||||||
#[cfg(feature = "cuda_tf32")]
|
#[cfg(feature = "cuda_tf32")]
|
||||||
{
|
{
|
||||||
cuda = cuda.with_tf32(true);
|
cuda = cuda.with_tf32(true);
|
||||||
@@ -101,11 +98,9 @@ pub fn synthesize(
|
|||||||
"noise_scale" => noise_scale,
|
"noise_scale" => noise_scale,
|
||||||
"noise_scale_w" => noise_scale_w,
|
"noise_scale_w" => noise_scale_w,
|
||||||
})?;
|
})?;
|
||||||
|
|
||||||
let audio_array = outputs["output"]
|
let audio_array = outputs["output"]
|
||||||
.try_extract_tensor::<f32>()?
|
.try_extract_array::<f32>()?
|
||||||
.into_dimensionality::<Ix3>()?
|
.into_dimensionality::<Ix3>()?
|
||||||
.to_owned();
|
.to_owned();
|
||||||
|
|
||||||
Ok(audio_array)
|
Ok(audio_array)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -25,6 +25,21 @@ static MORA_LIST_ADDITIONAL: Lazy<Vec<Mora>> = Lazy::new(|| {
|
|||||||
data.additional
|
data.additional
|
||||||
});
|
});
|
||||||
|
|
||||||
|
pub static MORA_PHONEMES_TO_MORA_KATA: Lazy<HashMap<String, String>> = Lazy::new(|| {
|
||||||
|
let mut map = HashMap::new();
|
||||||
|
for mora in MORA_LIST_MINIMUM.iter() {
|
||||||
|
map.insert(
|
||||||
|
format!(
|
||||||
|
"{}{}",
|
||||||
|
mora.consonant.clone().unwrap_or("".to_string()),
|
||||||
|
mora.vowel
|
||||||
|
),
|
||||||
|
mora.mora.clone(),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
map
|
||||||
|
});
|
||||||
|
|
||||||
pub static MORA_KATA_TO_MORA_PHONEMES: Lazy<HashMap<String, (Option<String>, String)>> =
|
pub static MORA_KATA_TO_MORA_PHONEMES: Lazy<HashMap<String, (Option<String>, String)>> =
|
||||||
Lazy::new(|| {
|
Lazy::new(|| {
|
||||||
let mut map = HashMap::new();
|
let mut map = HashMap::new();
|
||||||
@@ -37,4 +52,12 @@ pub static MORA_KATA_TO_MORA_PHONEMES: Lazy<HashMap<String, (Option<String>, Str
|
|||||||
map
|
map
|
||||||
});
|
});
|
||||||
|
|
||||||
|
pub static CONSONANTS: Lazy<Vec<String>> = Lazy::new(|| {
|
||||||
|
let consonants = MORA_KATA_TO_MORA_PHONEMES
|
||||||
|
.values()
|
||||||
|
.filter_map(|(consonant, _)| consonant.clone())
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
consonants
|
||||||
|
});
|
||||||
|
|
||||||
pub const VOWELS: [&str; 6] = ["a", "i", "u", "e", "o", "N"];
|
pub const VOWELS: [&str; 6] = ["a", "i", "u", "e", "o", "N"];
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
use crate::error::Result;
|
use crate::error::{Error, Result};
|
||||||
use ndarray::{s, Array1, Array2};
|
use ndarray::{s, Array1, Array2};
|
||||||
use serde::Deserialize;
|
use serde::Deserialize;
|
||||||
|
|
||||||
@@ -21,6 +21,18 @@ pub fn get_style_vector(
|
|||||||
style_id: i32,
|
style_id: i32,
|
||||||
weight: f32,
|
weight: f32,
|
||||||
) -> Result<Array1<f32>> {
|
) -> Result<Array1<f32>> {
|
||||||
|
if style_vectors.shape().len() != 2 {
|
||||||
|
return Err(Error::StyleError(
|
||||||
|
"Invalid shape for style vectors".to_string(),
|
||||||
|
));
|
||||||
|
}
|
||||||
|
if style_id < 0 || style_id >= style_vectors.shape()[0] as i32 {
|
||||||
|
return Err(Error::StyleError(format!(
|
||||||
|
"Invalid style ID: {}. Max ID: {}",
|
||||||
|
style_id,
|
||||||
|
style_vectors.shape()[0] - 1
|
||||||
|
)));
|
||||||
|
}
|
||||||
let mean = style_vectors.slice(s![0, ..]).to_owned();
|
let mean = style_vectors.slice(s![0, ..]).to_owned();
|
||||||
let style_vector = style_vectors.slice(s![style_id as usize, ..]).to_owned();
|
let style_vector = style_vectors.slice(s![style_id as usize, ..]).to_owned();
|
||||||
let diff = (style_vector - &mean) * weight;
|
let diff = (style_vector - &mean) * weight;
|
||||||
|
|||||||
@@ -41,7 +41,7 @@ pub struct TTSModelHolder {
|
|||||||
tokenizer: Tokenizer,
|
tokenizer: Tokenizer,
|
||||||
bert: Session,
|
bert: Session,
|
||||||
models: Vec<TTSModel>,
|
models: Vec<TTSModel>,
|
||||||
jtalk: jtalk::JTalk,
|
pub jtalk: jtalk::JTalk,
|
||||||
max_loaded_models: Option<usize>,
|
max_loaded_models: Option<usize>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -91,7 +91,7 @@ impl TTSModelHolder {
|
|||||||
}
|
}
|
||||||
let model = model::load_model(&aivmx_bytes, false)?;
|
let model = model::load_model(&aivmx_bytes, false)?;
|
||||||
let metadata = model.metadata()?;
|
let metadata = model.metadata()?;
|
||||||
if let Some(aivm_style_vectors) = metadata.custom("aivm_style_vectors")? {
|
if let Some(aivm_style_vectors) = metadata.custom("aivm_style_vectors") {
|
||||||
let aivm_style_vectors = BASE64_STANDARD.decode(aivm_style_vectors)?;
|
let aivm_style_vectors = BASE64_STANDARD.decode(aivm_style_vectors)?;
|
||||||
let style_vectors = Cursor::new(&aivm_style_vectors);
|
let style_vectors = Cursor::new(&aivm_style_vectors);
|
||||||
let reader = npyz::NpyFile::new(style_vectors)?;
|
let reader = npyz::NpyFile::new(style_vectors)?;
|
||||||
@@ -205,6 +205,24 @@ impl TTSModelHolder {
|
|||||||
) -> Result<(Array2<f32>, Array1<i64>, Array1<i64>, Array1<i64>)> {
|
) -> Result<(Array2<f32>, Array1<i64>, Array1<i64>, Array1<i64>)> {
|
||||||
crate::tts_util::parse_text_blocking(
|
crate::tts_util::parse_text_blocking(
|
||||||
text,
|
text,
|
||||||
|
None,
|
||||||
|
&self.jtalk,
|
||||||
|
&self.tokenizer,
|
||||||
|
|token_ids, attention_masks| {
|
||||||
|
crate::bert::predict(&mut self.bert, token_ids, attention_masks)
|
||||||
|
},
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
#[allow(clippy::type_complexity)]
|
||||||
|
pub fn parse_text_neo(
|
||||||
|
&mut self,
|
||||||
|
text: String,
|
||||||
|
given_tones: Option<Vec<i32>>,
|
||||||
|
) -> Result<(Array2<f32>, Array1<i64>, Array1<i64>, Array1<i64>)> {
|
||||||
|
crate::tts_util::parse_text_blocking(
|
||||||
|
&text,
|
||||||
|
given_tones,
|
||||||
&self.jtalk,
|
&self.jtalk,
|
||||||
&self.tokenizer,
|
&self.tokenizer,
|
||||||
|token_ids, attention_masks| {
|
|token_ids, attention_masks| {
|
||||||
@@ -222,39 +240,43 @@ impl TTSModelHolder {
|
|||||||
}
|
}
|
||||||
fn find_and_load_model<I: Into<TTSIdent>>(&mut self, ident: I) -> Result<bool> {
|
fn find_and_load_model<I: Into<TTSIdent>>(&mut self, ident: I) -> Result<bool> {
|
||||||
let ident = ident.into();
|
let ident = ident.into();
|
||||||
let (bytes, style_vectors) = {
|
// Locate target model entry
|
||||||
let model = self
|
let target_index = self
|
||||||
.models
|
|
||||||
.iter()
|
|
||||||
.find(|m| m.ident == ident)
|
|
||||||
.ok_or(Error::ModelNotFoundError(ident.to_string()))?;
|
|
||||||
if model.vits2.is_some() {
|
|
||||||
return Ok(true);
|
|
||||||
}
|
|
||||||
(model.bytes.clone().unwrap(), model.style_vectors.clone())
|
|
||||||
};
|
|
||||||
self.unload(ident.clone());
|
|
||||||
let s = model::load_model(&bytes, false)?;
|
|
||||||
if let Some(max) = self.max_loaded_models {
|
|
||||||
if self.models.iter().filter(|x| x.vits2.is_some()).count() >= max {
|
|
||||||
self.unload(self.models.first().unwrap().ident.clone());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
self.models.push(TTSModel {
|
|
||||||
bytes: Some(bytes.to_vec()),
|
|
||||||
vits2: Some(s),
|
|
||||||
style_vectors,
|
|
||||||
ident: ident.clone(),
|
|
||||||
});
|
|
||||||
let model = self
|
|
||||||
.models
|
.models
|
||||||
.iter()
|
.iter()
|
||||||
.find(|m| m.ident == ident)
|
.position(|m| m.ident == ident)
|
||||||
.ok_or(Error::ModelNotFoundError(ident.to_string()))?;
|
.ok_or(Error::ModelNotFoundError(ident.to_string()))?;
|
||||||
if model.vits2.is_some() {
|
|
||||||
|
// Already loaded
|
||||||
|
if self.models[target_index].vits2.is_some() {
|
||||||
return Ok(true);
|
return Ok(true);
|
||||||
}
|
}
|
||||||
Err(Error::ModelNotFoundError(ident.to_string()))
|
|
||||||
|
// Get bytes to build a Session
|
||||||
|
let bytes = self.models[target_index]
|
||||||
|
.bytes
|
||||||
|
.clone()
|
||||||
|
.ok_or(Error::ModelNotFoundError(ident.to_string()))?;
|
||||||
|
|
||||||
|
// Enforce max loaded models by evicting a different loaded model's session, not removing the entry
|
||||||
|
if let Some(max) = self.max_loaded_models {
|
||||||
|
let loaded_count = self.models.iter().filter(|m| m.vits2.is_some()).count();
|
||||||
|
if loaded_count >= max {
|
||||||
|
if let Some(evict_index) = self
|
||||||
|
.models
|
||||||
|
.iter()
|
||||||
|
.position(|m| m.vits2.is_some() && m.ident != ident)
|
||||||
|
{
|
||||||
|
// Drop only the session to free memory; keep bytes/style for future reload
|
||||||
|
self.models[evict_index].vits2 = None;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Build and set session in-place for the target model
|
||||||
|
let s = model::load_model(&bytes, false)?;
|
||||||
|
self.models[target_index].vits2 = Some(s);
|
||||||
|
Ok(true)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get style vector by style id and weight
|
/// Get style vector by style id and weight
|
||||||
@@ -347,6 +369,79 @@ impl TTSModelHolder {
|
|||||||
};
|
};
|
||||||
tts_util::array_to_vec(audio_array)
|
tts_util::array_to_vec(audio_array)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn easy_synthesize_neo<I: Into<TTSIdent> + Copy>(
|
||||||
|
&mut self,
|
||||||
|
ident: I,
|
||||||
|
text: &str,
|
||||||
|
given_tones: Option<Vec<i32>>,
|
||||||
|
style_id: i32,
|
||||||
|
speaker_id: i64,
|
||||||
|
options: SynthesizeOptions,
|
||||||
|
) -> Result<Vec<u8>> {
|
||||||
|
self.find_and_load_model(ident)?;
|
||||||
|
let style_vector = self.get_style_vector(ident, style_id, options.style_weight)?;
|
||||||
|
let audio_array = if options.split_sentences {
|
||||||
|
let texts: Vec<&str> = text.split('\n').collect();
|
||||||
|
let mut audios = vec![];
|
||||||
|
for (i, t) in texts.iter().enumerate() {
|
||||||
|
if t.is_empty() {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
let (bert_ori, phones, tones, lang_ids) =
|
||||||
|
self.parse_text_neo(t.to_string(), given_tones.clone())?;
|
||||||
|
|
||||||
|
let vits2 = self
|
||||||
|
.find_model(ident)?
|
||||||
|
.vits2
|
||||||
|
.as_mut()
|
||||||
|
.ok_or(Error::ModelNotFoundError(ident.into().to_string()))?;
|
||||||
|
let audio = model::synthesize(
|
||||||
|
vits2,
|
||||||
|
bert_ori.to_owned(),
|
||||||
|
phones,
|
||||||
|
Array1::from_vec(vec![speaker_id]),
|
||||||
|
tones,
|
||||||
|
lang_ids,
|
||||||
|
style_vector.clone(),
|
||||||
|
options.sdp_ratio,
|
||||||
|
options.length_scale,
|
||||||
|
0.677,
|
||||||
|
0.8,
|
||||||
|
)?;
|
||||||
|
audios.push(audio.clone());
|
||||||
|
if i != texts.len() - 1 {
|
||||||
|
audios.push(Array3::zeros((1, 1, 22050)));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
concatenate(
|
||||||
|
Axis(2),
|
||||||
|
&audios.iter().map(|x| x.view()).collect::<Vec<_>>(),
|
||||||
|
)?
|
||||||
|
} else {
|
||||||
|
let (bert_ori, phones, tones, lang_ids) = self.parse_text(text)?;
|
||||||
|
|
||||||
|
let vits2 = self
|
||||||
|
.find_model(ident)?
|
||||||
|
.vits2
|
||||||
|
.as_mut()
|
||||||
|
.ok_or(Error::ModelNotFoundError(ident.into().to_string()))?;
|
||||||
|
model::synthesize(
|
||||||
|
vits2,
|
||||||
|
bert_ori.to_owned(),
|
||||||
|
phones,
|
||||||
|
Array1::from_vec(vec![speaker_id]),
|
||||||
|
tones,
|
||||||
|
lang_ids,
|
||||||
|
style_vector,
|
||||||
|
options.sdp_ratio,
|
||||||
|
options.length_scale,
|
||||||
|
0.677,
|
||||||
|
0.8,
|
||||||
|
)?
|
||||||
|
};
|
||||||
|
tts_util::array_to_vec(audio_array)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Synthesize options
|
/// Synthesize options
|
||||||
|
|||||||
@@ -1,10 +1,22 @@
|
|||||||
use std::io::Cursor;
|
use std::io::Cursor;
|
||||||
|
|
||||||
use crate::error::Result;
|
use crate::error::Result;
|
||||||
|
use crate::jtalk::JTalkProcess;
|
||||||
|
use crate::mora::MORA_KATA_TO_MORA_PHONEMES;
|
||||||
|
use crate::norm::PUNCTUATIONS;
|
||||||
use crate::{jtalk, nlp, norm, tokenizer, utils};
|
use crate::{jtalk, nlp, norm, tokenizer, utils};
|
||||||
use hound::{SampleFormat, WavSpec, WavWriter};
|
use hound::{SampleFormat, WavSpec, WavWriter};
|
||||||
use ndarray::{concatenate, s, Array, Array1, Array2, Array3, Axis};
|
use ndarray::{concatenate, s, Array, Array1, Array2, Array3, Axis};
|
||||||
use tokenizers::Tokenizer;
|
use tokenizers::Tokenizer;
|
||||||
|
|
||||||
|
pub fn preprocess_parse_text(text: &str, jtalk: &jtalk::JTalk) -> Result<(String, JTalkProcess)> {
|
||||||
|
let text = jtalk.num2word(text)?;
|
||||||
|
let normalized_text = norm::normalize_text(&text);
|
||||||
|
|
||||||
|
let process = jtalk.process_text(&normalized_text)?;
|
||||||
|
Ok((normalized_text, process))
|
||||||
|
}
|
||||||
|
|
||||||
/// Parse text and return the input for synthesize
|
/// Parse text and return the input for synthesize
|
||||||
///
|
///
|
||||||
/// # Note
|
/// # Note
|
||||||
@@ -21,13 +33,9 @@ pub async fn parse_text(
|
|||||||
Box<dyn std::future::Future<Output = Result<ndarray::Array2<f32>>>>,
|
Box<dyn std::future::Future<Output = Result<ndarray::Array2<f32>>>>,
|
||||||
>,
|
>,
|
||||||
) -> Result<(Array2<f32>, Array1<i64>, Array1<i64>, Array1<i64>)> {
|
) -> Result<(Array2<f32>, Array1<i64>, Array1<i64>, Array1<i64>)> {
|
||||||
let text = jtalk.num2word(text)?;
|
let (normalized_text, process) = preprocess_parse_text(text, jtalk)?;
|
||||||
let normalized_text = norm::normalize_text(&text);
|
|
||||||
|
|
||||||
let process = jtalk.process_text(&normalized_text)?;
|
|
||||||
let (phones, tones, mut word2ph) = process.g2p()?;
|
let (phones, tones, mut word2ph) = process.g2p()?;
|
||||||
let (phones, tones, lang_ids) = nlp::cleaned_text_to_sequence(phones, tones);
|
let (phones, tones, lang_ids) = nlp::cleaned_text_to_sequence(phones, tones);
|
||||||
|
|
||||||
let phones = utils::intersperse(&phones, 0);
|
let phones = utils::intersperse(&phones, 0);
|
||||||
let tones = utils::intersperse(&tones, 0);
|
let tones = utils::intersperse(&tones, 0);
|
||||||
let lang_ids = utils::intersperse(&lang_ids, 0);
|
let lang_ids = utils::intersperse(&lang_ids, 0);
|
||||||
@@ -92,6 +100,7 @@ pub async fn parse_text(
|
|||||||
#[allow(clippy::type_complexity)]
|
#[allow(clippy::type_complexity)]
|
||||||
pub fn parse_text_blocking(
|
pub fn parse_text_blocking(
|
||||||
text: &str,
|
text: &str,
|
||||||
|
given_tones: Option<Vec<i32>>,
|
||||||
jtalk: &jtalk::JTalk,
|
jtalk: &jtalk::JTalk,
|
||||||
tokenizer: &Tokenizer,
|
tokenizer: &Tokenizer,
|
||||||
bert_predict: impl FnOnce(Vec<i64>, Vec<i64>) -> Result<ndarray::Array2<f32>>,
|
bert_predict: impl FnOnce(Vec<i64>, Vec<i64>) -> Result<ndarray::Array2<f32>>,
|
||||||
@@ -100,7 +109,10 @@ pub fn parse_text_blocking(
|
|||||||
let normalized_text = norm::normalize_text(&text);
|
let normalized_text = norm::normalize_text(&text);
|
||||||
|
|
||||||
let process = jtalk.process_text(&normalized_text)?;
|
let process = jtalk.process_text(&normalized_text)?;
|
||||||
let (phones, tones, mut word2ph) = process.g2p()?;
|
let (phones, mut tones, mut word2ph) = process.g2p()?;
|
||||||
|
if let Some(given_tones) = given_tones {
|
||||||
|
tones = given_tones;
|
||||||
|
}
|
||||||
let (phones, tones, lang_ids) = nlp::cleaned_text_to_sequence(phones, tones);
|
let (phones, tones, lang_ids) = nlp::cleaned_text_to_sequence(phones, tones);
|
||||||
|
|
||||||
let phones = utils::intersperse(&phones, 0);
|
let phones = utils::intersperse(&phones, 0);
|
||||||
@@ -161,8 +173,15 @@ pub fn parse_text_blocking(
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn array_to_vec(audio_array: Array3<f32>) -> Result<Vec<u8>> {
|
pub fn array_to_vec(audio_array: Array3<f32>) -> Result<Vec<u8>> {
|
||||||
|
// If SBV2_FORCE_STEREO is set ("1"/"true"), duplicate mono to stereo
|
||||||
|
let force_stereo = std::env::var("SBV2_FORCE_STEREO")
|
||||||
|
.ok()
|
||||||
|
.map(|v| matches!(v.as_str(), "1" | "true" | "TRUE" | "True"))
|
||||||
|
.unwrap_or(false);
|
||||||
|
|
||||||
|
let channels: u16 = if force_stereo { 2 } else { 1 };
|
||||||
let spec = WavSpec {
|
let spec = WavSpec {
|
||||||
channels: 1,
|
channels,
|
||||||
sample_rate: 44100,
|
sample_rate: 44100,
|
||||||
bits_per_sample: 32,
|
bits_per_sample: 32,
|
||||||
sample_format: SampleFormat::Float,
|
sample_format: SampleFormat::Float,
|
||||||
@@ -171,10 +190,38 @@ pub fn array_to_vec(audio_array: Array3<f32>) -> Result<Vec<u8>> {
|
|||||||
let mut writer = WavWriter::new(&mut cursor, spec)?;
|
let mut writer = WavWriter::new(&mut cursor, spec)?;
|
||||||
for i in 0..audio_array.shape()[0] {
|
for i in 0..audio_array.shape()[0] {
|
||||||
let output = audio_array.slice(s![i, 0, ..]).to_vec();
|
let output = audio_array.slice(s![i, 0, ..]).to_vec();
|
||||||
for sample in output {
|
if force_stereo {
|
||||||
writer.write_sample(sample)?;
|
for sample in output {
|
||||||
|
// Write to Left and Right channels
|
||||||
|
writer.write_sample(sample)?;
|
||||||
|
writer.write_sample(sample)?;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
for sample in output {
|
||||||
|
writer.write_sample(sample)?;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
writer.finalize()?;
|
writer.finalize()?;
|
||||||
Ok(cursor.into_inner())
|
Ok(cursor.into_inner())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn kata_tone2phone_tone(kata_tone: Vec<(String, i32)>) -> Vec<(String, i32)> {
|
||||||
|
let mut results = vec![("_".to_string(), 0)];
|
||||||
|
for (mora, tone) in kata_tone {
|
||||||
|
if PUNCTUATIONS.contains(&mora.as_str()) {
|
||||||
|
results.push((mora, 0));
|
||||||
|
continue;
|
||||||
|
} else {
|
||||||
|
let (consonant, vowel) = MORA_KATA_TO_MORA_PHONEMES.get(&mora).unwrap();
|
||||||
|
if let Some(consonant) = consonant {
|
||||||
|
results.push((consonant.to_string(), tone));
|
||||||
|
results.push((vowel.to_string(), tone));
|
||||||
|
} else {
|
||||||
|
results.push((vowel.to_string(), tone));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
results.push(("_".to_string(), 0));
|
||||||
|
results
|
||||||
|
}
|
||||||
|
|||||||
19
crates/sbv2_editor/Cargo.toml
Normal file
19
crates/sbv2_editor/Cargo.toml
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
[package]
|
||||||
|
name = "sbv2_editor"
|
||||||
|
version.workspace = true
|
||||||
|
edition.workspace = true
|
||||||
|
description.workspace = true
|
||||||
|
license.workspace = true
|
||||||
|
readme.workspace = true
|
||||||
|
repository.workspace = true
|
||||||
|
documentation.workspace = true
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
anyhow.workspace = true
|
||||||
|
axum = "0.8.1"
|
||||||
|
dotenvy.workspace = true
|
||||||
|
env_logger.workspace = true
|
||||||
|
log = "0.4.27"
|
||||||
|
sbv2_core = { version = "0.2.0-alpha6", path = "../sbv2_core", features = ["aivmx"] }
|
||||||
|
serde = { version = "1.0.219", features = ["derive"] }
|
||||||
|
tokio = { version = "1.44.1", features = ["full"] }
|
||||||
2
crates/sbv2_editor/README.md
Normal file
2
crates/sbv2_editor/README.md
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
# sbv2-voicevox
|
||||||
|
sbv2-apiをvoicevox化します。
|
||||||
226
crates/sbv2_editor/query2.json
Normal file
226
crates/sbv2_editor/query2.json
Normal file
@@ -0,0 +1,226 @@
|
|||||||
|
{
|
||||||
|
"accent_phrases": [
|
||||||
|
{
|
||||||
|
"moras": [
|
||||||
|
{
|
||||||
|
"text": "コ",
|
||||||
|
"consonant": "k",
|
||||||
|
"consonant_length": 0.10002632439136505,
|
||||||
|
"vowel": "o",
|
||||||
|
"vowel_length": 0.15740256011486053,
|
||||||
|
"pitch": 5.749961853027344
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "ン",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "N",
|
||||||
|
"vowel_length": 0.08265873789787292,
|
||||||
|
"pitch": 5.89122200012207
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "ニ",
|
||||||
|
"consonant": "n",
|
||||||
|
"consonant_length": 0.03657080978155136,
|
||||||
|
"vowel": "i",
|
||||||
|
"vowel_length": 0.1175866425037384,
|
||||||
|
"pitch": 5.969866752624512
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "チ",
|
||||||
|
"consonant": "ch",
|
||||||
|
"consonant_length": 0.09005842357873917,
|
||||||
|
"vowel": "i",
|
||||||
|
"vowel_length": 0.08666137605905533,
|
||||||
|
"pitch": 5.958892822265625
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "ワ",
|
||||||
|
"consonant": "w",
|
||||||
|
"consonant_length": 0.07833231985569,
|
||||||
|
"vowel": "a",
|
||||||
|
"vowel_length": 0.21250136196613312,
|
||||||
|
"pitch": 5.949411392211914
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"accent": 5,
|
||||||
|
"pause_mora": {
|
||||||
|
"text": "、",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "pau",
|
||||||
|
"vowel_length": 0.4723339378833771,
|
||||||
|
"pitch": 0.0
|
||||||
|
},
|
||||||
|
"is_interrogative": false
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"moras": [
|
||||||
|
{
|
||||||
|
"text": "オ",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "o",
|
||||||
|
"vowel_length": 0.22004225850105286,
|
||||||
|
"pitch": 5.6870927810668945
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "ン",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "N",
|
||||||
|
"vowel_length": 0.09161105751991272,
|
||||||
|
"pitch": 5.93472957611084
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "セ",
|
||||||
|
"consonant": "s",
|
||||||
|
"consonant_length": 0.08924821764230728,
|
||||||
|
"vowel": "e",
|
||||||
|
"vowel_length": 0.14142127335071564,
|
||||||
|
"pitch": 6.121850490570068
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "エ",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "e",
|
||||||
|
"vowel_length": 0.10636933892965317,
|
||||||
|
"pitch": 6.157896041870117
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "ゴ",
|
||||||
|
"consonant": "g",
|
||||||
|
"consonant_length": 0.07600915431976318,
|
||||||
|
"vowel": "o",
|
||||||
|
"vowel_length": 0.09598273783922195,
|
||||||
|
"pitch": 6.188933849334717
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "オ",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "o",
|
||||||
|
"vowel_length": 0.1079121008515358,
|
||||||
|
"pitch": 6.235202789306641
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "セ",
|
||||||
|
"consonant": "s",
|
||||||
|
"consonant_length": 0.09591838717460632,
|
||||||
|
"vowel": "e",
|
||||||
|
"vowel_length": 0.10286372154951096,
|
||||||
|
"pitch": 6.153214454650879
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "エ",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "e",
|
||||||
|
"vowel_length": 0.08992656320333481,
|
||||||
|
"pitch": 6.02571439743042
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "ノ",
|
||||||
|
"consonant": "n",
|
||||||
|
"consonant_length": 0.05660202354192734,
|
||||||
|
"vowel": "o",
|
||||||
|
"vowel_length": 0.09676017612218857,
|
||||||
|
"pitch": 5.711844444274902
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"accent": 5,
|
||||||
|
"pause_mora": null,
|
||||||
|
"is_interrogative": false
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"moras": [
|
||||||
|
{
|
||||||
|
"text": "セ",
|
||||||
|
"consonant": "s",
|
||||||
|
"consonant_length": 0.07805486768484116,
|
||||||
|
"vowel": "e",
|
||||||
|
"vowel_length": 0.09617523103952408,
|
||||||
|
"pitch": 5.774399280548096
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "カ",
|
||||||
|
"consonant": "k",
|
||||||
|
"consonant_length": 0.06712044775485992,
|
||||||
|
"vowel": "a",
|
||||||
|
"vowel_length": 0.148829385638237,
|
||||||
|
"pitch": 6.063965797424316
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "イ",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "i",
|
||||||
|
"vowel_length": 0.11061104387044907,
|
||||||
|
"pitch": 6.040698051452637
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "エ",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "e",
|
||||||
|
"vowel_length": 0.13046696782112122,
|
||||||
|
"pitch": 5.806027889251709
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"accent": 1,
|
||||||
|
"pause_mora": null,
|
||||||
|
"is_interrogative": false
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"moras": [
|
||||||
|
{
|
||||||
|
"text": "ヨ",
|
||||||
|
"consonant": "y",
|
||||||
|
"consonant_length": 0.07194744795560837,
|
||||||
|
"vowel": "o",
|
||||||
|
"vowel_length": 0.08622600883245468,
|
||||||
|
"pitch": 5.694094657897949
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "オ",
|
||||||
|
"consonant": null,
|
||||||
|
"consonant_length": null,
|
||||||
|
"vowel": "o",
|
||||||
|
"vowel_length": 0.10635452717542648,
|
||||||
|
"pitch": 5.787222385406494
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "コ",
|
||||||
|
"consonant": "k",
|
||||||
|
"consonant_length": 0.07077334076166153,
|
||||||
|
"vowel": "o",
|
||||||
|
"vowel_length": 0.09248624742031097,
|
||||||
|
"pitch": 5.793357849121094
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "ソ",
|
||||||
|
"consonant": "s",
|
||||||
|
"consonant_length": 0.08705667406320572,
|
||||||
|
"vowel": "o",
|
||||||
|
"vowel_length": 0.2238258570432663,
|
||||||
|
"pitch": 5.643765449523926
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"accent": 1,
|
||||||
|
"pause_mora": null,
|
||||||
|
"is_interrogative": false
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"speedScale": 1.0,
|
||||||
|
"pitchScale": 0.0,
|
||||||
|
"intonationScale": 1.0,
|
||||||
|
"volumeScale": 1.0,
|
||||||
|
"prePhonemeLength": 0.1,
|
||||||
|
"postPhonemeLength": 0.1,
|
||||||
|
"pauseLength": null,
|
||||||
|
"pauseLengthScale": 1.0,
|
||||||
|
"outputSamplingRate": 24000,
|
||||||
|
"outputStereo": false,
|
||||||
|
"kana": "コンニチワ'、オンセエゴ'オセエノ/セ'カイエ/ヨ'オコソ"
|
||||||
|
}
|
||||||
27
crates/sbv2_editor/src/error.rs
Normal file
27
crates/sbv2_editor/src/error.rs
Normal file
@@ -0,0 +1,27 @@
|
|||||||
|
use axum::{
|
||||||
|
http::StatusCode,
|
||||||
|
response::{IntoResponse, Response},
|
||||||
|
};
|
||||||
|
|
||||||
|
pub type AppResult<T> = std::result::Result<T, AppError>;
|
||||||
|
|
||||||
|
pub struct AppError(anyhow::Error);
|
||||||
|
|
||||||
|
impl IntoResponse for AppError {
|
||||||
|
fn into_response(self) -> Response {
|
||||||
|
(
|
||||||
|
StatusCode::INTERNAL_SERVER_ERROR,
|
||||||
|
format!("Something went wrong: {}", self.0),
|
||||||
|
)
|
||||||
|
.into_response()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<E> From<E> for AppError
|
||||||
|
where
|
||||||
|
E: Into<anyhow::Error>,
|
||||||
|
{
|
||||||
|
fn from(err: E) -> Self {
|
||||||
|
Self(err.into())
|
||||||
|
}
|
||||||
|
}
|
||||||
197
crates/sbv2_editor/src/main.rs
Normal file
197
crates/sbv2_editor/src/main.rs
Normal file
@@ -0,0 +1,197 @@
|
|||||||
|
use axum::extract::State;
|
||||||
|
use axum::{
|
||||||
|
extract::Query,
|
||||||
|
http::header::CONTENT_TYPE,
|
||||||
|
response::IntoResponse,
|
||||||
|
routing::{get, post},
|
||||||
|
Json, Router,
|
||||||
|
};
|
||||||
|
use sbv2_core::tts_util::kata_tone2phone_tone;
|
||||||
|
use sbv2_core::{
|
||||||
|
tts::{SynthesizeOptions, TTSModelHolder},
|
||||||
|
tts_util::preprocess_parse_text,
|
||||||
|
};
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
use tokio::{fs, net::TcpListener, sync::Mutex};
|
||||||
|
|
||||||
|
use std::env;
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use error::AppResult;
|
||||||
|
|
||||||
|
mod error;
|
||||||
|
|
||||||
|
#[derive(Deserialize)]
|
||||||
|
struct RequestCreateAudioQuery {
|
||||||
|
text: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
struct AudioQuery {
|
||||||
|
kana: String,
|
||||||
|
tone: i32,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize)]
|
||||||
|
struct ResponseCreateAudioQuery {
|
||||||
|
audio_query: Vec<AudioQuery>,
|
||||||
|
text: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn create_audio_query(
|
||||||
|
State(state): State<AppState>,
|
||||||
|
Query(request): Query<RequestCreateAudioQuery>,
|
||||||
|
) -> AppResult<impl IntoResponse> {
|
||||||
|
let (text, process) = {
|
||||||
|
let tts_model = state.tts_model.lock().await;
|
||||||
|
preprocess_parse_text(&request.text, &tts_model.jtalk)?
|
||||||
|
};
|
||||||
|
let kana_tone_list = process.g2kana_tone()?;
|
||||||
|
let audio_query = kana_tone_list
|
||||||
|
.iter()
|
||||||
|
.map(|(kana, tone)| AudioQuery {
|
||||||
|
kana: kana.clone(),
|
||||||
|
tone: *tone,
|
||||||
|
})
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
Ok(Json(ResponseCreateAudioQuery { audio_query, text }))
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Deserialize)]
|
||||||
|
pub struct RequestSynthesis {
|
||||||
|
text: String,
|
||||||
|
speaker_id: i64,
|
||||||
|
sdp_ratio: f32,
|
||||||
|
length_scale: f32,
|
||||||
|
style_id: i32,
|
||||||
|
audio_query: Vec<AudioQuery>,
|
||||||
|
ident: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn synthesis(
|
||||||
|
State(state): State<AppState>,
|
||||||
|
Json(request): Json<RequestSynthesis>,
|
||||||
|
) -> AppResult<impl IntoResponse> {
|
||||||
|
let phone_tone = request
|
||||||
|
.audio_query
|
||||||
|
.iter()
|
||||||
|
.map(|query| (query.kana.clone(), query.tone))
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
let phone_tone = kata_tone2phone_tone(phone_tone);
|
||||||
|
let tones = phone_tone.iter().map(|(_, tone)| *tone).collect::<Vec<_>>();
|
||||||
|
let buffer = {
|
||||||
|
let mut tts_model = state.tts_model.lock().await;
|
||||||
|
tts_model.easy_synthesize_neo(
|
||||||
|
&request.ident,
|
||||||
|
&request.text,
|
||||||
|
Some(tones),
|
||||||
|
request.style_id,
|
||||||
|
request.speaker_id,
|
||||||
|
SynthesizeOptions {
|
||||||
|
sdp_ratio: request.sdp_ratio,
|
||||||
|
length_scale: request.length_scale,
|
||||||
|
..Default::default()
|
||||||
|
},
|
||||||
|
)?
|
||||||
|
};
|
||||||
|
Ok(([(CONTENT_TYPE, "audio/wav")], buffer))
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone)]
|
||||||
|
struct AppState {
|
||||||
|
tts_model: Arc<Mutex<TTSModelHolder>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl AppState {
|
||||||
|
pub async fn new() -> anyhow::Result<Self> {
|
||||||
|
let mut tts_model = TTSModelHolder::new(
|
||||||
|
&fs::read(env::var("BERT_MODEL_PATH")?).await?,
|
||||||
|
&fs::read(env::var("TOKENIZER_PATH")?).await?,
|
||||||
|
env::var("HOLDER_MAX_LOADED_MODElS")
|
||||||
|
.ok()
|
||||||
|
.and_then(|x| x.parse().ok()),
|
||||||
|
)?;
|
||||||
|
let models = env::var("MODELS_PATH").unwrap_or("models".to_string());
|
||||||
|
let mut f = fs::read_dir(&models).await?;
|
||||||
|
let mut entries = vec![];
|
||||||
|
while let Ok(Some(e)) = f.next_entry().await {
|
||||||
|
let name = e.file_name().to_string_lossy().to_string();
|
||||||
|
if name.ends_with(".onnx") && name.starts_with("model_") {
|
||||||
|
let name_len = name.len();
|
||||||
|
let name = name.chars();
|
||||||
|
entries.push(
|
||||||
|
name.collect::<Vec<_>>()[6..name_len - 5]
|
||||||
|
.iter()
|
||||||
|
.collect::<String>(),
|
||||||
|
);
|
||||||
|
} else if name.ends_with(".sbv2") {
|
||||||
|
let entry = &name[..name.len() - 5];
|
||||||
|
log::info!("Try loading: {entry}");
|
||||||
|
let sbv2_bytes = match fs::read(format!("{models}/{entry}.sbv2")).await {
|
||||||
|
Ok(b) => b,
|
||||||
|
Err(e) => {
|
||||||
|
log::warn!("Error loading sbv2_bytes from file {entry}: {e}");
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
if let Err(e) = tts_model.load_sbv2file(entry, sbv2_bytes) {
|
||||||
|
log::warn!("Error loading {entry}: {e}");
|
||||||
|
};
|
||||||
|
log::info!("Loaded: {entry}");
|
||||||
|
} else if name.ends_with(".aivmx") {
|
||||||
|
let entry = &name[..name.len() - 6];
|
||||||
|
log::info!("Try loading: {entry}");
|
||||||
|
let aivmx_bytes = match fs::read(format!("{models}/{entry}.aivmx")).await {
|
||||||
|
Ok(b) => b,
|
||||||
|
Err(e) => {
|
||||||
|
log::warn!("Error loading aivmx bytes from file {entry}: {e}");
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
if let Err(e) = tts_model.load_aivmx(entry, aivmx_bytes) {
|
||||||
|
log::error!("Error loading {entry}: {e}");
|
||||||
|
}
|
||||||
|
log::info!("Loaded: {entry}");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for entry in entries {
|
||||||
|
log::info!("Try loading: {entry}");
|
||||||
|
let style_vectors_bytes =
|
||||||
|
match fs::read(format!("{models}/style_vectors_{entry}.json")).await {
|
||||||
|
Ok(b) => b,
|
||||||
|
Err(e) => {
|
||||||
|
log::warn!("Error loading style_vectors_bytes from file {entry}: {e}");
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
let vits2_bytes = match fs::read(format!("{models}/model_{entry}.onnx")).await {
|
||||||
|
Ok(b) => b,
|
||||||
|
Err(e) => {
|
||||||
|
log::warn!("Error loading vits2_bytes from file {entry}: {e}");
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
if let Err(e) = tts_model.load(&entry, style_vectors_bytes, vits2_bytes) {
|
||||||
|
log::warn!("Error loading {entry}: {e}");
|
||||||
|
};
|
||||||
|
log::info!("Loaded: {entry}");
|
||||||
|
}
|
||||||
|
Ok(Self {
|
||||||
|
tts_model: Arc::new(Mutex::new(tts_model)),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::main]
|
||||||
|
async fn main() -> anyhow::Result<()> {
|
||||||
|
dotenvy::dotenv_override().ok();
|
||||||
|
env_logger::init();
|
||||||
|
let app = Router::new()
|
||||||
|
.route("/", get(|| async { "Hello, world!" }))
|
||||||
|
.route("/audio_query", get(create_audio_query))
|
||||||
|
.route("/synthesis", post(synthesis))
|
||||||
|
.with_state(AppState::new().await?);
|
||||||
|
let listener = TcpListener::bind("0.0.0.0:8080").await?;
|
||||||
|
axum::serve(listener, app).await?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
@@ -17,4 +17,4 @@ sbv2_core = { path = "../sbv2_core", default-features = false, features = ["no_s
|
|||||||
once_cell.workspace = true
|
once_cell.workspace = true
|
||||||
js-sys = "0.3.70"
|
js-sys = "0.3.70"
|
||||||
ndarray.workspace = true
|
ndarray.workspace = true
|
||||||
wasm-bindgen-futures = "0.4.43"
|
wasm-bindgen-futures = "0.4.54"
|
||||||
|
|||||||
@@ -1,2 +1,2 @@
|
|||||||
# StyleBertVITS2 wasm
|
# StyleBertVITS2 wasm
|
||||||
refer to https://github.com/tuna2134/sbv2-api
|
refer to https://github.com/neodyland/sbv2-api
|
||||||
|
|||||||
@@ -11,6 +11,7 @@
|
|||||||
},
|
},
|
||||||
"keywords": [],
|
"keywords": [],
|
||||||
"author": "tuna2134",
|
"author": "tuna2134",
|
||||||
|
"contributes": ["neodyland"],
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@biomejs/biome": "^1.9.4",
|
"@biomejs/biome": "^1.9.4",
|
||||||
|
|||||||
8
crates/sbv2_wasm/pnpm-lock.yaml
generated
8
crates/sbv2_wasm/pnpm-lock.yaml
generated
@@ -23,7 +23,7 @@ importers:
|
|||||||
version: 0.25.0
|
version: 0.25.0
|
||||||
typescript:
|
typescript:
|
||||||
specifier: ^5.7.3
|
specifier: ^5.7.3
|
||||||
version: 5.7.3
|
version: 5.8.3
|
||||||
|
|
||||||
packages:
|
packages:
|
||||||
|
|
||||||
@@ -290,8 +290,8 @@ packages:
|
|||||||
resolution: {integrity: sha512-mRUWCc3KUU4w1jU8sGxICXH/gNS94DvI1gxqDvBzhj1JpcsimQkYiOJfwsPUykUI5ZaspFbSgmBLER8IrQ3tqw==}
|
resolution: {integrity: sha512-mRUWCc3KUU4w1jU8sGxICXH/gNS94DvI1gxqDvBzhj1JpcsimQkYiOJfwsPUykUI5ZaspFbSgmBLER8IrQ3tqw==}
|
||||||
engines: {node: '>=12.0.0'}
|
engines: {node: '>=12.0.0'}
|
||||||
|
|
||||||
typescript@5.7.3:
|
typescript@5.8.3:
|
||||||
resolution: {integrity: sha512-84MVSjMEHP+FQRPy3pX9sTVV/INIex71s9TL2Gm5FG/WG1SqXeKyZ0k7/blY/4FdOzI12CBy1vGc4og/eus0fw==}
|
resolution: {integrity: sha512-p1diW6TqL9L07nNxvRMM7hMMw4c5XOo/1ibL4aAIGmSAt9slTE1Xgw5KWuof2uTOvCg9BY7ZRi+GaF+7sfgPeQ==}
|
||||||
engines: {node: '>=14.17'}
|
engines: {node: '>=14.17'}
|
||||||
hasBin: true
|
hasBin: true
|
||||||
|
|
||||||
@@ -499,6 +499,6 @@ snapshots:
|
|||||||
'@types/node': 22.13.5
|
'@types/node': 22.13.5
|
||||||
long: 5.3.1
|
long: 5.3.1
|
||||||
|
|
||||||
typescript@5.7.3: {}
|
typescript@5.8.3: {}
|
||||||
|
|
||||||
undici-types@6.20.0: {}
|
undici-types@6.20.0: {}
|
||||||
|
|||||||
1
scripts/convert/.python-version
Normal file
1
scripts/convert/.python-version
Normal file
@@ -0,0 +1 @@
|
|||||||
|
3.11
|
||||||
@@ -1,5 +1,6 @@
|
|||||||
style-bert-vits2
|
git+https://github.com/neodyland/style-bert-vits2-ref
|
||||||
onnxsim
|
onnxsim
|
||||||
numpy<2
|
numpy<2
|
||||||
zstandard
|
zstandard
|
||||||
onnxruntime
|
onnxruntime
|
||||||
|
cmake<4
|
||||||
@@ -2,8 +2,16 @@ FROM rust AS builder
|
|||||||
WORKDIR /work
|
WORKDIR /work
|
||||||
COPY . .
|
COPY . .
|
||||||
RUN cargo build -r --bin sbv2_api
|
RUN cargo build -r --bin sbv2_api
|
||||||
FROM gcr.io/distroless/cc-debian12
|
FROM ubuntu AS upx
|
||||||
WORKDIR /work
|
WORKDIR /work
|
||||||
|
RUN apt update && apt-get install -y upx binutils
|
||||||
COPY --from=builder /work/target/release/sbv2_api /work/main
|
COPY --from=builder /work/target/release/sbv2_api /work/main
|
||||||
COPY --from=builder /work/target/release/*.so /work
|
COPY --from=builder /work/target/release/*.so /work
|
||||||
CMD ["/work/main"]
|
RUN upx --best --lzma /work/main
|
||||||
|
RUN find /work -maxdepth 1 -name "*.so" -exec strip --strip-unneeded {} +
|
||||||
|
RUN find /work -maxdepth 1 -name "*.so" -exec upx --best --lzma {} +
|
||||||
|
FROM gcr.io/distroless/cc-debian12
|
||||||
|
WORKDIR /work
|
||||||
|
COPY --from=upx /work/main /work/main
|
||||||
|
COPY --from=upx /work/*.so /work
|
||||||
|
CMD ["/work/main"]
|
||||||
|
|||||||
@@ -2,9 +2,16 @@ FROM rust AS builder
|
|||||||
WORKDIR /work
|
WORKDIR /work
|
||||||
COPY . .
|
COPY . .
|
||||||
RUN cargo build -r --bin sbv2_api -F cuda,cuda_tf32
|
RUN cargo build -r --bin sbv2_api -F cuda,cuda_tf32
|
||||||
FROM nvidia/cuda:12.3.2-cudnn9-runtime-ubuntu22.04
|
FROM ubuntu AS upx
|
||||||
WORKDIR /work
|
WORKDIR /work
|
||||||
|
RUN apt update && apt-get install -y upx binutils
|
||||||
COPY --from=builder /work/target/release/sbv2_api /work/main
|
COPY --from=builder /work/target/release/sbv2_api /work/main
|
||||||
COPY --from=builder /work/target/release/*.so /work
|
COPY --from=builder /work/target/release/*.so /work
|
||||||
|
RUN upx --best --lzma /work/main
|
||||||
|
RUN find /work -maxdepth 1 -name "*.so" -exec strip --strip-unneeded {} +
|
||||||
|
FROM nvidia/cuda:12.3.2-cudnn9-runtime-ubuntu22.04
|
||||||
|
WORKDIR /work
|
||||||
|
COPY --from=upx /work/main /work/main
|
||||||
|
COPY --from=upx /work/*.so /work
|
||||||
ENV LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/work
|
ENV LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/work
|
||||||
CMD ["/work/main"]
|
CMD ["/work/main"]
|
||||||
|
|||||||
@@ -1,3 +1,3 @@
|
|||||||
docker run -it --rm -p 3000:3000 --name sbv2 \
|
docker run -it --rm -p 3000:3000 --name sbv2 \
|
||||||
-v ./models:/work/models --env-file .env \
|
-v ./models:/work/models --env-file .env \
|
||||||
ghcr.io/tuna2134/sbv2-api:cpu
|
ghcr.io/neodyland/sbv2-api:cpu
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
docker run -it --rm -p 3000:3000 --name sbv2 \
|
docker run -it --rm -p 3000:3000 --name sbv2 \
|
||||||
-v ./models:/work/models --env-file .env \
|
-v ./models:/work/models --env-file .env \
|
||||||
--gpus all \
|
--gpus all \
|
||||||
ghcr.io/tuna2134/sbv2-api:cuda
|
ghcr.io/neodyland/sbv2-api:cuda
|
||||||
|
|||||||
19
test.py
Normal file
19
test.py
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
import requests
|
||||||
|
|
||||||
|
|
||||||
|
data = (requests.get("http://localhost:8080/audio_query", params={
|
||||||
|
"text": "こんにちは、今日はいい天気ですね。",
|
||||||
|
})).json()
|
||||||
|
print(data)
|
||||||
|
|
||||||
|
data = (requests.post("http://localhost:8080/synthesis", json={
|
||||||
|
"text": data["text"],
|
||||||
|
"ident": "tsukuyomi",
|
||||||
|
"speaker_id": 0,
|
||||||
|
"style_id": 0,
|
||||||
|
"sdp_ratio": 0.5,
|
||||||
|
"length_scale": 0.5,
|
||||||
|
"audio_query": data["audio_query"],
|
||||||
|
})).content
|
||||||
|
with open("test.wav", "wb") as f:
|
||||||
|
f.write(data)
|
||||||
Reference in New Issue
Block a user