spow12's picture
Update: clone from original and replace the model weights
104aac1
raw
history blame
4.83 kB
import argparse
import json
import shutil
from pathlib import Path
import yaml
from huggingface_hub import hf_hub_download
from style_bert_vits2.logging import logger
def download_bert_models():
with open("bert/bert_models.json", encoding="utf-8") as fp:
models = json.load(fp)
for k, v in models.items():
local_path = Path("bert").joinpath(k)
for file in v["files"]:
if not Path(local_path).joinpath(file).exists():
logger.info(f"Downloading {k} {file}")
hf_hub_download(v["repo_id"], file, local_dir=local_path)
def download_slm_model():
local_path = Path("slm/wavlm-base-plus/")
file = "pytorch_model.bin"
if not Path(local_path).joinpath(file).exists():
logger.info(f"Downloading wavlm-base-plus {file}")
hf_hub_download("microsoft/wavlm-base-plus", file, local_dir=local_path)
def download_pretrained_models():
files = ["G_0.safetensors", "D_0.safetensors", "DUR_0.safetensors"]
local_path = Path("pretrained")
for file in files:
if not Path(local_path).joinpath(file).exists():
logger.info(f"Downloading pretrained {file}")
hf_hub_download(
"litagin/Style-Bert-VITS2-1.0-base", file, local_dir=local_path
)
def download_jp_extra_pretrained_models():
files = ["G_0.safetensors", "D_0.safetensors", "WD_0.safetensors"]
local_path = Path("pretrained_jp_extra")
for file in files:
if not Path(local_path).joinpath(file).exists():
logger.info(f"Downloading JP-Extra pretrained {file}")
hf_hub_download(
"litagin/Style-Bert-VITS2-2.0-base-JP-Extra", file, local_dir=local_path
)
def download_default_models():
files = [
"jvnv-F1-jp/config.json",
"jvnv-F1-jp/jvnv-F1-jp_e160_s14000.safetensors",
"jvnv-F1-jp/style_vectors.npy",
"jvnv-F2-jp/config.json",
"jvnv-F2-jp/jvnv-F2_e166_s20000.safetensors",
"jvnv-F2-jp/style_vectors.npy",
"jvnv-M1-jp/config.json",
"jvnv-M1-jp/jvnv-M1-jp_e158_s14000.safetensors",
"jvnv-M1-jp/style_vectors.npy",
"jvnv-M2-jp/config.json",
"jvnv-M2-jp/jvnv-M2-jp_e159_s17000.safetensors",
"jvnv-M2-jp/style_vectors.npy",
]
for file in files:
if not Path(f"model_assets/{file}").exists():
logger.info(f"Downloading {file}")
hf_hub_download(
"litagin/style_bert_vits2_jvnv",
file,
local_dir="model_assets",
)
additional_files = {
"litagin/sbv2_koharune_ami": [
"koharune-ami/config.json",
"koharune-ami/style_vectors.npy",
"koharune-ami/koharune-ami.safetensors",
],
"litagin/sbv2_amitaro": [
"amitaro/config.json",
"amitaro/style_vectors.npy",
"amitaro/amitaro.safetensors",
],
}
for repo_id, files in additional_files.items():
for file in files:
if not Path(f"model_assets/{file}").exists():
logger.info(f"Downloading {file}")
hf_hub_download(
repo_id,
file,
local_dir="model_assets",
)
def main():
parser = argparse.ArgumentParser()
parser.add_argument("--skip_default_models", action="store_true")
parser.add_argument("--only_infer", action="store_true")
parser.add_argument(
"--dataset_root",
type=str,
help="Dataset root path (default: Data)",
default=None,
)
parser.add_argument(
"--assets_root",
type=str,
help="Assets root path (default: model_assets)",
default=None,
)
args = parser.parse_args()
download_bert_models()
if not args.skip_default_models:
download_default_models()
if not args.only_infer:
download_slm_model()
download_pretrained_models()
download_jp_extra_pretrained_models()
# If configs/paths.yml not exists, create it
default_paths_yml = Path("configs/default_paths.yml")
paths_yml = Path("configs/paths.yml")
if not paths_yml.exists():
shutil.copy(default_paths_yml, paths_yml)
if args.dataset_root is None and args.assets_root is None:
return
# Change default paths if necessary
with open(paths_yml, encoding="utf-8") as f:
yml_data = yaml.safe_load(f)
if args.assets_root is not None:
yml_data["assets_root"] = args.assets_root
if args.dataset_root is not None:
yml_data["dataset_root"] = args.dataset_root
with open(paths_yml, "w", encoding="utf-8") as f:
yaml.dump(yml_data, f, allow_unicode=True)
if __name__ == "__main__":
main()