import logging logging.getLogger('numba').setLevel(logging.WARNING) logging.getLogger('matplotlib').setLevel(logging.WARNING) logging.getLogger('urllib3').setLevel(logging.WARNING) import json import re import numpy as np import IPython.display as ipd import torch import commons import utils from models import SynthesizerTrn from text.symbols import symbols from text import text_to_sequence import gradio as gr import time import datetime import os import pickle import openai from scipy.io.wavfile import write import librosa from mel_processing import spectrogram_torch def is_japanese(string): for ch in string: if ord(ch) > 0x3040 and ord(ch) < 0x30FF: return True return False def is_english(string): import re pattern = re.compile('^[A-Za-z0-9.,:;!?()_*"\' ]+$') if pattern.fullmatch(string): return True else: return False def to_html(chat_history): chat_html = "" for item in chat_history: if item['role'] == 'user': chat_html += f"""
{item['content']}
""" else: chat_html += f"""
{item['content']}
""" output_html = f"""
{chat_html}
""" return output_html def extrac(text): text = re.sub("<[^>]*>","",text) result_list = re.split(r'\n', text) final_list = [] if not torch.cuda.is_available(): if len(final_list) > 10: return ['对不起,做不到'] for i in result_list: if is_english(i): i = romajitable.to_kana(i).katakana i = i.replace('\n','').replace(' ','') #Current length of single sentence: 20 if len(i)>1: if len(i) > 20: try: cur_list = re.split(r'。|!', i) for i in cur_list: if len(i)>1: final_list.append(i+'。') except: pass else: final_list.append(i) final_list = [x for x in final_list if x != ''] print(final_list) return final_list def to_numpy(tensor: torch.Tensor): return tensor.detach().cpu().numpy() if tensor.requires_grad \ else tensor.detach().numpy() def chatgpt(text): messages = [] try: with open('log.pickle', 'rb') as f: messages = pickle.load(f) messages.append({"role": "user", "content": text},) chat = openai.ChatCompletion.create(model="gpt-3.5-turbo", messages=messages) reply = chat.choices[0].message.content messages.append({"role": "assistant", "content": reply}) print(messages[-1]) if len(messages) == 12: messages[6:10] = messages[8:] del messages[-2:] with open('log.pickle', 'wb') as f: messages2 = [] pickle.dump(messages2, f) return reply,messages except: messages.append({"role": "user", "content": text},) chat = openai.ChatCompletion.create(model="gpt-3.5-turbo", messages=messages) reply = chat.choices[0].message.content messages.append({"role": "assistant", "content": reply}) print(messages[-1]) if len(messages) == 12: messages[6:10] = messages[8:] del messages[-2:] with open('log.pickle', 'wb') as f: pickle.dump(messages, f) return reply,messages def get_symbols_from_json(path): assert os.path.isfile(path) with open(path, 'r') as f: data = json.load(f) return data['symbols'] def sle(language,text): text = text.replace('\n', ' ').replace('\r', '').replace(" ", "") if language == "中文": tts_input1 = "[ZH]" + text + "[ZH]" return tts_input1 elif language == "自动": tts_input1 = f"[JA]{text}[JA]" if is_japanese(text) else f"[ZH]{text}[ZH]" return tts_input1 elif language == "日文": tts_input1 = "[JA]" + text + "[JA]" return tts_input1 elif language == "英文": tts_input1 = "[EN]" + text + "[EN]" return tts_input1 elif language == "手动": return text def get_text(text,hps_ms): text_norm = text_to_sequence(text,hps_ms.data.text_cleaners) if hps_ms.data.add_blank: text_norm = commons.intersperse(text_norm, 0) text_norm = torch.LongTensor(text_norm) return text_norm def create_vc_fn(net_g,hps): def vc_fn(text,language,n_scale,n_scale_w,l_scale,original_speaker, target_speaker, record_audio, upload_audio): input_audio = record_audio if record_audio is not None else upload_audio original_speaker_id = selection(original_speaker) target_speaker_id = selection(target_speaker) if input_audio is None: stn_tst = get_text(sle(language,text),hps) with torch.no_grad(): x_tst = stn_tst.unsqueeze(0).to(dev) x_tst_lengths = torch.LongTensor([stn_tst.size(0)]).to(dev) sid = torch.LongTensor([original_speaker_id]).to(dev) audio = net_g.infer(x_tst, x_tst_lengths, sid=sid, noise_scale=n_scale, noise_scale_w=n_scale_w, length_scale=l_scale)[0][0,0].data.cpu().float().numpy() sampling_rate = hps.data.sampling_rate else: sampling_rate, audio = input_audio audio = (audio / np.iinfo(audio.dtype).max).astype(np.float32) if len(audio.shape) > 1: audio = librosa.to_mono(audio.transpose(1, 0)) if sampling_rate != hps.data.sampling_rate: audio = librosa.resample(audio, orig_sr=sampling_rate, target_sr=hps.data.sampling_rate) with torch.no_grad(): y = torch.FloatTensor(audio) y = y / max(-y.min(), y.max()) / 0.99 y = y.to(dev) y = y.unsqueeze(0) spec = spectrogram_torch(y, hps.data.filter_length, hps.data.sampling_rate, hps.data.hop_length, hps.data.win_length, center=False).to(dev) spec_lengths = torch.LongTensor([spec.size(-1)]).to(dev) sid_src = torch.LongTensor([original_speaker_id]).to(dev) sid_tgt = torch.LongTensor([target_speaker_id]).to(dev) audio = net_g.voice_conversion(spec, spec_lengths, sid_src=sid_src, sid_tgt=sid_tgt)[0][ 0, 0].data.cpu().float().numpy() del y, spec, spec_lengths, sid_src, sid_tgt return "Success", (hps.data.sampling_rate, audio) return vc_fn def selection(speaker): if speaker == "高咲侑": spk = 0 return spk elif speaker == "歩夢": spk = 1 return spk elif speaker == "かすみ": spk = 2 return spk elif speaker == "しずく": spk = 3 return spk elif speaker == "果林": spk = 4 return spk elif speaker == "愛": spk = 5 return spk elif speaker == "彼方": spk = 6 return spk elif speaker == "せつ菜": spk = 7 return spk elif speaker == "エマ": spk = 8 return spk elif speaker == "璃奈": spk = 9 return spk elif speaker == "栞子": spk = 10 return spk elif speaker == "ランジュ": spk = 11 return spk elif speaker == "ミア": spk = 12 return spk elif speaker == "派蒙": spk = 16 return spk elif speaker == "c1": spk = 18 return spk elif speaker == "c2": spk = 19 return spk elif speaker == "華恋": spk = 21 return spk elif speaker == "まひる": spk = 22 return spk elif speaker == "なな": spk = 23 return spk elif speaker == "クロディーヌ": spk = 24 return spk elif speaker == "ひかり": spk = 25 return spk elif speaker == "純那": spk = 26 return spk elif speaker == "香子": spk = 27 return spk elif speaker == "真矢": spk = 28 return spk elif speaker == "双葉": spk = 29 return spk elif speaker == "ミチル": spk = 30 return spk elif speaker == "メイファン": spk = 31 return spk elif speaker == "やちよ": spk = 32 return spk elif speaker == "晶": spk = 33 return spk elif speaker == "いちえ": spk = 34 return spk elif speaker == "ゆゆ子": spk = 35 return spk elif speaker == "塁": spk = 36 return spk elif speaker == "珠緒": spk = 37 return spk elif speaker == "あるる": spk = 38 return spk elif speaker == "ララフィン": spk = 39 return spk elif speaker == "美空": spk = 40 return spk elif speaker == "静羽": spk = 41 return spk else: return 0 def check_text(input): if isinstance(input, str): return input else: with open(input.name, "r", encoding="utf-8") as f: return f.read() def create_tts_fn(net_g,hps,speaker_id): speaker_id = int(speaker_id) def tts_fn(is_gpt,api_key,is_audio,audiopath,repeat_time,text, language, extract, n_scale= 0.667,n_scale_w = 0.8, l_scale = 1 ): text = check_text(text) repeat_ime = int(repeat_time) if is_gpt: openai.api_key = api_key text,messages = chatgpt(text) htm = to_html(messages) else: messages = [] messages.append({"role": "assistant", "content": text}) htm = to_html(messages) if not extract: t1 = time.time() stn_tst = get_text(sle(language,text),hps) with torch.no_grad(): x_tst = stn_tst.unsqueeze(0).to(dev) x_tst_lengths = torch.LongTensor([stn_tst.size(0)]).to(dev) sid = torch.LongTensor([speaker_id]).to(dev) audio = net_g.infer(x_tst, x_tst_lengths, sid=sid, noise_scale=n_scale, noise_scale_w=n_scale_w, length_scale=l_scale)[0][0,0].data.cpu().float().numpy() t2 = time.time() spending_time = "推理时间为:"+str(t2-t1)+"s" print(spending_time) file_path = "subtitles.srt" try: write(audiopath + '.wav',22050,audio) if is_audio: for i in range(repeat_time): cmd = 'ffmpeg -y -i ' + audiopath + '.wav' + ' -ar 44100 '+ audiopath.replace('temp','temp'+str(i)) os.system(cmd) except: pass return (hps.data.sampling_rate, audio),file_path,htm else: a = ['【','[','(','('] b = ['】',']',')',')'] for i in a: text = text.replace(i,'<') for i in b: text = text.replace(i,'>') final_list = extrac(text.replace('“','').replace('”','')) audio_fin = [] c = 0 t = datetime.timedelta(seconds=0) for sentence in final_list: try: f1 = open("subtitles.srt",'w',encoding='utf-8') c +=1 stn_tst = get_text(sle(language,sentence),hps) with torch.no_grad(): x_tst = stn_tst.unsqueeze(0).to(dev) x_tst_lengths = torch.LongTensor([stn_tst.size(0)]).to(dev) sid = torch.LongTensor([speaker_id]).to(dev) t1 = time.time() audio = net_g.infer(x_tst, x_tst_lengths, sid=sid, noise_scale=n_scale, noise_scale_w=n_scale_w, length_scale=l_scale)[0][0,0].data.cpu().float().numpy() t2 = time.time() spending_time = "第"+str(c)+"句的推理时间为:"+str(t2-t1)+"s" print(spending_time) time_start = str(t).split(".")[0] + "," + str(t.microseconds)[:3] last_time = datetime.timedelta(seconds=len(audio)/float(22050)) t+=last_time time_end = str(t).split(".")[0] + "," + str(t.microseconds)[:3] print(time_end) f1.write(str(c-1)+'\n'+time_start+' --> '+time_end+'\n'+sentence+'\n\n') audio_fin.append(audio) except: pass try: write(audiopath + '.wav',22050,np.concatenate(audio_fin)) if is_audio: for i in range(repeat_time): cmd = 'ffmpeg -y -i ' + audiopath + '.wav' + ' -ar 44100 '+ audiopath.replace('temp','temp'+str(i)) os.system(cmd) except: pass file_path = "subtitles.srt" return (hps.data.sampling_rate, np.concatenate(audio_fin)),file_path,htm return tts_fn if __name__ == '__main__': hps = utils.get_hparams_from_file('checkpoints/tmp/config.json') dev = torch.device("cuda:0" if torch.cuda.is_available() else "cpu") models = [] schools_list = ["ShojoKageki-Nijigasaki","ShojoKageki","Nijigasaki"] schools = [] lan = ["中文","日文","自动","手动"] with open("checkpoints/info.json", "r", encoding="utf-8") as f: models_info = json.load(f) for i in models_info: school = models_info[i] speakers = school["speakers"] phone_dict = { symbol: i for i, symbol in enumerate(symbols) } checkpoint = models_info[i]["checkpoint"] net_g = SynthesizerTrn( len(symbols), hps.data.filter_length // 2 + 1, hps.train.segment_size // hps.data.hop_length, n_speakers=hps.data.n_speakers, **hps.model).to(dev) _ = net_g.eval() _ = utils.load_checkpoint(checkpoint , net_g) content = [] for j in speakers: sid = int(speakers[j]['sid']) title = school example = speakers[j]['speech'] name = speakers[j]["name"] content.append((sid, name, title, example, create_tts_fn(net_g,hps,sid))) models.append(content) schools.append((i,create_vc_fn(net_g,hps))) with gr.Blocks() as app: with gr.Tabs(): for (i,vc_fn) in schools: with gr.TabItem(i): idols = ["派蒙"] for (sid, name, title, example, tts_fn) in models[schools_list.index(i)]: idols.append(name) with gr.TabItem(name): with gr.Column(): with gr.Row(): with gr.Row(): gr.Markdown( '
' f'' '
' ) output_UI = gr.outputs.HTML() with gr.Row(): with gr.Column(scale=0.85): input1 = gr.TextArea(label="Text", value=example,lines = 1) with gr.Column(scale=0.15, min_width=0): btnVC = gr.Button("Send") output1 = gr.Audio(label="采样率22050") with gr.Accordion(label="Setting(TTS)", open=False): input2 = gr.Dropdown(label="Language", choices=lan, value="自动", interactive=True) input4 = gr.Slider(minimum=0, maximum=1.0, label="更改噪声比例(noise scale),以控制情感", value=0.6) input5 = gr.Slider(minimum=0, maximum=1.0, label="更改噪声偏差(noise scale w),以控制音素长短", value=0.668) input6 = gr.Slider(minimum=0.1, maximum=10, label="duration", value=1) with gr.Accordion(label="Advanced Setting(GPT3.5接口+小说合成,建议克隆本仓库后运行main.py)", open=False): input3 = gr.Checkbox(value=False, label="长句切割(小说合成)") inputxt = gr.File(label="Text") btnbook = gr.Button("小说合成") output2 = gr.outputs.File(label="字幕文件:subtitles.srt") api_input1 = gr.Checkbox(value=False, label="接入chatgpt") api_input2 = gr.TextArea(label="api-key",lines=1,value = '见 https://openai.com/blog/openai-api') audio_input1 = gr.Checkbox(value=False, label="修改音频路径(live2d)") audio_input2 = gr.TextArea(label="音频路径",lines=1,value = '#参考 D:/app_develop/live2d_whole/2010002/sounds/temp.wav') audio_input3 = gr.Dropdown(label="重复生成次数", choices=list(range(101)), value='0', interactive=True) btnbook.click(tts_fn, inputs=[api_input1,api_input2,audio_input1,audio_input2,audio_input3,inputxt,input2,input3,input4,input5,input6], outputs=[output1,output2,output_UI]) btnVC.click(tts_fn, inputs=[api_input1,api_input2,audio_input1,audio_input2,audio_input3,input1,input2,input3,input4,input5,input6], outputs=[output1,output2,output_UI]) with gr.Tab("Voice Conversion(类似sovits)"): gr.Markdown(""" 声线转化,使用模型中的说话人作为音源时效果更佳 """) with gr.Column(): with gr.Accordion(label="方法1:录制或上传声音,可进行歌声合成", open=False): record_audio = gr.Audio(label="record your voice", source="microphone") upload_audio = gr.Audio(label="or upload audio here", source="upload") with gr.Accordion(label="方法2:由原说话人先进行tts后套娃,适用于合成中文等特殊场景", open=True): text = gr.TextArea(label="Text", value='由源说话人进行语音转化',lines = 1) language = gr.Dropdown(label="Language", choices=lan, value="自动", interactive=True) n_scale = gr.Slider(minimum=0, maximum=1.0, label="更改噪声比例(noise scale),以控制情感", value=0.6) n_scale_w = gr.Slider(minimum=0, maximum=1.0, label="更改噪声偏差(noise scale w),以控制音素长短", value=0.668) l_scale = gr.Slider(minimum=0.1, maximum=10, label="duration", value=1.1) source_speaker = gr.Dropdown(choices=idols, value=idols[-2], label="source speaker") target_speaker = gr.Dropdown(choices=idols, value=idols[-3], label="target speaker") with gr.Column(): message_box = gr.Textbox(label="Message") converted_audio = gr.Audio(label='converted audio') btn = gr.Button("Convert!") btn.click(vc_fn, inputs=[text,language,n_scale,n_scale_w,l_scale,source_speaker, target_speaker, record_audio, upload_audio], outputs=[message_box, converted_audio]) with gr.Tab("说明"): gr.Markdown( "###
请不要生成会对个人以及企划造成侵害的内容,自觉遵守相关法律,静止商业使用或让他人产生困扰\n" "
从左到右分别是虹团,少歌中文特化版,以及五校混合版。这三个均为不同的模型,效果也有差异
\n" "
因为我会时不时地更新模型,所以会碰到平台抽风问题,大部分情况下一天就能恢复了。
\n" '
参数说明:这个十分玄学,我还没找到最合适的,如果效果不佳可以将噪声比例和噪声偏差调节至0,这回完全随机化音频源。按照经验,合成日语时也可以将噪声比例调节至0.2-0.3区间,语调会正常一些。duration代表整体语速,可视情况调至1.1或1.2
' '
建议只在平台上体验最基础的功能,强烈建议将该仓库克隆至本地或者于colab运行 main.py或app.py
') app.launch()