File "/root/autodl-tmp/CosyVoice/cosyvoice/cli/cosyvoice.py", line 29, ininit configs = load_hyperpyyaml(f) File "/root/miniconda3/envs/cosyvoice/lib/python3.8/site-packages/hyperpyyaml/core.py", line 188, in load_hyperpyyaml hparams = yaml.load(yaml_stream, Loader=loader) File "/...
you can use generator as input, this is useful when using text llm model as input# NOTE you should still have some basic sentence split logic because llm can not handle arbitrary sentence lengthdeftext_generator():yield'收到好友从远方寄来的生日礼物,'yield'那份意外的惊喜...
sampling:!name:cosyvoice.utils.common.ras_sampling top_p:0.8 top_k:25 win_size:10 tau_r:0.1 flow:!new:cosyvoice.flow.flow.MaskedDiffWithXvec input_size:512 output_size:80 spk_embed_dim:!ref<spk_embed_dim> output_type:'mel'
utils.common import fade_in_out import numpy as np class CosyVoiceModel: Expand Down Expand Up @@ -80,27 +80,27 @@ def load_onnx(self, flow_decoder_estimator_model): def llm_job(self, text, prompt_text, llm_prompt_speech_token, llm_embedding, uuid): with self.llm_context: for ...