-
Notifications
You must be signed in to change notification settings - Fork 1.9k
Description
Originally posted by zupu8 December 28, 2022
系统环境:win10
软件:Paddlespeech_Server
问题:非流式的tts_inference引擎下的模型调用tts_python下的模型报错
配置文件:
host: 0.0.0.0
port: 8099
protocol: 'http'
engine_list: ['tts_inference']
//################################### TTS #########################################//
//################### speech task: tts; engine_type: python #######################//
tts_python:
# am (acoustic model) choices=['speedyspeech_csmsc', 'fastspeech2_csmsc',
# 'fastspeech2_ljspeech', 'fastspeech2_aishell3',
# 'fastspeech2_vctk', 'fastspeech2_mix',
# 'tacotron2_csmsc', 'tacotron2_ljspeech']
am: 'fastspeech2_aishell3'
am_config:
am_ckpt:
am_stat:
phones_dict:
tones_dict:
speaker_dict:
# voc (vocoder) choices=['pwgan_csmsc', 'pwgan_ljspeech', 'pwgan_aishell3',
# 'pwgan_vctk', 'mb_melgan_csmsc', 'style_melgan_csmsc',
# 'hifigan_csmsc', 'hifigan_ljspeech', 'hifigan_aishell3',
# 'hifigan_vctk', 'wavernn_csmsc']
voc: 'hifigan_aishell3'
voc_config:
voc_ckpt:
voc_stat:
# others
lang: 'zh'
device: # set 'gpu:id' or 'cpu'
################### speech task: tts; engine_type: inference #######################
tts_inference:
# am (acoustic model) choices=['speedyspeech_csmsc', 'fastspeech2_csmsc']
am: 'fastspeech2_aishell3'
am_model: # the pdmodel file of your am static model (XX.pdmodel)
am_params: # the pdiparams file of your am static model (XX.pdipparams)
am_sample_rate: 24000
phones_dict:
tones_dict:
speaker_dict:
spk_id: 0
am_predictor_conf:
device: # set 'gpu:id' or 'cpu'
switch_ir_optim: True
glog_info: False # True -> print glog
summary: True # False -> do not show predictor config
# voc (vocoder) choices=['pwgan_csmsc', 'mb_melgan_csmsc','hifigan_csmsc']
voc: 'hifigan_aishell3'
voc_model: # the pdmodel file of your vocoder static model (XX.pdmodel)
voc_params: # the pdiparams file of your vocoder static model (XX.pdipparams)
voc_sample_rate: 24000
voc_predictor_conf:
device: # set 'gpu:id' or 'cpu'
switch_ir_optim: True
glog_info: False # True -> print glog
summary: True # False -> do not show predictor config
# others
lang: 'zh'
启动:paddlespeech_server start --config_file .paddlespeech\conf\application.yaml
报错:
[2022-12-28 16:46:57,668] [ INFO] - Initialize TTS server engine successfully on device: gpu:0.
Building prefix dict from the default dictionary ...
[2022-12-28 16:46:57] [DEBUG] [init.py:113] Building prefix dict from the default dictionary ...
Loading model from cache C:\Users\ADMINI1\AppData\Local\Temp\jieba.cache1\AppData\Local\Temp\jieba.cache
[2022-12-28 16:46:57] [DEBUG] [init.py:132] Loading model from cache C:\Users\ADMINI
Loading model cost 0.465 seconds.
[2022-12-28 16:46:58] [DEBUG] [init.py:164] Loading model cost 0.465 seconds.
Prefix dict has been built successfully.
[2022-12-28 16:46:58] [DEBUG] [init.py:166] Prefix dict has been built successfully.
[2022-12-28 16:46:58,240] [ ERROR] - Failed to warm up on tts engine.
[2022-12-28 16:46:58,240] [ ERROR] - local variable 'mel' referenced before assignment
Metadata
Metadata
Assignees
Labels
Type
Projects
Status