Skip to content

非流式的tts_inference引擎下,多说话人合成报错 #2780

@yt605155624

Description

@yt605155624

Originally posted by zupu8 December 28, 2022
系统环境:win10
软件:Paddlespeech_Server
问题:非流式的tts_inference引擎下的模型调用tts_python下的模型报错
配置文件:

host: 0.0.0.0
port: 8099

protocol: 'http'
engine_list: ['tts_inference']

//################################### TTS #########################################//
//################### speech task: tts; engine_type: python #######################//
tts_python:
# am (acoustic model) choices=['speedyspeech_csmsc', 'fastspeech2_csmsc',
# 'fastspeech2_ljspeech', 'fastspeech2_aishell3',
# 'fastspeech2_vctk', 'fastspeech2_mix',
# 'tacotron2_csmsc', 'tacotron2_ljspeech']
am: 'fastspeech2_aishell3'
am_config:
am_ckpt:
am_stat:
phones_dict:
tones_dict:
speaker_dict:

# voc (vocoder) choices=['pwgan_csmsc', 'pwgan_ljspeech', 'pwgan_aishell3',
#                        'pwgan_vctk', 'mb_melgan_csmsc', 'style_melgan_csmsc',
#                        'hifigan_csmsc', 'hifigan_ljspeech', 'hifigan_aishell3',
#                        'hifigan_vctk', 'wavernn_csmsc']
voc: 'hifigan_aishell3'
voc_config: 
voc_ckpt: 
voc_stat: 

# others
lang: 'zh'
device: # set 'gpu:id' or 'cpu'

################### speech task: tts; engine_type: inference #######################
tts_inference:
# am (acoustic model) choices=['speedyspeech_csmsc', 'fastspeech2_csmsc']
am: 'fastspeech2_aishell3'
am_model: # the pdmodel file of your am static model (XX.pdmodel)
am_params: # the pdiparams file of your am static model (XX.pdipparams)
am_sample_rate: 24000
phones_dict:
tones_dict:
speaker_dict:
spk_id: 0

am_predictor_conf:
    device:  # set 'gpu:id' or 'cpu'
    switch_ir_optim: True
    glog_info: False # True -> print glog
    summary: True  # False -> do not show predictor config

# voc (vocoder) choices=['pwgan_csmsc', 'mb_melgan_csmsc','hifigan_csmsc']
voc: 'hifigan_aishell3'
voc_model: # the pdmodel file of your vocoder static model (XX.pdmodel)
voc_params: # the pdiparams file of your vocoder static model (XX.pdipparams)
voc_sample_rate: 24000

voc_predictor_conf:
    device:  # set 'gpu:id' or 'cpu'  
    switch_ir_optim: True  
    glog_info: False # True -> print glog
    summary: True  # False -> do not show predictor config

# others
lang: 'zh'

启动:paddlespeech_server start --config_file .paddlespeech\conf\application.yaml
报错:
[2022-12-28 16:46:57,668] [ INFO] - Initialize TTS server engine successfully on device: gpu:0.
Building prefix dict from the default dictionary ...
[2022-12-28 16:46:57] [DEBUG] [init.py:113] Building prefix dict from the default dictionary ...
Loading model from cache C:\Users\ADMINI1\AppData\Local\Temp\jieba.cache
[2022-12-28 16:46:57] [DEBUG] [init.py:132] Loading model from cache C:\Users\ADMINI
1\AppData\Local\Temp\jieba.cache
Loading model cost 0.465 seconds.
[2022-12-28 16:46:58] [DEBUG] [init.py:164] Loading model cost 0.465 seconds.
Prefix dict has been built successfully.
[2022-12-28 16:46:58] [DEBUG] [init.py:166] Prefix dict has been built successfully.
[2022-12-28 16:46:58,240] [ ERROR] - Failed to warm up on tts engine.
[2022-12-28 16:46:58,240] [ ERROR] - local variable 'mel' referenced before assignment

Metadata

Metadata

Assignees

Labels

Type

No type

Projects

Status

Done

Milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions