From fca65485c283a152c8d255ed7b2e422b8ea775ca Mon Sep 17 00:00:00 2001 From: aky15 Date: Wed, 13 Sep 2023 14:53:41 +0800 Subject: [PATCH] Update asr_inference_pipeline.py add support for simulated streaming ASR inference. --- modelscope/pipelines/audio/asr_inference_pipeline.py | 9 +++++++-- 1 file changed, 7 insertions(+), 2 deletions(-) diff --git a/modelscope/pipelines/audio/asr_inference_pipeline.py b/modelscope/pipelines/audio/asr_inference_pipeline.py index 2379274c..cc3a53d8 100644 --- a/modelscope/pipelines/audio/asr_inference_pipeline.py +++ b/modelscope/pipelines/audio/asr_inference_pipeline.py @@ -160,6 +160,7 @@ class AutomaticSpeechRecognitionPipeline(Pipeline): token_num_relax=self.cmd['token_num_relax'], decoding_ind=self.cmd['decoding_ind'], decoding_mode=self.cmd['decoding_mode'], + simu_streaming=self.cmd['simu_streaming'], **kwargs, ) @@ -310,7 +311,8 @@ class AutomaticSpeechRecognitionPipeline(Pipeline): 'fs': { 'model_fs': None, 'audio_fs': None - } + }, + 'simu_streaming': False, } frontend_conf = None @@ -333,7 +335,9 @@ class AutomaticSpeechRecognitionPipeline(Pipeline): decoding_ind = root['decoding_ind'] if 'decoding_mode' in root: decoding_mode = root['decoding_mode'] - + if 'simu_streaming' in root: + simu_streaming = root['simu_streaming'] + cmd['beam_size'] = root['beam_size'] cmd['penalty'] = root['penalty'] cmd['maxlenratio'] = root['maxlenratio'] @@ -389,6 +393,7 @@ class AutomaticSpeechRecognitionPipeline(Pipeline): 'punc_model_file', 'punc_infer_config', 'param_dict', + 'simu_streaming', ] for user_args in user_args_dict: