pred_file=${output_dir}/1best_recog/text_tn ref_file=${output_dir}/1best_recog/label python ./demo_speech2text.py ${ckpt_dir} ${ckpt_id} ${jsonl} ${output_dir} ${device} cn_postprocess=false if [ $data_set = "common_voice_zh-CN_speech2text.jsonl" ];then cn_postprocess=...
mkdir -p ${inference_dir}/1best_recog for f in token score text; do if [ -f "${inference_dir}/${JOB}/1best_recog/${f}" ]; then for JOB in $(seq "${nj}"); do cat "${inference_dir}/${JOB}/1best_recog/${f}" done | sort -k1 >"${inference_dir}/1best_recog/${f...
${output_dir}/1best_recog/wer.txt awk '/utt:/ { utt=$2 } /WER:/ { print utt, $2 }' \ ${output_dir}/1best_recog/wer.txt > \ ${data_dir}/utt2wer \ ${data_dir}/utt2werfiif [ $stage -le 3 ] && [ $stop_stage -ge 3 ];then echo "Filter utt whose WER <= 5% ...
funasrruntime * add dec reset for batch * set batch default value * add argv for CutSplit * sort frame_queue * sorted msgs * fix FunOfflineInfer * add dynamic batch for fetch * fix FetchDynamic * update run_server.sh * update run_server.sh * cpp http post server support (model...
python utils/compute_wer.py ${inference_dir}/1best_recog/text.ref ${inference_dir}/1best_recog/text.proc ${inference_dir}/1best_recog/text.cer tail -n 3 ${inference_dir}/1best_recog/text.cer donefi 1 change: 1 addition & 0 deletions 1 examples/wenetspeech/conformer/utils Original ...
audio_clipper = VideoClipper(funasr_model) theme = gr.Theme.load("funclip/utils/theme.json") def audio_recog(audio_input, sd_switch, hotwords, output_dir): return audio_clipper.recog(audio_input, sd_switch, hotwords, output_dir=output_dir) def video_recog(video_input, sd_switch, hotword...
(funasr_model) audio_clipper.lang = 'en' if mode == 'audio': logging.warning("Recognizing audio file: {}".format(file)) wav, sr = librosa.load(file, sr=16000) res_text, res_srt, state = audio_clipper.recog((sr, wav), sd_switch) if mode == 'video': logging.warning("...
= null //重新识别 private var mNeedReRecognized = true fun init() { mEventManager = EventManagerFactory.create(App.context, "asr") mEventManager?.registerListener(this) } fun setRecognizedListener(recognizedListener: RecognizedListener) { mRecognizedListenerList.add(recognizedListener) } fun remove...
268 - recog_button1 = gr.Button("👂识别 Recognize") 269 - audio_text_output = gr.Textbox(label="✏️识别结果 Recognition Result") 270 - audio_srt_output = gr.Textbox(label="📖SRT字幕内容 RST Subtitles") 271 - with gr.Column(): 272 - audio_text_input = gr.Textbox(...
create_task(recog_coro)) if __name__ == "__main__": def init(): try: asyncio.run(main()) except KeyboardInterrupt: console.print(f'再见!') to_exit = True # 提醒录音线程,该关闭了 keyboard.unhook_all() # observer.stop() # 关闭文件监控 observer.stop() # 关闭文件监控 sys.exit...