Transcribing meetings containing overlapped speech with only a single distant microphone (SDM) has been one of the most challenging problems for automatic speech recognition (ASR). While various approaches have been proposed, all previous studies on the monaural overlapped speech recognition problem were based on either simulation data or small-scale real data. In this paper, we extensively investigate a two-step approach where we first pre-train a serialized output training (SOT)-based multi-talker ASR by using large-scale simulation data and then fine-tune the model with a small amount of real meeting data. Experiments are conducted by utilizing 75 thousand (K) hours of our internal single-talker recording to simulate a total of 900K hours of multi-talker audio segments for supervised pre-training. With fine-tuning on the 70 hours of the AMI-SDM training data, our SOT ASR model achieves a word error rate (WER) of 21.2% for the AMI-SDM evaluation set while automatically counting speakers in each test segment. This result is not only significantly better than the previous state-of-the-art WER of 36.4% with oracle utterance boundary information but also better than a result by a similarly fine-tuned single-talker ASR model applied to beamformed audio.
翻译:在本文中,我们广泛调查了一种两步方法,即我们首先使用大型模拟数据,对基于序列化产出培训(SOT)的多对话者ASR进行分级培训,然后用少量实际会议数据对模型进行微调。实验的进行方式是利用我们内部单讲机的75 000(K)小时进行内部单讲机记录,以模拟总共900K小时的多讲者音频段,进行监管前培训。在对AMI-SDM培训数据的70小时进行微调后,我们的SOT ASR模型在AMI-SDM评价组中实现了21.2%的字差率,同时在每个测试部分中自动计数演讲人。其结果不仅大大优于先前的SAR-SDM模型,而且比对ARC-A-R-AFA的微调结果要好得多。