TR2016-002

Deep Beamforming Networks for Multi-Channel Speech Recognition


    •  Xiao, X.; Watanabe, S.; Erdogan, H.; Lu, L.; Hershey, J.; Seltzer, M.; Chen, G.; Zhang, Y.; Mandel, M.; Yu, D., "Deep Beamforming Networks for Multi-Channel Speech Recognition", IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), DOI: 10.1109/ICASSP.2016.7472778, March 2016, pp. 5745-5749.
      BibTeX Download PDF
      • @inproceedings{Xiao2016mar,
      • author = {Xiao, X. and Watanabe, S. and Erdogan, H. and Lu, L. and Hershey, J. and Seltzer, M. and Chen, G. and Zhang, Y. and Mandel, M. and Yu, D.},
      • title = {Deep Beamforming Networks for Multi-Channel Speech Recognition},
      • booktitle = {IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP)},
      • year = 2016,
      • pages = {5745--5749},
      • month = mar,
      • doi = {10.1109/ICASSP.2016.7472778},
      • url = {http://www.merl.com/publications/TR2016-002}
      • }
  • MERL Contact:
  • Research Area:

    Speech & Audio


Despite the significant progress in speech recognition enabled by deep neural networks, poor performance persists in some scenarios. In this work, we focus on far-field speech recognition which remains challenging due to high levels of noise and reverberation in the captured speech signals. We propose to represent the stages of acoustic processing including beamforming, feature extraction, and acoustic modeling, as three components of a single unified computational network. The parameters of a frequency-domain beamformer are first estimated by a network based on features derived from the microphone channels. These filter coefficients are then applied to the array signals to form an enhanced signal. Conventional features are then extracted from this signal and passed to a second network that performs acoustic modeling for classification. The parameters of both the beamforming and acoustic modeling networks are trained jointly using back-propagation with a common crossentropy objective function. In experiments the AMI meeting corpus, we observed improvements by pre-training each sub-network with a network-specific objective function before joint training of both networks. The proposed method obtained a 3.2% absolute word error rate reduction compared to a conventional pipeline of independent processing stages.