Combination of probabilistic and possibilistic language models

In a previous paper we proposed Web-based language models relying on the possibility theory. These models explicitly represent the possibility of word sequences. In this paper we propose to find the best way of combining this kind of model with classical probabilistic models, in the context of automatic speech recognition. We propose several combination approaches, depending on the nature of the combined models. With respect to the baseline, the best combination provides an absolute word error rate reduction of about 1% on broadcast news transcription , and of 3.5% on domain-specific multimedia document transcription. Index Terms: language models, world wide web, possibility measure, automatic speech recognition

[1]  Didier Dubois,et al.  Possibility Theory - An Approach to Computerized Processing of Uncertainty , 1988 .

[2]  Joshua Goodman,et al.  A bit of progress in language modeling , 2001, Comput. Speech Lang..

[3]  Georges Linarès,et al.  Probabilistic and possibilistic language models based on the world wide web , 2009, INTERSPEECH.

[4]  Ciro Martins,et al.  Dynamic language modeling for a daily broadcast news transcription system , 2007, 2007 IEEE Workshop on Automatic Speech Recognition & Understanding (ASRU).

[5]  Didier Dubois,et al.  Possibility theory and statistical reasoning , 2006, Comput. Stat. Data Anal..

[6]  Robert Miller,et al.  Just-in-time language modelling , 1998, Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '98 (Cat. No.98CH36181).

[7]  Marcello Federico,et al.  Broadcast news LM adaptation over time , 2004, Comput. Speech Lang..