12
A New Bigram-PLSA Language Model for Speech Recognition Mohammad Bahrani and Hossein Sameti 報報報 報報報 2013/03/14 EURASIP 2010 Department of Computer Engineering, Sharif University of Technology

A New Bigram-PLSA Language Model for Speech Recognition

  • Upload
    vinny

  • View
    33

  • Download
    1

Embed Size (px)

DESCRIPTION

A New Bigram-PLSA Language Model for Speech Recognition. Mohammad Bahrani and Hossein Sameti. Department of Computer Engineering, Sharif University of Technology. EURASIP 2010. 報告者:郝柏翰. Outline. Introduction Review of the PLSA Model Combining Bigram and PLSA Models Experiments - PowerPoint PPT Presentation

Citation preview

Page 1: A New Bigram-PLSA  Language Model  for Speech Recognition

A New Bigram-PLSA Language Model for Speech Recognition

Mohammad Bahrani and Hossein Sameti

報告者:郝柏翰2013/03/14

EURASIP 2010

Department of Computer Engineering, Sharif University of Technology

Page 2: A New Bigram-PLSA  Language Model  for Speech Recognition

2

Outline

• Introduction

• Review of the PLSA Model

• Combining Bigram and PLSA Models

• Experiments

• Conclusion

Page 3: A New Bigram-PLSA  Language Model  for Speech Recognition

3

Review of the PLSA Model

𝑃 (𝑤𝑖|𝑑 𝑗 )=∑𝑘

𝑃 (𝑤𝑖∨𝑧𝑘 )𝑃 (𝑧𝑘∨𝑑 𝑗)

• Bag-of-words

• Conditional independent

Page 4: A New Bigram-PLSA  Language Model  for Speech Recognition

4

Combining Bigram and PLSA Models

1. Nie et al.’s Bigram-PLSA Model

2. Proposed Bigram-PLSA Model

we relax the assumption of independence between the latent topics and the context words and achieve a general form of the aspect model that considers the word history in the word document modeling.

l lijkil

iki

jik

zwwPdwzP

wdPwPwwdP

),|(),|(

)|()(),,(

Page 5: A New Bigram-PLSA  Language Model  for Speech Recognition

5

Parameter Estimation Using the EM Algorithm

𝑃 (𝑧𝑙|𝑑𝑘 ,𝑤 𝑖 ,𝑤 𝑗 )=𝑃 ( 𝑧𝑙 ,𝑑𝑘 ,𝑤 𝑖 ,𝑤 𝑗)

∑𝑙′𝑃 (𝑧 𝑙′ ,𝑑𝑘 ,𝑤𝑖 ,𝑤 𝑗)

• E-step

),|()|( kilik dwzPwdP

Page 6: A New Bigram-PLSA  Language Model  for Speech Recognition

6

Parameter Estimation Using the EM Algorithm

Let be the set of model parameters

apply Bayes’ rule

• M-step

Page 7: A New Bigram-PLSA  Language Model  for Speech Recognition

7

Parameter Estimation Using the EM Algorithm

• Using Jensen’s inequality

Page 8: A New Bigram-PLSA  Language Model  for Speech Recognition

8

Jensen’s inequality

)(1

)()('

k

ii

xP

xPxP

Page 9: A New Bigram-PLSA  Language Model  for Speech Recognition

9

Parameter Estimation Using the EM Algorithm

• appropriate Lagrange multipliers

Page 10: A New Bigram-PLSA  Language Model  for Speech Recognition

10

Comparison with Nie et al.’s Bigram-PLSA Model.

• The difference between our model and Nie et al.’s model is in the definition of the topic probability.

• we relax the assumption of independence between the latent topics and the context words and achieve a general form of the aspect model that considers the word history in the word-document modeling.

• The number of free parameters in our proposed model is

in Nie et al.’s model is

Page 11: A New Bigram-PLSA  Language Model  for Speech Recognition

11

Experiments

Page 12: A New Bigram-PLSA  Language Model  for Speech Recognition

12

Experiments