next up previous
Next: Prompt Selection Up: Implementation Previous: Model Training

Topic Classification

After the training data is collected and the class models are built, we run the system receiving the audio input from participants of the conversation. The audio input is processed by the speech recognition module which outputs a list of the word candidates as a discounted conversation model. The matching algorithm sequentially updates the conversation model in such a way that the words spoken most recently have the largest score, which is slowly decaying. For a conversation model ${\bf x}$, we compute the updates at each step after receiving a word wordk:


 \begin{displaymath}x_i^t = \alpha x_i^{t-1} + \delta(k, i)
\end{displaymath} (2)

where $\alpha$ is the decay parameter, $\delta(k, i)$ equals 1 if the the wordk is the same word as xi (i.e. i = k).

Having computed the conversation model for the time step t, its class-conditional probability is computed as follows:


 \begin{displaymath}P({\bf x}\vert c) = \prod_i P(word_i\vert c)^{x_i}
\end{displaymath} (3)

This probability is converted to the posterior probability of the topic c using the Bayes rule. Here, the prior probabilities P(c) of each class are estimated by cross-validation:


 \begin{displaymath}P(c\vert{\bf x}) = \frac{P({\bf x}\vert c) P(c)}{\sum_{k=1}^{C} P({\bf x}\vert k) P(k)}
\end{displaymath} (4)


  
Figure: Plot of class probabilities. The initial segment of spoken input related to the topic of ``intlcourtofjustice''. After the system settled, the correct topic was clearly prominent. The middle part shows change of topic to ``talk.religion.misc'' the model of which subsequently became dominant. The last part of the plot shows a switch to ``alt.jobs''.
\begin{figure}\psfig{figure=topics.eps,width=5in,height=3.1in}\end{figure}

Figure 1 shows class probabilities for the ongoing conversation. After these probabilities are computed for each class the most likely topic c is selected and the corresponding feedback is given to the users as described below.


next up previous
Next: Prompt Selection Up: Implementation Previous: Model Training
Tony Jebara
2000-02-24