WO1995028700A1 - Topic discriminator - Google Patents

Topic discriminator Download PDF

Info

Publication number
WO1995028700A1
WO1995028700A1 PCT/US1995/004673 US9504673W WO9528700A1 WO 1995028700 A1 WO1995028700 A1 WO 1995028700A1 US 9504673 W US9504673 W US 9504673W WO 9528700 A1 WO9528700 A1 WO 9528700A1
Authority
WO
WIPO (PCT)
Prior art keywords
topic
event
word
speech
die
Prior art date
Application number
PCT/US1995/004673
Other languages
French (fr)
Inventor
John W. Mcdonough
Herbert Gish
Jan Robin Rohlicek
Kenney Ng
Phillippe Jeanrenaud
Original Assignee
Bolt Beranek And Newman Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Bolt Beranek And Newman Inc. filed Critical Bolt Beranek And Newman Inc.
Priority to JP7527135A priority Critical patent/JPH08512148A/en
Priority to EP95915682A priority patent/EP0708960B1/en
Priority to AU22480/95A priority patent/AU688030B2/en
Priority to CA002162696A priority patent/CA2162696C/en
Priority to DE69524036T priority patent/DE69524036T2/en
Publication of WO1995028700A1 publication Critical patent/WO1995028700A1/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1822Parsing for meaning understanding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/10Speech classification or search using distance or distortion measures between unknown speech and reference templates
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1815Semantic context, e.g. disambiguation of the recognition hypotheses based on word meaning

Definitions

  • the present invention relates generally to speech recognition systems, and more particularly to an improved topic discri-minator of a spoken message.
  • a topic discriminator for spoken data is used to classify the data into one of a set of known topics or to discriminate between data belonging to a known topic from data belonging to other topics.
  • the topic discrimination is usually performed using only features extracted from the speech data itself.
  • Applications similar, but not identical, to topic discrimination have been disclosed previously in the art, and have been designated by terms such as “gisting”, “topic identification”, and as " automatic acquisition of language .
  • An example of a prior use of a speech topic discriminator includes classification of recording of air-traffic-control dialogs by whether the flight is landing, taking off, or neither landing or taking off, as was described in Rohlicek, Ayuso et al. (1992) (J.R. Rohlicek and D.
  • Implementing a topic discriminator generally involves a training cycle in which a human operator selects the topic categories of interest. Selected topics may be, for example, (1) the weather, (2) the arts, and (3) sports. As part of the training cycle, the operator also provides a set of recorded speech messages that exemplify each of the selected topic categories. In the above example, the operator would provide a set of recorded speech messages about the weather, a set of recorded speech messages about the arts, and a set of recorded speech messages about sports. The set of all the recorded speech messages used in trairiing is generally known as a traiiiing corpus.
  • a training corpus is generally developed by recording speech samples of one or more people, as for example, where one or more people have been directed to speak about a specific topic (e.g., the weather).
  • a good training corpus typically contains speech messages recorded from a large number of people.
  • a training corpus may contain written transcripts of the speech messages, acoustically recorded speech messages, or both.
  • a topic discriminator based on the determined content of the message the discriminator will attempt to recognize whether the determined subject matter of the input speech message is more similar to the subject matter of those speech messages of the training corpus in one of the categories than to the subject matter of those speech messages of the training corpus in the other categories.
  • the basic approach to the problem has been to treat topic classification as a text classification problem with the text being created by a speech recognizer. For example, Farrell, et al., (K. Farrell, R.J. Mammone and A.L.
  • an unknown word can be identified in an incoming message.
  • the new word is then added to the vocabulary through the creation of a new input node and its association with the desired action is learned through an iterative training process.
  • the training process attempts to increase the rate of learning for new words appearing in messages that were initially mis-classified. This learning process, however, requires that the system be able to query the user as to the correctness of the action it proposes (e.g., "Would you like to be connected with the furniture department?"), and subsequently re-learn those messages which produce undesirable recommendations .
  • the word recognizer comprising one component of the gisting system is able to make use of finite state networks specifically designed to model each of a number of commonly occurring words and phrases; less commonly occurring words are not as explicitly modeled, but instead are represented by a phoneme or "filler" loop.
  • Message classification is performed in the gisting system by forming a binary vector representing each word or phrase present in a recognized utterance, which may well be errorful. This vector is taken as the input to a classification tree that has been previously constructed based on some amount of recognized training data. See Breimin, et al. (L. Breiman, J.H. Friedman, R.A. Olshen and C.J.
  • This system Because this system is intended to be used on general, unconstrained speech, it uses a large vocabulary and a bigram or stochastic "language" model.
  • the system employs a set of "keywords" that are relevant to a given topic. These words are found by taking text transcripts, compiled by human transcribers, and building contingency tables for each possible keyword; a contingency table tabulates the number of conversations in which a given word appeared seldom or often and can be used as the basis of a hypothesis test as to whether the frequency of occurrence of a word is significantly different across two or more topics.
  • the system of Gillick et al. also uses text transcripts to construct topic models, which in this case are unigram or multi-nomial models of topic- conditioned keyword-frequency.
  • Topic classification is performed by running the large vocabulary word recognizer on an input speech message and scoring the resulting errorful transcript against each competing topic model—the conversation is classified as belonging to that topic whose model scores highest. In this system, no attempt is made to associate a score indicative of the accuracy of the recognizer output with any word or phrase; i.e., none of the statistics generated during the recognition process contribute to the subsequent topic classification process.
  • techniques for discrimination of naturally spoken speech messages by topic have been described in the prior art. Several simply use a speech recognizer to produce an hypothesized transcription of the spoken data which is men input to a text-based topic discrimination system trained only on correctly transcribed text.
  • Rose et al (1991) use text training data but also incorporate some characteristics of their word spotter in the design of their topic discriminator.
  • the prior techniques may be applicable in certain situations, there are limitations that are addressed by the current invention. In particular, all the prior techniques require either transcribed speech data for training me topic discriminator, do not make use of a phrase spotter as a detector for events useful for topic discrimination, do not use word or phrase spotting confidence measure to improve performance, or require some sort of user feedback for training or during actual operation.
  • Objects of the Invention are to provide a topic discriminator which addresses situations for which the prior art is not applicable and to provide substantially improved topic discrimination performance relative to that possible using the prior art.
  • a more specific object of the present invention is to provide a topic discriminator capable of working when little or no transcribed training data is available for the topic modeling component of the discriminator.
  • Another specific object of the present invention is to address situations in which user feedback is undesirable or impossible, such as when using prerecorded speech data.
  • Yet another specific object of the present invention is to provide an improved topic discriminator, which can use the occurrence of complex phrases, for example any dollar amount, in addition to words or simple word sequences as input events on which to base topic discrimination decisions.
  • Still another object of the present invention is to improve performance of a topic discriminator by making direct use of confidence scores from a word or phrase spotter, rather than using binary detection decisions or hypothesized transcriptions, in the topic discriminator.
  • Another object of the present invention is to provide a means of determining a set of potential speech events to be detected by the speech event detector of a topic discriminator using transcribed training data if it is available.
  • an improved topic discriminator that includes an integrated speech recognizer or word and phrase spotter as part of a speech event detector, and a topic classifier trained on topic-dependent event frequencies.
  • the event frequencies are determined from either or both transcribed data with a text event detector and untranscribed speech data with a speech event detector.
  • the phrase spotter is used to detect the presence of phrases without the need of parsing the output of a speech recognizer's hypothesized transcription.
  • an improved technique is used to select a subset of the potential speech events on whose event frequencies the topic discrimination decision is made.
  • Figure 1 is a block diagram of the preferred embodiment of the topic discriminator of the present invention for use in processing new spoken data
  • Figure 2 is a block diagram of the components that are used to train the system of Figure 1, i.e., how to develop the model parameters of the Figure 1 system
  • Figure 3 is a block diagram of further components that are used in a preliminary phase of training, i.e., how one obtains the potential speech events
  • Figure 4a shows one type of contingency table used for selecting keywords for use in a topic discriminator according to the invention
  • Figure 4b shows another type of contingency table used for selecting keywords for use in a topic discriminator according to the invention.
  • Figure 1 is a block diagram of the components that are used to process a spoken message, or other speech data input, indicated at 10, and ultimately produce the topic classifier output 18.
  • the spoken message 10 is processed by a speech event frequency detector 12, which in turn is coupled to a predetermined set of potential speech events 20, e.g., a vocabulary of predetermined words and phrases.
  • the speech event frequency detector produces a signal representative of a set of event frequencies 14 for the potential speech events.
  • the potential speech events 20 can include individual words, multiword phrases, and complex phrases specified in a form such as a regular expression or a context-free grammar.
  • the event frequencies are preferably estimates of the frequency of occurrence of the potential speech events in the spoken data.
  • the speech event frequency detector preferably includes a speech recognizer or word and phrase spotter.
  • the frequency of occurrence of the specified potential speech events is determined by processing the output of me speech recognizer or word spotter although such processing could be integrated into the speech recognizer or word and phrase spotter.
  • the event frequencies are processed by the topic classifier 16 to produce the topic classifier output 18.
  • the output can take the form of a choice from a preselected set of known topics, a choice of either presence or absence of a particular known topic, or a confidence score that a particular known topic is present.
  • the topic classifier 16 makes use of topic model parameters 22 that are determined during a prior, or potentially ongoing, training procedure.
  • FIG. 2 is a block diagram of the primary components used in the framing procedure for training the system of Figure 1.
  • the input to the training procedure is either or both a transcribed traii-ing data 30 and untranscribed speech data 36.
  • Transcribed training data corresponds to a set of text sequences each identified with the true topic.
  • Untranscribed speech data corresponds to a set of recorded messages each labeled with its true topic but which are otherwise unannotated.
  • each text sequence is converted into a set of event frequencies using the text event detector 32. For each of a set of potential text events 34, the text event detector scans the text and determines the frequency of occurrence of that event.
  • potential text events can include individual words, multiword phrases, and complex phrases specified in a form such as a regular expression or a context-free grammar.
  • each speech data sequence is processed using the speech event frequency detector 38 into a set of event frequencies in the same manner as will be used for new unknown topic messages using the speech event detector 12 shown in Figure 1.
  • the topic modeling component 42 uses as input the output 40 representative set of event frequencies along with the true topic labels.
  • topic modeling comprises first selection of a subset of the available events. Then, parametric probabilistic models for the event frequency of each of the selected events are estimated.
  • the parametric models take the form of multinomial distributions or mixtures of multinomial distributions, although other distributions can be used as described in greater detail hereinafter.
  • the topic model parameters 22 are then comprised of the selected subsets of events and the parameter values for the individual event frequency distributions.
  • Figure 3 is a block diagram of further components that are used in a preliminary phase of fraining to determine the potential speech events 20.
  • the set of potential speech events can be provided as an input to the system based, for example, on a human specification 58. Alternatively, the set of potential speech events can be determined from a set of transcribed text data 50.
  • transcribed fraining data 50 can be distinct from the transcribed fraining data 30, shown in Figure 2, used in the training of the topic classifier.
  • a text event detector 52 is used to generate at output 54 a signal representative of event frequencies from which the set of potential speech events 20 is selected by the potential speech event selector 56.
  • Event Frequency Detectors The speech event frequency detectors 12 and 38 of Figures 1 and 2, which are used eid er in processing new speech data or in fraining the system, are each designed to extract relevant features from the speech data. These features are the estimated frequency of occurrence of each event of a selected set of events. Events include presence of individual words, multiword phrases, and complex phrases specified in a form such as a regular expression or a context-free grammar. An example of a multiword phrase would be "credit card" or a brand name card such as "American Express".
  • Each event frequency detector 12 and 38 can be constructed by first detecting a hypothesized string of events and then computing the event frequencies from the string of events.
  • the hypothesized string can be a hypothesized word sequence produced by a speech recognizer, a set of putative word or phrase occurrences produced by a word spotter, or a set of putative word or phrase occurrences with associated confidence scores.
  • the event frequencies can be computed by simple counting of words in the case of the speech recognizer, by counting putative word and phrase events or by combining (e.g., summing) the confidence scores in the case of the word and phrase spotter, or by first detecting phrases in the word sequence (i.e., parsing) produced by the speech recognizer followed by counting the number of word and phrase events.
  • Other related techniques for generating speech events are also possible, as well as integrating the counting of combination steps directly into the speech recognition or word spotting phases.
  • a hidden Markov model (HMM) based word and phrase spotter is used.
  • is defined as:
  • the estimate of the number of occurrences of a word or phrase can be obtained by summing over time as follows:
  • n(w) ⁇ p(wjt) t
  • Topic Model Training Topic model training as shown in Figure 2 is composed of two steps. First, a subset of events are selected. Second, probabilistic models for the event frequencies, conditional on the topic, are estimated. The second step, topic modeling, is described first followed by the event selection step.
  • Topic Modeling Topic modeling is the process of constructing a probability model for the word or event occurrence patterns observed in speech messages about a given topic.
  • one such probability model is trained for every topic in a known set; optionally, a "background" topic may be defined so as to include die universe of topics not of interest to the potential user.
  • These models are trained using a set of word occurrence statistics that vary according to the specific topic on which they are based.
  • word occurrence statistics can be derived from text transcriptions indicated at 30 in Figure 2, where each transcription corresponds to a speech message in a training set.
  • such occurrence statistics can be derived from the output of event detector 38 when the speech messages 36 in the fraining set are used as input.
  • the second method is to be preferred, as it has proven more effective in capturing the specific characteristics of both the input speech, and me event detector that comprises one component of the topic discrimination system.
  • me event detector that comprises one component of the topic discrimination system.
  • There are several classes of probability models mat have proven effective in the topic modeling application.
  • One preferred class is predicated on an independent, identically-distributed (i.i.d.) assumption about the words or events observed in an event sequence w, and issues in a multi-nomial model.
  • the resulting sequence-likelihood, as conditioned on topic T t can be expressed as:
  • the out-of- vocabulary (OOV) event, V Q which denotes the occurrence of ' 'none of the above, ' ' is assigned probability l-J ⁇ ⁇ CT,-) and included in event vocabulary V.
  • Topic identification can equation (6).
  • the topic-conditional mixture density can be expressed as
  • Speech Event Subset Selection The prior art has taught that all events should be used in me development and training of a topic discrimination system. While tiiis may be the case in some instances, in accordance widi the teachings of the present invention improved performance is often obtained by restricting the set of events considered relevant to a topic.
  • event selection is necessary to determine the subset of words or omer acoustic events which, when observed or hypothesized in a speech message, best serve to indicate the topic membership of that message. As indicated in Figure 2, this type of selection can be made on the basis of either text transcriptions of speech messages or the output produced by an event detector when such speech messages are taken as input. The latter method is generally preferred for reasons given previously.
  • One preferred method of event selection is founded on me concept of hypothesis testing.
  • hypothesis testing is used to dete ⁇ nine whether or not there is sufficient evidence to establish that the occurrence pattern for a given keyword or event is dependent on die topic membership of the speech messages in which the event is observed. If such a dependence is established, d e event is assumed to be a good indicator of topic membership.
  • the ⁇ 2 -test is well-known in the art (see Alexander M. Mood and Franklin A. Graybill, Introduction to the Theory of Statistics, McGraw-Hill, NY, NY, 1950) and useful for testing dependencies of this type.
  • One possible formulation of the ⁇ 2 -test makes use of contingency tables of the type, shown in Figures 4a and 4b.
  • each cell of me contingency table contains the number of speech messages, drawn from a topic or set of topics, for which the candidate word or event was observed to occur witiiin a specified frequency range.
  • Two variations on the analysis are introduced by me manner in which the messages from each topic are grouped. In the table of Figure 4a, each topic is allotted its own row and the relevance of each word is tested across all topics simultaneously. This is referred to as the All- against-All (AAA) test.
  • AAA All- against-All
  • a single topic i.e., me topic of interest
  • the final keyword set in this variation which is referred to as Topic- against-AU-Other (TAO) test, is comprised of me union of all keyword subsets, each of which is chosen by designating a different topic as the topic of interest.
  • TAO Topic- against-AU-Other
  • the raw ⁇ 2 statistic can be converted into a corresponding significance value by means well-known in the art (see, for example, William H. Press, Saul A. Teukolsky, et al, Numerical Recipes in C: The Art of Scientific Computing, Cambridge University Press, Cambridge, England, 1992).
  • Event selection is accomplished by considering as relevant all those words or events whose significance values exceed a predefined threshold.
  • a second preferred method for event selection is based on me determination of a mutual information measure between the ensemble of event occurrences and die ensemble of message topic membership. Such measures are well-known in the art (see Thomas M. Cover and Joy A. Thomas, Elements of Information Theory, John Wiley and Sons, NY, NY, 1991).
  • the desired mutual information measure I(T,W) can be expressed as
  • T (n) ) can be estimated by resorting to any of the probability models previously discussed; a preferred method makes use of the multi-nomial model in combination with an event occurrence probability characterized by a mixture density.
  • a diird preferred method for event selection employs a Kullback-Liebler distance measure. Such a distance measure is well-known in the art (see, for example, (see Thomas M. Cover and Joy A. Thomas, Elements of Information Theory, John Wiley and Sons, NY, NY, 1991) and provides a measure of me dissimilarity of the occurrence patterns of an event for a given topic, as opposed to all other topics.
  • the Kullback-Liebler distance can be estimated according to where N(T) is the total number of conversations or messages relating to the topic of interest, and N(T) is the number relating to all other topics.
  • T (n) ) can be estimated using any of the probability models previously mentioned.
  • a generic phrase refers to a segment of speech that can be described by a formalized grammar. Examples of generic phrases are a phone number, a dollar amount, a date, and a time. Each generic phrase encompasses a large class of specific phrases. For example, a dollar amount might include the specific phrases "five dollars and ten cents" and "twenty-two dollars and diirty cents.” Another way to define a generic phrase is to enumerate every specific phrase in a generic category. For example, the generic phrase "dog breeds" could encompass a prescribed set of different kinds of dogs. Recent advances in the art (Jeanrenaud et al) provide for the construction of event detectors capable of modeling and detecting such generalized phrases.
  • the present invention can be used for sorting speech data in any one of a number of applications.
  • od er examples include sorting of speech data, such as radio news recordings, by the topic of the news story. This type of sorting can be used for automatic detection of speech data of interest or can be used as part of an automatic indexing mechanism for improved access to audio archives.
  • Still other examples include automatic response to or routing of phone calls based on die topic or subject matter in a spoken message from the caller.
  • Still other applications similar to the automatic indexing and data detection applications described above include classification of stored or incoming voice messages in a voice mail system.
  • the topic disc- ninator of the present invention thus addresses situations for which the prior art is not applicable; and provides substantially improved topic discrimination performance compared to me above-noted prior art techniques.
  • the topic discriminator of the present invention is capable of working when little or no transcribed training data is available for the topic modeling component of the discriminator; is capable of being used is situations in which user feedback is undesirable or impossible, such as when using prerecorded speech data; and can use die occurrence of complex phrases, for example any dollar amount, in addition to words or simple word sequences as input events on which to base topic discrimination decisions.
  • the performance of a topic discriminator is improved by making direct use of confidence scores from a word or phrase spotter, rather than using binary detection decisions or hypothesized transcriptions, and by using tiiese scores to derive an expected number of occurrences for every event in a given set.
  • the topic discriminator of the present invention allows one to determine a set of potential speech events to be detected by the text event detector of a topic discriminator using transcribed trai-ning data if it is available. Since certain changes may be made in the above apparatus and method, without departing from the scope of the invention herein involved, it is dierefore intended tiiat all matter contained in the above description or shown in the accompanying drawings shall be illustrative and not in a limiting sense.

Abstract

An improved topic discriminator includes an integrated speech recognizer or word and phrase spotter as part of a speech event detector, and a topic classifier (block 16) trained on topic-dependent event frequencies. The event frequencies (block 14) are determined from either or both transcribed data with a text event detector and untranscribed speech data with a speech event detector (block 12). In accordande with another aspect of the present invention, the phrase spotter is used to detect the presence of phrases without the need of parsing the output of a speech recognizer's hypothesized transcription. Finally, in accordance with another aspect of the invention, an improved method of topic modeling (block 20) is used to improve the performance of the topic discriminator.

Description

TOPIC DISCRIMINATOR
Field of the Invention The present invention relates generally to speech recognition systems, and more particularly to an improved topic discri-minator of a spoken message.
Background of the Invention A topic discriminator for spoken data is used to classify the data into one of a set of known topics or to discriminate between data belonging to a known topic from data belonging to other topics. The topic discrimination is usually performed using only features extracted from the speech data itself. Applications similar, but not identical, to topic discrimination have been disclosed previously in the art, and have been designated by terms such as "gisting", "topic identification", and as " automatic acquisition of language . ' ' An example of a prior use of a speech topic discriminator includes classification of recording of air-traffic-control dialogs by whether the flight is landing, taking off, or neither landing or taking off, as was described in Rohlicek, Ayuso et al. (1992) (J.R. Rohlicek and D. Ayuso, et al.; "Gisting Conversational Speech"; IEEE ICASSP; 1992; Volume II, pp. 113-116). Implementing a topic discriminator generally involves a training cycle in which a human operator selects the topic categories of interest. Selected topics may be, for example, (1) the weather, (2) the arts, and (3) sports. As part of the training cycle, the operator also provides a set of recorded speech messages that exemplify each of the selected topic categories. In the above example, the operator would provide a set of recorded speech messages about the weather, a set of recorded speech messages about the arts, and a set of recorded speech messages about sports. The set of all the recorded speech messages used in trairiing is generally known as a traiiiing corpus. A training corpus is generally developed by recording speech samples of one or more people, as for example, where one or more people have been directed to speak about a specific topic (e.g., the weather). A good training corpus typically contains speech messages recorded from a large number of people. A training corpus may contain written transcripts of the speech messages, acoustically recorded speech messages, or both. Once a topic discriminator has been provided with a training corpus, the discriminator attempts to determine which of the preselected topics is the most likely subject matter of each speech message received. In keeping with the above example, if a topic discriminator is provided with an input speech message, based on the determined content of the message the discriminator will attempt to recognize whether the determined subject matter of the input speech message is more similar to the subject matter of those speech messages of the training corpus in one of the categories than to the subject matter of those speech messages of the training corpus in the other categories. Several approaches to topic classification have been attempted in the past. The basic approach to the problem has been to treat topic classification as a text classification problem with the text being created by a speech recognizer. For example, Farrell, et al., (K. Farrell, R.J. Mammone and A.L. Gorin; "Adaptive Language Acquisition Using Incremental Learning"; IEEE ICASSP; 1993; Volume I; pp.501-504) have investigated the pairing of spoken phone messages with desired "actions". The actions considered are the routing of messages to one of several departments of a retail store. This system is based on a one-layer neural network whose connection weights are related to the "association" between a word known to the system, with each word represented by a node at the input layer of the neural network, and a desired action, each action being represented by a node at the output layer. While it is assumed that all possible actions are known, the system has the capacity to interactively learn new vocabulary words as it is being used by a customer. Using acoustic similarity measures between words spoken and the system's current vocabulary, an unknown word can be identified in an incoming message. The new word is then added to the vocabulary through the creation of a new input node and its association with the desired action is learned through an iterative training process. The training process attempts to increase the rate of learning for new words appearing in messages that were initially mis-classified. This learning process, however, requires that the system be able to query the user as to the correctness of the action it proposes (e.g., "Would you like to be connected with the furniture department?"), and subsequently re-learn those messages which produce undesirable recommendations . Additionally , the system presently under discussion cannot be used in applications where the user speaks "naturally" or without making a special effort to be understood—it is dependent on each word being spoken in isolation. Related research is described in Gorin, et al. (A.L. Gorin, L.G. Miller and S.E. Levinson; " Some Experiments in Spoken Language Acquisition" ; IEEE ICASSP; 1993 ; Volume I, pp. 505-508). A system similar to that proposed by Farrell, et al., and Gorin, et al. and apparently motivated by it has been described by Rose, et al. (R.C. Rose, E.I. Chang and R.P. Lippmann; "Techniques for Information Retrieval from Voice Messages"; IEEE ICASSP; 1991, Volume I, pp. 317-320). The latter group proposed the use of a word spotting system in conjunction with a one-layer neural network classifier whose weights are trained to minimize classification error. This system uses the spotting score associated with each putative hit as an indication of the "accuracy" of a given event. Unlike the Farrell, et al. and Gorin, et al. system, however, it does not have the capacity to learn new words through interactive use. J.R. Rohlicek and D. Ayuso, et al. (1992), supra; and Denenberg, et al. (L. Denenberg and H. Gish; "Gisting Conversational Speech in Real Time"; IEEE ICASSP; 1993, Volume II; pp. 131-134) have proposed and built a system for "gisting" conversational speech. The application to which this system was addressed was two-way communication between air traffic controllers and airplane pilots. The system attempts to determine approximately what the controller or pilot has said in each transmission; i.e., get the "gist" of the speech, defined as the flight scenario, such as take-off or landing, a given aircraft is in. This task is made tractable by the constrained nature of the dialogue between pilots and controllers. Typically each transmission must begin with a flight identification and then contain one or more instructions whose number is known in advance. For this reason, the word recognizer comprising one component of the gisting system is able to make use of finite state networks specifically designed to model each of a number of commonly occurring words and phrases; less commonly occurring words are not as explicitly modeled, but instead are represented by a phoneme or "filler" loop. Message classification is performed in the gisting system by forming a binary vector representing each word or phrase present in a recognized utterance, which may well be errorful. This vector is taken as the input to a classification tree that has been previously constructed based on some amount of recognized training data. See Breimin, et al. (L. Breiman, J.H. Friedman, R.A. Olshen and C.J. Stone; "Classification and Regression Trees"; Wadsworth International Group, Belmont, CA, 1984). The tree performs the desired classification based on an optimal set of "questions" about the absence or presence of sets of words and phrases. A variation on the basic approach attempts to reduce the effect of recognition errors by using for classification the N-best or most likely word sequences instead of only the single best. Gillick, et al. (L. Gillick and J. Baker, et al.; "Application of Large Vocabulary Continuous Speech Recognition to Topic and Speaker Identification Using Telephone Speech"; IEEE ICASSP; 1993, Volume II, pp. 471-474) have developed a system for topic identification for conversational speech over the telephone, as provided by the NIST Switchboard Corpus. Because this system is intended to be used on general, unconstrained speech, it uses a large vocabulary and a bigram or stochastic "language" model. The system employs a set of "keywords" that are relevant to a given topic. These words are found by taking text transcripts, compiled by human transcribers, and building contingency tables for each possible keyword; a contingency table tabulates the number of conversations in which a given word appeared seldom or often and can be used as the basis of a hypothesis test as to whether the frequency of occurrence of a word is significantly different across two or more topics. The system of Gillick et al. also uses text transcripts to construct topic models, which in this case are unigram or multi-nomial models of topic- conditioned keyword-frequency. Topic classification is performed by running the large vocabulary word recognizer on an input speech message and scoring the resulting errorful transcript against each competing topic model—the conversation is classified as belonging to that topic whose model scores highest. In this system, no attempt is made to associate a score indicative of the accuracy of the recognizer output with any word or phrase; i.e., none of the statistics generated during the recognition process contribute to the subsequent topic classification process. In summary, techniques for discrimination of naturally spoken speech messages by topic have been described in the prior art. Several simply use a speech recognizer to produce an hypothesized transcription of the spoken data which is men input to a text-based topic discrimination system trained only on correctly transcribed text. Rose et al (1991) use text training data but also incorporate some characteristics of their word spotter in the design of their topic discriminator. Although the prior techniques may be applicable in certain situations, there are limitations that are addressed by the current invention. In particular, all the prior techniques require either transcribed speech data for training me topic discriminator, do not make use of a phrase spotter as a detector for events useful for topic discrimination, do not use word or phrase spotting confidence measure to improve performance, or require some sort of user feedback for training or during actual operation.
Objects of the Invention General objects of the present invention are to provide a topic discriminator which addresses situations for which the prior art is not applicable and to provide substantially improved topic discrimination performance relative to that possible using the prior art. A more specific object of the present invention is to provide a topic discriminator capable of working when little or no transcribed training data is available for the topic modeling component of the discriminator. Another specific object of the present invention is to address situations in which user feedback is undesirable or impossible, such as when using prerecorded speech data. Yet another specific object of the present invention is to provide an improved topic discriminator, which can use the occurrence of complex phrases, for example any dollar amount, in addition to words or simple word sequences as input events on which to base topic discrimination decisions. Still another object of the present invention is to improve performance of a topic discriminator by making direct use of confidence scores from a word or phrase spotter, rather than using binary detection decisions or hypothesized transcriptions, in the topic discriminator. Another object of the present invention is to provide a means of determining a set of potential speech events to be detected by the speech event detector of a topic discriminator using transcribed training data if it is available.
Summary of the Invention These and other objects are provided by an improved topic discriminator that includes an integrated speech recognizer or word and phrase spotter as part of a speech event detector, and a topic classifier trained on topic-dependent event frequencies. The event frequencies are determined from either or both transcribed data with a text event detector and untranscribed speech data with a speech event detector. In accordance with another aspect of the present invention, the phrase spotter is used to detect the presence of phrases without the need of parsing the output of a speech recognizer's hypothesized transcription. In accordance with another aspect of the invention, an improved technique is used to select a subset of the potential speech events on whose event frequencies the topic discrimination decision is made. Finally, in accordance with another aspect of the invention, an improved method of topic modeling is used to improve the performance of the topic discriminator. Still other objects and advantages of the present invention will become readily apparent to those skilled in the art from the following detailed description wherein several embodiments are shown and described, simply by way of illustration of the best mode of the invention. As will be realized, the invention is capable of other and different embodiments, and its several details are capable of modifications in various respects, all without departing from the invention. Accordingly, the drawings and description are to be regarded as illustrative in nature, and not a restrictive or limiting sense, with the scope of the application being indicated in the claims.
Brief Description of the Drawings For a fuller understanding of the nature and objects of the present invention, reference should be had to the following detailed description taken in connection with the accompanying drawings in which the same reference numerals are used to indicate the same or similar parts wherein: Figure 1 is a block diagram of the preferred embodiment of the topic discriminator of the present invention for use in processing new spoken data; Figure 2 is a block diagram of the components that are used to train the system of Figure 1, i.e., how to develop the model parameters of the Figure 1 system; Figure 3 is a block diagram of further components that are used in a preliminary phase of training, i.e., how one obtains the potential speech events; and Figure 4a shows one type of contingency table used for selecting keywords for use in a topic discriminator according to the invention; and Figure 4b shows another type of contingency table used for selecting keywords for use in a topic discriminator according to the invention.
Detailed Description of the Drawings Block diagrams of an improved topic discriminator according to the invention and its development are shown in Figures 1, 2 and 3. Figure 1 is a block diagram of the components that are used to process a spoken message, or other speech data input, indicated at 10, and ultimately produce the topic classifier output 18. The spoken message 10 is processed by a speech event frequency detector 12, which in turn is coupled to a predetermined set of potential speech events 20, e.g., a vocabulary of predetermined words and phrases. The speech event frequency detector produces a signal representative of a set of event frequencies 14 for the potential speech events. The potential speech events 20 can include individual words, multiword phrases, and complex phrases specified in a form such as a regular expression or a context-free grammar. The event frequencies are preferably estimates of the frequency of occurrence of the potential speech events in the spoken data. The speech event frequency detector preferably includes a speech recognizer or word and phrase spotter. The frequency of occurrence of the specified potential speech events is determined by processing the output of me speech recognizer or word spotter although such processing could be integrated into the speech recognizer or word and phrase spotter. The event frequencies are processed by the topic classifier 16 to produce the topic classifier output 18. The output can take the form of a choice from a preselected set of known topics, a choice of either presence or absence of a particular known topic, or a confidence score that a particular known topic is present. The topic classifier 16 makes use of topic model parameters 22 that are determined during a prior, or potentially ongoing, training procedure. \ Figure 2 is a block diagram of the primary components used in the framing procedure for training the system of Figure 1. The input to the training procedure is either or both a transcribed traii-ing data 30 and untranscribed speech data 36. Transcribed training data corresponds to a set of text sequences each identified with the true topic. Untranscribed speech data corresponds to a set of recorded messages each labeled with its true topic but which are otherwise unannotated. In the case where transcribed traimng data is available as indicated at output 40, each text sequence is converted into a set of event frequencies using the text event detector 32. For each of a set of potential text events 34, the text event detector scans the text and determines the frequency of occurrence of that event. As is the case with the potential speech events 20, potential text events can include individual words, multiword phrases, and complex phrases specified in a form such as a regular expression or a context-free grammar. In the case where untranscribed speech data is available, as indicated by the output 40, each speech data sequence is processed using the speech event frequency detector 38 into a set of event frequencies in the same manner as will be used for new unknown topic messages using the speech event detector 12 shown in Figure 1. The topic modeling component 42 uses as input the output 40 representative set of event frequencies along with the true topic labels. In the preferred embodiment, topic modeling comprises first selection of a subset of the available events. Then, parametric probabilistic models for the event frequency of each of the selected events are estimated. In the preferred embodiment, the parametric models take the form of multinomial distributions or mixtures of multinomial distributions, although other distributions can be used as described in greater detail hereinafter. The topic model parameters 22 are then comprised of the selected subsets of events and the parameter values for the individual event frequency distributions. Figure 3 is a block diagram of further components that are used in a preliminary phase of fraining to determine the potential speech events 20. The set of potential speech events can be provided as an input to the system based, for example, on a human specification 58. Alternatively, the set of potential speech events can be determined from a set of transcribed text data 50. Note that the transcribed fraining data 50 can be distinct from the transcribed fraining data 30, shown in Figure 2, used in the training of the topic classifier. When text data is used, a text event detector 52 is used to generate at output 54 a signal representative of event frequencies from which the set of potential speech events 20 is selected by the potential speech event selector 56.
Event Frequency Detectors The speech event frequency detectors 12 and 38 of Figures 1 and 2, which are used eid er in processing new speech data or in fraining the system, are each designed to extract relevant features from the speech data. These features are the estimated frequency of occurrence of each event of a selected set of events. Events include presence of individual words, multiword phrases, and complex phrases specified in a form such as a regular expression or a context-free grammar. An example of a multiword phrase would be "credit card" or a brand name card such as "American Express". An example of a complex phrase would be a syntactically correct flight identification in an air-ttaffic-conttol command, or a time including any form such as "twelve o'clock", "noon", or "five minutes to ten." Each event frequency detector 12 and 38 can be constructed by first detecting a hypothesized string of events and then computing the event frequencies from the string of events. In this form of event frequency detector, the hypothesized string can be a hypothesized word sequence produced by a speech recognizer, a set of putative word or phrase occurrences produced by a word spotter, or a set of putative word or phrase occurrences with associated confidence scores. The event frequencies can be computed by simple counting of words in the case of the speech recognizer, by counting putative word and phrase events or by combining (e.g., summing) the confidence scores in the case of the word and phrase spotter, or by first detecting phrases in the word sequence (i.e., parsing) produced by the speech recognizer followed by counting the number of word and phrase events. Other related techniques for generating speech events are also possible, as well as integrating the counting of combination steps directly into the speech recognition or word spotting phases. In a preferred embodiment of the event frequency detector, a hidden Markov model (HMM) based word and phrase spotter is used. Confidence scores which are probabilities of occurrence of the putative words and phrase are computed using the techniques described in Rohlicek, et al. (1989) (J.R. Rohlicek, W.Russell, S. Roukos, and H. Gish; "Continuous Hidden Markov Modeling for Speaker- Independent Word Spotting", IEEE ICASSP, 1989, pp. 627-630), Rohlicek, et al. (1993) (J.R. Rohlicek, P. Jeanrenaud, K. Ng, H. Gish, et al. "Phonetic Training and Language Modeling for Word Spotting" IEEE ICASSP, 1993, volume II, pp. 459- 462), and Jeanrenaud, et al., (1993) (P. Jeanrenaud, M. Siu, K. Ng. R. Rohlicek, and H. Gish; "Phonetic-based Word Spotter: Various Configurations and Application to Event Spotting"; in ESCA Eurospeech, 1993, Volume II, pp 1057-1060). Summing of these scores for each event over the speech data directly provides the expected number of occurrences of the event given the data. In the cited approaches to word spotting, the scoring statistic computed is either the probability of ending a keyword at time t given input data up to that time O ...,0„ or else the probability of ending the keyword given all the data in the message up to the ending time T. The former is computed as: (1) p(w,t) = — -^
Y all states saM rf wherein ew is the last state of me keyword and a is defined as:
(2) ct(s,t) = P(state at t = s, O ...,Ot)
and can be computed using a well-known interation, e.g. , L. R. Rabiner in Readings in Speech Recognition, pp. 267-296. A Waibel, 2nd, K.F. Lee (eds.) 1990. The latter probability is computed as
a(ew,t)β e ) (3) p(wjt)
∑ oil states s a(sj)β(sf)
wherein β is defined as:
(4) β(s,t) = P(Ol+1,...,Oτ I state at t = s)
and can be computed with another well-known iteration. The probability of a phrase event is computed in the same manner except that the end state ew is replaced with the ending state of a subnetwork allowing all valid word sequences for the phrase. Using these score definitions, the estimate of the number of occurrences of a word or phrase can be obtained by summing over time as follows:
(5) n(w) = ∑p(wjt) t
In the case where p(w,t) is computed from the a and β terms, n(w) is the conditional expectation of the number of occurrences given all the data 0ι,...,Or. The potential speech events can be used to configure the speech recognizer or word and phrase spotter so that these events are reported in the hypothesized event string. Topic Model Training Topic model training as shown in Figure 2 is composed of two steps. First, a subset of events are selected. Second, probabilistic models for the event frequencies, conditional on the topic, are estimated. The second step, topic modeling, is described first followed by the event selection step.
Topic Modeling Topic modeling is the process of constructing a probability model for the word or event occurrence patterns observed in speech messages about a given topic. In order to provide a basis for distinguishing between or amongst topics, one such probability model is trained for every topic in a known set; optionally, a "background" topic may be defined so as to include die universe of topics not of interest to the potential user. These models are trained using a set of word occurrence statistics that vary according to the specific topic on which they are based. Such word occurrence statistics can be derived from text transcriptions indicated at 30 in Figure 2, where each transcription corresponds to a speech message in a training set. Alternatively, such occurrence statistics can be derived from the output of event detector 38 when the speech messages 36 in the fraining set are used as input. Generally, the second method is to be preferred, as it has proven more effective in capturing the specific characteristics of both the input speech, and me event detector that comprises one component of the topic discrimination system. There are several classes of probability models mat have proven effective in the topic modeling application. One preferred class is predicated on an independent, identically-distributed (i.i.d.) assumption about the words or events observed in an event sequence w, and issues in a multi-nomial model. The resulting sequence-likelihood, as conditioned on topic Tt , can be expressed as:
Figure imgf000014_0001
where r . is the number of times event vk is hypothesized or observed in sequence w; N=∑ n, is the total number of events observed in w; and pk(T)=P(vk\T^ is the k=0 probability of the occurrence of event vk conditioned on topic Tt . It is me set of event occurrence probabilities {/>Λ(-r£)} -0 that must be estimated using either text transcriptions of ttai-ning messages, or the output of an event detector corresponding to such messages. To provide for a non-exhaustive event set V, the out-of- vocabulary (OOV) event, VQ , which denotes the occurrence of ' 'none of the above, ' ' is assigned probability l-J^ ^CT,-) and included in event vocabulary V. Topic identification can
Figure imgf000015_0001
equation (6). A second class of preferred topic models is mat based on a multi-nomial topic model, in which the event occurrence probabilities Θk=P(vk) are not assumed to be known or deterministic, but rather are represented with a probabilistic mixture density. For the complete vector of occurrence parameters, θ={θk}k=Q, the topic-conditional mixture density can be expressed as
Figure imgf000015_0002
where^ isthepriorprobabilityoftheselectionofmixturemconditionedontopic Tt, f^iθ) is the conditional multi-variate probability density function of θ ; and Mi is the total number of mixtures associated with topic Ti . When the mixture density of equation (7) is used in conjunction with the multi-nomial model of equation (6), the resulting sequence likelihood can be expressed as
(8) lo P(w limUΘ)
Figure imgf000015_0003
Two preferred density functions f^θ) are the Dirac delta function, and the beta-density function, given by
Figure imgf000015_0004
Ctøg.ώ 'ϊff)" where C(nf n^ '-••• ,n^) is a normalization constant chosen to ensure f^iθ) is a proper density function. Both such densities are well-known to those skilled in the art (see, for example, Alexander M. Mood and Franklin A. Graybill, Introduction to the Theory of Statistics, McGraw-Hill, NY, NY, 1950). The parameters associated with mixture densities such as that represented in equation (7) can be estimated using the Estimation-Maximization algorithm, which is also well-known to those skilled in the art (see, for example, A. P. Dempster, N. M. Laird and D. B. Rubin, "Maximum Likelihood from Incomplete Data via the EM Algorithm," Journal Royal Statistical Society, Series B, 39:1-38,1977).
Speech Event Subset Selection The prior art has taught that all events should be used in me development and training of a topic discrimination system. While tiiis may be the case in some instances, in accordance widi the teachings of the present invention improved performance is often obtained by restricting the set of events considered relevant to a topic. Hence, event selection is necessary to determine the subset of words or omer acoustic events which, when observed or hypothesized in a speech message, best serve to indicate the topic membership of that message. As indicated in Figure 2, this type of selection can be made on the basis of either text transcriptions of speech messages or the output produced by an event detector when such speech messages are taken as input. The latter method is generally preferred for reasons given previously. One preferred method of event selection is founded on me concept of hypothesis testing. According to one aspect of the invention, hypothesis testing is used to deteπnine whether or not there is sufficient evidence to establish that the occurrence pattern for a given keyword or event is dependent on die topic membership of the speech messages in which the event is observed. If such a dependence is established, d e event is assumed to be a good indicator of topic membership. The χ2-test is well-known in the art (see Alexander M. Mood and Franklin A. Graybill, Introduction to the Theory of Statistics, McGraw-Hill, NY, NY, 1950) and useful for testing dependencies of this type. One possible formulation of the χ2-test makes use of contingency tables of the type, shown in Figures 4a and 4b. To implement the test or analysis, a table is constructed for each candidate word or event by first deteπnining how frequently it appears in either the text transcripts or event detector output associated witii each speech message, and men "binning" the messages according to tiiis frequency. Hence, each cell of me contingency table contains the number of speech messages, drawn from a topic or set of topics, for which the candidate word or event was observed to occur witiiin a specified frequency range. Two variations on the analysis are introduced by me manner in which the messages from each topic are grouped. In the table of Figure 4a, each topic is allotted its own row and the relevance of each word is tested across all topics simultaneously. This is referred to as the All-Against-All (AAA) test. In the table of Figure 4b, a single topic (i.e., me topic of interest) is allotted the first row and the remaining topics are grouped together in the second. The final keyword set in this variation, which is referred to as Topic-Against-AU-Other (TAO) test, is comprised of me union of all keyword subsets, each of which is chosen by designating a different topic as the topic of interest. Once a contingency table has been constructed, die corresponding χ2 value can be determined in the following manner. The number of conversations Ntj assigned to cell (i,j) is found for each cell in die table and tiiese conversation counts are used to find me row and column totals
(10) N, = ∑Nij
and
(11) Ν,. = ∑JVf
respectively. The χ2-value is then provided by
Figure imgf000017_0001
where is the expected number of counts for each cell.
Figure imgf000018_0001
The raw χ2 statistic can be converted into a corresponding significance value by means well-known in the art (see, for example, William H. Press, Saul A. Teukolsky, et al, Numerical Recipes in C: The Art of Scientific Computing, Cambridge University Press, Cambridge, England, 1992). Event selection is accomplished by considering as relevant all those words or events whose significance values exceed a predefined threshold. A second preferred method for event selection is based on me determination of a mutual information measure between the ensemble of event occurrences and die ensemble of message topic membership. Such measures are well-known in the art (see Thomas M. Cover and Joy A. Thomas, Elements of Information Theory, John Wiley and Sons, NY, NY, 1991). The desired mutual information measure I(T,W) can be expressed as
Figure imgf000018_0002
where p(wk \T(μ)) is the likelihood of die occurrence pattern for word or event v k, observed in fraining sequence w (n) conditioned on its true topic T n) , and P(wk"))=^/p(wk n) \T)p(T) is the likelihood for the same occurrence pattern
T independent of topic. The topic-conditional sequence likelihoods p(wk n) |T(n)) can be estimated by resorting to any of the probability models previously discussed; a preferred method makes use of the multi-nomial model in combination with an event occurrence probability characterized by a mixture density. A diird preferred method for event selection employs a Kullback-Liebler distance measure. Such a distance measure is well-known in the art (see, for example, (see Thomas M. Cover and Joy A. Thomas, Elements of Information Theory, John Wiley and Sons, NY, NY, 1991) and provides a measure of me dissimilarity of the occurrence patterns of an event for a given topic, as opposed to all other topics. The Kullback-Liebler distance can be estimated according to
Figure imgf000019_0001
where N(T) is the total number of conversations or messages relating to the topic of interest, and N(T) is the number relating to all other topics. --nth--si-n--tance,as-ntheprevious,tte | T(n)) can be estimated using any of the probability models previously mentioned.
Potential Speech Event Selection Before an acoustic event detector can be constructed, some initial set of possible words or events must be defined. This is necessary as these words or events must be modeled in me internal strucmre of the event detector. Such an initial selection can accomplished in a number of ways. One preferred method is represented schematically in Figure 3 by boxes 50, 52, 54 and 56. It entails using any of the previously discussed techniques for selecting keywords, taking as input some small imtial set of text transcriptions of speech messages. From these latter, all word frequency statistics needed to conduct hypothesis tests or calculate distance metrics can be extracted. According to anotiier preferred metiiod of imtial event selection, additional keywords or events corresponding to generic phrases may be selected by a human operator. This process is illustrated by box 58 of Figure 3. A generic phrase refers to a segment of speech that can be described by a formalized grammar. Examples of generic phrases are a phone number, a dollar amount, a date, and a time. Each generic phrase encompasses a large class of specific phrases. For example, a dollar amount might include the specific phrases "five dollars and ten cents" and "twenty-two dollars and diirty cents." Another way to define a generic phrase is to enumerate every specific phrase in a generic category. For example, the generic phrase "dog breeds" could encompass a prescribed set of different kinds of dogs. Recent advances in the art (Jeanrenaud et al) provide for the construction of event detectors capable of modeling and detecting such generalized phrases. The present invention can be used for sorting speech data in any one of a number of applications. For example, in addition to classifying recordings of air- traffic control dialogs, od er examples include sorting of speech data, such as radio news recordings, by the topic of the news story. This type of sorting can be used for automatic detection of speech data of interest or can be used as part of an automatic indexing mechanism for improved access to audio archives. Still other examples include automatic response to or routing of phone calls based on die topic or subject matter in a spoken message from the caller. Still other applications similar to the automatic indexing and data detection applications described above include classification of stored or incoming voice messages in a voice mail system. The topic disc- ninator of the present invention thus addresses situations for which the prior art is not applicable; and provides substantially improved topic discrimination performance compared to me above-noted prior art techniques. The topic discriminator of the present invention is capable of working when little or no transcribed training data is available for the topic modeling component of the discriminator; is capable of being used is situations in which user feedback is undesirable or impossible, such as when using prerecorded speech data; and can use die occurrence of complex phrases, for example any dollar amount, in addition to words or simple word sequences as input events on which to base topic discrimination decisions. The performance of a topic discriminator is improved by making direct use of confidence scores from a word or phrase spotter, rather than using binary detection decisions or hypothesized transcriptions, and by using tiiese scores to derive an expected number of occurrences for every event in a given set. In using the topic discriminator of the present invention, it allows one to determine a set of potential speech events to be detected by the text event detector of a topic discriminator using transcribed trai-ning data if it is available. Since certain changes may be made in the above apparatus and method, without departing from the scope of the invention herein involved, it is dierefore intended tiiat all matter contained in the above description or shown in the accompanying drawings shall be illustrative and not in a limiting sense.

Claims

What is Claimed is: 1. A topic discriminator for determining the topic, represented by a prescribed set of speech events, of input speech data, said discriminator comprising, in combination: means for receiving an input signal representative of said input speech data; means for estimating the frequency of occurrence of each prescribed speech event associated with the topic and potentially contained in me speech data; and a topic classification means, responsive to the estimated event frequencies, for discriminating the topic of the input speech data.
2. A topic discriminator according to claim 1, wherein the topic classification means includes means for selecting from a set of preselected topics.
3. A topic discriminator according to claim 1, wherein the topic classification means includes means for making a binary determination whether the speech data belongs to said topic.
4. A topic discriminator according to claim 1 , further includes means for providing a confidence measure according to die certainty that the input speech data belongs to said topic.
5. A topic discriminator according to claim 1 , wherein the speech events includes word events, and said means for estimating the frequency of occurrence of each prescribed speech event associated widi die topic and potentially contained in d e speech data comprises: a speech recognizer for producing a hypothesized word sequence corresponding to die input speech data; and counting means for determining me number of each word event present in the output of the speech recognizer.
6. A topic discriminator according to claim 1 , wherein die speech events includes word and phrase events, and said means for estimating the frequency of occurrence of each prescribed speech event associated with die topic and potentially contained in the speech data comprises: a speech recognizer for producing a hypotiiesized word sequence corresponding to the input speech data; a parser to detect die presence of phrase events in the recognized word sequence; and counting means for determining the number of each word or phrase event present in the output of the speech recognizer.
7. A topic discriminator according to claim 1 , wherein the speech events includes word and phrase events, and die means for estimating the frequency of occurrence of each prescribed speech event associated widi the topic and potentially contained in the speech data comprises: a word and phrase spotter producing a set of hypotiiesized event occurrences corresponding to die input speech data; and counting means, responsive to the output of the word and phrase spotter, for estimating the number of each word or phrase event present in the speech data.
8. A topic discriminator according to claim 7, wherein said word and phrase spotter includes means for associating confidence scores with each hypodiesized event, and die counting means includes means for combining the confidence scores to determine me estimate of the number of each word or phrase event present in the input speech data.
9. A topic discriminator according to claim 8, wherein the word and phrase spotter includes a hidden Markov model for determining confidence scores based on die probability of occurrence of each presecribed speech event.
10. A topic discriminator according to claim 9, wherein the counting means comprises means for accumulating the probabilities of occurrence of each prescribed event.
11. A topic discriminator according to claim 1 , wherein at least a portion of the prescribed set of speech events representing said topic is determined by trained topic model parameters, which in turn are determined from a set of framing event frequencies each associated with said topic.
12. A topic discriminator according to claim 11, wherein the training event frequencies are determined from a set of text fraining data.
13. A topic discriminator according to claim 11, wherein the fraining event frequencies are determined from a set of untranscribed speech fraining data.
14. A topic discriminator according to claim 11, wherein at least anomer portion of prescribed events is selected.
15. A topic discriminator according to claim 11, wherein parametric probability models for determing the frequency of selected events are estimated from the training event frequencies.
16. A topic discriminator according to claim 15, wherein said parametric probability models include a single multi-nomial model of event probabilities for each topic of a known set.
17. A topic discriminator according to claim 16, wherein said single multi-nomial model is defined as follows:
Figure imgf000023_0001
wherein Tj is one of i topics; nk is the number of times event vk is hypothesized or observed in sequence w;
K N=∑ ni is the total number of events observed in w; and k=0 pk( ^=P(yk \T^ is die probability of the occurrence of event vk conditioned on topic r..
18. A topic discriminator according to claim 15, wherein said parametric probability models include a multi-nomial model of event probabilities in combination with a topic-specific mixture of word probability densities for each topic of a known set.
19. A topic discriminator according to claim 18, wherein said word-probability mixture densities have the following form:
P(fl\ '∑ q Hflm(fl
wherein T; is one of i topics; q^ is the prior probability of the selection of mixture m conditioned on topic T,;
Figure imgf000024_0001
is the conditional multi-variate probability density function of θ ; and Mi is the total number of mixtures associated with topic -T, .
20. A topic discriminator according to claim 18, wherein the parameters corresponding to said mixtare densities are estimated witii an Estimation-Maximization algorithm.
21. A method of selecting a vocabulary of important keywords or speech events, me presence or absence of which in a hypodiesis string associated with a speech message serves to indicate the topic membership of the message, said memod comprising the steps of: generating hypothesis strings for a plurality of tta-ning messages, each of which is known to be about one of a finite set of topics; and deriving word or event frequency statistics from said hypomesis strings; and using said event frequency statistics to assign topic-conditional significance or association scores to all possible words or events; and selecting diose words or events having topic-conditional significance or association scores above a predetermined direshold.
22. The method according to claim 21, wherein said step of generating said hypodiesis strings for a plurality of framing messages includes me step of generating said hypomesis strings so mat each indicates only the binary-valued presence or absence of a word or event.
23. The method according to claim 22, wherein said step of generating said hypomesis strings for a plurality of fraining messages includes d e step of deriving said hypomesis strings from a text transcription of me speech message.
24. The method according to claim 22, wherein said step of generating said hypomesis strings for a plurality of fraining messages includes me steps of (a) obtaining said hypodiesis strings by applying the speech messages to the input of an event detector of me type providing a hypomesis string output in response to a speech message; and (b) detecting me output of die event detector.
25. The method according to claim 24, wherein the event detector is a speech recognizer and said step of detecting me output of the event detector includes me step of detecting the output of said speech recognizer.
26. The method according to claim 21, wherein said step of generating hypomesis strings for a plurality of training messages includes me step of generating each hypomesis string so that said hypomesis string includes a probability or confidence score for every word or event in me string, said score indicating me overall likelihood die word or event occurred in the actual speech message associated with the string.
27. The method according to claim 26, wherein said step of generating hypomesis strings for a plurality of training messages includes die steps of (a) generating said hypodiesis sfrings by applying me speech messages to the input of an event detector of die type providing hypodiesis strings and associated word or events scores in response to speech messages; and (b) detecting d e output of the event detector.
28. The method according to claim 27, wherein the event detector is a word or event spotter and said step of detecting die output of the event detector includes die step of detecting me output of said word or event spotter.
29. The method according to claim 21 , wherein said step of deriving word or event frequency statistics from said hypodiesis strings includes die step of counting the number of keyword or event occurrences in the hypodiesis strings.
30. The method according to claim 21 , wherein said step of deriving word or event frequency statistics from said hypothesis sfrings includes the step of obtaining said word-frequency statistics by summing the confidence scores for all words or events in the hypothesis sfrings.
31. The method according to claim 21, wherein said step of using said event frequency statistics to assign topic-conditional significance or association scores to all possible words or events includes die step of obtaining topic-conditional significance or association scores by a hypothesis test.
32. The method according to claim 31, wherein said hypodiesis test is. a chi-squared test, and die step of obtaining topic-conditional significance or association scores by said chi-squared test includes d e step of calculating the chi-squared (χ2) values as follows:
Figure imgf000027_0001
and storing said chi-squared values in a table having cells defined by i rows and j columns;
wherein is the expected number of counts for each cell;
Figure imgf000027_0002
Figure imgf000027_0003
N.j = ∑Nij ; and
N^ equals die number of conversations assigned to cell (i,j) found for each cell in the table.
33. The method according to claim 32, wherein said step of obtaining topic-conditional significance or association scores by a chi-squared test includes die step of generating at least one contingency table from the word-frequency statistics derived from the message hypothesis strings and basing said chi-squared test on said contingency tables.
34. The method according to claim 33, wherein each topic is allotted its own row in the contingency table and said step of obtaining topic-conditional significance or association scores by a chi-squared test includes the step of conducting a hypodiesis test across all topics simultaneously.
35. The method according to claim 33, wherein a single topic is allotted its own row in the contingency table while all other topics are grouped togetiier in a second row of said table, and said step of obtaining topic-conditional significance or association scores by a chi-squared test includes me step conducting said chi-squared test for only a single word and topic at a time.
36. The method according to claim 21 , wherein said step of deriving word or event frequency statistics from said hypodiesis sfrings includes the steps of (a) using said word or event frequency statistics to construct topic-conditional mixtare densities for each possible word or event, and (b) using said topic-conditional mixtare densities to calculate an association score for each word.
37. The memod according to claim 36 , wherein the mixtare densities have
q fJβ)
Figure imgf000028_0001
wherein T; is one of i topics; q^ is the prior probability of the selection of mixture m conditioned on topic Tj5 f n ) is the conditional multi-variate probability density function of θ ; and Mt is the total number of mixtures associated witii topic Ti .
38. The method according to claim 37, wherein said mixtare densities are comprised of the weighted sum of delta-dirac functions.
39. The method according to claim 37, wherein said mixure densities are comprised of the weighted sum of beta-density functions of the form:
Figure imgf000028_0002
wherein C nj ,nj£ ,»■ ,nj^) is a normalization constant chosen to ensure f^θ) is a proper density function.
40. The method according to claim 39, wherein further including die steps of estimating the parameters of said mixtare densities witii an Estimation-Maximization algorithm.
41. The method according to claim 39, wherein said association score is a mutual information measure occurrences and tiiat of topic membership, of the form: The desired mutual information measure I(T,W) can be expressed as:
Figure imgf000029_0001
where p(wk i) \T(μ)) is die likelihood of the occurrence pattern for word or event V kj observed in fraining sequence w ) conditioned on its true topic -T(π) , and P(Wk n))=∑P(w k n) \T)PCT) is the likelihood for die same occurrence pattern
T independent of topic.
42. The method according to claim 39, wherein said association score is a Kullback-Liebler distance between the probability density of the word conditioned on a given topic, and the density of the word conditioned on all other topics, said distance measure having the form:
Figure imgf000029_0002
where N(7) is the total number of conversations or messages relating to the topic of interest, and N( ) is the number relating to all other topics.
PCT/US1995/004673 1994-04-18 1995-04-14 Topic discriminator WO1995028700A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
JP7527135A JPH08512148A (en) 1994-04-18 1995-04-14 Topic discriminator
EP95915682A EP0708960B1 (en) 1994-04-18 1995-04-14 Topic discriminator
AU22480/95A AU688030B2 (en) 1994-04-18 1995-04-14 Topic discriminator
CA002162696A CA2162696C (en) 1994-04-18 1995-04-14 Topic discriminator
DE69524036T DE69524036T2 (en) 1994-04-18 1995-04-14 DEVICE FOR DETECTING TALK TOPICS

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/228,755 1994-04-18
US08/228,755 US5625748A (en) 1994-04-18 1994-04-18 Topic discriminator using posterior probability or confidence scores

Publications (1)

Publication Number Publication Date
WO1995028700A1 true WO1995028700A1 (en) 1995-10-26

Family

ID=22858461

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1995/004673 WO1995028700A1 (en) 1994-04-18 1995-04-14 Topic discriminator

Country Status (7)

Country Link
US (1) US5625748A (en)
EP (1) EP0708960B1 (en)
JP (1) JPH08512148A (en)
AU (1) AU688030B2 (en)
CA (1) CA2162696C (en)
DE (1) DE69524036T2 (en)
WO (1) WO1995028700A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0788090A3 (en) * 1996-02-02 1998-08-19 International Business Machines Corporation Transcription of speech data with segments from acoustically dissimilar environments

Families Citing this family (141)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2751856B2 (en) * 1995-02-03 1998-05-18 日本電気株式会社 Pattern adaptation method using tree structure
US5717826A (en) * 1995-08-11 1998-02-10 Lucent Technologies Inc. Utterance verification using word based minimum verification error training for recognizing a keyboard string
US6064959A (en) * 1997-03-28 2000-05-16 Dragon Systems, Inc. Error correction in speech recognition
US7714778B2 (en) 1997-08-20 2010-05-11 Tracbeam Llc Wireless location gateway and applications therefor
US9134398B2 (en) 1996-09-09 2015-09-15 Tracbeam Llc Wireless location using network centric location estimators
US7903029B2 (en) 1996-09-09 2011-03-08 Tracbeam Llc Wireless location routing applications and architecture therefor
US6236365B1 (en) 1996-09-09 2001-05-22 Tracbeam, Llc Location of a mobile station using a plurality of commercial wireless infrastructures
WO1998010307A1 (en) 1996-09-09 1998-03-12 Dennis Jay Dupray Location of a mobile station
US5797123A (en) * 1996-10-01 1998-08-18 Lucent Technologies Inc. Method of key-phase detection and verification for flexible speech understanding
US6185531B1 (en) * 1997-01-09 2001-02-06 Gte Internetworking Incorporated Topic indexing method
US6035017A (en) * 1997-01-24 2000-03-07 Lucent Technologies Inc. Background speech recognition for voice messaging applications
GB2323693B (en) * 1997-03-27 2001-09-26 Forum Technology Ltd Speech to text conversion
US6009392A (en) * 1998-01-15 1999-12-28 International Business Machines Corporation Training speech recognition by matching audio segment frequency of occurrence with frequency of words and letter combinations in a corpus
US6104989A (en) * 1998-07-29 2000-08-15 International Business Machines Corporation Real time detection of topical changes and topic identification via likelihood based methods
US6192337B1 (en) * 1998-08-14 2001-02-20 International Business Machines Corporation Apparatus and methods for rejecting confusible words during training associated with a speech recognition system
US6185530B1 (en) 1998-08-14 2001-02-06 International Business Machines Corporation Apparatus and methods for identifying potential acoustic confusibility among words in a speech recognition system
US6269335B1 (en) 1998-08-14 2001-07-31 International Business Machines Corporation Apparatus and methods for identifying homophones among words in a speech recognition system
GB9824552D0 (en) * 1998-11-09 1999-01-06 Royal Holloway University Of L Data classification apparatus and method thereof
US6571210B2 (en) * 1998-11-13 2003-05-27 Microsoft Corporation Confidence measure system using a near-miss pattern
US20030146871A1 (en) * 1998-11-24 2003-08-07 Tracbeam Llc Wireless location using signal direction and time difference of arrival
US8135413B2 (en) 1998-11-24 2012-03-13 Tracbeam Llc Platform and applications for wireless location and other complex services
US6275803B1 (en) * 1999-02-12 2001-08-14 International Business Machines Corp. Updating a language model based on a function-word to total-word ratio
US6304833B1 (en) * 1999-04-27 2001-10-16 The United States Of America As Represented By The Secretary Of The Navy Hypothesis selection for evidential reasoning systems
US6711585B1 (en) * 1999-06-15 2004-03-23 Kanisa Inc. System and method for implementing a knowledge management system
JP2001160067A (en) * 1999-09-22 2001-06-12 Ddi Corp Method for retrieving similar document and recommended article communication service system using the method
WO2002000316A1 (en) 1999-09-24 2002-01-03 Goldberg Sheldon F Geographically constrained network services
US6539353B1 (en) 1999-10-12 2003-03-25 Microsoft Corporation Confidence measures using sub-word-dependent weighting of sub-word confidence scores for robust speech recognition
US7050977B1 (en) * 1999-11-12 2006-05-23 Phoenix Solutions, Inc. Speech-enabled server for internet website and method
US9076448B2 (en) 1999-11-12 2015-07-07 Nuance Communications, Inc. Distributed real time speech recognition system
US7725307B2 (en) * 1999-11-12 2010-05-25 Phoenix Solutions, Inc. Query engine for processing voice based queries including semantic decoding
US7392185B2 (en) 1999-11-12 2008-06-24 Phoenix Solutions, Inc. Speech based learning/training system using semantic decoding
US8108218B1 (en) * 1999-12-13 2012-01-31 Avaya Inc. Methods and apparatus for voice recognition for call treatment modification on messaging
US6411683B1 (en) * 2000-02-09 2002-06-25 At&T Corp. Automated telephone call designation system
US7505921B1 (en) 2000-03-03 2009-03-17 Finali Corporation System and method for optimizing a product configuration
US7263484B1 (en) 2000-03-04 2007-08-28 Georgia Tech Research Corporation Phonetic searching
EP1275042A2 (en) * 2000-03-06 2003-01-15 Kanisa Inc. A system and method for providing an intelligent multi-step dialog with a user
US6912498B2 (en) * 2000-05-02 2005-06-28 Scansoft, Inc. Error correction in speech recognition by correcting text around selected area
WO2001090921A2 (en) * 2000-05-25 2001-11-29 Kanisa, Inc. System and method for automatically classifying text
US10684350B2 (en) 2000-06-02 2020-06-16 Tracbeam Llc Services and applications for a communications network
US9875492B2 (en) 2001-05-22 2018-01-23 Dennis J. Dupray Real estate transaction system
US10641861B2 (en) 2000-06-02 2020-05-05 Dennis J. Dupray Services and applications for a communications network
US6856956B2 (en) * 2000-07-20 2005-02-15 Microsoft Corporation Method and apparatus for generating and displaying N-best alternatives in a speech recognition system
US7216077B1 (en) * 2000-09-26 2007-05-08 International Business Machines Corporation Lattice-based unsupervised maximum likelihood linear regression for speaker adaptation
US7162422B1 (en) * 2000-09-29 2007-01-09 Intel Corporation Apparatus and method for using user context information to improve N-best processing in the presence of speech recognition uncertainty
US7191133B1 (en) 2001-02-15 2007-03-13 West Corporation Script compliance using speech recognition
US8180643B1 (en) 2001-02-15 2012-05-15 West Corporation Script compliance using speech recognition and compilation and transmission of voice and text records to clients
US7966187B1 (en) 2001-02-15 2011-06-21 West Corporation Script compliance and quality assurance using speech recognition
US7664641B1 (en) 2001-02-15 2010-02-16 West Corporation Script compliance and quality assurance based on speech recognition and duration of interaction
US7739115B1 (en) 2001-02-15 2010-06-15 West Corporation Script compliance and agent feedback
US20020133392A1 (en) * 2001-02-22 2002-09-19 Angel Mark A. Distributed customer relationship management systems and methods
US6996531B2 (en) * 2001-03-30 2006-02-07 Comverse Ltd. Automated database assistance using a telephone for a speech based or text based multimedia communication mode
US6895083B1 (en) * 2001-05-02 2005-05-17 Verizon Corporate Services Group Inc. System and method for maximum benefit routing
US6839667B2 (en) * 2001-05-16 2005-01-04 International Business Machines Corporation Method of speech recognition by presenting N-best word candidates
US8082096B2 (en) 2001-05-22 2011-12-20 Tracbeam Llc Wireless location routing applications and architecture therefor
JP2002366166A (en) * 2001-06-11 2002-12-20 Pioneer Electronic Corp System and method for providing contents and computer program for the same
US7206778B2 (en) 2001-12-17 2007-04-17 Knova Software Inc. Text search ordered along one or more dimensions
US20030115191A1 (en) * 2001-12-17 2003-06-19 Max Copperman Efficient and cost-effective content provider for customer relationship management (CRM) or other applications
KR20030069377A (en) * 2002-02-20 2003-08-27 대한민국(전남대학교총장) Apparatus and method for detecting topic in speech recognition system
US7343372B2 (en) * 2002-02-22 2008-03-11 International Business Machines Corporation Direct navigation for information retrieval
US7340466B2 (en) * 2002-02-26 2008-03-04 Kang Jo Mgmt. Limited Liability Company Topic identification and use thereof in information retrieval systems
US20060004732A1 (en) * 2002-02-26 2006-01-05 Odom Paul S Search engine methods and systems for generating relevant search results and advertisements
US7716207B2 (en) * 2002-02-26 2010-05-11 Odom Paul S Search engine methods and systems for displaying relevant topics
US20030220917A1 (en) * 2002-04-03 2003-11-27 Max Copperman Contextual search
DE10220522B4 (en) * 2002-05-08 2005-11-17 Sap Ag Method and system for processing voice data using voice recognition and frequency analysis
DE10220521B4 (en) * 2002-05-08 2005-11-24 Sap Ag Method and system for processing voice data and classifying calls
DE10220524B4 (en) * 2002-05-08 2006-08-10 Sap Ag Method and system for processing voice data and recognizing a language
EP1363271A1 (en) * 2002-05-08 2003-11-19 Sap Ag Method and system for processing and storing of dialogue speech data
DE10220520A1 (en) * 2002-05-08 2003-11-20 Sap Ag Method of recognizing speech information
US7403967B1 (en) 2002-06-18 2008-07-22 West Corporation Methods, apparatus, and computer readable media for confirmation and verification of shipping address data associated with a transaction
US7228275B1 (en) * 2002-10-21 2007-06-05 Toyota Infotechnology Center Co., Ltd. Speech recognition system having multiple speech recognizers
WO2004049192A2 (en) 2002-11-28 2004-06-10 Koninklijke Philips Electronics N.V. Method to assign word class information
US20040128132A1 (en) * 2002-12-30 2004-07-01 Meir Griniasty Pronunciation network
WO2004072780A2 (en) 2003-02-05 2004-08-26 Verint Systems, Inc. Method for automatic and semi-automatic classification and clustering of non-deterministic texts
DE10306022B3 (en) * 2003-02-13 2004-02-19 Siemens Ag Speech recognition method for telephone, personal digital assistant, notepad computer or automobile navigation system uses 3-stage individual word identification
US7292982B1 (en) * 2003-05-29 2007-11-06 At&T Corp. Active labeling for spoken language understanding
US7487094B1 (en) 2003-06-20 2009-02-03 Utopy, Inc. System and method of call classification with context modeling based on composite words
US7454336B2 (en) * 2003-06-20 2008-11-18 Microsoft Corporation Variational inference and learning for segmental switching state space models of hidden speech dynamics
US7650282B1 (en) * 2003-07-23 2010-01-19 Nexidia Inc. Word spotting score normalization
WO2005010866A1 (en) * 2003-07-23 2005-02-03 Nexidia Inc. Spoken word spotting queries
US8170873B1 (en) * 2003-07-23 2012-05-01 Nexidia Inc. Comparing events in word spotting
US6930759B2 (en) * 2003-09-23 2005-08-16 Eastman Kodak Company Method and apparatus for exposing a latent watermark on film
JP4860265B2 (en) * 2004-01-16 2012-01-25 日本電気株式会社 Text processing method / program / program recording medium / device
US8874438B2 (en) * 2004-03-12 2014-10-28 Siemens Aktiengesellschaft User and vocabulary-adaptive determination of confidence and rejecting thresholds
US20050282563A1 (en) * 2004-06-17 2005-12-22 Ixi Mobile (R&D) Ltd. Message recognition and display system and method for a mobile communication device
EP1836555A4 (en) * 2004-08-02 2009-04-22 Scientigo Inc Search engine methods and systems for generating relevant search results and advertisements
US20060122834A1 (en) * 2004-12-03 2006-06-08 Bennett Ian M Emotion detection device & method for use in distributed systems
US7627638B1 (en) * 2004-12-20 2009-12-01 Google Inc. Verbal labels for electronic messages
US7865362B2 (en) * 2005-02-04 2011-01-04 Vocollect, Inc. Method and system for considering information about an expected response when performing speech recognition
US7949533B2 (en) * 2005-02-04 2011-05-24 Vococollect, Inc. Methods and systems for assessing and improving the performance of a speech recognition system
US8200495B2 (en) * 2005-02-04 2012-06-12 Vocollect, Inc. Methods and systems for considering information about an expected response when performing speech recognition
US7827032B2 (en) 2005-02-04 2010-11-02 Vocollect, Inc. Methods and systems for adapting a model for a speech recognition system
US7895039B2 (en) * 2005-02-04 2011-02-22 Vocollect, Inc. Methods and systems for optimizing model adaptation for a speech recognition system
US20060245641A1 (en) * 2005-04-29 2006-11-02 Microsoft Corporation Extracting data from semi-structured information utilizing a discriminative context free grammar
WO2007064639A2 (en) * 2005-11-29 2007-06-07 Scientigo, Inc. Methods and systems for providing personalized contextual search results
US8509563B2 (en) * 2006-02-02 2013-08-13 Microsoft Corporation Generation of documents from images
US7539616B2 (en) * 2006-02-20 2009-05-26 Microsoft Corporation Speaker authentication using adapted background models
US7599861B2 (en) 2006-03-02 2009-10-06 Convergys Customer Management Group, Inc. System and method for closed loop decisionmaking in an automated care system
US8364467B1 (en) * 2006-03-31 2013-01-29 Google Inc. Content-based classification
US8379830B1 (en) 2006-05-22 2013-02-19 Convergys Customer Management Delaware Llc System and method for automated customer service with contingent live interaction
US7809663B1 (en) 2006-05-22 2010-10-05 Convergys Cmg Utah, Inc. System and method for supporting the utilization of machine language
US8275613B2 (en) * 2006-08-21 2012-09-25 Unifiedvoice Corporation All voice transaction data capture—dictation system
JP5177561B2 (en) * 2007-02-06 2013-04-03 日本電気株式会社 Recognizer weight learning device, speech recognition device, and system
US8051372B1 (en) * 2007-04-12 2011-11-01 The New York Times Company System and method for automatically detecting and extracting semantically significant text from a HTML document associated with a plurality of HTML documents
US8285539B2 (en) * 2007-06-18 2012-10-09 International Business Machines Corporation Extracting tokens in a natural language understanding application
US9342588B2 (en) * 2007-06-18 2016-05-17 International Business Machines Corporation Reclassification of training data to improve classifier accuracy
US8521511B2 (en) 2007-06-18 2013-08-27 International Business Machines Corporation Information extraction in a natural language understanding system
US9058319B2 (en) * 2007-06-18 2015-06-16 International Business Machines Corporation Sub-model generation to improve classification accuracy
US20090099847A1 (en) * 2007-10-10 2009-04-16 Microsoft Corporation Template constrained posterior probability
US7437291B1 (en) * 2007-12-13 2008-10-14 International Business Machines Corporation Using partial information to improve dialog in automatic speech recognition systems
US20090326947A1 (en) * 2008-06-27 2009-12-31 James Arnold System and method for spoken topic or criterion recognition in digital media and contextual advertising
US8654963B2 (en) 2008-12-19 2014-02-18 Genesys Telecommunications Laboratories, Inc. Method and system for integrating an interaction management system with a business rules management system
US8463606B2 (en) 2009-07-13 2013-06-11 Genesys Telecommunications Laboratories, Inc. System for analyzing interactions and reporting analytic results to human-operated and system interfaces in real time
JP2011033680A (en) * 2009-07-30 2011-02-17 Sony Corp Voice processing device and method, and program
US20110184736A1 (en) * 2010-01-26 2011-07-28 Benjamin Slotznick Automated method of recognizing inputted information items and selecting information items
JP5533042B2 (en) * 2010-03-04 2014-06-25 富士通株式会社 Voice search device, voice search method, program, and recording medium
US9538493B2 (en) 2010-08-23 2017-01-03 Finetrak, Llc Locating a mobile station and applications therefor
US10248960B2 (en) 2010-11-16 2019-04-02 Disney Enterprises, Inc. Data mining to determine online user responses to broadcast messages
US8688453B1 (en) * 2011-02-28 2014-04-01 Nuance Communications, Inc. Intent mining via analysis of utterances
US8914290B2 (en) 2011-05-20 2014-12-16 Vocollect, Inc. Systems and methods for dynamically improving user intelligibility of synthesized speech in a work environment
US8965763B1 (en) 2012-02-02 2015-02-24 Google Inc. Discriminative language modeling for automatic speech recognition with a weak acoustic model and distributed training
US8543398B1 (en) 2012-02-29 2013-09-24 Google Inc. Training an automatic speech recognition system using compressed word frequencies
US8473293B1 (en) * 2012-04-17 2013-06-25 Google Inc. Dictionary filtering using market data
US8374865B1 (en) * 2012-04-26 2013-02-12 Google Inc. Sampling training data for an automatic speech recognition system based on a benchmark classification distribution
CA2869530A1 (en) * 2012-04-27 2013-10-31 Aravind GANAPATHIRAJU Negative example (anti-word) based performance improvement for speech recognition
US8571859B1 (en) 2012-05-31 2013-10-29 Google Inc. Multi-stage speaker adaptation
US8805684B1 (en) 2012-05-31 2014-08-12 Google Inc. Distributed speaker adaptation
US8880398B1 (en) 2012-07-13 2014-11-04 Google Inc. Localized speech recognition with offload
US9123333B2 (en) 2012-09-12 2015-09-01 Google Inc. Minimum bayesian risk methods for automatic speech recognition
US9912816B2 (en) 2012-11-29 2018-03-06 Genesys Telecommunications Laboratories, Inc. Workload distribution with resource awareness
US9542936B2 (en) 2012-12-29 2017-01-10 Genesys Telecommunications Laboratories, Inc. Fast out-of-vocabulary search in automatic speech recognition systems
US20140278357A1 (en) * 2013-03-14 2014-09-18 Wordnik, Inc. Word generation and scoring using sub-word segments and characteristic of interest
US9978395B2 (en) 2013-03-15 2018-05-22 Vocollect, Inc. Method and system for mitigating delay in receiving audio stream during production of sound from audio stream
US9645703B2 (en) 2014-05-14 2017-05-09 International Business Machines Corporation Detection of communication topic change
US10304440B1 (en) * 2015-07-10 2019-05-28 Amazon Technologies, Inc. Keyword spotting using multi-task configuration
US9984677B2 (en) * 2015-09-30 2018-05-29 Nice Ltd. Bettering scores of spoken phrase spotting
US10714121B2 (en) 2016-07-27 2020-07-14 Vocollect, Inc. Distinguishing user speech from background speech in speech-dense environments
KR101934280B1 (en) * 2016-10-05 2019-01-03 현대자동차주식회사 Apparatus and method for analyzing speech meaning
US10235365B2 (en) * 2017-06-20 2019-03-19 Microsoft Technology Licensing, Llc Transforming spoken thoughts to a visual representation
US10963510B2 (en) * 2018-08-09 2021-03-30 Bank Of America Corporation Dynamic natural language processing tagging
US11405506B2 (en) * 2020-06-29 2022-08-02 Avaya Management L.P. Prompt feature to leave voicemail for appropriate attribute-based call back to customers
US20220269934A1 (en) * 2021-02-22 2022-08-25 Tencent America LLC Dialogue model training based on reference-free discriminators

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4718094A (en) * 1984-11-19 1988-01-05 International Business Machines Corp. Speech recognition system
US4975959A (en) * 1983-11-08 1990-12-04 Texas Instruments Incorporated Speaker independent speech recognition process

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS6219899A (en) * 1985-07-18 1987-01-28 松下電器産業株式会社 Sentence voice recognition equipment
JPS6273391A (en) * 1985-09-27 1987-04-04 Toshiba Corp Pattern recognition learning device
JPH0634207B2 (en) * 1987-07-24 1994-05-02 日本電気株式会社 Topic prediction device
GB9220404D0 (en) * 1992-08-20 1992-11-11 Nat Security Agency Method of identifying,retrieving and sorting documents

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4975959A (en) * 1983-11-08 1990-12-04 Texas Instruments Incorporated Speaker independent speech recognition process
US4718094A (en) * 1984-11-19 1988-01-05 International Business Machines Corp. Speech recognition system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
PROCEEDINGS OF ICASSP, MILLER AND GORIN, March 1992: "A Structured Network for Adaptive Language Acquisition", pp. 201-204, see abstract, sec. 5.1, I-203, sec 5, I-203, sec 6, I-203. *
See also references of EP0708960A4 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0788090A3 (en) * 1996-02-02 1998-08-19 International Business Machines Corporation Transcription of speech data with segments from acoustically dissimilar environments
US6067517A (en) * 1996-02-02 2000-05-23 International Business Machines Corporation Transcription of speech data with segments from acoustically dissimilar environments

Also Published As

Publication number Publication date
CA2162696C (en) 2000-06-20
AU2248095A (en) 1995-11-10
JPH08512148A (en) 1996-12-17
DE69524036T2 (en) 2002-07-18
EP0708960B1 (en) 2001-11-21
AU688030B2 (en) 1998-03-05
US5625748A (en) 1997-04-29
CA2162696A1 (en) 1995-10-26
EP0708960A4 (en) 1998-06-03
DE69524036D1 (en) 2002-01-03
EP0708960A1 (en) 1996-05-01

Similar Documents

Publication Publication Date Title
EP0708960B1 (en) Topic discriminator
CN109151218B (en) Call voice quality inspection method and device, computer equipment and storage medium
US10453117B1 (en) Determining domains for natural language understanding
US9514126B2 (en) Method and system for automatically detecting morphemes in a task classification system using lattices
US5797123A (en) Method of key-phase detection and verification for flexible speech understanding
US9646605B2 (en) False alarm reduction in speech recognition systems using contextual information
EP1800293B1 (en) Spoken language identification system and methods for training and operating same
DE10111056B4 (en) Method and apparatus for identifying a non-target language in a speech recognition system
US8069043B2 (en) System and method for using meta-data dependent language modeling for automatic speech recognition
McDonough et al. Approaches to topic identification on the switchboard corpus
Wright et al. Automatic acquisition of salient grammar fragments for call-type classification.
US20130289987A1 (en) Negative Example (Anti-Word) Based Performance Improvement For Speech Recognition
Gorin Processing of semantic information in fluently spoken language
Gillick et al. Application of large vocabulary continuous speech recognition to topic and speaker identification using telephone speech
Kawahara et al. Key-phrase detection and verification for flexible speech understanding
Rose et al. Integration of utterance verification with statistical language modeling and spoken language understanding
Rose Word spotting from continuous speech utterances
JPH07261785A (en) Voice recognition method and voice recognition device
Williams A study of the use and evaluation of confidence measures in automatic speech recognition
EP2948943B1 (en) False alarm reduction in speech recognition systems using contextual information
WO2002029612A1 (en) Method and system for generating and searching an optimal maximum likelihood decision tree for hidden markov model (hmm) based speech recognition
Nagarajan et al. A pairwise multiple codebook approach to implicit language identification
Wallace Fast and accurate phonetic spoken term detection
KR100382473B1 (en) Speech recognition method
Rose et al. Integrating multiple knowledge sources for utterance verification in a large vocabulary speech understanding system

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AU CA JP

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE

WWE Wipo information: entry into national phase

Ref document number: 2162696

Country of ref document: CA

WWE Wipo information: entry into national phase

Ref document number: 1995915682

Country of ref document: EP

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWP Wipo information: published in national office

Ref document number: 1995915682

Country of ref document: EP

WWG Wipo information: grant in national office

Ref document number: 1995915682

Country of ref document: EP