Добавил:
Upload Опубликованный материал нарушает ваши авторские права? Сообщите нам.
Вуз: Предмет: Файл:
Скачиваний:
192
Добавлен:
08.06.2015
Размер:
3.63 Mб
Скачать

578 Daniel Jurafsky

26Pragmatics and Computational Linguistics

DANIEL JURAFSKY

1 Introduction

These days there’s a computational version of everything. Computational biology, computational musicology, computational archaeology, and so on, ad infinitum. Even movies are going digital. This chapter, as you might have guessed by now, thus explores the computational side of pragmatics. Computational pragmatics might be defined as the computational study of the relation between utterances and context. Like other kinds of pragmatics, this means that computational pragmatics is concerned with indexicality, with the relation between utterances and action, with the relation between utterances and discourse, and with the relationship between utterances and the place, time, and environmental context of their being uttered.

As Bunt and Black (2000) point out, computational pragmatics, like pragmatics in general, is especially concerned with INFERENCE. Four core inferential problems in pragmatics have received the most attention in the computational community: REFERENCE RESOLUTION, the interpretation and generation of SPEECH ACTS, the interpretation and generation of DISCOURSE STRUCTURE AND COHER- ENCE RELATIONS, and ABDUCTION. Each of these four problems can be cast as an inference task, one of somehow filling in information that isn’t actually present in the utterance at hand. Two of these tasks are addressed in other chapters of this volume; abduction in Hobbs (this volume), and discourse structure and coherence in Kehler (this volume). Reference resolution is covered in Kehler (2000b). I have therefore chosen the interpretation of speech acts as the topic of this chapter.

Speech act interpretation, a classic pragmatic problem, is a good choice for this overview chapter for many reasons. First, the early computational work drew very strongly from the linguistics literature of the period. This enables us to closely compare the ways that computational linguistic and non-computational linguistic approaches differ in their methodology. Second, there are two distinct computational paradigms in speech act interpretation: a logic-based approach

Pragmatics and Computational Linguistics 579

and a probabilistic approach. I see these two approaches as good vehicles for motivating the two dominant paradigms in computational linguistics: one based on logic, logical inference, feature-structures, and unification, and the other based on probabilistic approaches. Third, speech act interpretation provides a good example of pragmatic inference: inferring a kind of linguistic structure which is not directly present in the input utterance. Finally, speech act interpretation is a problem that applies very naturally both to written and spoken genres. This allows us to discuss the computational processing of speech input, and in general talk about the way that computational linguistics has dealt with the differences between spoken and written inputs.

I like to think of the role of computational models in linguistics as a kind of musical conversation among three melodic voices. The base melody is the role of computational linguistics as a core of what we sometimes call “mathematical foundations” of linguistics, the study of the formal underpinnings of models such as rules or trees, features or unification, indices or optimality. The middle line is the attempt to do what we sometimes call language engineering. One futuristic goal of this research is the attempt to build artificial agents that can carry on conversations with humans in order to perform tasks like answering questions, keeping schedules, or giving directions. The third strain is what is usually called “computational psycholinguistics”: the use of computational techniques to build processing models of human psycholinguistic performance. All of these melodic lines appear in computational pragmatics, although in this overview chapter we will focus more on the first two roles; linguistic foundations and language engineering.

The problem with focusing on speech act interpretation, of course, is that we will not be able to address the breadth of work in computational pragmatics. As suggested above above, the interested reader should turn to other chapters in this volume (especially Kehler and Hobbs) and also to Jurafsky and Martin (2000), which covers a number of computational pragmatic issues from a pedagogical perspective. Indeed, this chapter itself began as an expansion of, and meditation on, the section on dialogue act interpretation in Jurafsky and Martin (2000).

2Speech Act Interpretation: the Problem, and a Quick Historical Overview

The problem of speech act interpretation is to determine, given an utterance, which speech act it realizes. Of course, some speech acts have surface cues to their form; some questions, for example, begin with wh-words or with auxinversion. The Literal Meaning Hypothesis (Gazdar 1981), also called the Literal Force Hypothesis (Levinson 1983), is a strong version of this hypothesis, suggesting that every utterance has an illocutionary force which is built into its surface form. According to this hypothesis, aux-inverted sentences in English

580 Daniel Jurafsky

have QUESTION force; subject-deleted sentences have IMPERATIVE force, and so on (see Sadock, this volume).

But it has long been known that many or even most sentences do not seem to have the speech act type associated with their syntactic form. Consider two kinds of examples of this phenomenon. One example is INDIRECT REQUESTS, in which what looks on the surface like a question is actually a polite form of a directive or a request to perform an action. The sentence:

(1) Can you pass the salt?

looks on the surface like a yes-no question asking about the hearer’s ability to pass the salt, but functions actually as a polite directive to pass the salt.

There are other examples where the surface form of an utterance doesn’t match its speech act form. For example, what looks on the surface like a statement can really be a question. A very common kind of question, called a CHECK question (Labov and Fanshel 1977, Carletta et al. 1997b) is used to ask the other participant to confirm something that this other participant has privileged knowledge about. These checks are questions, but they have declarative word order, as in the bold-faced utterance in the following snippet from a travel agent conversation:

(2)A: I was wanting to make some arrangements for a trip that I’m going to be taking uh to LA uh beginning of the week after next.

B:OK uh let me pull up your profile and I’ll be right with you here. [pause]

B:And you said you wanted to travel next week?

A:Uh, yes.

There are two computational models of the interpretation of speech acts. The first class of models was originally motivated by indirect requests of the “pass the salt” type. Gordon and Lakoff (1971), and then Searle (1975a), proposed the seeds of this INFERENTIAL approach. Their intuition was that a sentence like Can you pass the salt? is unambiguous, having the literal meaning of a question: Do you have the ability to pass me the salt? The request speech act

Pass me the salt is inferred by the hearer in a later step of understanding after processing the literal question. Computational implementations of this idea focus on using belief logics to model this inference chain.

The second class of models has been called CUE-BASED or PROBABILISTIC ( Jurafsky and Martin 2000). The name CUE-BASED draws on the key role of cues in such psychological models as the Competition Model of Bates and MacWhinney (MacWhinney et al. 1984, MacWhinney 1987). These models are motivated more by indirect requests like CHECK questions. Here the problem is to figure out that what looks on the surface like a statement is really a question. Cue-based models think of the surface form of the sentence as a set of CUES to

Pragmatics and Computational Linguistics 581

the speaker’s intentions. Figuring out these intentions does require inference, but not of the type that chains through literal meanings.

These two models also differ in another important way. The inferential models are based on belief logics and use logical inference to reason about the speaker’s intentions. The cue-based models tend to be probabilistic machine learning models. They see interpretation as a classification task, and solve it by training statistical classifiers on labeled examples of speech acts.

Despite their differences, these models have in common the use of a kind of abductive inference. In each case, the hearer infers something that was not contained directly in the semantics of the input utterance. That makes them an excellent pair of examples of these two different ways of looking at computational linguistics. The next section introduces a version of the inferential model called the PLAN INFERENCE or BDI model, and the following section the CUE- BASED model.

3The Plan Inference (or BDI) Model of Speech Act Interpretation

The first approach to speech act interpretation we will consider is generally called the BDI (belief, desire, and intention) or PLAN-BASED model, proposed by Allen, Cohen, and Perrault and their colleagues (e.g. Allen 1995). Bunt and Black (2000: 15) define this line of inquiry as follows:

to apply the principles of rational agenthood to the modeling of a (computerbased) dialogue participant, where a rational communicative agent is endowed not only with certain private knowledge and the logic of belief, but is considered to also assume a great deal of common knowledge/beliefs with an interlocutor, and to be able to update beliefs about the interlocutor’s intentions and beliefs as a dialogue progresses.

The earliest papers, such as Cohen and Perrault (1979), offered an AI planning model for how speech acts are generated. One agent, seeking to find out some information, could use standard planning techniques to come up with the plan of asking the hearer to tell the speaker the information. Perrault and Allen (1980) and Allen and Perrault (1980) also applied this BDI approach to comprehension, specifically the comprehension of indirect speech effects.

Their application of the BDI model to comprehension draws on the planinference approach to dialogue act interpretation, first proposed by Gordon and Lakoff (1971) and Searle (1975a). Gordon, Lakoff, and Searle noticed that there was a structure to what kind of things a speaker could do to make an indirect request. In particular, they noticed that a speaker could mention or question various quite specific properties of the desired activity to make an indirect request. For example, the air travel request, “Give me certain flight information” can be realized as many different kinds of indirect requests. Here

582 Daniel Jurafsky

is a partial list from Jurafsky and Martin (2000) with examples from the ATIS1 corpus of sentences spoken to a computerized speech understanding system for planning air travel:

1The speaker can question the hearer’s ability to perform the activity:

Can you give me a list of the flights from Atlanta to Boston?

Could you tell me if Delta has a hub in Boston?

Would you be able to, uh, put me on a flight with Delta?

2The speaker can mention speaker’s wish or desire about the activity:

I want to fly from Boston to San Francisco.

I would like to stop somewhere else in between.

I’m looking for one-way flights from Tampa to Saint Louis.

I need that for Tuesday.

I wonder if there are any flights from Boston to Dallas.

3The speaker can mention the hearer’s doing the action:

Would you please repeat that information?

Will you tell me the departure time and arrival time on this American flight?

4The speaker can question the speaker’s having permission to receive results of the action:

May I get a lunch on flight UA 21 instead of breakfast?

Could I have a listing of flights leaving Boston?

Based on the realization that there were certain systemic ways of making indirect requests, Searle (1975a: 73) proposed that the hearer’s chain of reasoning upon hearing Can you give me a list of the flights from Atlanta to Boston? might be something like the following (Searle’s sentence was actually different; I’ve modified it to this ATIS example):

1X has asked me a question about whether I have the ability to give a list of flights.

2I assume that X is being cooperative in the conversation (in the Gricean sense) and that his utterance therefore has some aim.

3X knows I have the ability to give such a list, and there is no alternative reason why X should have a purely theoretical interest in my list-giving ability.

4Therefore X’s utterance probably has some ulterior illocutionary point. What can it be?

5A preparatory condition for a directive is that the hearer have the ability to perform the directed action.

6Therefore X has asked me a question about my preparedness for the action of giving X a list of flights.

Pragmatics and Computational Linguistics 583

7Furthermore, X and I are in a conversational situation in which giving lists of flights is a common and expected activity.

8Therefore, in the absence of any other plausible illocutionary act, X is probably requesting me to give him a list of flights.

The inferential approach thus explains why Can you give me a list of flights from Boston? is a reasonable way of making an indirect request in a way that Boston is in New England is not: the former mentions a precondition for the desired activity, and there is a reasonable inferential chain from the precondition to the activity itself.

As we suggested above, Perrault and Allen (1980) and Allen and Perrault (1980) applied this BDI approach to the comprehension of indirect speech effects, essentially cashing out Searle’s (1975a) promissory note in a computational formalism.

I’ll begin by summarizing Perrault and Allen’s formal definitions of belief and desire in the predicate calculus. I’ll represent “S believes the proposition P” as the two-place predicate B(S,P). Reasoning about belief is done with a number of axiom schemas inspired by Hintikka (1969) (such as B(A,P) B(A, Q) B(A,P Q); see Perrault and Allen 1980 for details). Knowledge is defined as “true belief”; S knows that P will be represented as KNOW(S,P), defined as follows:

KNOW(S,P) ≡ P B(S,P)

In addition to knowing that, we need to define knowing whether. S knows whether (KNOWIF) a proposition P is true if S KNOWs that P or S KNOWs that ←P:

KNOWIF(S,P) KNOW(S,P) KNOW(S, ←P)

The theory of desire relies on the predicate WANT. If an agent S wants P to be true, we say WANT(S,P), or W(S,P) for short. P can be a state or the execution of some action. Thus if ACT is the name of an action, W(S,ACT (H)) means that S wants H to do ACT. The logic of WANT relies on its own set of axiom schemas just like the logic of belief.

The BDI models also require an axiomatization of actions and planning; the simplest of these is based on a set of ACTION SCHEMAs similar to the AI planning model STRIPS (Fikes and Nilsson 1971). Each action schema has a set of parameters with CONSTRAINTS about the type of each variable, and three parts:

PRECONDITIONS: Conditions that must already be true in order to successfully perform the action.

EFFECTS: Conditions that become true as a result of successfully performing the action.

BODY: A set of partially ordered goal states that must be achieved in performing the action.

584 Daniel Jurafsky

In the travel domain, for example, the action of agent A booking flight F for client C might have the following simplified definition:

BOOK-FLIGHT(A,C,F):

Constraints: Agent(A) Flight(F) Client(C)

Precondition: Know(A,departure-date(F)) Know(A,departure-time(F)) Know(A,origin-city(F)) Know(A,destination-city(F)) Know(A,flight-type(F)) Has-Seats(F) W(C,(Book(A,C,F))). . .

Effect: Flight-Booked(A,C,F) Body: Make-Reservation(A,F,C)

Cohen and Perrault (1979) and Perrault and Allen (1980) use this kind of action specification for speech acts. For example, here is Perrault and Allen’s definition for three speech acts relevant to indirect requests. INFORM is the speech act of informing the hearer of some proposition (the Austin/Searle ASSERTIVE). The definition of INFORM is based on Grice’s 1957 idea that a speaker informs the hearer of something merely by causing the hearer to believe that the speaker wants them to know something:

INFORM(S,H,P):

Constraints: Speaker(S) Hearer(H) Proposition(P)

Precondition: Know(S,P) W(S,INFORM(S,H,P))

Effect: Know(H,P)

Body: B(H,W(S,Know(H,P)))

INFORMIF is the act used to inform the hearer whether a proposition is true or not; like INFORM, the speaker INFORMIFs the hearer by causing the hearer to believe the speaker wants them to KNOWIF something:

INFORMIF(S,H,P):

Constraints:

Speaker(S) Hearer(H) Proposition(P)

Precondition:

KnowIf(S,P) W(S, INFORMIF(S,H,P))

Effect:

KnowIf(H,P)

Body:

B(H, W(S, KnowIf(H,P)))

REQUEST is the directive speech act for requesting the hearer to perform some action:

REQUEST(S,H,ACT):

Constraints:

Speaker(S) Hearer(H) ACT(A) H is agent of ACT

Precondition:

W(S,ACT(H))

Effect:

W(H,ACT(H))

Body:

B(H,W(S,ACT(H)))

Pragmatics and Computational Linguistics 585

Perrault and Allen’s theory also requires what are called SURFACE-LEVEL ACTS. These correspond to the “literal meanings” of the imperative, interrogative, and declarative structures. For example the “surface-level” act S.REQUEST produces imperative utterances:

S.REQUEST(S,H,ACT):

Effect: B(H,W(S,ACT(H)))

The effects of s.REQUEST match the body of a regular REQUEST, since this is the default or standard way of doing a request (but not the only way). This “default” or “literal” meaning is the start of the hearer’s inference chain. The hearer will be given an input which indicates that the speaker is requesting the hearer to inform the speaker whether the hearer is capable of giving the speaker

alist:

S.REQUEST(S,H,InformIf(H,S,CanDo(H,Give(H,S,LIST))))

The hearer must figure out that the speaker is actually making a request: REQUEST(H,S,Give(H,S,LIST))

The inference chain from the request-to-inform-if-cando to the request-to- give is based on a chain of PLAUSIBLE INFERENCE, based on heuristics called PLAN INFERENCE (PI) rules. We will use the following subset of the rules that Perrault and Allen (1980) propose:

(PI.AE) Action–Effect Rule: For all agents S and H, if Y is an effect of action X and if H believes that S wants X to be done, then it is plausible that H believes that S wants Y to obtain.

(PI.PA) Precondition–Action Rule: For all agents S and H, if X is a precondition of action Y and if H believes S wants X to obtain, then it is plausible that H believes that S wants Y to be done.

(PI.BA) Body–Action Rule: For all agents S and H, if X is part of the body of Y and if H believes that S wants X done, then it is plausible that H believes that S wants Y done.

(PI.KD) Know–Desire Rule: For all agents S and H, if H believes S wants to KNOWIF(P), then H believes S wants P to be true:

plausible

B(H,W(S,KNOWIF(S,P))) rB(H,W(S,P))

plausible

(EI.1) Extended Inference Rule: if B(H,W(S,X)) rB(H,W(S,Y)) is a PI rule, then

plausible

B(H,W(S,B(H,(W,(S,X))))) rB(H,W(S,B(H,W(S,Y))))

is a PI rule (i.e. you can prefix B(H,W(S)) to any plan inference rule).

586 Daniel Jurafsky

Let’s see how to use these rules to interpret the indirect speech act in Can you give me a list of flights from Atlanta? Step 0 in the table below shows the speaker’s initial speech act, which the hearer initially interprets literally as a question. Step 1 then uses Plan Inference rule ACTION-EFFECT, which suggests that if the speaker asked for something (in this case information), they probably want it. Step 2 again uses the ACTION-EFFECT rule, here suggesting that if the Speaker wants an INFORMIF, and KNOWIF is an effect of INFORMIF, then the speaker probably also wants KNOWIF.

(3)

Rule Step

Result

 

 

 

 

0

S.REQUEST(S,H,InformIf(H,S,CanDo(H,Give(H,S,LIST))))

PI.AE

1

B(H,W(S,InformIf(H,S,CanDo(H,Give(H,S,LIST)))))

PI.AE/EI

2

B(H,W(S,KnowIf(H,S,CanDo(H,Give(H,S,LIST)))))

PI.KP/EI

3

B(H,W(S,CanDo(H,Give(H,S,LIST))))

PI.PA/EI

4

B(H,W(S,Give(H,S,LIST)))

PI.BA

5

REQUEST(H,S,Give(H,S,LIST))

Step 3 adds the crucial inference that people don’t usually ask about things they aren’t interested in; thus if the speaker asks whether something is true (in this case CanDo), the speaker probably wants it (CanDo) to be true. Step 4 makes use of the fact that CanDo(ACT) is a precondition for (ACT), making the inference that if the speaker wants a precondition (CanDo) for an action (Give), the speaker probably also wants the action (Give). Finally, step 5 relies on the definition of REQUEST to suggest that if the speaker wants someone to know that the speaker wants them to do something, then the speaker is probably REQUESTing them to do it.

In summary, the BDI model of speech act interpretation is based on three components:

1An axiomatization of belief, of desire, of action and of planning inspired originally by the work of Hintikka (1969)

2A set of plan inference rules, which codify the abductive heuristics of the understanding system

3A theorem prover

Given these three components and an input sentence, a plan-inference system can interpret the correct speech act to assign to the utterance by simulating the inference chain suggested by Searle (1975a).

The BDI model has many advantages. It is an explanatory model, in that its plan-inference rules explain why people make certain inferences rather than others. It is a rich and deep model of the knowledge that humans use in interpretation; thus in addition to its basis for building a conversational agent, the BDI model might be used as a formalization of a cognitive model of human

Pragmatics and Computational Linguistics 587

interpretation. The BDI model also shows how linguistic knowledge can be integrated with non-linguistic knowledge in building a model of cognition. Finally, the BDI model is a clear example of the role of computational linguistics as a foundational tool in formalizing linguistic models.

In giving this summary of the plan-inference approach to indirect speech act comprehension, I have left out many details, including many necessary axioms, as well as mechanisms for deciding which inference rule to apply. The interested reader should consult Perrault and Allen (1980).

4The Cue-based Model of Speech Act Interpretation

The plan-inference approach to dialogue act comprehension is extremely powerful; by using rich knowledge structures and powerful planning techniques the algorithm is designed to address even subtle indirect uses of dialogue acts. Furthermore, the BDI model incorporates knowledge about speaker and hearer intentions, actions, knowledge, and belief that is essential for any complete model of dialogue. But although the BDI model itself has crucial advantages, there are a number of disadvantages to the way the BDI model attempts to solve the speech act interpretation problem.

Perhaps the largest drawback is that the BDI model of speech act interpretation requires that each utterance have a single literal meaning, which is operated on by plan inference rules to produce a final non-literal interpretation. Much recent work has argued against this literal-first non-literal-second model of interpretation. As Levinson (1983) suggests, for example, the speech act force of most utterances does not match their surface form. Levinson points out, for example, that the imperative is very rarely used to issue requests in English. He also notes another problem: that indirect speech acts often manifest surface syntactic reflexes associated with their indirect force as well as their putative “literal force.”

The psycholinguistic literature, similarly, has not found evidence for the temporal primacy of literal interpretation. Swinney and Cutler (1979), just to give one example, found that literal and figurative meanings of idioms are accessed in parallel by the human sentence processor.

Finally, for many speech act types that are less well studied than the “big three” (question, statement, request), it’s not clear what the “literal” force would be. Consider, for example, utterances like “yeah” which can function as YES- ANSWERS, AGREEMENTS, and BACKCHANNELS. It’s not clear why any one of these should necessarily be the literal speech act and the others be the inferred act.

An alternative way of looking at disambiguation is to downplay the role of a “literal meaning.” In this alternate CUE model, we think of the listener as using different cues in the input to help decide how to build an interpretation. Thus the surface input to the interpretive algorithm provides clues to structurebuilding, rather than providing a literal meaning which must be modified by

588 Daniel Jurafsky

purely inferential processes. What characterizes a cue-based model is the use of different sources of knowledge (cues) for detecting a speech act, such as lexical, collocational, syntactic, prosodic, or conversational-structure cues.

The cue-based approach is based on metaphors from a different set of linguistic literature than the plan-inference approach. Where the plan-inference approach relies on Searle-like intuitions about logical inference from literal meaning, the cue-based approach draws from the conversational analytic tradition. In particular, it draws from intuitions about what Goodwin (1996) called MICROGRAMMAR (specific lexical, collocation, and prosodic features which are characteristic of particular conversational moves), as well as from the British pragmatics tradition on conversational games and moves (Power 1979). In addition, where the plan-inference model draws most heavily from analysis of written text, the cue-based literature is grounded much more in the analysis of spoken language. Thus, for example, a cue-based approach might use cues from many domains to recognize a true question, including lexical and syntactic knowledge like aux-inversion, prosodic cues like rising intonation, and conversational structure clues, like the neighboring discourse structure, turn boundaries, etc.

4.1Speech acts and dialogue acts

Before I give the cue-based algorithm for speech act interpretation, I need to digress a bit to give some examples of the kind of speech acts that these algorithms will be addressing. This section summarizes a number of computational tag sets of possible speech acts. The next section chooses one such act, CHECK, to discuss in more detail.

While speech acts provide a useful characterization of one kind of pragmatic force, more recent work, especially computational work in building dialogue systems, has significantly expanded this core notion, modeling more kinds of conversational functions that an utterance can perform. The resulting enriched acts are often called DIALOGUE ACTS (Bunt 1994) or CONVERSATIONAL MOVES (Power 1979, Carletta et al. 1997b).

The phrase dialogue act is unfortunately ambiguous. As Bunt and Black (2000) point out, it has been variously used to loosely mean “speech act, in the context of a dialogue” (Bunt 1994), to mean a combination of the speech act and semantic force of an utterance (Bunt 2000), or to mean an act with internal structure related specifically to its dialogue function (Allen and Core 1997). The third usage is perhaps the most common in the cue-based literature, and I will rely on it here.

In the remainder of this section, I discuss various examples of dialogue acts and dialogue act structures. A recent ongoing effort to develop dialogue act tagging schemes is the DAMSL (Dialogue Act Markup in Several Layers) architecture (Walker et al. 1996, Allen and Core 1997, Carletta et al. 1997a, Core et al. 1999), which codes various kinds of dialogue information about utterances. As we suggested above, DAMSL and other such computational efforts

Pragmatics and Computational Linguistics 589

to build practical descriptions of dialogue acts, like cue-based models in general, all draw on a number of research areas outside of the philosophical traditions that first defined speech acts. Perhaps the most important source has been work in conversation analysis and related fields. These include work on REPAIR (Schegloff et al. 1977), work on GROUNDING (Clark and Schaefer 1989), and work on the relation of utterances to the preceding and succeeding discourse (Schegloff 1968, 1988, Allwood et al. 1992, Allwood 1995).

For example, drawing on Allwood’s work, the DAMSL tagset distinguishes between the FORWARD LOOKING and BACKWARD LOOKING function of an utterance. The forward looking function of an utterance corresponds to something like the Searle/Austin speech act. The DAMSL tag set is more complex in having a hierarchically structured representation that I won’t discuss here and differs also from the Searle/Austin speech act in being focused somewhat on the kind of dialogue acts that tend to occur in task-oriented dialogue:

(4)

STATEMENT

a claim made by the speaker

INFO-REQUEST

a question by the speaker

CHECK

a question for confirming information

INFLUENCE-ON-ADDRESSEE

(= Searle’s directives)

OPEN-OPTION

a weak suggestion or listing of options

ACTION-DIRECTIVE

an actual command

INFLUENCE-ON-SPEAKER

(= Austin’s commissives)

OFFER

speaker offers to do something,

 

(subject to confirmation)

COMMIT

speaker is committed to doing something

CONVENTIONAL

other

OPENING

greetings

CLOSING

farewells

THANKING

thanking and responding to thanks

 

 

The backward-looking function of DAMSL focuses on the relationship of an utterance to previous utterances by the other speaker. These include accepting and rejecting proposals (since DAMSL is focused on task-oriented dialogue), as well as acts involved in grounding and repair:

(5)

AGREEMENT

speaker’s response to previous proposal

ACCEPT

accepting the proposal

ACCEPT-PART

accepting some part of the proposal

MAYBE

neither accepting nor rejecting the proposal

REJECT-PART

rejecting some part of the proposal

REJECT

rejecting the proposal

HOLD

putting off response, usually via subdialogue

590 Daniel Jurafsky

ANSWER

answering a question

UNDERSTANDING

whether speaker understood previous

SIGNALNONUNDER

speaker didn’t understand

SIGNAL-UNDER

speaker did understand

ACKNOWLEDGEMENT

demonstrated via backchannel or assessment

REPEAT-REPHRASE

demonstrated via repetition or reformulation

COMPLETION

demonstrated via collaborative completion

 

 

DAMSL and DAMSL-like sets of dialogue acts have been applied both to task-oriented dialogue and to non-task-oriented casual conversational speech. We give examples of two dialogue act tagsets designed for task-oriented dialogue and one for casual speech.

The task-oriented corpora are the Map Task and Verbmobil corpora. The Map Task corpus (Anderson et al. 1991) consists of conversations between two speakers with slightly different maps of an imaginary territory. Their task is to help one speaker reproduce a route drawn only on the other speaker’s map, all without being able to see each other’s maps. The Verbmobil corpus consists of two-party scheduling dialogues, in which the speakers were asked to plan a meeting at some future date. Tables 26.1 and 26.2 show the most commonly used versions of the tagsets from those two tasks.

Switchboard is a large collection of 2,400 six-minute telephone conversations between strangers who were asked to informally chat about certain topics (cars, children, crime). The SWBD-DAMSL tagset ( Jurafsky et al. 1997b) was developed from the DAMSL tagset in an attempt to label the kind of non-task- oriented dialogues that occur in Switchboard. The tagset was multidimensional, with approximately 50 basic tags (QUESTION, STATEMENT, etc.) and various diacritics. A labeling project described in Jurafsky et al. (1997b) labeled every utterance in about 1,200 of the Switchboard conversations; approximately 200,000 utterances were labeled. Approximately 220 of the many possible unique combinations of the SWBD-DAMSL codes were used by the coders. To obtain a system with somewhat higher inter-labeler agreement, as well as enough data per class for statistical modeling purposes, a less fine-grained tagset was devised, distinguishing 42 mutually exclusive utterance types ( Jurafsky et al. 1998a; Stolcke et al. 2000). Table 26.3 shows the 42 categories with examples and relative frequencies.

None of these various sets of dialogue acts are meant to be an exhaustive list. Each was designed with some particular computational task in mind, and hence will have domain-specific inclusions or absences. I have included them here mainly to show the kind of delimited task that the computational modeling community has set for itself. As is clear from the examples above, the various tagsets do include commonly studied speech acts like QUESTION and REQUEST. They also, however, include acts that have not been studied in the speech act literature. In the next section, I summarize one of these dialogue acts, the CHECK, in order to give the reader a more in-depth view of at least one of these various “minor” acts.

Pragmatics and Computational Linguistics 591

Table 26.1 The 18 high-level dialogue acts used in Verbmobil-1, abstracted over a total of 43 more specific dialogue acts: examples are from Jekat et al. (1995)

Tag

Example

 

 

THANK

Thanks

GREET

Hello Dan

INTRODUCE

It’s me again

BYE

Allright bye

REQUEST-COMMENT

How does that look?

SUGGEST

from thirteenth through seventeenth June

REJECT

No Friday I’m booked all day

ACCEPT

Saturday sounds fine

REQUEST-SUGGEST

What is a good day of the week for you?

INIT

I wanted to make an appointment with you

GIVE_REASON

Because I have meetings all afternoon

FEEDBACK

Okay

DELIBERATE

Let me check my calendar here

CONRM

Okay, that would be wonderful

CLARIFY

Okay, do you mean Tuesday the 23rd?

DIGRESS

[we could meet for lunch] and eat lots of ice cream

MOTIVATE

We should go to visit our subsidiary in Munich

GARBAGE

Oops, I-

 

 

Table 26.2 The 12 move types used in the Map Task: examples are from Taylor et al. (1998)

Tag

Example

 

 

INSTRUCT

Go round, ehm horizontally underneath diamond mind

EXPLAIN

I don’t have a ravine

ALIGN

Okay?

CHECK

So going down to Indian Country?

QUERY-YN

Have you got the graveyard written down?

QUERY-W

In where?

ACKNOWLEDGE

Okay

CLARIFY

{you want to go . . . diagonally} Diagonally down

REPLY-Y

I do.

REPLY-N

No, I don’t

REPLY-W

{And across to?} The pyramid.

READY

Okay

 

 

592 Daniel Jurafsky

Table 26.3 The 42 dialogue act labels, from Stolcke et al. (2000): dialogue act frequencies are given as percentages of the total number of utterances in the corpus

Tag

Example

%

 

 

 

STATEMENT

Me, I’m in the legal department.

36

BACKCHANNEL/ACKNOWLEDGE

Uh-huh.

19

OPINION

I think it’s great.

13

ABANDONED/UNINTERPRETABLE

So, -/

6

AGREEMENT/ACCEPT

That’s exactly it.

5

APPRECIATION

I can imagine.

2

YES-NO-QUESTION

Do you have to have any special training?

2

NON-VERBAL

Laughter , Throat_clearing

2

YES ANSWERS

Yes.

1

CONVENTIONAL-CLOSING

Well, it’s been nice talking to you.

1

WH-QUESTION

What did you wear to work today?

1

NO ANSWERS

No.

1

RESPONSE ACKNOWLEDGMENT

Oh, okay.

1

HEDGE

I don’t know if I’m making any sense or not.

1

DECLARATIVE YES-NO-QUESTION

So you can afford to get a house?

1

OTHER

Well give me a break, you know.

1

BACKCHANNEL-QUESTION

Is that right?

1

QUOTATION

You can’t be pregnant and have cats.

0.5

SUMMARIZE/REFORMULATE

Oh, you mean you switched schools for the kids.

0.5

AFRMATIVE NON-YES ANSWERS

It is.

0.4

ACTION-DIRECTIVE

Why don’t you go first.

0.4

COLLABORATIVE COMPLETION

Who aren’t contributing.

0.4

REPEAT-PHRASE

Oh, fajitas.

0.3

OPEN-QUESTION

How about you?

0.3

RHETORICAL-QUESTIONS

Who would steal a newspaper?

0.2

HOLD BEFORE ANSWER/

I’m drawing a blank.

0.3

AGREEMENT

 

 

REJECT

Well, no.

0.2

NEGATIVE NON-NO ANSWERS

Uh, not a whole lot.

0.1

SIGNAL-NON-UNDERSTANDING

Excuse me?

0.1

OTHER ANSWERS

I don’t know.

0.1

CONVENTIONAL-OPENING

How are you?

0.1

OR-CLAUSE

or is it more of a company?

0.1

DISPREFERRED ANSWERS

Well, not so much that.

0.1

3RD-PARTY-TALK

My goodness, Diane, get down from there.

0.1

OFFERS, OPTIONS and COMMITS

I’ll have to check that out.

0.1

SELF-TALK

What’s the word I’m looking for?

0.1

DOWNPLAYER

That’s all right.

0.1

MAYBE/ACCEPT-PART

Something like that.

<0.1

TAG-QUESTION

Right?

<0.1

DECLARATIVE WH-QUESTION

You are what kind of buff?

<0.1

APOLOGY

I’m sorry.

<0.1

THANKING

Hey thanks a lot.

<0.1

 

 

 

Pragmatics and Computational Linguistics 593

4.2 The dialogue act check

We saw in previous sections that the motivating example for the plan-based approach was based on indirect requests (surface questions with the illocutionary force of a REQUEST). In this section we’ll look at a different kind of indirect speech act, one that has motivated some of the cue-based literature. The speech act we will look at, introduced very briefly above, is often called a CHECK or a CHECK QUESTION (Labov and Fanshel 1977, Carletta et al. 1997b). A CHECK is a subtype of question which requests the interlocutor to confirm some information; the information may have been mentioned explicitly in the preceding dialogue (as in the example below), or it may have been inferred from what the interlocutor has said:

(6)

A:I was wanting to make some arrangements for a trip that I’m going to be taking uh to LA uh beginning of the week after next.

B:OK uh let me pull up your profile and I’ll be right with you here. [pause]

B:And you said you wanted to travel next week?

A:Uh yes.

Here are some sample realizations of CHECKs in English from various corpora, showing their various surface forms:

(7) As tag questions (example from the Trains corpus; Allen and Core 1997):

U:and it’s gonna take us also an hour to load boxcars right?

S:Right

(8)As declarative questions, usually with rising intonation (Quirk et al. 1985: 814) (example from the Switchboard corpus; Godfrey et al. 1992):

A:and we have a powerful computer down at work.

B:Oh (laughter)

B:so, you don’t need a personal one (laughter)?

A:No

(9)As fragment questions (subsentential unit, e.g. words, noun phrases, clauses) (Weber 1993) (example from the Map Task corpus; Carletta et al. 1997b):

G:Ehm, curve round slightly to your right.

F:To my right?

G:Yes

The next section will discuss the kind of cues that are used to detect CHECKS and other dialogue acts.

594 Daniel Jurafsky

4.3Cues

A CUE is a surface feature that is probabilistically associated with some speech or dialogue act. Commonly studied features include lexical, syntactic, prosodic, and discourse factors, but cues may also involve more sophisticated and complex knowledge, such as speaker-specific or dyad-specific modeling.

4.3.1Lexical or syntactic cues

Lexical and syntactic cues have been widely described, at least for the most commonly studied speech acts. In a useful typological study, Sadock and Zwicky (1985) mention the existence of such cues for DECLARATIVE acts as declarative particles (in Welsh or Hidatsa), or different inflectional forms used specifically in declarative acts (Greenlandic).

Cross-linguistically common lexical or syntactic cues for imperatives include sentence-initial or sentence-final particles, verbal clitics, special verb morphology in the verb stem, subject deletion, and special subject pronoun forms that are used specifically in the imperative (Sadock and Zwicky 1985).

A similar inventory of cue types applies to lexical or syntactic cues for yes-no QUESTIONS, including sentence-initial or sentence-final particles, special verb morphology, and word order.

In addition to these cross-linguistic universals for the major acts, more recent work has begun to examine lexical and syntactic cues for minor acts. Michaelis (2001) shows that EXCLAMATIVES, for example, are characterized crosslinguistically by anaphoric degree adverbs, as well as various surface cues associated with information questions. Michaelis and Lambrecht (1996) discuss the wide variety of surface syntactic features which can characterize EXCLAMATIVES in English, including extraposition, bare complements, and certain kinds of definite noun phrases.

I have seen these same kinds of cues in my own work and that of my colleagues. Studies of CHECKs, for example, have shown that, like the examples above, they are most often realized with declarative structure (i.e. no auxinversion), and they often have a following question tag, usually right (Quirk et al. 1985: 810–14), as in example (7) above. They also are often realized as fragments, i.e. subsentential words or phrases (Weber 1993).

In the Switchboard corpus, a very common type of check is the REFORMULA- TION. A reformulation, by repeating back some summarized or rephrased version of the interlocutor’s talk, is one way to ask “is this an acceptable summary of your talk?” Our examination of 960 reformulations in Switchboard ( Jurafsky and Martin 2000) shows that they have a very specific micro-grammar. They generally have declarative word order, often with you as the subject (31 percent of the cases), often beginning with so (20 percent) or oh, and sometimes ending with then. Some examples:

(10) Oh so you’re from the Midwest too.

Pragmatics and Computational Linguistics 595

(11)So you can steady it.

(12)You really rough it then.

This kind of micro-grammar was originally noted by Goodwin (1996), in his discussion of ASSESSMENTS. Assessments are a particular kind of evaluative act, used to ascribe positive or negative properties:

(13)That’s good.

(14)Oh that’s nice.

(15)It’s great.

Goodwin (1996) found that assessments often display the following format:

(16) Pro Term + Copula + (Intensifier) + Assessment Adjective

Jurafsky et al. (1998b) found an even more constrained, and more lexicalized, micro-grammar for the 1,150 assessments with overt subjects in Switchboard. They found that the vast majority (80 percent) of the Pro Terms were that, that only two types of intensifiers occurred (really and pretty), and that the range of assessment adjective was quite small, consisting only of the following: great, good, nice, wonderful, cool, fun, terrible, exciting, interesting, wild, scary, hilarious, neat, funny, amazing, tough, incredible, and awful.

4.3.2Prosodic cues

Prosody is another important cue for dialogue act identity. The final pitch rise of yes-no questions in American English (Sag and Liberman 1975, Pierrehumbert 1980) as well as cross-linguistically (Sadock and Zwicky 1985) is well known. Similarly well studied is the realization of final lowering in declaratives and wh- questions in English (the H*L L% tune) (Pierrehumbert and Hirschberg 1990).

Prosody plays an important role in other dialogue acts. Shriberg et al. (1998) and Weber (1993), for example, found that CHECKs, like other questions, are also most likely to have rising intonation. Curl and Bell (2001) examined the dialogue-act coded portion of Switchboard for three dialogue acts which can all

be realized by the word yeah: AGREEMENTS, YES-ANSWERS, and BACKCHANNELS.

They found that yeah agreements are associated with high falling contour, and yeah backchannels with low falling or level contours.

Sadock and Zwicky (1985) mention various other types of prosodic cues that occur cross-linguistically, including special stress in the first word of a yes-no QUESTION in Hopi, and a glottal stop in the last word of a yes-no QUESTION in Hidatsa.

Pierrehumbert and Hirschberg (1990) offer a much more compositional kind of cue-based theory for the role of prosody in utterance interpretation in general. In their model, pitch accents convey information about such things as the

596 Daniel Jurafsky

status of discourse referents, phrase accents convey information about the semantic relationship between intermediate phrases, and boundary tones convey information about the directionality of interpretation. Presumably these kinds of intonational meaning cues, and others such as, for example, the rejection contour of Sag and Liberman (1975) or the uncertainty/incredulity contour of Ward and Hirschberg (1985, 1988), could be used to build a model of prosodic cues specifically for dialogue acts. (See also Hirschberg, this volume.)

4.3.3Discourse cues and summary

Finally, discourse structure is obviously an important cue for dialogue act identity. A dialogue act which functions as the second part of an adjacency pair (for example the YES-ANSWER), obviously depends on the presence of the first part (in this case a QUESTION). This is even true for sequences that aren’t clearly adjacency pairs. Allwood (1995) points out that the utterance “No it isn’t” is an AGREEMENT after a negative statement like “It isn’t raining” but a DISAGREEMENT after a positive statement like “It is raining.”

The importance of this contextual role of discourse cues has been a main focus of the conversation analysis tradition. For example Schegloff (1988) focuses on the way that the changing discourse context and the changing understanding of the hearer affects his interpretation of the discourse function of the utterance. Schegloff gives the following example utterance:

(17) Do you know who’s going to that meeting?

which occurs in the following dialogue:

Mother: Do you know who’s going to that meeting?

Russ: Who?

Mother: I don’t kno:w

Russ: Oh:: Prob’ly Missiz McOwen . . .

Mother had meant her first utterance as a REQUEST. But Russ misinterprets it as a PRE-ANNOUNCEMENT, and gives an appropriate response to such pre-announcements, by asking the question word which was included in the pre-announcement (“Who?”). Mother’s response (“I don’t know”) makes it clear that her utterance was a REQUEST rather than a PRE-ANNOUNCEMENT. In Russ’s second utterance, he uses this information to reanalyze and re-respond to Mother’s utterance.

This example shows that complex discourse information, such as the fact that an interlocutor has displayed a problem with a previous dialogue act interpretation, can play a role in future dialogue act interpretation.

In summary, we have seen three kinds of cues that can be used to help determine the dialogue act type of an utterance: prosodic cues, lexical and grammatical cues, and discourse structure cues. The next section discusses how cue-based algorithms make use of these cues to recognize dialogue acts.

Pragmatics and Computational Linguistics 597

4.4 The cue-based algorithms

The cue-based algorithm for speech act interpretation is given as input an utterance, and produces as output the most probable dialogue act for that utterance. In a sense, the idea behind the cue-based models is to treat every utterance as if it had no literal force. Determining the correct force is treated as a task of probabilistic reasoning, in which different cues at different levels supply the evidence.

In other words, I and other proponents of the cue-based model believe that the literal force hypothesis is simply wrong, i.e. that there is not a literal force for each surface sentence type. Certainly it is the case that some surface cues are more commonly associated with certain dialogue act types. But rather than model this commonality as a fact about literal meaning (the “Literal Force Hypothesis”) the cue-based models treat it as a fact about a probabilistic relationship between cue and dialogue act; the probability of a given dialogue act may simply be quite high given some particular cue.

In discussing these cue-based approaches, I will draw particularly on research in which I have participated and hence with which I am familiar, such as Stolcke et al. (2000) and Shriberg et al. (1998). As we will see, these algorithms are mostly designed to work directly from input speech waveforms. This means that they are of necessity based on heuristic approximations to the available cues. For example, a useful prosodic cue might come from a perfect ToBI phonological parse of an input utterance. But the computational problem of deriving a perfect ToBI parse from speech input is unsolved. So we see very simplistic approximations to the syntactic, discourse, and prosodic knowledge that we will someday have better models of.

The models we will describe generally use supervised machine-learning algorithms, trained on a corpus of dialogues that is hand-labeled with dialogue acts for each utterance. That is, these algorithms are statistical classifiers. We train a QUESTION-classifier on many instances of QUESTIONS, and it learns to recognize the combination of features (prosodic, lexical, syntactic, and discourse) which suggest the presence of a question. We train a REQUEST-classifier on many instances of REQUESTs, a BACKCHANNEL-classifier on many instances of

BACKCHANNELs, and so on.

Let’s begin with lexical and syntactic features. The simplest way to build a probabilistic model which detects lexical and phrasal cues is simply to look at which words and phrases occur more often in one dialogue act than another. Many scholars, beginning with Nagata and Morimoto (1994), realized that simple statistical grammars based on words and short phrases could serve to detect local structures indicative of particular dialogue acts. They implemented this intuition by modeling each dialogue act as having its own separate N-gram grammar (see e.g. Suhm and Waibel 1994, Mast et al. 1996, Jurafsky et al. 1997a, Warnke et al. 1997, Reithinger and Klesen 1997, Taylor et al. 1998). An N-gram grammar is a simple Markov model which stores, for each word, what its probability of occurrence is given one or more particular previous words.

598 Daniel Jurafsky

These systems create a separate mini-corpus from all the utterances which realize the same dialogue act, and then train a separate N-gram grammar on each of these mini-corpora. (In practice, more sophisticated N-gram models are generally used, such as backoff, interpolated, or class N-gram language models.) Given an input utterance consisting of a sequence of words W, they then choose the dialogue act d whose N-gram grammar assigns the highest likelihood to W. Technically, the formula for this maximization problem is as follows (although the non-probabilistic reader can safely ignore the formulas):

(18) d* = argmaxP(d|W) d

(19) = argmaxP(d)P(W|d) d

Equation (18) says that our estimate of the best dialogue act d* for an utterance is the dialogue act d which has the highest probability given the string W. By Bayes’ Rule, that can be rewritten as equation (19). This says that the dialogue act which is most probable given the input is the one which maximizes the product of two factors: the prior probability of a particular dialogue act P(d) and the probability P(W| d), which expresses, given that we had picked a certain dialogue act d, the probability it would be realized as the string of words W.

This N-gram approach, while only a local heuristic to more complex syntactic constraints, does indeed capture much of the micro-grammar. For example yes-no QUESTIONS often have bigram pairs indicative of aux-inversion (do you, are you, was he, etc.). Similarly, the most common bigrams in REFORMULATIONS are very indicative pairs like so you, sounds like, so you’re, oh so, you mean, so they, and so it’s.

While this N-gram model of micro-grammar has proved successful in practical implementations of dialogue act detection, it is obviously a gross simplification of micro-grammar. It is possible to keep the idea of separate, statistically trained micro-grammars for each dialogue act while extending the simple N-gram model to more sophisticated probabilistic grammars. For example Jurafsky et al. (1998b) show that the grammar of some dialogue acts, like APPRECIATIONS, can be captured by building probabilistic grammars of lexical category sequences. Alexandersson and Reithinger (1997) propose even more linguistically sophisticated grammars for each dialogue act, such as probabilistic context-free grammars. To reiterate this point: the idea of cuebased processing does not require that the cues be simplistic Markov models. A complex phrase-structural or configurational feature is just as good a cue. The model merely requires that these features be defined probabilistically.

Prosodic models of dialogue act micro-grammar rely on phonological features like pitch or accent, or their acoustic correlates like f0, duration, and energy. We mentioned above that features like final pitch rise are commonly used for

Pragmatics and Computational Linguistics 599

questions and fall for assertions. Indeed, computational approaches to dialogue act prosody modeling have mostly focused on f0. Many studies have successfully shown an increase in the ability to detect yes-no questions by combining lexical cues with these pitch-based cues (Waibel 1988, Daly and Zue 1992, Kompe et al. 1993, Taylor et al. 1998).

One such system, Shriberg et al. (1998), trained CART-style decision trees on simple acoustically based prosodic features such as the slope of f0 at the end of the utterance, the average energy at different places in the utterance, and various duration measures. They found that these features were useful, for example, in distinguishing four broad clusters of dialogue acts, STATEMENTS

(S), yes–no QUESTIONS (QY), DECLARATIVE-QUESTIONS like CHECKS (QD) and

wh-QUESTIONS (QW), from each other. Figure 26.1 shows the decision tree which gives the posterior probability P(d|F) of a dialogue act d type given a sequence of acoustic features F. Each node in the tree shows four probabilities, one for each of the four dialogue acts in the order S, QY, QW, QD; the most likely of the four is shown as the label for the node. Via Bayes’ Rule, this probability can be used to compute the likelihood of the acoustic features given the dialogue act: P(f |d).

In general, most such systems use phonetic rather than phonological cues, modeling f0 patterns with techniques such as vector quantization and Gaussian classifiers on acoustic input (Kießling et al. 1993, Kompe et al. 1995, Yoshimura et al. 1996). But some more recent systems actually attempt to directly model phonological cues such as pitch accent and boundary tone sequence (Taylor et al. 1997).

A final important cue for dialogue act interpretation is conversational structure. One simple way to model conversational structure, drawing on the idea of adjacency pairs (Schegloff 1968, Sacks et al. 1974) introduced above, is as a probabilistic sequence of dialogue acts. As first proposed by Nagata (1992), and in a follow-up paper (Nagata and Morimoto 1994), the identity of the previous dialogue acts can be used to help predict upcoming dialogue acts. For example, BACKCHANNELS or AGREEMENTS might be very likely to follow

STATEMENTS. ACCEPTS or REJECTS might be more likely to follow REQUESTS, and

so on. Woszczyna and Waibel (1994) give the dialogue automaton shown in figure 26.2, which models simple N-gram probabilities of dialogue act sequences for a Verbmobil-like appointment scheduling task.

Of course this idea of modeling dialogue act sequences as “N-grams” of dialogue acts only captures the effects of simple local discourse context. As I mentioned earlier in my discussion concerning syntactic cues, a more sophisticated model will need to take into account hierarchical discourse structure of various kinds. Indeed, the deficiencies of an N-gram model of dialogue structure are so great and so obvious that it might have been a bad idea for me to start this dialogue section of the chapter with them. But the fact is that the recent work on dialogue act interpretation that I describe here relies only on such simple cues. Once again, the fact that the examples we give all involve simple Markov models of dialogue structure should not be taken to imply that

 

 

 

 

 

S QY QW QD

 

 

 

 

 

0.25 0.25 0.25 0.25

 

 

 

 

 

 

cont_speech_frames < 196.5 cont_speech_frames >= 196.5

 

 

 

 

QW

 

S

 

 

 

 

 

0.2561 0.1642

 

0.2357 0.4508

cont_speech_frames_n >= 98.334

 

 

 

 

0.2732 0.3065

 

0.1957 0.1178

 

 

 

 

 

 

 

 

 

end_grad < 32.345

end_grad >= 32.345

cont_speech_frames_n < 98.334

 

 

QW

 

 

 

 

S

 

 

 

 

QY

 

S

 

 

0.2327 0.2018

0.2978 0.09721

 

0.2581 0.2984

0.2191 0.5637

 

 

0.1919 0.3735

0.4181 0.1869

 

0.2796 0.164

0.1335 0.08367

 

f0_mean_zcv < 0.76806

f0_mean_zcv >= 0.76806

 

norm_f0_diff < 0.064562 norm_f0_diff >= 0.064562

 

S

 

QW

 

 

S

 

QY

 

0.276 0.2811

0.1859 0.116

 

 

0.3089 0.3387

0.1857 0.241

 

0.1747 0.2683

0.2106 0.4875

 

 

0.1419 0.2105

0.4756 0.09772

 

cont_speech_frames_n < 98.388

cont_speech_frames_n >= 98.388

f0_mean_zcv < 0.76197

f0_mean_zcv >= 0.76197

 

QW

 

S

 

 

S

 

QW

 

0.2935 0.1768

0.2438 0.4729

0.3253 0.4315

0.2759 0.1517

 

0.2017 0.328

0.125 0.1583

0.1062 0.137

0.2138 0.3586

 

utt_grad < –36.113 utt_grad >= –36.113

 

stdev_enr_utt < 0.02903 stdev_enr_utt >= 0.02903

 

 

QW

QD

QW

 

S

 

 

 

0.2044 0.1135

0.3316 0.2038

0.3069 0.08995

0.2283 0.5668

 

 

 

0.1362 0.5459

0.2297 0.2349

0.1799 0.4233

0.1115 0.09339

 

 

 

Figure 26.1 Decision tree for the classification of STATEMENT (S), yes–no QUESTIONS (QY), wh-QUESTIONS (QW) and DECLARATIVE

QUESTIONS (QD), after Shriberg et al. (1998). Note that the difference between S and QY toward the right of the tree is based on the

feature norm_f0_diff (normalized difference between mean f0 of end and penultimate regions), while the difference between QW

and QD at the bottom left is based on utt_grad, which measures f0 slope across the whole utterance.

 

Jurafsky Daniel 600

Pragmatics and Computational Linguistics 601

 

 

opening

 

 

.76

 

 

.23

 

 

 

.63

 

suggest

.19

.99

constraint

.18

.36

 

 

.77

 

accept

.18

.22

reject

.46

closing

Figure 26.2 A dialogue act HMM for simple appointment scheduling conversations (after Woszczyna and Waibel 1994)

cue-based models of dialogue structure have to be simple. As the field progresses, presumably we will develop more complex probabilistic models of how speakers act in dialogue situations. Indeed, many others have already begun to enhance this N-gram approach (Nagata and Morimoto 1994, Suhm and Waibel 1994, Warnke et al. 1997, Stolcke et al. 1998, Taylor et al. 1998). ChuCarroll (1998), for example, has shown how to model subdialogue structure in a cue-based model. Her model deals with hierarchical dialogue structures like insertion sequences (in which a question is followed by another question) and other kinds of complex structure. It’s also important to note that a cue-based model doesn’t disallow non-probabilistic knowledge sources; certainly not all dialogue structural information is probabilistic. For example, a REJECTION (a “no” response) is a dispreferred response to a REQUEST. I suspect this isn’t a probabilistic fact; rejections may be always dispreferred. Studying how to integrate non-probabilistic knowledge of this sort into a cue-based model is a key problem that I return to in the conclusion.

I have now talked about simple statistical implementations of detectors for three kinds of cues for dialogue acts: lexical/syntactic, prosodic, and discourse structural. How can a dialogue act interpreter combine these different cues to find the most likely correct sequence of correct dialogue acts given a conversation?

One way to combine these statistical cues into a single probabilistic cuebased model is to treat a conversation as a Hidden Markov Model (HMM), an idea that seems to have been first suggested by Woszczyna and Waibel (1994) and Suhm and Waibel (1994). A Hidden Markov Model is a kind of probabilistic automaton in which a series of states in an automaton probabilistically generate sequences of symbols. Since the output is probabilistic, it is not possible to be certain from a given output symbol which state generated it; hence the states are “hidden.” The intuition behind using an HMM for a dialogue is that the dialogue acts play the role of the hidden states. The words, syntax, and prosody act as observed output symbols.

602 Daniel Jurafsky

HMMs can be viewed as generative or as interpretive models. As a generative model, given that the automaton is about to generate a particular dialogue act, the probabilistic cue-models give the probabilities of the different words, syntax, and prosody being produced. As an interpretive model, given a known sequence of words, syntax and prosody for an utterance, the HMM can be used to choose the single dialogue act which was most likely to have generated that sequence.

Stolcke et al. (2000) and Taylor et al. (1998) apply the HMM intuition of Woszczyna and Waibel (1994) to treat the dialogue act detection process as HMM-parsing. Given all available cues C about a conversation, the goal is to find the dialogue act sequence D = {d1, d2 . . . , dN} that has the highest posterior probability P(D|C) given those cues (here we are using capital letters to mean sequences of things). Applying Bayes’ Rule we get:

(20) D* = argmaxP(D|C)

D

= P(D)P(C|D) argmax—————

DP(C)

=argmaxP(D)P(C|D)

D

Equation (20) should remind the reader of equation (19). It says that we can estimate the best series of dialogue acts for an entire conversation by choosing that dialogue act sequence which maximizes the product of two probabilities, P(D) and P(C|D).

The first, P(D), is the probability of a sequence of dialogue acts. Sequences of dialogue acts which are more coherent will tend to occur more often than incoherent sequences of dialogue acts, and will hence be more probable. Thus P(D) essentially acts as a model of conversational structure. One simple way to compute an approximation to this probability is via the dialogue act N-grams introduced by Nagata and Morimoto (1994).

The second probability which must be considered is the likelihood P(C|D). This is the probability, given that we have a particular dialogue act sequence D, of observing a particular set of observed surface cues C. This likelihood P(C|D) can be computed from two sets of cues. First, the micro-syntax models (for example the different word-N-gram grammars for each dialogue act) can be used to estimate P(W|D), the probability of the sequence of words W given a particular sequence of dialogue acts D. Next, the micro-prosody models (for example the decision tree for the prosodic features of each dialogue act) can be used to estimate P(F|D), the probability of the sequence of prosodic features F. If we make the simplifying (but of course incorrect) assumption that the prosody and the words are independent, we can thus estimate the cue likelihood for a sequence of dialogue acts D as follows:

Pragmatics and Computational Linguistics 603

(21) P(C|D) = P(F|D)P(W|D)

We can compute the most likely sequence of dialogue acts D* by substituting equation (21) into equation (20), thus choosing the dialogue act sequence which maximizes the product of the three knowledge sources (conversational structure, prosody, and lexical/syntactic knowledge):

(22) D* = argmaxP(D)P(F|D)P(W|D)

D

Standard HMM-parsing techniques (like Viterbi) can then be used to search for this most-probable sequence of dialogue acts given the sequence of input utterances.

The HMM method is only one way of solving the problem of cue-based dialogue act identification. The link with HMM tagging suggests another approach, treating dialogue acts as tags, and applying other part-of-speech tagging methods based on various cues in the input. Samuel et al. (1998), for example, applied Transformation-Based Learning to dialogue act tagging.

As we conclude this section on the cue-based approach, it’s worth taking a moment to distinguish the cue-based approach from what has been called the IDIOM or CONVENTIONAL approach. The idiom approach assumes that a sentence structure like Can you give me a list? or Can you pass the salt? is ambiguous between a literal meaning as a yes–no QUESTION and an idiomatic meaning as a REQUEST. The grammar of English would simply list REQUEST as one meaning of Can you X? The cue-based model does share some features of the idiom model; certain surface cues are directly linked to certain discourse functions. The difference is that the pure idiom model is by definition non-compositional and non-probabilistic. A certain surface sentence type is linked with a certain set of discourse functions, one of which must be chosen. The cue-based model can capture some generalizations which the idiom approach cannot; certain cues for questions, say, may play a role also in requests. We can thus capture the link between questions and requests by saying that a certain cue plays a role in both dialogue acts.

5 Conclusion

In summary, the BDI and cue-based models of computational pragmatics are both important, and both will continue to play a role in future computational modeling. The BDI model focuses on the kind of rich, sophisticated knowledge and reasoning that is clearly necessary for building conversational agents that can interact. Agents have to know why they are asking questions, and have to be able to reason about complex pragmatic and world-knowledge issues. But the depth and richness of this model comes at the expense of breadth; current models only deal with a small number of speech acts and

604 Daniel Jurafsky

situations. The cue-based model focuses on statistical examination of the surface cues to the realization of dialogue acts. Agents have to be able to make use of the rich lexical, prosodic, and grammatical cues to interpretation. But the breadth and coverage of this model come at the expense of depth; current algorithms are able to model only very simplistic and local heuristics for cues.

As I mentioned earlier, I chose speech act interpretation as the topic for this chapter because I think of it as a touchstone task. Thus this same dialectic between logical models based on knowledge-based reasoning and probabilistic models based on statistical interpretation applies in other computational pragmatic areas like reference resolution and discourse structure interpretation.

This dialectic is also important for the field of linguistics as well. Although linguistics has traditionally embraced the symbolic, structural, and philosophical paradigm implicit in the BDI model, it has only recently begun to flirt with the probabilistic paradigm. The cue-based model shows one way in which the probabilistic paradigm can inform our understanding of the relationship between linguistic form and linguistic function.

It is clear that both these models of computational pragmatics are in their infancy. I expect significant progress in both areas in the near future, and I look forward to a comprehensive and robust integration of the two methods.

ACKNOWLEDGMENTS

Work on this chapter was partially supported by NSF CAREER-IIS-9733067. I’m very grateful to the editors and to Andy Kehler, Laura Michaelis, Barbara Fox, Brad Miller, and Jim Martin for advice on and discussion about this material.

NOTE

1ATIS, or Air Travel Information System, is a corpus of sentences spoken to a computerized speech understanding system for planning air travel. It is available as part of the Penn Treebank Project (Marcus et al. 1999).

Part IV Pragmatics and

Cognition

HH