Workshop at the 38th Annual Conference of the German Linguistics Society (DGfS) in Konstanz, February 24-26, 2016.
Computational pragmatics can be understood in two different senses. First, it can be seen as a subfield of computational linguistics, in which it has a longer tradition. Example phenomena addressed in this tradition are: computational models of implicature, dialogue act planning, discourse structuring, coreference resolution (Bunt & Black 2000, and others). Second, it can refer to a rapidly growing field at the interface between linguistics, cognitive science and artificial intelligence. An example is the rational speech act model (Frank & Goodman 2012) which uses Bayesian methods for modeling cognitive aspects of the interpretation of sentence fragments and implicatures. Computational pragmatics is of growing interest to linguistic pragmatics, first, due to the availability of theories that are precise enough to form the basis of NLP systems (e.g. game theoretic pragmatics, SDRT, RST), and second, due to the additional opportunities which computational pragmatics provides for advanced experimental testing of pragmatic theories. As such, it enhances theoretical, experimental and corpus-based approaches to pragmatics.
In this workshop, we want to bring together researchers from both branches of computational linguistics, as well as linguists with an interest in formal approaches to pragmatics. Topics of the workshop include, but are not limited to, the following issues:
Wednesday, 24.2. |
|
14:00 - 15:00 |
As a sequel to Bunt and Black (2000), which presented a characterization of the field of computational pragmatics and a survey of its main issues, this paper discusses some of the most interesting developments in the field in the last 15 years. Current research is dependent on large-scale annotated corpora. The paper includes an overview of such corpora and accompanying software tools. Of the pragmatic phenomena that have received attention in such corpora, the use of dialogue acts in spoken interaction stands out. Dialogue acts, which have become popular for modeling the use of language as the performance of actions in context, are realized by ‘functional segments’ of communicative behavior; these may be discontinuous, may overlap, and may contain parts contributed by different speakers. |
15:00 - 16:00 |
Mixed motives represent a mixture of congruent, i.e., joint motives as well as incongruent, partially conflictive motives of interlocutors in dialogues. Motives refer to objectives or situations that interlocutors would like to accomplish in the sense of a motivational state. As mixed-motive dialogues we describe all grades between collaborative dialogues with exclusively congruent interlocutors’ motives, e.g., when solving a PC problem together, and non-collaborative dialogues with purely incongruent motives of dialogue participants, e.g., in a pro/contra debate. Adopting the idea of mixed-motive games by Schelling, we consider these dialogues as situations in which participants are faced with a conflict between their motives to cooperate and to compete with each other, e.g., in sales conversation, where bargainers have to make concessions to establish a compromise agreement, but at the same time, they must compete to achieve a good bargain. In everyday life, interlocutors are able to solve this conflict between cooperation and competition with trade-offs between selfishness and fair play for creating dialogues perceived as fair. |
16:00 - 16:30 |
Coffee break |
16:30 - 17:30 |
In this talk, we will report on our research on interactive natural language generation (NLG) in the context of situated dialogue systems. In situated communication, the meaning of a sentence is always relative to the environment in which it is uttered, requiring us to model both in parallel. Within this task, we are particularly interested in generating referring expressions (REs), i.e. of noun phrases that identify a given object effectively within the scene. More specifically, our research on situated NLG focuses on generating instructions that help a human user solve a given task in a virtual 3D environment. This domain has the advantage of a technical complexity and reliability that is greatly reduced compared to situated communication in real-life environments. Furthermore, data collection and evaluation can be done with experimental subjects that are recruited over the Internet. We will report on the GIVE Challenge, an NLG evaluation challenge organized by our group which is built on top of this idea. Since 2009 we have developed a set of tools capable of recording, analyzing, and modeling user behavior around this challenge scenario, and have collected hundreds of hours of interactions between NLG systems and experiment subjects, which we can use to train and evaluate our systems. Crowdsourcing, the practice of collecting data from participants all over the world, allows us today to test new hypotheses in a cheap and efficient manner. Next to this training and evaluation setting, we will also report on our work on the interactive, situated generation of REs. Generating REs, reacting to misunderstandings and establishing common ground are some of the pragmatic phenomena that we must take into account. We developed a data-driven approach that allows us to generate the "best" RE for any given situation. Unlike some earlier research, we take "best" to mean the RE that maximizes the chance that the listener will understand the RE correctly. We then exploit the interactivity of the environment by tracking the listener's behavior in the virtual environment in real time. We have implemented a system that detects automatically whether the listener has understood the RE correctly, and generates corrective feedback if a misunderstanding occurred.
|
17:30 - 18:30 |
In this talk I will, firstly, summarise the state of the art of the Generation of Referring Expressions, viewed as the construction of computational models of human reference production; in this first part of the talk, I will ask what algorithms in this area are able to do well and what it is that they still struggle to do. In the second part of the talk, I will argue that the most difficult problems for the Generation of Referring Expressions arise from situations in which reference is something other than the ''simple'' identification of a referent by means of knowledge that the speaker shares with the hearer; I will give examples of these epistemically problematic situations and of the generation algorithms that try to address them. The talk offers a sneak preview of my book ''Computational Models of Referring: a Study in Cognitive Science'', which is soon to appear with MIT Press.
|
Thursday, 25.2. |
|
09:00 - 10:00 |
Consider the problem of generating and intepreting non-literal utterances in the context of (1), where "Rewe'' and "Edeka'' are supermarkets.
|
10:00 - 11:00 |
The question whether and when pragmatic enrichments, like scalar
implicatures, can occur in nonmatrix position is crucial for understanding
pragmatic inferences and processing in general. Here, we would like to
address the associated disambiguation problem (c.f. Chemla and Singh,
2014): any theory of implicature-like meaning enrichments should ideally
specify, for any sentence and context pair, which candidate readings are
preferred, and to what extent even dispreferred readings
may be selected. References Chemla, Emmanuel and Raij Singh (2014). “Remarks on the Experimental Turn in the Study of Scalar Implicature (Part I & II)”. In: Language and Linguistics Compass 8.9, pp. 373–386, 387–399. Frank, Michael C. and Noah D. Goodman (2012). “Predicting Pragmatic Reasoning in Language Games”. In: Science 336.6084, p. 998. Franke, Michael et al. (2015). “Embedded Scalars, Preferred Readings and Intonation: An Experimental Revisit”. Under review, Journal of Semantics.
|
11:00 - 11:30 |
|
11:30 - 12:30 |
Probabilistic models of human cognition have been widely successful at capturing the ways that people represent and reason with uncertain knowledge. The Rational Speech Act framework uses probabilistic modeling tools to formalize natural language understanding as social reasoning: literal sentence meaning arises through probabilistic conditioning, and pragmatic enrichment is the result of listeners reasoning about cooperative speakers. I will consider how this framework provides a theory of the role of context in language understanding. In particular I will show that when uncertainty about the speaker is included in the pragmatic inference several of the most subtle aspects of language emerge: vagueness (in scalar adjectives and generics) and presupposition accommodation.
|
Friday, 26.2. |
|
11:30 - 12:00 |
We present an extension of LFG’s Abstract Knowledge Representation
(AKR) component that integrates a model of the Common Ground (CG) and
allows for the calculation of pragmatic inferences. The system uses a rule
set based on Gunglogson’s (2002) discourse model. We illustrate our
implementation with respect to a set of German discourse particles. These
particles arguably contribute information that is pertinent for the CG (e.g.,
Zimmerman 2011). References: Asher, N. & A. Lascarides. 2003. Logics of Conversation Bobrow, D. G., B. Cheslow, C. Condoravdi, L. Karttunen, T. H. King, R. Nairn, & A. Zaenen. 2007. PARC’s bridge and question answering system. GEAF 2007 Workshop. Condoravdi, Cleo, D. Crouch, R. Stolle, V. de Paiva, & D. G. Bobrow. 2003. Entailment, intensionality and text understanding. Human Language Technology Conference Crouch, D. und T. Holloway King. 2006. Semantics via f-structure rewriting LFG06 Gunglogson, C. 2002. Declarative questions. SALT XII Zimmermann, M. 2011. Discourse particles. In P. Portner, C. Maienborn, und K. von Heusinger, edt., Semantics. HSK 33.2
|
12:00 - 12:30 |
The work in progress presented here is a contribution to argumentation
mining in the German legal text domain. Focused in this abstract is the
building of a corpus of argumentative sequences and argumentation
structures of German legal decisions that will later provide models of these
layers for conditional random field-based sequence labelling and treekernels
for structure classification. Most related works are Mochales and
Moens 2011 and Stab et al. 2014. However, there is no corpus of German
legal decisions available and building a gold-standard corpus of this genre
will be an important addition to all related fields of research. The data
collection has been compiled from a free online service and consists of 100
private law decisions. For pre-processing, a genre-specific sentence
tokenizer has been trained. The annotation framework chosen for the
study is Webanno (Yimam et al. 2013). References R. Mochales and M. F. Moens. Argumentation mining. Artificial Intelligence and Law, 19(1):1–22, 2011. C. Stab, C. Kirschner, J. Eckle-Kohler, and I. Gurevych. Argumentation mining in persuasive essays and scientific articles from the discourse structure perspective. Frontiers and Connections between Argumentation Theory and Natural Language Processing, Bertinoro, Italy, 2014. S. M. Yimam, I. Gurevych, R. E. de Castilho, and C. Biemann. Webanno: A flexible, web-based and visually supported system for distributed annotations. In ACL (Conference System Demonstrations), pages 1–6, 2013.
|
12:30 - 13:00 |
Every morning, while reading the newspaper, we are faced with a lot of different omissions, which we do not always even realize. We read headlines as
In the first case, we consider an (def. or indef.) article is missing, as there is an obligatory article before singular nouns in German. In 1 b., the structure additionally lacks some copula verb, like "(Ein) Kampfjet ist in Bayern abgestürzt''. These kinds of ellipsis are found not only in headlines. We claim that we can get a profile of text types on the basis of their distribution of ellipses. Hereto, we built a corpus containing more than 10 different text types (spoken and written language) to compare the patterns. A big challenge was the right detection and annotation of the missing elements. How can we more or less automatically find the missing article (<art>)?
Since the STTS doesn't distinguish between singular and plural nouns, the query for patterns as in 2 a. gives no satisfying output. Cases like in 2(b) further challenge the task. Finally, we did the annotation by hand for having a reliable annotation. Secondly, I want to discuss the different forms of article omissions in the light of Information Theory. A central claim is that such omissions are a way to "densify'' an utterance in order to reduce redundancy. For this purpose we train Language Models on different text types and calculate Information Density (i.e. -log2 P(w|c)) like in (3) for headlines.
In (3), the ID of the noun without preceding article (11.6582) is higher and hence much more "dense'' than in the case of article realization (7.7953). The ID of the article itself is quite low (4.6328). Thus, a puzzle I want to address in the talk is, why the article sometimes is realized instead of its omission -- and vice versa. Furthermore, which role plays the Uniform Information Hypothesis (e.g. Jaeger 2010) here? A further aim of this (ongoing) work is to compare the values in different text types. The claim of a certain "profile'' for each text type should be reflected in different probability values, different ID profiles respectively. The aim is to show first profiles and to discuss further possibilities in CompPrag -- since there are some other ellipsis on hold.
|
13:00 - 13:30 |
In this paper, we use the phenomenon of 'embarrassed laughter' as a case study of one approach to corpus pragmatics. We construct a set of interlinked ontologies by comparing the transcription practice of various collections of data as summarised by Hepburn and Varney (2013), making explicit the implied knowledge underlying those transcription practices about the characteristics of laughter which have been treated as interactionally relevant. These ontologies allow us to see the essentially combinatorial nature of certain pragmatic phenomena and therefore also allow us to develop strategies for searching for relevant data. We then proceed to illustrate how such search strategies can work with the example of 'embarrassed laughter'. Such laughter often occurs early in an interaction (especially first encounters) and following long pauses. We can therefore establish a set of search criteria (laughter AND (start of interaction OR long pause) to try to find possible instances of this phenomenon in varied collections of data such as those which form part of the Australian National Corpus. Our approach acknowledges the complexity of the factors which may be relevant to the identification of any pragmatic phenomenon without relying on the prior identification of instances in any specific dataset and has the capability to generate candidate sets of examples across varied data sets while relying on features which are annotated in standard practice. We suggest that looking for clusters of features which characterize pragmatic phenomena and organizing our knowledge of the features with ontologies constitutes a very promising approach in the field of corpus pragmatics. Reference: Hepburn, Alexa & Scott Varney. 2013. Beyond ((Laughter)): Some Notes on Transcription. In Phillip Glenn & Elizabeth Holt (eds.), Studies of Laughter in Interaction, 25–38. London: Bloomsbury Academic. http://www.bloomsburycollections.com/book/studies-of-laughter-ininteraction (16 August, 2015). |
13:30 - 14:00 |
Concluding remarks |