Evaluating Answers to Reading Comprehension Questions in Context: Results for German and the Role of Information Structure

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

Detmar Meurers, Ramon Ziai, Niels Ott & Janina Kopp

Experiment Overall results Detailed evaluation

¨ Tubingen ¨ Universitat SFB 833, Projekt A4

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

EMNLP TextInfer-Workshop 2011 July 30, 2011 1 / 27

Overview

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure

Introduction

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction

Our Corpus

Our Corpus Data sets used

CoMiC Approach Annotation

CoMiC Approach

Alignment Classification Features

Experiment Overall results

Experiment

Detailed evaluation

Information Structure Givenness filter

Information Structure

Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion

Conclusion

References SFB 833

2 / 27

Long-term research questions

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction

I

What linguistic representations can be used robustly and efficiently in automatic meaning comparison?

Our Corpus Data sets used

CoMiC Approach Annotation

I

What is the role of context and how can we utilize knowledge about it in comparing meaning automatically?

Alignment Classification Features

Experiment Overall results

I

Context here means questions and reading texts in reading comprehension tasks.

Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

3 / 27

Aims of this talk

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction

I

Present first content assessment approach for German

Our Corpus

I

Explore impact of

CoMiC Approach

Data sets used

I I

question types and ways of encoding information in the text

Annotation Alignment Classification Features

Experiment

I

Discuss the importance of explicit language-based context I

here: information structure of answers given questions and text

Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

4 / 27

Connection to RTE and Textual Inference

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

I

What is Content Assessment? I

I

The task of determining whether a response actually answers a given question about a specific text.

Two possible perspectives in connection with RTE:

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

1. Decide whether reading text T supports student answer SA , i.e., whether SA is entailed by T .

Experiment

2. Decide whether student answer SA is paraphrase of target answer TA . ⇒ bi-directional entailment

Information Structure

Overall results Detailed evaluation

Givenness filter Alternative question problem From Givenness to Focus

In this talk, we focus on the second perspective.

Towards annotating focus

Conclusion References SFB 833

5 / 27

Example from our corpus

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure

T:

(Reading comprehension text)

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

¨ ¨ Hamburg außern? Q: Was sind die Kritikpunkte, die Leute uber ‘What are the objections people have about Hamburg?’ TA: Der Gestank von Fisch und Schiffsdiesel an den Kais . The stink of fish and fuel on the quays .

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter

SA: Der Geruch zon Fish und Schiffsdiesel beim Hafen . The smell oferr fisherr and fuel at the port .

Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

6 / 27

Data source: CREG Corpus of Reading Comprehension Exercises in German I

Consists of I I I I

reading texts, reading comprehension questions, target answers formulated by teachers, student answers to the questions.

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

7 / 27

Data source: CREG Corpus of Reading Comprehension Exercises in German I

Consists of I I I I

reading texts, reading comprehension questions, target answers formulated by teachers, student answers to the questions.

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment

I

Is being collected in two large German programs in US I I

The Ohio State University (Prof. Kathryn Corl) Kansas University (Prof. Nina Vyatkina)

Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

7 / 27

Data source: CREG Corpus of Reading Comprehension Exercises in German I

Consists of I I I I

reading texts, reading comprehension questions, target answers formulated by teachers, student answers to the questions.

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment

I

Is being collected in two large German programs in US I I

The Ohio State University (Prof. Kathryn Corl) Kansas University (Prof. Nina Vyatkina)

Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter

I

Two research assistants independently rate each student answer with respect to meaning. I I I

Did student provide meaningful answer to question? Binary categories: adequate/inadequate Annotators also identify target answer for student answer

Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

7 / 27

Data sets used I

From the corpus in development, we took a snapshot I I

with full agreement in binary ratings, and with half of the answers being rated as inadequate (random base line = 50%).

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

I

Resulted in one data set for each of the two sites I

No overlap in exercise material

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

8 / 27

Evaluating Answers to RC Questions in Context:

Data sets used I

From the corpus in development, we took a snapshot I I

with full agreement in binary ratings, and with half of the answers being rated as inadequate (random base line = 50%).

Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

I

Resulted in one data set for each of the two sites I

CoMiC Approach Annotation

No overlap in exercise material

Alignment Classification Features

Target Answers Questions Student Answers # of Students SAs per question avg. Token #

KU data set 136 117 610 141 5.21 9.71

OSU data set 87 60 422 175 7.03 15.00

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

8 / 27

General CoMiC Approach (Bailey & Meurers 2008; Meurers, Ziai, Ott & Bailey 2011) The overall approach has three phases:

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction

1. Annotation uses NLP to enrich the student and target answers, as well as the question text, with linguistic information on different levels and types of abstraction.

Our Corpus Data sets used

CoMiC Approach Annotation Alignment

2. Alignment maps elements of the learner answer to elements of the target response using annotation. I

Global alignment solution computed by Traditional Marriage Algorithm (Gale & Shapley 1962)

Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem

3. Classification analyzes the possible alignments and labels the learner response with a binary content assessment and a detailed diagnosis code.

From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

9 / 27

Evaluating Answers to RC Questions in Context:

Annotation NLP Components Annotation Task Sentence Detection

Results for German and the Role of Information Structure

NLP Component OpenNLP http://incubator.apache.org/opennlp

Tokenization Lemmatization Spell Checking

Part-of-speech Tagging Noun Phrase Chunking Lexical Relations Similarity Scores Dependency Relations

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

OpenNLP TreeTagger (Schmid 1994) Edit distance (Levenshtein 1966) igerman98 word list

CoMiC Approach

http://www.j3e.de/ispell/igerman98

Information Structure

TreeTagger (Schmid 1994) OpenNLP GermaNet (Hamp & Feldweg 1997) PMI-IR (Turney 2001) MaltParser (Nivre et al. 2007)

Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

10 / 27

Alignment Example

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

11 / 27

Evaluating Answers to RC Questions in Context:

Classification Features I

Results for German and the Role of Information Structure

Content Assessment is based on 13 features: % of Overlapping Matches: I

keyword (head)

I

target/learner token

I

target/learner chunk

I

target/learner triple

Nature of Matches: I % token matches I % lemma matches I % synonym matches I % similarity matches I % sem. type matches I match variety

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

12 / 27

Evaluating Answers to RC Questions in Context:

Classification Features I

Results for German and the Role of Information Structure

Content Assessment is based on 13 features: % of Overlapping Matches: I

keyword (head)

I

target/learner token

I

target/learner chunk

I

target/learner triple

Nature of Matches: I % token matches I % lemma matches I % synonym matches I % similarity matches I % sem. type matches I match variety

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure

I

We combined the evidence with memory-based learning (TiMBL, Daelemans et al. 2007) I

Trained seven classifiers using different distance metrics, overall outcome obtained through majority voting.

I

Used leave-one-out testing: For each test item train on all answer pairs except the test item itself.

Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

12 / 27

Experiment

Evaluating Answers to RC Questions in Context:

Overall results

Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

# of answers Accuracy

KU data set 610 84.6%

OSU data set 422 84.6%

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

I

I

I

Remarkable similarity of results across completely different data sets Same overall results when macro-averaging over individual questions Competitive with results obtained for English (78%) in Bailey & Meurers (2008) and related results of C-Rater for short answer scoring (Leacock & Chodorow 2003).

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

13 / 27

Detailed Evaluation

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure

I

I

Global accuracy scores do not tell us how well the system fares, e.g., in terms of question types. First step towards deeper analysis of results: manual annotation of reading comprehension question properties

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation

I

Annotation scheme follows Day & Park (2005) guidelines for development of reading comprehension questions I

Comprehension Types: I

I

nature & depth of comprehension required by learner to answer the question in our data: “Literal”, “Reorganization” and “Inference”

Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

I

Question Forms: I

Surface-based question classes such as “yes/no” or “who” questions

Conclusion References SFB 833

14 / 27

Evaluating Answers to RC Questions in Context:

Detailed Evaluation Accuracy by question form and comprehension Literal

Reorganization

Inference

Results for German and the Role of Information Structure

Total

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

0 (1)

− (0)

66.67 (6)

57.14 (7)

Introduction

How

85.71 (126)

83.33 (12)

100 (7)

86.21 (145)

Our Corpus

What

87.04 (247)

74.19 (31)

83.33 (6)

85.56 (284)

When

85.71 (7)

− (0)

− (0)

85.71 (7)

Where

88.89 (9)

− (0)

− (0)

88.89 (9)

Which

92.35 (183)

100 (14)

83.33 (6)

92.61 (203)

Who

73.91 (23)

94.44 (18)

− (0)

82.93 (41)

Why

Alternative

80.47 (128)

57.14 (14)

84.38 (32)

79.31 (174)

Yes/No

− (0)

100 (5)

− (0)

100 (5)

Several

82.11 (95)

68.42 (38)

75 (24)

77.71 (157)

85.96 (819)

78.03 (132)

81.48 (81)

84.59 (1032)

Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter

Total

Alternative question problem From Givenness to Focus Towards annotating focus

I

Answer counts shown in brackets

I

Error bars indicate 95% confidence intervals

Conclusion References SFB 833

15 / 27

Evaluating Answers to RC Questions in Context:

Most important results Comprehension types Literal

Results for German and the Role of Information Structure

Reorganization

Inference

Total

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

0 (1)

− (0)

66.67 (6)

57.14 (7)

Introduction

How

85.71 (126)

83.33 (12)

100 (7)

86.21 (145)

Our Corpus

What

87.04 (247)

74.19 (31)

83.33 (6)

85.56 (284)

When

85.71 (7)

− (0)

− (0)

85.71 (7)

Where

88.89 (9)

− (0)

− (0)

88.89 (9)

Which

92.35 (183)

100 (14)

83.33 (6)

92.61 (203)

Who

73.91 (23)

94.44 (18)

− (0)

82.93 (41)

Why

Alternative

80.47 (128)

57.14 (14)

84.38 (32)

79.31 (174)

Yes/No

− (0)

100 (5)

− (0)

100 (5)

Several

82.11 (95)

68.42 (38)

75 (24)

77.71 (157)

85.96 (819)

78.03 (132)

81.48 (81)

84.59 (1032)

Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter

Total

Alternative question problem From Givenness to Focus Towards annotating focus

I

“Literal” questions (86.0%) seem to be easier than “Reorganization” (78.0%) and “Inference” (81.5%).

Conclusion References SFB 833

16 / 27

Evaluating Answers to RC Questions in Context:

Most important results Question forms: easy case Literal

Reorganization

Results for German and the Role of Information Structure

Inference

Total

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

0 (1)

− (0)

66.67 (6)

57.14 (7)

Introduction

How

85.71 (126)

83.33 (12)

100 (7)

86.21 (145)

Our Corpus

What

87.04 (247)

74.19 (31)

83.33 (6)

85.56 (284)

When

85.71 (7)

− (0)

− (0)

85.71 (7)

Where

88.89 (9)

− (0)

− (0)

88.89 (9)

Which

92.35 (183)

100 (14)

83.33 (6)

92.61 (203)

Who

73.91 (23)

94.44 (18)

− (0)

82.93 (41)

Why

Alternative

80.47 (128)

57.14 (14)

84.38 (32)

79.31 (174)

Yes/No

− (0)

100 (5)

− (0)

100 (5)

Several

82.11 (95)

68.42 (38)

75 (24)

77.71 (157)

85.96 (819)

78.03 (132)

81.48 (81)

84.59 (1032)

Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter

Total

Alternative question problem From Givenness to Focus Towards annotating focus

I

Accuracy for wh-questions based on concrete information from text is rather high, e.g., 92.6% for “which” questions.

Conclusion References SFB 833

17 / 27

Evaluating Answers to RC Questions in Context:

Most important results Question forms: hard case Literal

Reorganization

Results for German and the Role of Information Structure

Inference

Total

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

0 (1)

− (0)

66.67 (6)

57.14 (7)

Introduction

How

85.71 (126)

83.33 (12)

100 (7)

86.21 (145)

Our Corpus

What

87.04 (247)

74.19 (31)

83.33 (6)

85.56 (284)

When

85.71 (7)

− (0)

− (0)

85.71 (7)

Where

88.89 (9)

− (0)

− (0)

88.89 (9)

Which

92.35 (183)

100 (14)

83.33 (6)

92.61 (203)

Who

73.91 (23)

94.44 (18)

− (0)

82.93 (41)

Why

Alternative

80.47 (128)

57.14 (14)

84.38 (32)

79.31 (174)

Yes/No

− (0)

100 (5)

− (0)

100 (5)

Several

82.11 (95)

68.42 (38)

75 (24)

77.71 (157)

85.96 (819)

78.03 (132)

81.48 (81)

84.59 (1032)

Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter

Total

Alternative question problem From Givenness to Focus Towards annotating focus

I

“why” questions are difficult (79.3%): Asking for reasons/causes supports more answer variation.

Conclusion References SFB 833

18 / 27

Evaluating Answers to RC Questions in Context:

Most important results Question forms: a puzzle Literal

Reorganization

Results for German and the Role of Information Structure

Inference

Total

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

0 (1)

− (0)

66.67 (6)

57.14 (7)

Introduction

How

85.71 (126)

83.33 (12)

100 (7)

86.21 (145)

Our Corpus

What

87.04 (247)

74.19 (31)

83.33 (6)

85.56 (284)

When

85.71 (7)

− (0)

− (0)

85.71 (7)

Where

88.89 (9)

− (0)

− (0)

88.89 (9)

Which

92.35 (183)

100 (14)

83.33 (6)

92.61 (203)

Who

73.91 (23)

94.44 (18)

− (0)

82.93 (41)

Why

Alternative

80.47 (128)

57.14 (14)

84.38 (32)

79.31 (174)

Yes/No

− (0)

100 (5)

− (0)

100 (5)

Several

82.11 (95)

68.42 (38)

75 (24)

77.71 (157)

85.96 (819)

78.03 (132)

81.48 (81)

84.59 (1032)

Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter

Total

Alternative question problem From Givenness to Focus Towards annotating focus

I

“Alternative” questions are near random level (57.1%). I

Why?

Conclusion References SFB 833

19 / 27

Information Structure I

Information Structure (IS) research investigates: I

How is the meaning of a sentence integrated into the discourse?

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction

I

One relevant notion is Givenness: I

“A constituent C counts as Given if there is a salient antecedent A for C, such that A either I I I

co-refers with C, is a synonym of C or ¨ is a hyponym of C.” (Buring 2006)

Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

I

Our system as a first approximation excludes all words from alignment that appear in the question. I

I

Motivation: Mentioned lexical material typically does not contain new information answering the question.

However, in some interesting cases, the answer to a question does include given information. I

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

Example: “Alternative” questions 20 / 27

Alternative question example

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction

Q: Ist die Wohnung in einem Neubau oder einem Altbau?

Our Corpus

‘Is the flat in a new building or in an old building?’

CoMiC Approach

Data sets used

Annotation Alignment Classification Features

TA: Die Wohnung ist in einem Neubau . The flat is in a new building

Experiment Overall results Detailed evaluation

Information Structure Givenness filter

SA: Die Wohnung ist in einem Neubau The flat is in a new building

Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

21 / 27

From Givenness to Focus

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure

I

The IS notion of a Focus as the expression which addresses an explicit or implicit question under discussion (Krifka 2004) helps address the issue. → Given information is relevant when it is part of the focus.

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

I

Making the focus explicit can also help in cases such as:

Experiment Overall results

Q: Was muss die Meerjungfrau erleiden, wenn sie Menschenbeine haben will?

Detailed evaluation

Information Structure

‘What must the mermaid suffer if she wants to have human legs?’

Givenness filter Alternative question problem From Givenness to Focus

TA: Die Meerjungfrau muss schreckliche Qualen erleiden bei jedem Schritt . The mermaid must horrible torment suffer with every step .

Towards annotating focus

Conclusion References

SA: Sie erleidt bei jedem Schritt. She suffer with every step.

SFB 833

22 / 27

Towards annotating focus

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

I

Idea: Integrate an automatic focus identification component into CoMiC.

Introduction Our Corpus Data sets used

I

Approach should be informed by manual approaches to annotating information structure aspects: I

I

CoMiC Approach Annotation Alignment Classification Features

Those targeting focus are moderately successful (Dipper et al. 2007; Calhoun et al. 2010).

Experiment

In the CREG corpus, the explicit linguistic context (text, question) may support more reliable focus identification.

Information Structure

Overall results Detailed evaluation

Givenness filter Alternative question problem

I

Information Status (Given vs. New) of referential expressions (Riester et al. 2010) may help as “backbone”.

From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

23 / 27

Conclusion I

We presented the first content assessment system for German, CoMiC-DE I I

accuracy of 84.6% on authentic classroom data competitive with results for English

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

I

Detailed evaluation by question form and comprehension type I I

clear differences in performance identifies avenues for future research improving analysis for specific question forms and comprehension types

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter

I

To identify which parts of an answer are most relevant for content assessment, information structure distinctions should be integrated. I I

manual annotation of the focus of an answer is a first step explicit language-based context of task is crucial

Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

24 / 27

Evaluating Answers to RC Questions in Context:

The End

Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation

Thank you!

Alignment Classification Features

Experiment Overall results Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

25 / 27

References I Bailey, S. & D. Meurers (2008). Diagnosing meaning errors in short answers to reading comprehension questions. In J. Tetreault, J. Burstein & R. D. Felice (eds.), Proceedings of the 3rd Workshop on Innovative Use of NLP for Building Educational Applications (BEA-3) at ACL’08. Columbus, Ohio, pp. 107–115. URL http://aclweb.org/anthology/W08-0913. ¨ ¨ Buring, D. (2006). Intonation und Informationsstruktur. In H. Bluhdorn, E. Breindl & U. H. Waßner (eds.), Text ¨ — Verstehen. Grammatik und daruber hinaus. Berlin/New York: de Gruyter, pp. 144–163. URL http://semanticsarchive.net/Archive/jI0OTk3O/buring.ids2005.pdf. Calhoun, S., J. Carletta, J. Brenier, N. Mayo, D. Jurafsky, M. Steedman & D. Beaver (2010). The NXT-format Switchboard Corpus: A Rich Resource for Investigating the Syntax, Semantics, Pragmatics and Prosody of Dialogue. Language Resources and Evaluation 44, 387–419. Daelemans, W., J. Zavrel, K. van der Sloot & A. van den Bosch (2007). TiMBL: Tilburg Memory-Based Learner Reference Guide, ILK Technical Report ILK 07-03. Induction of Linguistic Knowledge Research Group Department of Communication and Information Sciences, Tilburg University, P.O. Box 90153, NL-5000 LE, Tilburg, The Netherlands, version 6.0 ed. URL http://ilk.uvt.nl/downloads/pub/papers/ilk.0703.pdf.

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

Experiment Overall results

Day, R. R. & J.-S. Park (2005). Developing Reading Comprehension Questions. Reading in a Foreign Language 17(1), 60–73. URL http://nflrc.hawaii.edu/rfl/april2005/day/day.pdf. ¨ Dipper, S., M. Gotze & S. Skopeteas (eds.) (2007). Information Structure in Cross-Linguistic Corpora: Annotation Guidelines for Phonology, Morphology, Syntax, Semantics and Information Structure, vol. 7 of ¨ Interdisciplinary Studies on Information Structure. Potsdam, Germany: Universitatsverlag Potsdam. Gale, D. & L. S. Shapley (1962). College Admissions and the Stability of Marriage. American Mathematical Monthly 69, 9–15. URL http://www.econ.ucsb.edu/∼tedb/Courses/Ec100C/galeshapley.pdf. Hamp, B. & H. Feldweg (1997). GermaNet – a Lexical-Semantic Net for German. In Proceedings of ACL workshop Automatic Information Extraction and Building of Lexical Semantic Resources for NLP Applications. Madrid. URL http://aclweb.org/anthology/W97-0802. ¨ Krifka, M. (2004). The semantics of questions and the focusation of answers. In C. Lee, M. Gordon & D. Buring (eds.), Topic and Focus: A Cross-Linguistic Perspective, Dordrecht: Kluwer Academic Publishers, pp. 139–151.

Detailed evaluation

Information Structure Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

26 / 27

References II

Evaluating Answers to RC Questions in Context: Results for German and the Role of Information Structure

Leacock, C. & M. Chodorow (2003). C-rater: Automated Scoring of Short-Answer Questions. Computers and the Humanities 37, 389–405. URL http://www.ingentaconnect.com/content/klu/chum/2003/00000037/00000004/05144721?crawler=true. Levenshtein, V. I. (1966). Binary Codes Capable of Correcting Deletions, Insertions, and Reversals. Soviet Physics Doklady 10(8), 707–710. URL http://www.mendeley.com/research/binary-codes-capable-of-correcting-insertions-and-reversals/. Meurers, D., R. Ziai, N. Ott & S. Bailey (2011). Integrating Parallel Analysis Modules to Evaluate the Meaning of Answers to Reading Comprehension Questions. IJCEELL. Special Issue on Automatic Free-text Evaluation 21(4), 355–369. URL http://purl.org/dm/papers/meurers-ziai-ott-bailey-11.html.

Detmar Meurers, Ramon Ziai, Niels Ott, Janina Kopp

Introduction Our Corpus Data sets used

CoMiC Approach Annotation Alignment Classification Features

¨ Nivre, J., J. Nilsson, J. Hall, A. Chanev, G. Eryigit, S. Kubler, S. Marinov & E. Marsi (2007). MaltParser: A Language-Independent System for Data-Driven Dependency Parsing. Natural Language Engineering ∼ 13(1), 1–41. URL http://w3.msi.vxu.se/ nivre/papers/nle07.pdf.

Experiment

Riester, A., D. Lorenz & N. Seemann (2010). A Recursive Annotation Scheme for Referential Information Status. In Proceedings of the 7th International Conference on Language Resources and Evaluation. Valletta, Malta.

Information Structure

Schmid, H. (1994). Probabilistic Part-of-Speech Tagging Using Decision Trees. In Proceedings of the International Conference on New Methods in Language Processing. Manchester, UK, pp. 44–49. URL http://www.ims.uni-stuttgart.de/ftp/pub/corpora/tree-tagger1.pdf. Turney, P. (2001). Mining the Web for Synonyms: PMI-IR Versus LSA on TOEFL. In Proceedings of the Twelfth European Conference on Machine Learning (ECML-2001). Freiburg, Germany, pp. 491–502.

Overall results Detailed evaluation

Givenness filter Alternative question problem From Givenness to Focus Towards annotating focus

Conclusion References SFB 833

27 / 27

Evaluating Answers to Reading Comprehension ...

References. SFB 833. Long-term research questions .... Two research assistants independently rate each .... 2007). ▷ Trained seven classifiers using different distance metrics, ..... 'Is the flat in a new building or in an old building?' TA: Die. The.

839KB Sizes 1 Downloads 228 Views

Recommend Documents

Reading comprehension practice.pdf
Sign in. Loading… Whoops! There was a problem loading more pages. Retrying... Whoops! There was a problem previewing this document. Retrying.

Comprehension reading test.pdf
I was turning a sharp corner, the back type skidded. That's the last. thing I remember, says David. David lost control of the bike and. when he woke up, he was at ...

Reading-Comprehension-Questions.pdf
How did Govind acquire his dream of gold finally? (1) The sage gave him gold as a reward for his hard work was in. vain. (2) His wife diligently saved the gold he had received from his. father. (3) By selling the banana plantation. (4) His wife sold

Reading Comprehension Worksheet-Otzi the Iceman.pdf ...
Reading Comprehension Worksheet-Otzi the Iceman.pdf. Reading Comprehension Worksheet-Otzi the Iceman.pdf. Open. Extract. Open with. Sign In.

new-car-fourth-grade-reading-comprehension-worksheet.pdf ...
Bobby sighed as his mother pulled up in the car loop to get him after school. ... Displaying new-car-fourth-grade-reading-comprehension-worksheet.pdf. Page 1 ...

Reading Stories for Comprehension Success
Ms. in Elementary Education from the University of Texas at El Paso. ... L. Hall Book] Reading Stories for Comprehension Success: Primary Level, Grades 1-3 Online PDF free ... The Inventor's Desktop Companion: The Guide to Successfully Marketing ...

PrepTest 52 Reading Comprehension Explanations Sample.pdf
9. Common theme. The two passages are alike in that they suggest that both professions focus. too little on storytelling, and too much on ideas. See lines 9-14, ...

PrepTest 52 Reading Comprehension Explanations Sample.pdf
storytelling, not just West African tradition. (C) would weaken the argument by suggesting that Sembene's technique was explicitly Marxist. (D) is appealing, but ...