Leveraging Newswire Treebanks for Parsing ... - Semantic Scholar

3 downloads 0 Views 209KB Size Report
Hungarian ... permute all the nodes of verbal projections in gold dependency trees preserve .... Data Augmentation: train parsing model on the union of multiple ...
Leveraging Newswire Treebanks for Parsing Conversational Data with Argument Scrambling

Riyaz Ahmad Bhat, Irshad Ahmad Bhat, Dipti Misra Sharma International Institute of Information Technology, India

Table of contents

1. Introduction 2. Argument Scrambling 3. Tree Transformations 4. Data 5. Parsing Framework 6. Experiments and Results

1

Introduction

Formal vs. Informal Language Use i • Formal Language  adherence to defining typological properties of a language 1. 2. 3. 4.

the the the the

order order order order

of of of of

subject, object and verb, possessive (genitive) and head noun, adposition and noun, and adjective and noun.

 examples: newswire, academic writing, religious sermons etc. • Informal Language  less adherence to standard grammar 1. variation in word-order 2. implicit information such as ellipsis, null co-ordination etc. 3. argument scrambling (morphologically rich languages)  examples: colloquial language, conversations, social media, movie scripts etc. 2

Formal vs. Informal Language Use ii

• Current NLP is mostly build on Newswire texts:  newswire texts greatly adhere to standard grammar  more passive sentences, lesser imperative and interrogative sentences

3

Formal vs. Informal Language Use iii News Anc. Greek Arabic Basque Bulgarian Catalan Chinese Croatian Czech Danish Dutch English Estonian Finnish French Galician German Gothic Greek Hebrew Hindi Hungarian Indonesian Irish Italian Kazakh Latin Latvian Norwegian O.Slavonic Persian Polish Portuguese Romanian Russian Slovenian Spanish Swedish Tamil Turkish

X X X X X X X X X X X X X

Fiction X

Non Fict. X

Blog

Bible X

X X

Legal

Medical

Social

Spoken

Wiki

Web

Reviews

X X X

X

X X

X

X

X X

X X X X X

X

X

X X

X

X

X

X

X

X

X

X X

X

X

X

X

X X X X X X X X

X

X

X X X X

X X

X X X X

X X X

X X X

X X X X X X X X X X

X X

X X

X

X

X

X

X

X

X X X X

X X X

X X X

X X

X

X

X

X

X

4

Formal vs. Informal Language Use iv Challenges in Parsing Informal Texts using Newswire Annotations • Sampling Bias: limited-sized newswire treebanks only represent a subset of possible structures S.No.

Order

Percentage

1

SOV

91.83

2

OSV

7.80

3

OVS

0.19

4

SVO

0.19

5

VOS

0.0

6

VSO

0.0

Our Goal • to fix the sampling bias • to exploit existing newswire annotations for parsing informal texts

5

Argument Scrambling

Argument Scrambling i

• movement of arguments for pragmatic reasons (change in information structure) • default or unmarked constituent order e.g. Hindi → SOV • common in day-to-day conversation • word-order flexibility exploited for information structure • free word-order languages allow n factorial (n!) permutations (where n is the number of verb arguments and/or adjuncts)

6

Argument Scrambling ii

nsubj iobj dobj

koi

meri

didi

ko

bhejde

ye

tasveer

.

someone

my

sister

DAT

send

this

picture

.

nsubj dobj

yeh

achanak

kya

ho

gaya

hai

tumhen

.

this

suddenly

what

happened

PST

has

you

.

iobj dobj

nsubj

Ab

kya

jawab

doon

mein

sahib

ko

.

now

what

answer

give

I

boss

DAT

.

7

Argument Scrambling iii

nmod:tmod

aaye

nahi

aap

teen

din

se

.

come

not

you

three

days

since

.

Relation Subject Direct Object Indirect Object Time Place

Newswire (Head Final) ∼100 75 100 100 ∼100

Conversation (Head Final) 67 64 52 58 53

8

Argument Scrambling iv What triggers scrambling? • Agreement

ti

Ghar

pura

hojata

apnai

α

house+M.Sg

complete

become

our+M.Sg

seekha hai

tu-nei

• Case Marking

ti

Itne

α these

saal-mein years-in

kya

what learned be you-ERG

9

Tree Transformations

Tree Transformations i • Transformational grammar: syntactic variations derivable from basic structures called kernels Chomsky [1]  statements to questions e.g. subject-auxiliary inversion CP C

CP

IP

C

Bill

Will

I0

NP

=⇒ I

0

Bill

VP

will

V

IP

I

0

VP

ti

PP

go

I0

NP

to the movies

V

PP

go

to the movies

 topicalization CP NP

CP IP

I

NP I0

NP I0 won’t

=⇒

that pizza

IP

I

VP V

NP

eat

that pizza

I0

NP I

0

won’t

VP V

NP

eat

ti

10

Tree Transformations ii

• Dependency Tree Transformation:  permute all the nodes of verbal projections in gold dependency trees  preserve non-projective arcs by pseudo-projectivizing a tree before transformations  only alter the linear precedence relations between arguments of a verbal projection  t x 10! (more than 3 million) possible permutations for ‘t’ syntactic trees containing an average of 10 nodes

11

Tree Transformations iii  restrict permutations ◦ permute a subset of the training data (representative of the newswire domain) 95 90 85 80

00

00 00 70 00 80 00 90 0 10 0 00 11 0 00 12 0 00 13 0 00 14 0 00 15 0 00 0 60

00

50

40

30

20

10

70

0 00 00

75

◦ take only k permutations with lowest perplexity assigned by a language model ? k is the number of nodes permuted for each projection ? language model is trained on a large and diverse data set (newswire, entertainment, social media, stories, etc.)

12

Tree Transformations iv d¯i

Gopal

Ram ne

d¯i .

kit¯ ab

=⇒

Ram

kit¯ ab

ko

ko

d¯i

Gopal

Ram ne

d¯i .

kit¯ ab

=⇒

kit¯ ab

Gopal ne

.

Gopal

Ram ne

ko

ko

d¯i

Ram

.

Gopal

ne

kit¯ ab ko

d¯i .

=⇒

Ram

kit¯ ab ne

.

Gopal ko

13

Data

Data i

Newswire Training and Evaluation Data Training

Testing

Development

13304

1684

1659

Newswire Transformed(Scrambled) Data • generated 9K trees from 4K representative sentences

14

Data ii Movie Scripts and Twitter Data • Bollywood movie scripts • Twitter posts of Hindi monolingual speakers  select Hindi only tweets using a language identification system  only those tweets that contain a minimum of one argument scrambling  CNN-based sentence classification using the canonical and transformed treebank data (∼98%), Yoon Kim [2] Training

Testing

Development

-

250

250

S.No. 1 2 3 4 5 6

Order SOV OSV OVS SVO VOS VSO

Percentage 33.07 23.62 17.32 14.17 9.45 2.36

15

Parsing Framework

Transition Systems i • Arc-eager algorithm:  defines 2n configurations for w1 ,...,wn ; C = (S, B, A); where S is a stack, B a buffer, and A a set of arcs  Initialization: S = [ROOT], B = [w1 ,...,wn ] and A = ∅  Termination: the buffer is empty and the stack contains ROOT Types of transitions (t): l

1. A LEFT-ARC(l) adds an arc Bj → Si to A & removes Si from the stack ◦ Precondition: Si does not already have a head l

2. A RIGHT-ARC(l) adds an arc Si → Bj to A & pushes Bj onto the stack 3. The REDUCE transition removes the top node in the stack ◦ Precondition: S0 has a head 4. The SHIFT transition moves B0 from buffer to stack 16

Transition Systems ii Configuration:

s2

s1

s0

b0

b1

b2

ram

gopal

kitab

di

.

ROOT

ne

Scoring:

ko

(ScoreLeftArc , ScoreRightArc , ScoreShift , ScoreReduce )

MLP

Vram

Vne

Vgopal

Vko

Vkitab

Vdi

V.

VROOT

concat

concat

concat

concat

concat

concat

concat

concat

LSTMb

LSTMf

xram

LSTMb

LSTMf

xne

LSTMb

LSTMf

xgopal

LSTMb

LSTMf

xko

LSTMb

LSTMf

xkitab

LSTMb

LSTMf

xdi

LSTMb

LSTMf

x.

LSTMb

LSTMf

xROOT

17

Experiments and Results

Experiments and Results i

• Training procedure:  Data Augmentation: train parsing model on the union of multiple views of the treebank  Crosslingual Model: train parsing model on a union of crosslingual treebanks (Hindi+English)

18

Experiments and Results ii

NewswirePG /UD

NewswirePG /UD +Transformed NewswirePG /UD

UAS

LAS

LAS

96.41

92.08

LAS 91.75−0.33

UAS

NewswirePG ConversationPG

UAS 96.07−0.34

-

-

74.03

64.30

NewswireUD ConversationUD

95.04

92.65

84.68+10.65 94.59−0.45

73.94+9.64 92.03−0.62

94.56−0.48

73.23

64.77

83.97+10.74

74.61+9.84

77.73+4.5

91.87−0.78 68.12+3.35

Data-set

Table 1:

NewswireUD +EnglishUD

Parsing results with gold POS-tag

NewswirePG /UD

NewswirePG /UD +Transformed NewswirePG /UD

UAS

LAS

LAS

94.55

89.51

LAS 89.28−0.23

UAS

NewswirePG ConversationPG

UAS 94.29−0.26

-

-

69.52

58.91

NewswireUD ConversationUD

93.85

90.59

79.07+9.55 93.32−0.53

67.41+8.5 89.98−0.61

93.22−0.63

68.81

59.43

78.38+9.57

67.98+8.55

71.29+2.48

89.72−0.87 62.46+3.03

Data-set

Table 2: •

NewswireUD +EnglishUD

Parsing results with auto POS-tag

Improvement over Canonical: 8.5 LAS

19

Thank You

20

References i

Noam Chomsky. Aspects of the Theory of Syntax, volume 11. MIT press, 2014. Yoon Kim. Convolutional neural networks for sentence classification. arXiv preprint arXiv:1408.5882, 2014.

21