SlideShare a Scribd company logo
1 of 116
Download to read offline
Neural AMR: Sequence-to-Sequence
Models for Parsing and Generation
Ιοannis Konstas
joint work with Srinivasan Iyer, Mark Yatskar,
Yejin Choi, Luke Zettlemoyer
AMR graph
text
Generate from AMR
Attention
Encoder Decodergraph text
AMR graph
text
Generate from AMRParse to AMR
Attention
Encoder Decodergraph text
Attention
Encoder Decodertext graph
AMR graph
text
Generate from AMRParse to AMR
Attention
Encoder Decodergraph text
Attention
Encoder Decodertext graph
Paired
Training
AMR graph
text
Generate from AMRParse to AMR
Attention
Encoder Decodergraph text
Attention
Encoder Decodertext graph
SOTA
Paired
Training
Abstract Meaning Representation
(Banarescu et al., 2013)
know
I planet
lazy
ARG0 ARG1
inhabit
man
ARG1-of
ARG0
mod
I have known a planet that was inhabited by
a lazy man.
‣ Rooted Directed Acyclic Graph
‣ Nodes: concepts (nouns, verbs, named entities, etc)
‣ Edges: Semantic Role Labels
Abstract Meaning Representation
(Banarescu et al., 2013)
know
I planet
lazy
ARG0 ARG1
inhabit
man
ARG1-of
ARG0
mod
I have known a planet that was inhabited by
a lazy man.
know
‣ Rooted Directed Acyclic Graph
‣ Nodes: concepts (nouns, verbs, named entities, etc)
‣ Edges: Semantic Role Labels
Abstract Meaning Representation
(Banarescu et al., 2013)
know
I planet
lazy
ARG0 ARG1
inhabit
man
ARG1-of
ARG0
mod
I have known a planet that was inhabited by
a lazy man.
know
I
‣ Rooted Directed Acyclic Graph
‣ Nodes: concepts (nouns, verbs, named entities, etc)
‣ Edges: Semantic Role Labels
Abstract Meaning Representation
(Banarescu et al., 2013)
know
I planet
lazy
ARG0 ARG1
inhabit
man
ARG1-of
ARG0
mod
I have known a planet that was inhabited by
a lazy man.
know
I planet
‣ Rooted Directed Acyclic Graph
‣ Nodes: concepts (nouns, verbs, named entities, etc)
‣ Edges: Semantic Role Labels
Abstract Meaning Representation
(Banarescu et al., 2013)
Input: AMR Graph
know
I planet
lazy
ARG0 ARG1
inhabit
man
ARG1-of
ARG0
mod
I have known a planet that was inhabited by
a lazy man.
I knew a planet that was inhabited by a lazy
man.
I know a planet. It is inhabited by a lazy
man.
know
I planet
Generate from AMR
‣ Rooted Directed Acyclic Graph
‣ Nodes: concepts (nouns, verbs, named entities, etc)
‣ Edges: Semantic Role Labels
Abstract Meaning Representation
(Banarescu et al., 2013)
know
I planet
lazy
ARG0 ARG1
inhabit
man
ARG1-of
ARG0
mod
I have known a planet that was inhabited by
a lazy man.
know
I planet
Parse to AMR
Input: Text
‣ Rooted Directed Acyclic Graph
‣ Nodes: concepts (nouns, verbs, named entities, etc)
‣ Edges: Semantic Role Labels
Applications
‣ Text Summarization (Liu et al., 2015)
Applications
‣ Text Summarization (Liu et al., 2015)
sentences:
Applications
‣ Text Summarization (Liu et al., 2015)
sentences:
Parse
sentence
AMR graphs:
Applications
‣ Text Summarization (Liu et al., 2015)
summary
AMR graph:
sentences:
Parse
sentence
AMR graphs:
Applications
‣ Text Summarization (Liu et al., 2015)
summary
AMR graph:
sentences:
Parse
sentence
AMR graphs:
Generate
summary:
Applications
‣ Text Summarization (Liu et al., 2015)
summary
AMR graph:
sentences:
Parse
sentence
AMR graphs:
The children told that lie
Source
その うそ は ⼦子供 たち が つい た
sono uso-wa kodomo-tachi-ga tsui-ta
that lie-TOP child-and others-NOM breathe out-PAST
Target
Generate
summary:
Applications
‣ Text Summarization (Liu et al., 2015)
summary
AMR graph:
sentences:
Parse
sentence
AMR graphs:
The children told that lie
Source
その うそ は ⼦子供 たち が つい た
sono uso-wa kodomo-tachi-ga tsui-ta
that lie-TOP child-and others-NOM breathe out-PAST
Target
tell
child lie
that
ARG0 ARG1
ARG0-of
‣ Machine Translation (Jones et al., 2012)
Generate
summary:
Parse AMR
graph:
Applications
‣ Text Summarization (Liu et al., 2015)
summary
AMR graph:
sentences:
Parse
sentence
AMR graphs:
The children told that lie
Source
その うそ は ⼦子供 たち が つい た
sono uso-wa kodomo-tachi-ga tsui-ta
that lie-TOP child-and others-NOM breathe out-PAST
Target
Graph-to-graph
transformation:
tell
child lie
that
ARG0 ARG1
ARG0-of
tsuku
kodomo
tachi
sono
ARG1 ARG0
ARG0-of
‣ Machine Translation (Jones et al., 2012)
Generate
summary:
Parse AMR
graph:
Applications
‣ Text Summarization (Liu et al., 2015)
summary
AMR graph:
sentences:
Parse
sentence
AMR graphs:
The children told that lie
Source
その うそ は ⼦子供 たち が つい た
sono uso-wa kodomo-tachi-ga tsui-ta
that lie-TOP child-and others-NOM breathe out-PAST
Target
Graph-to-graph
transformation:
tell
child lie
that
ARG0 ARG1
ARG0-of
tsuku
kodomo
tachi
sono
ARG1 ARG0
ARG0-of
‣ Machine Translation (Jones et al., 2012)
Generate
summary:
Parse AMR
graph:
Generate
translation:
Existing Approaches
‣ MT-based
‣ Flanigan et al. 2016, Pourdamaghani and Knight 2016, Song et al. 2016
‣ Grammar-based
‣ Lampouras and Vlachos 2017, Mille et al. 2017
Generate from AMR
Existing Approaches
‣ MT-based
‣ Flanigan et al. 2016, Pourdamaghani and Knight 2016, Song et al. 2016
‣ Grammar-based
‣ Lampouras and Vlachos 2017, Mille et al. 2017
Parse to AMR
‣ Alignment-based
‣ Flanigan et al. 2014, 2017 (JAMR)
‣ Grammar-based
‣ Wang et al. 2016 (CAMR), Pust et al. 2015, Artzi et al. 2015, Damonte et al. 2017,
Goodman et al. 2016, Puzikov et al. 2016, Brandt et al. 2017, Nguyen et al. 2017
‣ Neural-based
‣ Barzdins and Gosko 2016, Peng et al. 2017, Noord and Bos 2017, Buys and Blunsom
2017
Generate from AMR
Overview
‣ Sequence-to-sequence architecture
‣ End-to-end model w/o intermediate representations
‣ Linearisation of AMR graph to string
‣ Pre-process
‣ Paired Training
‣ Scalable data augmentation algorithm
Sequence-to-sequence model
Encoderinput
Sequence-to-sequence model
Encoderinput
know
ARG0
I
ARG1
(
h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
[ ] [ ] [ ] [ ] [ ]h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
Sequence-to-sequence model
Encoder Decoderinput output
I
The
A
…
know
knew
planet
…
a
planet
man
…
…
inhabit
inhabited
was
…
ˆw = argmax
w
Y
i
p wi|w<i, h(s)
know
ARG0
I
ARG1
(
h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
[ ] [ ] [ ] [ ] [ ]h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
Sequence-to-sequence model
Attention
Encoder Decoderinput output
know ARG0 I ARG1 ( planet ARG1-of inhabit
<s>
I
know
the
planet
of
I
The
A
…
know
knew
planet
…
a
planet
man
…
…
inhabit
inhabited
was
…
ˆw = argmax
w
Y
i
p wi|w<i, h(s)
know
ARG0
I
ARG1
(
h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
[ ] [ ] [ ] [ ] [ ]h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
Linearization
Graph —> Depth First Search (Human-authored annotation)
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
Linearization
Graph —> Depth First Search (Human-authored annotation)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
Linearization
Graph —> Depth First Search (Human-authored annotation)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
person
Linearization
Graph —> Depth First Search (Human-authored annotation)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
person
Linearization
Graph —> Depth First Search (Human-authored annotation)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
person
have-role
country
“United
States”
official
Linearization
Graph —> Depth First Search (Human-authored annotation)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
person
have-role
country
“United
States”
official
Pre-processing
Linearization —> Anonymization
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
Pre-processing
Linearization —> Anonymization
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
country
“United
States”
Pre-processing
Linearization —> Anonymization
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
country
“United
States”
2002 1
Pre-processing
Linearization —> Anonymization
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
country
“United
States”
2002 1 “New York”
city
Pre-processing
Linearization —> Anonymization
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
hold
person
meet
group
ARG0 ARG1
person
expert
ARG1-of
have-role
country
“United
States”
official
date-entity city
“New York”2002 1
time location
name
ARG1
name
ARG2-of
ARG0-of
ARG2
year month
ARG0
US officials held an expert group meeting in January 2002 in New York .
country
“United
States”
2002 1 “New York”
city
loc_0 officials held an expert group meeting in month_0 year_0 in loc_1 .
Experimental Setup
AMR LDC2015E86 (SemEval-2016 Task 8)
‣ Hand annotated MR graphs: newswire, forums
‣ ~16k training / 1k development / 1k test pairs
Train
‣ Optimize cross-entropy loss
Evaluation
‣ BLEU n-gram precision (Generation)
(Papineni et al., 2002)
‣ SMATCH score (Parsing)
(Cai and Knight, 2013)
Experiments
‣ Vanilla experiment
‣ Limited Language Model Capacity
‣ Paired Training
‣ Data augmentation algorithm
First Attempt (Generation)
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
First Attempt (Generation)BLEU
0
5.8
11.6
17.4
23.2
29
TreeToStr TSP PBMT NeuralAMR
26.9
22.423
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
First Attempt (Generation)BLEU
0
5.8
11.6
17.4
23.2
29
TreeToStr TSP PBMT NeuralAMR
22
26.9
22.423
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
First Attempt (Generation)BLEU
0
5.8
11.6
17.4
23.2
29
TreeToStr TSP PBMT NeuralAMR
22
26.9
22.423
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
All systems use a
Language Model
trained on a very
large corpus.
We will emulate via
data augmentation.
(Sennrich et al., ACL 2016)
What went wrong?
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
US officials held an expert group meeting in January
2002 in New York .
United States officials held held a meeting in
January 2002 .
Reference
Prediction
44.26%
74.85%
What went wrong?
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
US officials held an expert group meeting in January
2002 in New York .
United States officials held held a meeting in
January 2002 .
Reference
Prediction
‣ Repetition
44.26%
74.85%
What went wrong?
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
US officials held an expert group meeting in January
2002 in New York .
United States officials held held a meeting in
January 2002 .
Reference
Prediction
‣ Repetition
‣ Coverage
44.26%
74.85%
What went wrong?
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
US officials held an expert group meeting in January
2002 in New York .
United States officials held held a meeting in
January 2002 .
Reference
Prediction
‣ Repetition
‣ Coverage
a) Sparsity
Tokens
0
4500
9000
13500
18000
Total OOV@1 OOV@5
44.26%
74.85%
What went wrong?
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
US officials held an expert group meeting in January
2002 in New York .
United States officials held held a meeting in
January 2002 .
Reference
Prediction
‣ Repetition
‣ Coverage
a) Sparsity
b) Avg sent length: 20 words
c) Limited Language
Modeling capacity
Tokens
0
4500
9000
13500
18000
Total OOV@1 OOV@5
44.26%
74.85%
Data Augmentation
Original Dataset: ~16k graph-sentence pairs
Data Augmentation
Original Dataset: ~16k graph-sentence pairs
Gigaword: ~183M sentences *only*
Data Augmentation
Original Dataset: ~16k graph-sentence pairs
Gigaword: ~183M sentences *only*
Sample sentences with vocabulary overlap
%
0
20
40
60
80
OOV@1 OOV@5
Original Giga-200k
Giga-2M Giga-20M
Data Augmentation
graph
text
Generate from AMR
Attention
Encoder Decodergraph text
Data Augmentation
graph
text
Generate from AMRParse to AMR
Attention
Encoder Decodergraph text
Attention
Encoder Decodertext graph
Data Augmentation
graph
text
Generate from AMRParse to AMR
Attention
Encoder Decodergraph text
Attention
Encoder Decodertext graph
Data Augmentation
graph
text
Generate from AMRParse to AMR
Attention
Encoder Decodergraph text
Attention
Encoder Decodertext graph
Re-train
Semi-supervised Learning
‣Self-training
‣ McClosky et al. 2006
‣Co-training
‣ Yarowski 1995, Blum and Mitchell 1998, Sarkar 2001
‣ Sogaard and Rishoj, 2010
Paired Training
Paired Training
Train AMR Parser P on Original Dataset ( , )
Paired Training
for i = 0 … N
Train AMR Parser P on Original Dataset ( , )
Paired Training
for i = 0 … N
Si =Sample k 10i sentences from Gigaword
Train AMR Parser P on Original Dataset ( , )
Paired Training
for i = 0 … N
Parse Si sentences with P
Si =Sample k 10i sentences from Gigaword
Train AMR Parser P on Original Dataset ( , )
Paired Training
Self-trainParser
for i = 0 … N
Parse Si sentences with P
Si =Sample k 10i sentences from Gigaword
Re-train AMR Parser P on Si
Train AMR Parser P on Original Dataset ( , )
Paired Training
Self-trainParser
for i = 0 … N
Parse Si sentences with P
Si =Sample k 10i sentences from Gigaword
Re-train AMR Parser P on Si
Train AMR Parser P on Original Dataset ( , )
Paired Training
Self-trainParser
for i = 0 … N
Parse Si sentences with P
Si =Sample k 10i sentences from Gigaword
Re-train AMR Parser P on Si
Train AMR Parser P on Original Dataset ( , )
Train Generator G on SN ( , )
Training AMR Parser
Train P on
Original Dataset
Training AMR Parser
Train P on
Original Dataset
Training AMR Parser
Sample S1=200k
sentences
from Gigaword
Train P on
Original Dataset
200k
Training AMR Parser
Sample S1=200k
sentences
from Gigaword
Parse S1 with P
Train P on
Original Dataset
200k
200k
( , )
Training AMR Parser
Sample S1=200k
sentences
from Gigaword
Train P on
S1=200k
Parse S1 with P
Train P on
Original Dataset
200k
200k 200k
( , )
Training AMR Parser
Sample S1=200k
sentences
from Gigaword
Train P on
S1=200k
Fine-tune P on
Original Dataset
Parse S1 with P
Train P on
Original Dataset
200k
200k
Fine-tune: init parameters from previous
step and train on Original Dataset
200k
( , )
Training AMR Parser
Train P on
S2=2M
Fine-tune P on
Original Dataset
Sample S2=2M
sentences
from Gigaword
Parse S2 with P
200k
200k 200k
Fine-tune: init parameters from previous
step and train on Original Dataset
( , )
Training AMR Parser
Train P on
S2=2M
Fine-tune P on
Original Dataset
Sample S2=2M
sentences
from Gigaword
Parse S2 with P
00k
2M
2M 2M
Fine-tune: init parameters from previous
step and train on Original Dataset
( , )
Training AMR Parser
Train P on
S3=20M
Fine-tune P on
Original Dataset
Sample S3=20M
sentences
from Gigaword
Parse S3 with P
2M
2M 2M
Fine-tune: init parameters from previous
step and train on Original Dataset
( , )
Training AMR Parser
Train P on
S3=20M
Fine-tune P on
Original Dataset
Sample S3=20M
sentences
from Gigaword
Parse S3 with P
2M
Fine-tune: init parameters from previous
step and train on Original Dataset
( , )
20M
20M 20M
Training AMR Parser
Train P on
S3=20M
Fine-tune P on
Original Dataset
Sample S3=20M
sentences
from Gigaword
Parse S3 with P
2M
Fine-tune: init parameters from previous
step and train on Original Dataset
( , )
20M
20M 20M
Training AMR Generator
Train G on
S4=20M
Fine-tune G on
Original Dataset
Sample S4=20M
sentences
from Gigaword
Parse S4 with P
Fine-tune: init parameters from previous
step and train on Original Dataset
( , )
2M
20M
20M 20M
Training AMR Generator
Train G on
S4=20M
Fine-tune G on
Original Dataset
Sample S4=20M
sentences
from Gigaword
Parse S4 with P
Fine-tune: init parameters from previous
step and train on Original Dataset
( , )
20M
20M 20M
G G
20M
Training AMR Generator
Train G on
S4=20M
Fine-tune G on
Original Dataset
Sample S4=20M
sentences
from Gigaword
Parse S4 with P
Fine-tune: init parameters from previous
step and train on Original Dataset
( , )
20M
20M 20M
G G
20M
Final Results (Generation)
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
Final Results (Generation)BLEU
0
7
14
21
28
35
TreeToStr TSP PBMT
NeuralAMR NeuralAMR-200k NeuralAMR-2M
NeuralAMR-20M
22
26.9
22.423
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
Final Results (Generation)BLEU
0
7
14
21
28
35
TreeToStr TSP PBMT
NeuralAMR NeuralAMR-200k NeuralAMR-2M
NeuralAMR-20M
27.4
22
26.9
22.423
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
Final Results (Generation)BLEU
0
7
14
21
28
35
TreeToStr TSP PBMT
NeuralAMR NeuralAMR-200k NeuralAMR-2M
NeuralAMR-20M
32.3
27.4
22
26.9
22.423
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
Final Results (Generation)BLEU
0
7
14
21
28
35
TreeToStr TSP PBMT
NeuralAMR NeuralAMR-200k NeuralAMR-2M
NeuralAMR-20M
33.8
32.3
27.4
22
26.9
22.423
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
Final Results (Generation)BLEU
0
7
14
21
28
35
TreeToStr TSP PBMT
NeuralAMR NeuralAMR-200k NeuralAMR-2M
NeuralAMR-20M
33.8
32.3
27.4
22
26.9
22.423
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
Final Results (Generation)BLEU
0
7
14
21
28
35
TreeToStr TSP PBMT
NeuralAMR NeuralAMR-200k NeuralAMR-2M
NeuralAMR-20M
33.8
32.3
27.4
22
26.9
22.423
TreeToStr: Flanigan et al, NAACL 2016
TSP: Song et al, EMNLP 2016
PBMT: Pourdamaghani and Knight, INLG 2016
Final Results (Parsing)
SBMT: Pust et al, 2015
CharLSTM+CAMR: Noord and Bos, 2017
Seq2Seq: Peng et al., 2017
Final Results (Parsing)SMATCH
0
14
28
42
56
70
SBMT CharLSTM+CAMR Seq2Seq NeuralAMR-20M
62.1
52
67.367.1
SBMT: Pust et al, 2015
CharLSTM+CAMR: Noord and Bos, 2017
Seq2Seq: Peng et al., 2017
Final Results (Parsing)SMATCH
0
14
28
42
56
70
SBMT CharLSTM+CAMR Seq2Seq NeuralAMR-20M
62.1
52
67.367.1
SBMT: Pust et al, 2015
CharLSTM+CAMR: Noord and Bos, 2017
Seq2Seq: Peng et al., 2017
Final Results (Parsing)SMATCH
0
14
28
42
56
70
SBMT CharLSTM+CAMR Seq2Seq NeuralAMR-20M
62.1
52
67.367.1
SBMT: Pust et al, 2015
CharLSTM+CAMR: Noord and Bos, 2017
Seq2Seq: Peng et al., 2017
Final Results (Parsing)SMATCH
0
14
28
42
56
70
SBMT CharLSTM+CAMR Seq2Seq NeuralAMR-20M
62.1
52
67.367.1
SBMT: Pust et al, 2015
CharLSTM+CAMR: Noord and Bos, 2017
Seq2Seq: Peng et al., 2017
How did we do? (Generation)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
US officials held an expert group meeting in
January 2002 in New York .
In January 2002 United States officials held a
meeting of the group experts in New York .
Reference
Prediction
44.26%
74.85%
Errors: Disfluency Coverage
How did we do? (Generation)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 loc_0
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity year_0 month_0)
:location loc_1
US officials held an expert group meeting in
January 2002 in New York .
In January 2002 United States officials held a
meeting of the group experts in New York .
Reference
Prediction
44.26%
74.85%
The report stated British government must
help to stabilize weak states and push for
international regulations that would stop
terrorists using freely available information to
create and unleash new forms of biological
warfare such as a modified version of the
influenza virus.
Reference
The report stated that the Britain government
must help stabilize the weak states and push
international regulations to stop the use of freely
available information to create a form of new
biological warfare such as the modified version
of the influenza .
Prediction
Errors: Disfluency Coverage
Summary
‣ Sequence-to-sequence models for Parsing and Generation
‣ Paired Training: scalable data augmentation algorithm
‣ Achieve state-of-the-art performance on generating from AMR
‣ Best-performing Neural AMR Parser
‣ Demo, Code and Pre-trained Models: http://ikonstas.net
Summary
Thank You
‣ Sequence-to-sequence models for Parsing and Generation
‣ Paired Training: scalable data augmentation algorithm
‣ Achieve state-of-the-art performance on generating from AMR
‣ Best-performing Neural AMR Parser
‣ Demo, Code and Pre-trained Models: http://ikonstas.net
thank-01
you
ARG1
Bonus Slides
Encoding
Linearize —> RNN encoding
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
Encoding
Linearize —> RNN encoding
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
Encoding
Linearize —> RNN encoding
hold
ARG0
(
person
ARG0-of
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
- Token embeddings
Encoding
Linearize —> RNN encoding
hold
ARG0
(
person
ARG0-of
h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
- Token embeddings
- Recurrent Neural Network (RNN)
Encoding
Linearize —> RNN encoding
hold
ARG0
(
person
ARG0-of
h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
- Token embeddings
- Recurrent Neural Network (RNN)
- Bi-directional RNN
Encoding
Linearize —> RNN encoding
hold
ARG0
(
person
ARG0-of
h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
[ ] [ ] [ ] [ ] [ ]h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
- Token embeddings
- Recurrent Neural Network (RNN)
- Bi-directional RNN
Encoding
Linearize —> RNN encoding
hold
ARG0
(
person
ARG0-of
h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
[ ] [ ] [ ] [ ] [ ]h1
(s) h2
(s) h3
(s) h4
(s) h5
(s)
hold
:ARG0 (person
:ARG0-of (have-role
:ARG1 United_States
:ARG2 official)
)
:ARG1 (meet
:ARG0 (person
:ARG1-of expert
:ARG2-of group)
)
:time (date-entity 2002 1)
:location New_York
- Token embeddings
- Recurrent Neural Network (RNN)
- Bi-directional RNN
Decoding
h1hN
(s)
RNN Encoding —> RNN Decoding (Beam search)
Decoding
h1hN
(s)
;
RNN Encoding —> RNN Decoding (Beam search)
- init h(s)
Decoding
h1hN
(s)
;
Holding
Held
US
…
RNN Encoding —> RNN Decoding (Beam search)
- softmax
- init h(s)
Decoding
h1hN
(s)
;
Holding
Held
US
…
h2
a
the
meeting
…
…
w11: Holding
Heldsw12:
Holdw13:
USw14:
RNN Encoding —> RNN Decoding (Beam search)
- softmax
- p wi|w<i, h(s)
- init h(s)
Decoding
h1hN
(s)
;
Holding
Held
US
…
h2
a
the
meeting
…
h3
US
person
expert
…
…
…
w11: Holding
Heldsw12:
Holdw13:
USw14:
…
Hold aw21:
Hold thew22:
Held aw23:
Held thew24:
RNN Encoding —> RNN Decoding (Beam search)
- softmax
- p wi|w<i, h(s)
- init h(s)
Decoding
h1hN
(s)
;
Holding
Held
US
…
h2
a
the
meeting
…
h3
US
person
expert
…
hk…
…
w11: Holding
Heldsw12:
Holdw13:
USw14:
…
Hold aw21:
Hold thew22:
Held aw23:
Held thew24:
wk1: The US officials held
wk2: US officials held a
wk3: US officials hold the
wk4: US officials will hold a
…
meeting
meetings
meet
…
RNN Encoding —> RNN Decoding (Beam search)
- softmax
- p wi|w<i, h(s)
- init h(s)
Attention
h2 h3
a
the
meeting
…
w2: held
Attention
h3
a
the
meeting
…
w2: held
hold
ARG0
(
person
ARG0-of
[][][][][]h1
(s)h2
(s)h3
(s)h4
(s)h5
(s)
c3
[ ]
Attention
h3
a
the
meeting
…
w2: held
hold
ARG0
(
person
ARG0-of
[][][][][]h1
(s)h2
(s)h3
(s)h4
(s)h5
(s)
c3
[ ]
ci =
X
i
aijh
(s)
j
ai = soft max fi h(s)
, hi
Attention
hold ARG0 ( person role US official ) ARG1 ( meet expert group )
US
officials
held
an
expert
group
meeting
in
January
2002
h3
a
the
meeting
…
w2: held
hold
ARG0
(
person
ARG0-of
[][][][][]h1
(s)h2
(s)h3
(s)h4
(s)h5
(s)
c3
[ ]
ci =
X
i
aijh
(s)
j
ai = soft max fi h(s)
, hi
Attention
hold ARG0 ( person role US official ) ARG1 ( meet expert group )
US
officials
held
an
expert
group
meeting
in
January
2002
h3
a
the
meeting
…
w2: held
hold
ARG0
(
person
ARG0-of
[][][][][]h1
(s)h2
(s)h3
(s)h4
(s)h5
(s)
c3
[ ]
ci =
X
i
aijh
(s)
j
ai = soft max fi h(s)
, hi
Attention
hold ARG0 ( person role US official ) ARG1 ( meet expert group )
US
officials
held
an
expert
group
meeting
in
January
2002
h3
a
the
meeting
…
w2: held
hold
ARG0
(
person
ARG0-of
[][][][][]h1
(s)h2
(s)h3
(s)h4
(s)h5
(s)
c3
[ ]
ci =
X
i
aijh
(s)
j
ai = soft max fi h(s)
, hi
Attention
hold ARG0 ( person role US official ) ARG1 ( meet expert group )
US
officials
held
an
expert
group
meeting
in
January
2002
h3
a
the
meeting
…
w2: held
hold
ARG0
(
person
ARG0-of
[][][][][]h1
(s)h2
(s)h3
(s)h4
(s)h5
(s)
c3
[ ]
ci =
X
i
aijh
(s)
j
ai = soft max fi h(s)
, hi

More Related Content

More from Association for Computational Linguistics

Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...Association for Computational Linguistics
 
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...Association for Computational Linguistics
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopAssociation for Computational Linguistics
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...Association for Computational Linguistics
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...Association for Computational Linguistics
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Association for Computational Linguistics
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Association for Computational Linguistics
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopAssociation for Computational Linguistics
 
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...Association for Computational Linguistics
 
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...Association for Computational Linguistics
 
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...Association for Computational Linguistics
 
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...Association for Computational Linguistics
 
Toshiaki Nakazawa - 2015 - Overview of the 2nd Workshop on Asian Translation
Toshiaki Nakazawa  - 2015 - Overview of the 2nd Workshop on Asian TranslationToshiaki Nakazawa  - 2015 - Overview of the 2nd Workshop on Asian Translation
Toshiaki Nakazawa - 2015 - Overview of the 2nd Workshop on Asian TranslationAssociation for Computational Linguistics
 
Hua Shan - 2015 - A Dependency-to-String Model for Chinese-Japanese SMT System
Hua Shan - 2015 - A Dependency-to-String Model for Chinese-Japanese SMT SystemHua Shan - 2015 - A Dependency-to-String Model for Chinese-Japanese SMT System
Hua Shan - 2015 - A Dependency-to-String Model for Chinese-Japanese SMT SystemAssociation for Computational Linguistics
 
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...Association for Computational Linguistics
 
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...Association for Computational Linguistics
 
Katsuhito Sudoh - 2015 Chinese-to-Japanese Patent Machine Translation based o...
Katsuhito Sudoh - 2015 Chinese-to-Japanese Patent Machine Translation based o...Katsuhito Sudoh - 2015 Chinese-to-Japanese Patent Machine Translation based o...
Katsuhito Sudoh - 2015 Chinese-to-Japanese Patent Machine Translation based o...Association for Computational Linguistics
 

More from Association for Computational Linguistics (20)

Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
 
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
 
Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
 
Hyoung-Gyu Lee - 2015 - NAVER Machine Translation System for WAT 2015
Hyoung-Gyu Lee - 2015 - NAVER Machine Translation System for WAT 2015Hyoung-Gyu Lee - 2015 - NAVER Machine Translation System for WAT 2015
Hyoung-Gyu Lee - 2015 - NAVER Machine Translation System for WAT 2015
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
 
Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015
 
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
 
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
 
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
 
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
Terumasa Ehara - 2015 - System Combination of RBMT plus SPE and Preordering p...
 
Toshiaki Nakazawa - 2015 - Overview of the 2nd Workshop on Asian Translation
Toshiaki Nakazawa  - 2015 - Overview of the 2nd Workshop on Asian TranslationToshiaki Nakazawa  - 2015 - Overview of the 2nd Workshop on Asian Translation
Toshiaki Nakazawa - 2015 - Overview of the 2nd Workshop on Asian Translation
 
Hua Shan - 2015 - A Dependency-to-String Model for Chinese-Japanese SMT System
Hua Shan - 2015 - A Dependency-to-String Model for Chinese-Japanese SMT SystemHua Shan - 2015 - A Dependency-to-String Model for Chinese-Japanese SMT System
Hua Shan - 2015 - A Dependency-to-String Model for Chinese-Japanese SMT System
 
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
 
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
Wei Yang - 2015 - Sampling-based Alignment and Hierarchical Sub-sentential Al...
 
Katsuhito Sudoh - 2015 Chinese-to-Japanese Patent Machine Translation based o...
Katsuhito Sudoh - 2015 Chinese-to-Japanese Patent Machine Translation based o...Katsuhito Sudoh - 2015 Chinese-to-Japanese Patent Machine Translation based o...
Katsuhito Sudoh - 2015 Chinese-to-Japanese Patent Machine Translation based o...
 

Recently uploaded

Satirical Depths - A Study of Gabriel Okara's Poem - 'You Laughed and Laughed...
Satirical Depths - A Study of Gabriel Okara's Poem - 'You Laughed and Laughed...Satirical Depths - A Study of Gabriel Okara's Poem - 'You Laughed and Laughed...
Satirical Depths - A Study of Gabriel Okara's Poem - 'You Laughed and Laughed...HetalPathak10
 
Objectives n learning outcoms - MD 20240404.pptx
Objectives n learning outcoms - MD 20240404.pptxObjectives n learning outcoms - MD 20240404.pptx
Objectives n learning outcoms - MD 20240404.pptxMadhavi Dharankar
 
Man or Manufactured_ Redefining Humanity Through Biopunk Narratives.pptx
Man or Manufactured_ Redefining Humanity Through Biopunk Narratives.pptxMan or Manufactured_ Redefining Humanity Through Biopunk Narratives.pptx
Man or Manufactured_ Redefining Humanity Through Biopunk Narratives.pptxDhatriParmar
 
ClimART Action | eTwinning Project
ClimART Action    |    eTwinning ProjectClimART Action    |    eTwinning Project
ClimART Action | eTwinning Projectjordimapav
 
Unraveling Hypertext_ Analyzing Postmodern Elements in Literature.pptx
Unraveling Hypertext_ Analyzing  Postmodern Elements in  Literature.pptxUnraveling Hypertext_ Analyzing  Postmodern Elements in  Literature.pptx
Unraveling Hypertext_ Analyzing Postmodern Elements in Literature.pptxDhatriParmar
 
DBMSArchitecture_QueryProcessingandOptimization.pdf
DBMSArchitecture_QueryProcessingandOptimization.pdfDBMSArchitecture_QueryProcessingandOptimization.pdf
DBMSArchitecture_QueryProcessingandOptimization.pdfChristalin Nelson
 
CLASSIFICATION OF ANTI - CANCER DRUGS.pptx
CLASSIFICATION OF ANTI - CANCER DRUGS.pptxCLASSIFICATION OF ANTI - CANCER DRUGS.pptx
CLASSIFICATION OF ANTI - CANCER DRUGS.pptxAnupam32727
 
Q-Factor General Quiz-7th April 2024, Quiz Club NITW
Q-Factor General Quiz-7th April 2024, Quiz Club NITWQ-Factor General Quiz-7th April 2024, Quiz Club NITW
Q-Factor General Quiz-7th April 2024, Quiz Club NITWQuiz Club NITW
 
An Overview of the Calendar App in Odoo 17 ERP
An Overview of the Calendar App in Odoo 17 ERPAn Overview of the Calendar App in Odoo 17 ERP
An Overview of the Calendar App in Odoo 17 ERPCeline George
 
BÀI TẬP BỔ TRỢ TIẾNG ANH 11 THEO ĐƠN VỊ BÀI HỌC - CẢ NĂM - CÓ FILE NGHE (GLOB...
BÀI TẬP BỔ TRỢ TIẾNG ANH 11 THEO ĐƠN VỊ BÀI HỌC - CẢ NĂM - CÓ FILE NGHE (GLOB...BÀI TẬP BỔ TRỢ TIẾNG ANH 11 THEO ĐƠN VỊ BÀI HỌC - CẢ NĂM - CÓ FILE NGHE (GLOB...
BÀI TẬP BỔ TRỢ TIẾNG ANH 11 THEO ĐƠN VỊ BÀI HỌC - CẢ NĂM - CÓ FILE NGHE (GLOB...Nguyen Thanh Tu Collection
 
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITWQ-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITWQuiz Club NITW
 
Scientific Writing :Research Discourse
Scientific  Writing :Research  DiscourseScientific  Writing :Research  Discourse
Scientific Writing :Research DiscourseAnita GoswamiGiri
 
How to Uninstall a Module in Odoo 17 Using Command Line
How to Uninstall a Module in Odoo 17 Using Command LineHow to Uninstall a Module in Odoo 17 Using Command Line
How to Uninstall a Module in Odoo 17 Using Command LineCeline George
 
Indexing Structures in Database Management system.pdf
Indexing Structures in Database Management system.pdfIndexing Structures in Database Management system.pdf
Indexing Structures in Database Management system.pdfChristalin Nelson
 
4.9.24 Social Capital and Social Exclusion.pptx
4.9.24 Social Capital and Social Exclusion.pptx4.9.24 Social Capital and Social Exclusion.pptx
4.9.24 Social Capital and Social Exclusion.pptxmary850239
 
Mythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITWMythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITWQuiz Club NITW
 

Recently uploaded (20)

Spearman's correlation,Formula,Advantages,
Spearman's correlation,Formula,Advantages,Spearman's correlation,Formula,Advantages,
Spearman's correlation,Formula,Advantages,
 
Satirical Depths - A Study of Gabriel Okara's Poem - 'You Laughed and Laughed...
Satirical Depths - A Study of Gabriel Okara's Poem - 'You Laughed and Laughed...Satirical Depths - A Study of Gabriel Okara's Poem - 'You Laughed and Laughed...
Satirical Depths - A Study of Gabriel Okara's Poem - 'You Laughed and Laughed...
 
Objectives n learning outcoms - MD 20240404.pptx
Objectives n learning outcoms - MD 20240404.pptxObjectives n learning outcoms - MD 20240404.pptx
Objectives n learning outcoms - MD 20240404.pptx
 
Man or Manufactured_ Redefining Humanity Through Biopunk Narratives.pptx
Man or Manufactured_ Redefining Humanity Through Biopunk Narratives.pptxMan or Manufactured_ Redefining Humanity Through Biopunk Narratives.pptx
Man or Manufactured_ Redefining Humanity Through Biopunk Narratives.pptx
 
ClimART Action | eTwinning Project
ClimART Action    |    eTwinning ProjectClimART Action    |    eTwinning Project
ClimART Action | eTwinning Project
 
Unraveling Hypertext_ Analyzing Postmodern Elements in Literature.pptx
Unraveling Hypertext_ Analyzing  Postmodern Elements in  Literature.pptxUnraveling Hypertext_ Analyzing  Postmodern Elements in  Literature.pptx
Unraveling Hypertext_ Analyzing Postmodern Elements in Literature.pptx
 
prashanth updated resume 2024 for Teaching Profession
prashanth updated resume 2024 for Teaching Professionprashanth updated resume 2024 for Teaching Profession
prashanth updated resume 2024 for Teaching Profession
 
CARNAVAL COM MAGIA E EUFORIA _
CARNAVAL COM MAGIA E EUFORIA            _CARNAVAL COM MAGIA E EUFORIA            _
CARNAVAL COM MAGIA E EUFORIA _
 
DBMSArchitecture_QueryProcessingandOptimization.pdf
DBMSArchitecture_QueryProcessingandOptimization.pdfDBMSArchitecture_QueryProcessingandOptimization.pdf
DBMSArchitecture_QueryProcessingandOptimization.pdf
 
CLASSIFICATION OF ANTI - CANCER DRUGS.pptx
CLASSIFICATION OF ANTI - CANCER DRUGS.pptxCLASSIFICATION OF ANTI - CANCER DRUGS.pptx
CLASSIFICATION OF ANTI - CANCER DRUGS.pptx
 
Q-Factor General Quiz-7th April 2024, Quiz Club NITW
Q-Factor General Quiz-7th April 2024, Quiz Club NITWQ-Factor General Quiz-7th April 2024, Quiz Club NITW
Q-Factor General Quiz-7th April 2024, Quiz Club NITW
 
An Overview of the Calendar App in Odoo 17 ERP
An Overview of the Calendar App in Odoo 17 ERPAn Overview of the Calendar App in Odoo 17 ERP
An Overview of the Calendar App in Odoo 17 ERP
 
BÀI TẬP BỔ TRỢ TIẾNG ANH 11 THEO ĐƠN VỊ BÀI HỌC - CẢ NĂM - CÓ FILE NGHE (GLOB...
BÀI TẬP BỔ TRỢ TIẾNG ANH 11 THEO ĐƠN VỊ BÀI HỌC - CẢ NĂM - CÓ FILE NGHE (GLOB...BÀI TẬP BỔ TRỢ TIẾNG ANH 11 THEO ĐƠN VỊ BÀI HỌC - CẢ NĂM - CÓ FILE NGHE (GLOB...
BÀI TẬP BỔ TRỢ TIẾNG ANH 11 THEO ĐƠN VỊ BÀI HỌC - CẢ NĂM - CÓ FILE NGHE (GLOB...
 
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITWQ-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
 
Introduction to Research ,Need for research, Need for design of Experiments, ...
Introduction to Research ,Need for research, Need for design of Experiments, ...Introduction to Research ,Need for research, Need for design of Experiments, ...
Introduction to Research ,Need for research, Need for design of Experiments, ...
 
Scientific Writing :Research Discourse
Scientific  Writing :Research  DiscourseScientific  Writing :Research  Discourse
Scientific Writing :Research Discourse
 
How to Uninstall a Module in Odoo 17 Using Command Line
How to Uninstall a Module in Odoo 17 Using Command LineHow to Uninstall a Module in Odoo 17 Using Command Line
How to Uninstall a Module in Odoo 17 Using Command Line
 
Indexing Structures in Database Management system.pdf
Indexing Structures in Database Management system.pdfIndexing Structures in Database Management system.pdf
Indexing Structures in Database Management system.pdf
 
4.9.24 Social Capital and Social Exclusion.pptx
4.9.24 Social Capital and Social Exclusion.pptx4.9.24 Social Capital and Social Exclusion.pptx
4.9.24 Social Capital and Social Exclusion.pptx
 
Mythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITWMythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITW
 

Neural AMR: Sequence-to-Sequence Models for Parsing and Generation

  • 1. Neural AMR: Sequence-to-Sequence Models for Parsing and Generation Ιοannis Konstas joint work with Srinivasan Iyer, Mark Yatskar, Yejin Choi, Luke Zettlemoyer
  • 2. AMR graph text Generate from AMR Attention Encoder Decodergraph text
  • 3. AMR graph text Generate from AMRParse to AMR Attention Encoder Decodergraph text Attention Encoder Decodertext graph
  • 4. AMR graph text Generate from AMRParse to AMR Attention Encoder Decodergraph text Attention Encoder Decodertext graph Paired Training
  • 5. AMR graph text Generate from AMRParse to AMR Attention Encoder Decodergraph text Attention Encoder Decodertext graph SOTA Paired Training
  • 6. Abstract Meaning Representation (Banarescu et al., 2013) know I planet lazy ARG0 ARG1 inhabit man ARG1-of ARG0 mod I have known a planet that was inhabited by a lazy man. ‣ Rooted Directed Acyclic Graph ‣ Nodes: concepts (nouns, verbs, named entities, etc) ‣ Edges: Semantic Role Labels
  • 7. Abstract Meaning Representation (Banarescu et al., 2013) know I planet lazy ARG0 ARG1 inhabit man ARG1-of ARG0 mod I have known a planet that was inhabited by a lazy man. know ‣ Rooted Directed Acyclic Graph ‣ Nodes: concepts (nouns, verbs, named entities, etc) ‣ Edges: Semantic Role Labels
  • 8. Abstract Meaning Representation (Banarescu et al., 2013) know I planet lazy ARG0 ARG1 inhabit man ARG1-of ARG0 mod I have known a planet that was inhabited by a lazy man. know I ‣ Rooted Directed Acyclic Graph ‣ Nodes: concepts (nouns, verbs, named entities, etc) ‣ Edges: Semantic Role Labels
  • 9. Abstract Meaning Representation (Banarescu et al., 2013) know I planet lazy ARG0 ARG1 inhabit man ARG1-of ARG0 mod I have known a planet that was inhabited by a lazy man. know I planet ‣ Rooted Directed Acyclic Graph ‣ Nodes: concepts (nouns, verbs, named entities, etc) ‣ Edges: Semantic Role Labels
  • 10. Abstract Meaning Representation (Banarescu et al., 2013) Input: AMR Graph know I planet lazy ARG0 ARG1 inhabit man ARG1-of ARG0 mod I have known a planet that was inhabited by a lazy man. I knew a planet that was inhabited by a lazy man. I know a planet. It is inhabited by a lazy man. know I planet Generate from AMR ‣ Rooted Directed Acyclic Graph ‣ Nodes: concepts (nouns, verbs, named entities, etc) ‣ Edges: Semantic Role Labels
  • 11. Abstract Meaning Representation (Banarescu et al., 2013) know I planet lazy ARG0 ARG1 inhabit man ARG1-of ARG0 mod I have known a planet that was inhabited by a lazy man. know I planet Parse to AMR Input: Text ‣ Rooted Directed Acyclic Graph ‣ Nodes: concepts (nouns, verbs, named entities, etc) ‣ Edges: Semantic Role Labels
  • 13. Applications ‣ Text Summarization (Liu et al., 2015) sentences:
  • 14. Applications ‣ Text Summarization (Liu et al., 2015) sentences: Parse sentence AMR graphs:
  • 15. Applications ‣ Text Summarization (Liu et al., 2015) summary AMR graph: sentences: Parse sentence AMR graphs:
  • 16. Applications ‣ Text Summarization (Liu et al., 2015) summary AMR graph: sentences: Parse sentence AMR graphs: Generate summary:
  • 17. Applications ‣ Text Summarization (Liu et al., 2015) summary AMR graph: sentences: Parse sentence AMR graphs: The children told that lie Source その うそ は ⼦子供 たち が つい た sono uso-wa kodomo-tachi-ga tsui-ta that lie-TOP child-and others-NOM breathe out-PAST Target Generate summary:
  • 18. Applications ‣ Text Summarization (Liu et al., 2015) summary AMR graph: sentences: Parse sentence AMR graphs: The children told that lie Source その うそ は ⼦子供 たち が つい た sono uso-wa kodomo-tachi-ga tsui-ta that lie-TOP child-and others-NOM breathe out-PAST Target tell child lie that ARG0 ARG1 ARG0-of ‣ Machine Translation (Jones et al., 2012) Generate summary: Parse AMR graph:
  • 19. Applications ‣ Text Summarization (Liu et al., 2015) summary AMR graph: sentences: Parse sentence AMR graphs: The children told that lie Source その うそ は ⼦子供 たち が つい た sono uso-wa kodomo-tachi-ga tsui-ta that lie-TOP child-and others-NOM breathe out-PAST Target Graph-to-graph transformation: tell child lie that ARG0 ARG1 ARG0-of tsuku kodomo tachi sono ARG1 ARG0 ARG0-of ‣ Machine Translation (Jones et al., 2012) Generate summary: Parse AMR graph:
  • 20. Applications ‣ Text Summarization (Liu et al., 2015) summary AMR graph: sentences: Parse sentence AMR graphs: The children told that lie Source その うそ は ⼦子供 たち が つい た sono uso-wa kodomo-tachi-ga tsui-ta that lie-TOP child-and others-NOM breathe out-PAST Target Graph-to-graph transformation: tell child lie that ARG0 ARG1 ARG0-of tsuku kodomo tachi sono ARG1 ARG0 ARG0-of ‣ Machine Translation (Jones et al., 2012) Generate summary: Parse AMR graph: Generate translation:
  • 21. Existing Approaches ‣ MT-based ‣ Flanigan et al. 2016, Pourdamaghani and Knight 2016, Song et al. 2016 ‣ Grammar-based ‣ Lampouras and Vlachos 2017, Mille et al. 2017 Generate from AMR
  • 22. Existing Approaches ‣ MT-based ‣ Flanigan et al. 2016, Pourdamaghani and Knight 2016, Song et al. 2016 ‣ Grammar-based ‣ Lampouras and Vlachos 2017, Mille et al. 2017 Parse to AMR ‣ Alignment-based ‣ Flanigan et al. 2014, 2017 (JAMR) ‣ Grammar-based ‣ Wang et al. 2016 (CAMR), Pust et al. 2015, Artzi et al. 2015, Damonte et al. 2017, Goodman et al. 2016, Puzikov et al. 2016, Brandt et al. 2017, Nguyen et al. 2017 ‣ Neural-based ‣ Barzdins and Gosko 2016, Peng et al. 2017, Noord and Bos 2017, Buys and Blunsom 2017 Generate from AMR
  • 23. Overview ‣ Sequence-to-sequence architecture ‣ End-to-end model w/o intermediate representations ‣ Linearisation of AMR graph to string ‣ Pre-process ‣ Paired Training ‣ Scalable data augmentation algorithm
  • 25. Sequence-to-sequence model Encoderinput know ARG0 I ARG1 ( h1 (s) h2 (s) h3 (s) h4 (s) h5 (s) [ ] [ ] [ ] [ ] [ ]h1 (s) h2 (s) h3 (s) h4 (s) h5 (s)
  • 26. Sequence-to-sequence model Encoder Decoderinput output I The A … know knew planet … a planet man … … inhabit inhabited was … ˆw = argmax w Y i p wi|w<i, h(s) know ARG0 I ARG1 ( h1 (s) h2 (s) h3 (s) h4 (s) h5 (s) [ ] [ ] [ ] [ ] [ ]h1 (s) h2 (s) h3 (s) h4 (s) h5 (s)
  • 27. Sequence-to-sequence model Attention Encoder Decoderinput output know ARG0 I ARG1 ( planet ARG1-of inhabit <s> I know the planet of I The A … know knew planet … a planet man … … inhabit inhabited was … ˆw = argmax w Y i p wi|w<i, h(s) know ARG0 I ARG1 ( h1 (s) h2 (s) h3 (s) h4 (s) h5 (s) [ ] [ ] [ ] [ ] [ ]h1 (s) h2 (s) h3 (s) h4 (s) h5 (s)
  • 28. Linearization Graph —> Depth First Search (Human-authored annotation) hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York .
  • 29. Linearization Graph —> Depth First Search (Human-authored annotation) hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York .
  • 30. Linearization Graph —> Depth First Search (Human-authored annotation) hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York . person
  • 31. Linearization Graph —> Depth First Search (Human-authored annotation) hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York . person
  • 32. Linearization Graph —> Depth First Search (Human-authored annotation) hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York . person have-role country “United States” official
  • 33. Linearization Graph —> Depth First Search (Human-authored annotation) hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York . person have-role country “United States” official
  • 34. Pre-processing Linearization —> Anonymization hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York .
  • 35. Pre-processing Linearization —> Anonymization hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York . country “United States”
  • 36. Pre-processing Linearization —> Anonymization hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York . country “United States” 2002 1
  • 37. Pre-processing Linearization —> Anonymization hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York . country “United States” 2002 1 “New York” city
  • 38. Pre-processing Linearization —> Anonymization hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 hold person meet group ARG0 ARG1 person expert ARG1-of have-role country “United States” official date-entity city “New York”2002 1 time location name ARG1 name ARG2-of ARG0-of ARG2 year month ARG0 US officials held an expert group meeting in January 2002 in New York . country “United States” 2002 1 “New York” city loc_0 officials held an expert group meeting in month_0 year_0 in loc_1 .
  • 39. Experimental Setup AMR LDC2015E86 (SemEval-2016 Task 8) ‣ Hand annotated MR graphs: newswire, forums ‣ ~16k training / 1k development / 1k test pairs Train ‣ Optimize cross-entropy loss Evaluation ‣ BLEU n-gram precision (Generation) (Papineni et al., 2002) ‣ SMATCH score (Parsing) (Cai and Knight, 2013)
  • 40. Experiments ‣ Vanilla experiment ‣ Limited Language Model Capacity ‣ Paired Training ‣ Data augmentation algorithm
  • 41. First Attempt (Generation) TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 42. First Attempt (Generation)BLEU 0 5.8 11.6 17.4 23.2 29 TreeToStr TSP PBMT NeuralAMR 26.9 22.423 TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 43. First Attempt (Generation)BLEU 0 5.8 11.6 17.4 23.2 29 TreeToStr TSP PBMT NeuralAMR 22 26.9 22.423 TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 44. First Attempt (Generation)BLEU 0 5.8 11.6 17.4 23.2 29 TreeToStr TSP PBMT NeuralAMR 22 26.9 22.423 TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016 All systems use a Language Model trained on a very large corpus. We will emulate via data augmentation. (Sennrich et al., ACL 2016)
  • 45. What went wrong? hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 US officials held an expert group meeting in January 2002 in New York . United States officials held held a meeting in January 2002 . Reference Prediction 44.26% 74.85%
  • 46. What went wrong? hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 US officials held an expert group meeting in January 2002 in New York . United States officials held held a meeting in January 2002 . Reference Prediction ‣ Repetition 44.26% 74.85%
  • 47. What went wrong? hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 US officials held an expert group meeting in January 2002 in New York . United States officials held held a meeting in January 2002 . Reference Prediction ‣ Repetition ‣ Coverage 44.26% 74.85%
  • 48. What went wrong? hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 US officials held an expert group meeting in January 2002 in New York . United States officials held held a meeting in January 2002 . Reference Prediction ‣ Repetition ‣ Coverage a) Sparsity Tokens 0 4500 9000 13500 18000 Total OOV@1 OOV@5 44.26% 74.85%
  • 49. What went wrong? hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 US officials held an expert group meeting in January 2002 in New York . United States officials held held a meeting in January 2002 . Reference Prediction ‣ Repetition ‣ Coverage a) Sparsity b) Avg sent length: 20 words c) Limited Language Modeling capacity Tokens 0 4500 9000 13500 18000 Total OOV@1 OOV@5 44.26% 74.85%
  • 50. Data Augmentation Original Dataset: ~16k graph-sentence pairs
  • 51. Data Augmentation Original Dataset: ~16k graph-sentence pairs Gigaword: ~183M sentences *only*
  • 52. Data Augmentation Original Dataset: ~16k graph-sentence pairs Gigaword: ~183M sentences *only* Sample sentences with vocabulary overlap % 0 20 40 60 80 OOV@1 OOV@5 Original Giga-200k Giga-2M Giga-20M
  • 53. Data Augmentation graph text Generate from AMR Attention Encoder Decodergraph text
  • 54. Data Augmentation graph text Generate from AMRParse to AMR Attention Encoder Decodergraph text Attention Encoder Decodertext graph
  • 55. Data Augmentation graph text Generate from AMRParse to AMR Attention Encoder Decodergraph text Attention Encoder Decodertext graph
  • 56. Data Augmentation graph text Generate from AMRParse to AMR Attention Encoder Decodergraph text Attention Encoder Decodertext graph Re-train
  • 57. Semi-supervised Learning ‣Self-training ‣ McClosky et al. 2006 ‣Co-training ‣ Yarowski 1995, Blum and Mitchell 1998, Sarkar 2001 ‣ Sogaard and Rishoj, 2010
  • 59. Paired Training Train AMR Parser P on Original Dataset ( , )
  • 60. Paired Training for i = 0 … N Train AMR Parser P on Original Dataset ( , )
  • 61. Paired Training for i = 0 … N Si =Sample k 10i sentences from Gigaword Train AMR Parser P on Original Dataset ( , )
  • 62. Paired Training for i = 0 … N Parse Si sentences with P Si =Sample k 10i sentences from Gigaword Train AMR Parser P on Original Dataset ( , )
  • 63. Paired Training Self-trainParser for i = 0 … N Parse Si sentences with P Si =Sample k 10i sentences from Gigaword Re-train AMR Parser P on Si Train AMR Parser P on Original Dataset ( , )
  • 64. Paired Training Self-trainParser for i = 0 … N Parse Si sentences with P Si =Sample k 10i sentences from Gigaword Re-train AMR Parser P on Si Train AMR Parser P on Original Dataset ( , )
  • 65. Paired Training Self-trainParser for i = 0 … N Parse Si sentences with P Si =Sample k 10i sentences from Gigaword Re-train AMR Parser P on Si Train AMR Parser P on Original Dataset ( , ) Train Generator G on SN ( , )
  • 66. Training AMR Parser Train P on Original Dataset
  • 67. Training AMR Parser Train P on Original Dataset
  • 68. Training AMR Parser Sample S1=200k sentences from Gigaword Train P on Original Dataset 200k
  • 69. Training AMR Parser Sample S1=200k sentences from Gigaword Parse S1 with P Train P on Original Dataset 200k 200k ( , )
  • 70. Training AMR Parser Sample S1=200k sentences from Gigaword Train P on S1=200k Parse S1 with P Train P on Original Dataset 200k 200k 200k ( , )
  • 71. Training AMR Parser Sample S1=200k sentences from Gigaword Train P on S1=200k Fine-tune P on Original Dataset Parse S1 with P Train P on Original Dataset 200k 200k Fine-tune: init parameters from previous step and train on Original Dataset 200k ( , )
  • 72. Training AMR Parser Train P on S2=2M Fine-tune P on Original Dataset Sample S2=2M sentences from Gigaword Parse S2 with P 200k 200k 200k Fine-tune: init parameters from previous step and train on Original Dataset ( , )
  • 73. Training AMR Parser Train P on S2=2M Fine-tune P on Original Dataset Sample S2=2M sentences from Gigaword Parse S2 with P 00k 2M 2M 2M Fine-tune: init parameters from previous step and train on Original Dataset ( , )
  • 74. Training AMR Parser Train P on S3=20M Fine-tune P on Original Dataset Sample S3=20M sentences from Gigaword Parse S3 with P 2M 2M 2M Fine-tune: init parameters from previous step and train on Original Dataset ( , )
  • 75. Training AMR Parser Train P on S3=20M Fine-tune P on Original Dataset Sample S3=20M sentences from Gigaword Parse S3 with P 2M Fine-tune: init parameters from previous step and train on Original Dataset ( , ) 20M 20M 20M
  • 76. Training AMR Parser Train P on S3=20M Fine-tune P on Original Dataset Sample S3=20M sentences from Gigaword Parse S3 with P 2M Fine-tune: init parameters from previous step and train on Original Dataset ( , ) 20M 20M 20M
  • 77. Training AMR Generator Train G on S4=20M Fine-tune G on Original Dataset Sample S4=20M sentences from Gigaword Parse S4 with P Fine-tune: init parameters from previous step and train on Original Dataset ( , ) 2M 20M 20M 20M
  • 78. Training AMR Generator Train G on S4=20M Fine-tune G on Original Dataset Sample S4=20M sentences from Gigaword Parse S4 with P Fine-tune: init parameters from previous step and train on Original Dataset ( , ) 20M 20M 20M G G 20M
  • 79. Training AMR Generator Train G on S4=20M Fine-tune G on Original Dataset Sample S4=20M sentences from Gigaword Parse S4 with P Fine-tune: init parameters from previous step and train on Original Dataset ( , ) 20M 20M 20M G G 20M
  • 80. Final Results (Generation) TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 81. Final Results (Generation)BLEU 0 7 14 21 28 35 TreeToStr TSP PBMT NeuralAMR NeuralAMR-200k NeuralAMR-2M NeuralAMR-20M 22 26.9 22.423 TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 82. Final Results (Generation)BLEU 0 7 14 21 28 35 TreeToStr TSP PBMT NeuralAMR NeuralAMR-200k NeuralAMR-2M NeuralAMR-20M 27.4 22 26.9 22.423 TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 83. Final Results (Generation)BLEU 0 7 14 21 28 35 TreeToStr TSP PBMT NeuralAMR NeuralAMR-200k NeuralAMR-2M NeuralAMR-20M 32.3 27.4 22 26.9 22.423 TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 84. Final Results (Generation)BLEU 0 7 14 21 28 35 TreeToStr TSP PBMT NeuralAMR NeuralAMR-200k NeuralAMR-2M NeuralAMR-20M 33.8 32.3 27.4 22 26.9 22.423 TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 85. Final Results (Generation)BLEU 0 7 14 21 28 35 TreeToStr TSP PBMT NeuralAMR NeuralAMR-200k NeuralAMR-2M NeuralAMR-20M 33.8 32.3 27.4 22 26.9 22.423 TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 86. Final Results (Generation)BLEU 0 7 14 21 28 35 TreeToStr TSP PBMT NeuralAMR NeuralAMR-200k NeuralAMR-2M NeuralAMR-20M 33.8 32.3 27.4 22 26.9 22.423 TreeToStr: Flanigan et al, NAACL 2016 TSP: Song et al, EMNLP 2016 PBMT: Pourdamaghani and Knight, INLG 2016
  • 87. Final Results (Parsing) SBMT: Pust et al, 2015 CharLSTM+CAMR: Noord and Bos, 2017 Seq2Seq: Peng et al., 2017
  • 88. Final Results (Parsing)SMATCH 0 14 28 42 56 70 SBMT CharLSTM+CAMR Seq2Seq NeuralAMR-20M 62.1 52 67.367.1 SBMT: Pust et al, 2015 CharLSTM+CAMR: Noord and Bos, 2017 Seq2Seq: Peng et al., 2017
  • 89. Final Results (Parsing)SMATCH 0 14 28 42 56 70 SBMT CharLSTM+CAMR Seq2Seq NeuralAMR-20M 62.1 52 67.367.1 SBMT: Pust et al, 2015 CharLSTM+CAMR: Noord and Bos, 2017 Seq2Seq: Peng et al., 2017
  • 90. Final Results (Parsing)SMATCH 0 14 28 42 56 70 SBMT CharLSTM+CAMR Seq2Seq NeuralAMR-20M 62.1 52 67.367.1 SBMT: Pust et al, 2015 CharLSTM+CAMR: Noord and Bos, 2017 Seq2Seq: Peng et al., 2017
  • 91. Final Results (Parsing)SMATCH 0 14 28 42 56 70 SBMT CharLSTM+CAMR Seq2Seq NeuralAMR-20M 62.1 52 67.367.1 SBMT: Pust et al, 2015 CharLSTM+CAMR: Noord and Bos, 2017 Seq2Seq: Peng et al., 2017
  • 92. How did we do? (Generation) hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 US officials held an expert group meeting in January 2002 in New York . In January 2002 United States officials held a meeting of the group experts in New York . Reference Prediction 44.26% 74.85% Errors: Disfluency Coverage
  • 93. How did we do? (Generation) hold :ARG0 (person :ARG0-of (have-role :ARG1 loc_0 :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity year_0 month_0) :location loc_1 US officials held an expert group meeting in January 2002 in New York . In January 2002 United States officials held a meeting of the group experts in New York . Reference Prediction 44.26% 74.85% The report stated British government must help to stabilize weak states and push for international regulations that would stop terrorists using freely available information to create and unleash new forms of biological warfare such as a modified version of the influenza virus. Reference The report stated that the Britain government must help stabilize the weak states and push international regulations to stop the use of freely available information to create a form of new biological warfare such as the modified version of the influenza . Prediction Errors: Disfluency Coverage
  • 94. Summary ‣ Sequence-to-sequence models for Parsing and Generation ‣ Paired Training: scalable data augmentation algorithm ‣ Achieve state-of-the-art performance on generating from AMR ‣ Best-performing Neural AMR Parser ‣ Demo, Code and Pre-trained Models: http://ikonstas.net
  • 95. Summary Thank You ‣ Sequence-to-sequence models for Parsing and Generation ‣ Paired Training: scalable data augmentation algorithm ‣ Achieve state-of-the-art performance on generating from AMR ‣ Best-performing Neural AMR Parser ‣ Demo, Code and Pre-trained Models: http://ikonstas.net thank-01 you ARG1
  • 97. Encoding Linearize —> RNN encoding hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York
  • 98. Encoding Linearize —> RNN encoding hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York
  • 99. Encoding Linearize —> RNN encoding hold ARG0 ( person ARG0-of hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York - Token embeddings
  • 100. Encoding Linearize —> RNN encoding hold ARG0 ( person ARG0-of h1 (s) h2 (s) h3 (s) h4 (s) h5 (s) hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York - Token embeddings - Recurrent Neural Network (RNN)
  • 101. Encoding Linearize —> RNN encoding hold ARG0 ( person ARG0-of h1 (s) h2 (s) h3 (s) h4 (s) h5 (s) hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York - Token embeddings - Recurrent Neural Network (RNN) - Bi-directional RNN
  • 102. Encoding Linearize —> RNN encoding hold ARG0 ( person ARG0-of h1 (s) h2 (s) h3 (s) h4 (s) h5 (s) [ ] [ ] [ ] [ ] [ ]h1 (s) h2 (s) h3 (s) h4 (s) h5 (s) hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York - Token embeddings - Recurrent Neural Network (RNN) - Bi-directional RNN
  • 103. Encoding Linearize —> RNN encoding hold ARG0 ( person ARG0-of h1 (s) h2 (s) h3 (s) h4 (s) h5 (s) [ ] [ ] [ ] [ ] [ ]h1 (s) h2 (s) h3 (s) h4 (s) h5 (s) hold :ARG0 (person :ARG0-of (have-role :ARG1 United_States :ARG2 official) ) :ARG1 (meet :ARG0 (person :ARG1-of expert :ARG2-of group) ) :time (date-entity 2002 1) :location New_York - Token embeddings - Recurrent Neural Network (RNN) - Bi-directional RNN
  • 104. Decoding h1hN (s) RNN Encoding —> RNN Decoding (Beam search)
  • 105. Decoding h1hN (s) ; RNN Encoding —> RNN Decoding (Beam search) - init h(s)
  • 106. Decoding h1hN (s) ; Holding Held US … RNN Encoding —> RNN Decoding (Beam search) - softmax - init h(s)
  • 108. Decoding h1hN (s) ; Holding Held US … h2 a the meeting … h3 US person expert … … … w11: Holding Heldsw12: Holdw13: USw14: … Hold aw21: Hold thew22: Held aw23: Held thew24: RNN Encoding —> RNN Decoding (Beam search) - softmax - p wi|w<i, h(s) - init h(s)
  • 109. Decoding h1hN (s) ; Holding Held US … h2 a the meeting … h3 US person expert … hk… … w11: Holding Heldsw12: Holdw13: USw14: … Hold aw21: Hold thew22: Held aw23: Held thew24: wk1: The US officials held wk2: US officials held a wk3: US officials hold the wk4: US officials will hold a … meeting meetings meet … RNN Encoding —> RNN Decoding (Beam search) - softmax - p wi|w<i, h(s) - init h(s)
  • 113. Attention hold ARG0 ( person role US official ) ARG1 ( meet expert group ) US officials held an expert group meeting in January 2002 h3 a the meeting … w2: held hold ARG0 ( person ARG0-of [][][][][]h1 (s)h2 (s)h3 (s)h4 (s)h5 (s) c3 [ ] ci = X i aijh (s) j ai = soft max fi h(s) , hi
  • 114. Attention hold ARG0 ( person role US official ) ARG1 ( meet expert group ) US officials held an expert group meeting in January 2002 h3 a the meeting … w2: held hold ARG0 ( person ARG0-of [][][][][]h1 (s)h2 (s)h3 (s)h4 (s)h5 (s) c3 [ ] ci = X i aijh (s) j ai = soft max fi h(s) , hi
  • 115. Attention hold ARG0 ( person role US official ) ARG1 ( meet expert group ) US officials held an expert group meeting in January 2002 h3 a the meeting … w2: held hold ARG0 ( person ARG0-of [][][][][]h1 (s)h2 (s)h3 (s)h4 (s)h5 (s) c3 [ ] ci = X i aijh (s) j ai = soft max fi h(s) , hi
  • 116. Attention hold ARG0 ( person role US official ) ARG1 ( meet expert group ) US officials held an expert group meeting in January 2002 h3 a the meeting … w2: held hold ARG0 ( person ARG0-of [][][][][]h1 (s)h2 (s)h3 (s)h4 (s)h5 (s) c3 [ ] ci = X i aijh (s) j ai = soft max fi h(s) , hi