MRP 2020: Cross-Framework and Cross-Lingual Meaning Representation Parsing - Daniel Hershcovich

Page created by Ray Leonard
 
CONTINUE READING
MRP 2020:
Cross-Framework and Cross-Lingual
 Meaning Representation Parsing
                             http://mrp.nlpl.eu

 Stephan Oepen♣ , Omri Abend♠ , Lasha Abzianidze♥ , Johan Bos♦ ,
    Jan Hajič◦ , Daniel Hershcovich? , Bin Li• , Tim O’Gorman ,
                 Nianwen Xue∗ , and Daniel Zeman◦
                       ♣ University of Oslo, Department of Informatics
     ♠ The Hebrew University of Jerusalem, School of Computer Science and Engineering
                    ♥ Utrecht University, Logic, Language, and Information
                 ♦ University of Groningen, Center for Language and Cognition
           ◦ Charles University, Prague, Institute of Formal and Applied Linguistics
                 ? University of Copenhagen, Department of Computer Science
           • Nanjing Normal University, School of Chinese Language and Literature
    University of Massachusetts at Amherst, College of Information and Computer Sciences
                    ∗ Brandeis University, Department of Computer Science

                         mrp-organizers@nlpl.eu
10,000-Meter Perspective: Parsing into Semantic Graphs

  A similar technique is almost impossible to apply to other crops.

                                                                      2
10,000-Meter Perspective: Parsing into Semantic Graphs

                                         almost

                                          domain

                                 possible-01
                                 polarity -

                                      ARG1

           resemble-01            apply-02            other

                     ARG1        ARG1        ARG2     domain

                     technique                 crop

                                                               2
10,000-Meter Perspective: Parsing into Semantic Graphs

                                        A   F   D     F   P         A        U

                              ⟨20:22⟩           ⟨41:43⟩       ⟨44:49⟩                 ⟨65:66⟩

                F   E                   E   C                           R     E        C

    ⟨0:1⟩               C     ⟨23:29⟩       ⟨30:40⟩       ⟨50:52⟩           ⟨53:58⟩             ⟨59:65⟩

            S       A

    ⟨2:9⟩           ⟨10:19⟩

                                                                                                          2
10,000-Meter Perspective: Parsing into Semantic Graphs

                                                         PRED

                                                        be
                                                      ⟨20:22⟩
                                              sentmod          enunc
                                              sempos               v
                                              frame   en-v#ev-w218f2

                                                   ACT            PAT

                                     apply
                                                                     possible
                                ⟨41:43⟩ ⟨44:49⟩
                                                                     ⟨30:40⟩
                           sempos               v
                                                                 sempos adj.denot
                           frame en-v#ev-w119f2

                          ADDR          PAT                               BEN         EXT

             crop                   technique                                                 almost
                                                                   #Benef
        ⟨50:52⟩ ⟨59:64⟩            ⟨0:1⟩ ⟨10:19⟩          ACT                                 ⟨23:29⟩
                                                                  sempos x
       sempos n.denot            sempos n.denot                                      sempos adv.denot.grad.neg

              RSTR                      RSTR                            coref.gram

           other                     similar
                                                                  #Gen
          ⟨53:58⟩                     ⟨2:9⟩
                                                                sempos x
      sempos adj.denot           sempos adj.denot

                                                                                                                 2
10,000-Meter Perspective: Parsing into Semantic Graphs

                                                   in                                          ATTRIBUTION        PRESUPPOSITION

                                          impossible.a.01

                    in          in                                              in                                in

                         Time          Topic                        Degree               in                  apply.v.01

                                                        Theme                                                                      in

     in           time.n.08     proposition.n.01                       almost.r.01            Theme              Goal                   in

                    EQU                                                      technique.n.01                            crop.n.01

                   "now"                                Attribute                                                         NEQ

   similar.a.01                                                                                                                 crop.n.01

                                                                                                                                             2
10,000-Meter Perspective: Parsing into Semantic Graphs

                                           _almost_a_1
                                              ⟨23:29⟩

                                                 ARG1

                                         _impossible_a_for
          comp
                                             ⟨30:40⟩
          ⟨2:9⟩
                                         TENSE       pres

             ARG1                                ARG1

       _similar_a_to          _a_q         _apply_v_to               udef_q      _other_a_1
           ⟨2:9⟩              ⟨0:1⟩          ⟨44:49⟩                ⟨53:100⟩       ⟨53:58⟩

                       ARG1     BV        ARG2               ARG3      BV      ARG1

                        _technique_n_1                          _crop_n_1
                            ⟨10:19⟩                              ⟨59:65⟩
                        NUM         sg                          NUM pl

                                                                                              2
Why Graph-Based Meaning Representation?
   I saw Joe’s dog, which was running in the garden.
              The dog was chasing a cat.

                                                       3
Why Graph-Based Meaning Representation?
   I saw Joe’s dog, which was running in the garden.
              The dog was chasing a cat.

                           semantic parsing

            see-01
           ARG0    ARG1

       I                  dog                        chase-01
                    poss        ARG0                ARG0   ARG1

                  joe            run-02       dog                 cat
                                   location

                                 garden

                                                                        3
Why Graph-Based Meaning Representation?
   I saw Joe’s dog, which was running in the garden.
              The dog was chasing a cat.

                             semantic parsing

            see-01
           ARG0    ARG1

       I                  dog                                            chase-01
                    poss        ARG0                                 ARG0      ARG1

                  joe              run-02                     dog                     cat
                                   location

                                   garden

                                          merge

                             see-01                  chase-01
                            ARG0       ARG1         ARG0          ARG1

                        I                     dog                        cat
                                        poss        ARG0

                                   joe               run-02
                                                       location

                                                     garden

                                                                                            3
Why Graph-Based Meaning Representation?
   I saw Joe’s dog, which was running in the garden.
              The dog was chasing a cat.

                             semantic parsing

            see-01
           ARG0    ARG1

       I                  dog                                            chase-01
                    poss        ARG0                                 ARG0      ARG1

                  joe              run-02                     dog                     cat
                                   location

                                   garden

                                          merge

                             see-01                  chase-01
                                                                                                            chase-01
                            ARG0       ARG1         ARG0          ARG1

                                                                                                         ARG0    location   ARG1
                        I                     dog                        cat
                                                                                                                                   cat
                                        poss        ARG0
                                                                                      summarize   dog           garden
                                   joe               run-02                                       poss

                                                       location
                                                                                                  joe
                                                     garden

                                                                                                                                         3
Why Graph-Based Meaning Representation?
   I saw Joe’s dog, which was running in the garden.
              The dog was chasing a cat.

                             semantic parsing

            see-01
           ARG0    ARG1

       I                  dog                                            chase-01
                    poss        ARG0                                 ARG0      ARG1

                  joe              run-02                     dog                     cat
                                   location

                                   garden
                                                                                            Joe’s dog was chasing a cat in the garden.

                                          merge                                                       surface realisation
                             see-01                  chase-01
                                                                                                                 chase-01
                            ARG0       ARG1         ARG0          ARG1

                                                                                                              ARG0    location   ARG1
                        I                     dog                        cat
                                                                                                                                        cat
                                        poss        ARG0
                                                                                      summarize        dog           garden
                                   joe               run-02                                            poss

                                                       location
                                                                                                       joe
                                                     garden

                                                                                                                                              3
Why Graph-Based Meaning Representation?
    I saw Joe’s dog, which was running in the garden.
               The dog was chasing a cat.

                              semantic parsing

             see-01
            ARG0    ARG1

        I                  dog                                            chase-01
                     poss        ARG0                                 ARG0      ARG1

                   joe              run-02                     dog                     cat
                                    location

                                    garden
                                                                                             Joe’s dog was chasing a cat in the garden.

                                           merge                                                       surface realisation
                              see-01                  chase-01
                                                                                                                  chase-01
                             ARG0       ARG1         ARG0          ARG1

                                                                                                               ARG0    location   ARG1
                         I                     dog                        cat
                                                                                                                                         cat
                                         poss        ARG0
                                                                                       summarize        dog           garden
                                    joe               run-02                                            poss

                                                        location
                                                                                                        joe
                                                      garden

Hardy & Vlachos (2018): 2+ ROUGE points over strong encoder–decoder.
                                                                                                                                               3
Syntactic Trees vs. Semantic Graphs
                       root
                                             p
      nmod                    prd                           pmod
        nmod     sbj           amod   amod   im   adv         nmod

 A similar technique is almost impossible to apply      to other crops .

                                                                           4
Syntactic Trees vs. Semantic Graphs
                            root
                                                            p
      nmod                          prd                                      pmod
         nmod         sbj             amod          amod    im     adv         nmod

 A similar technique is almost impossible to apply                       to other crops .
                                    top

         BV                               ARG2                               ARG3
               ARG1                          ARG1          ARG1                     ARG1

 A   similar     technique         almost      impossible         apply     other      crops

                                                                                               4
Syntactic Trees vs. Semantic Graphs
                            root
                                                            p
      nmod                          prd                                      pmod
         nmod         sbj             amod          amod    im     adv         nmod

 A similar technique is almost impossible to apply                       to other crops .
                                    top

         BV                               ARG2                               ARG3
               ARG1                          ARG1          ARG1                     ARG1

 A   similar     technique         almost      impossible         apply     other      crops
        ∃x : technique’(x) ∧ similar’(x); ∃y : crop’(y) ∧ other’(y)
                      → almost’(¬possible’(apply’(e, x, y)))

                                                                                               4
Syntactic Trees vs. Semantic Graphs
                            root
                                                            p
      nmod                          prd                                      pmod
         nmod         sbj             amod          amod    im     adv         nmod

 A similar technique is almost impossible to apply                       to other crops .
                                    top

         BV                               ARG2                               ARG3
               ARG1                          ARG1          ARG1                     ARG1

 A   similar     technique         almost      impossible         apply     other      crops
        ∃x : technique’(x) ∧ similar’(x); ∃y : crop’(y) ∧ other’(y)
                      → almost’(¬possible’(apply’(e, x, y)))

                                                                                               4
Syntactic Trees vs. Semantic Graphs
                            root
                                                            p
      nmod                          prd                                      pmod
         nmod         sbj             amod          amod    im     adv         nmod

 A similar technique is almost impossible to apply                       to other crops .
                                    top

         BV                               ARG2                               ARG3
               ARG1                          ARG1          ARG1                     ARG1

 A   similar     technique         almost      impossible         apply     other      crops
        ∃x : technique’(x) ∧ similar’(x); ∃y : crop’(y) ∧ other’(y)
                      → almost’(¬possible’(apply’(e, x, y)))

                                                                                               4
Syntactic Trees vs. Semantic Graphs
                            root
                                                            p
      nmod                          prd                                      pmod
         nmod         sbj             amod          amod    im     adv         nmod

 A similar technique is almost impossible to apply                       to other crops .
                                    top

         BV                               ARG2                               ARG3
               ARG1                          ARG1          ARG1                     ARG1

 A   similar     technique         almost      impossible         apply     other      crops
        ∃x : technique’(x) ∧ similar’(x); ∃y : crop’(y) ∧ other’(y)
                      → almost’(¬possible’(apply’(e, x, y)))

                                                                                               4
Syntactic Trees vs. Semantic Graphs
                             root
                                                             p
       nmod                          prd                                      pmod
          nmod         sbj             amod          amod    im     adv         nmod

  A similar technique is almost impossible to apply                       to other crops .
                                     top

          BV                               ARG2                               ARG3
                ARG1                          ARG1          ARG1                     ARG1

 A    similar     technique         almost      impossible         apply     other      crops
         ∃x : technique’(x) ∧ similar’(x); ∃y : crop’(y) ∧ other’(y)
                       → almost’(¬possible’(apply’(e, x, y)))

Different Desiderata and Levels of Abstraction
I Grammaticality (e.g. subject–verb agreement) vs. relational structure.
                                                                                                4
Semi-Formally: Trees vs. Graphs

Structural Wellformedness Conditions on Trees
I Unique root, connected, single parent, free of cycles; maybe projective;
→ all nodes (but the root) reachable by unique directed path from root.

                                                                             5
Semi-Formally: Trees vs. Graphs

Structural Wellformedness Conditions on Trees
I Unique root, connected, single parent, free of cycles; maybe projective;
→ all nodes (but the root) reachable by unique directed path from root.

                            top
         BV                       ARG2                    ARG3
          ARG1                    ARG1     ARG1                  ARG1

  A similar technique is almost impossible to apply to other crops .

                                                                             5
Semi-Formally: Trees vs. Graphs

Structural Wellformedness Conditions on Trees
I Unique root, connected, single parent, free of cycles; maybe projective;
→ all nodes (but the root) reachable by unique directed path from root.

                            top
         BV                       ARG2                    ARG3
          ARG1                    ARG1     ARG1                  ARG1

  A similar technique is almost impossible to apply to other crops .

Beyond Trees: General Graphs
I Argument sharing: nodes with multiple incoming edges (in-degree > 1);

                                                                             5
Semi-Formally: Trees vs. Graphs

Structural Wellformedness Conditions on Trees
I Unique root, connected, single parent, free of cycles; maybe projective;
→ all nodes (but the root) reachable by unique directed path from root.

                            top
         BV                       ARG2                    ARG3
          ARG1                    ARG1     ARG1                  ARG1

  A similar technique is almost impossible to apply to other crops .

Beyond Trees: General Graphs
I Argument sharing: nodes with multiple incoming edges (in-degree > 1);
I some surface tokens do not contribute (as nodes; many function words);

                                                                             5
Semi-Formally: Trees vs. Graphs

Structural Wellformedness Conditions on Trees
I Unique root, connected, single parent, free of cycles; maybe projective;
→ all nodes (but the root) reachable by unique directed path from root.

                            top
         BV                       ARG2                    ARG3
          ARG1                    ARG1     ARG1                  ARG1

  A similar technique is almost impossible to apply to other crops .

Beyond Trees: General Graphs
I Argument sharing: nodes with multiple incoming edges (in-degree > 1);
I some surface tokens do not contribute (as nodes; many function words);
I (structurally) multi-rooted: more than one node with zero in-degree;

                                                                             5
Semi-Formally: Trees vs. Graphs

Structural Wellformedness Conditions on Trees
I Unique root, connected, single parent, free of cycles; maybe projective;
→ all nodes (but the root) reachable by unique directed path from root.

                            top
         BV                       ARG2                    ARG3
          ARG1                    ARG1     ARG1                  ARG1

  A similar technique is almost impossible to apply to other crops .

Beyond Trees: General Graphs
I Argument sharing: nodes with multiple incoming edges (in-degree > 1);
I some surface tokens do not contribute (as nodes; many function words);
I (structurally) multi-rooted: more than one node with zero in-degree;
→ massive growth in modeling and algorithmic complexity (NP-complete).
                                                                             5
High-Level Goals of the Shared Task
Cross-Framework Comparability and Interoperability
I Vast, complex landscape of representing natural language meaning;
I diverse linguistic traditions, modeling assumptions, levels of ambition;

→ clarify concepts and terminology; unify representations and evaluation.

                                                                             6
High-Level Goals of the Shared Task
Cross-Framework Comparability and Interoperability
I Vast, complex landscape of representing natural language meaning;
I diverse linguistic traditions, modeling assumptions, levels of ambition;

→ clarify concepts and terminology; unify representations and evaluation.

Parsing into Graph-Structured Representations
I Cottage industry of parsers with output structures beyond rooted trees;
I distinct techniques, e.g. based on transitions, composition, ‘translation’;

→ evaluate across frameworks; learning from complementary knowledge.

                                                                                6
High-Level Goals of the Shared Task
Cross-Framework Comparability and Interoperability
I Vast, complex landscape of representing natural language meaning;
I diverse linguistic traditions, modeling assumptions, levels of ambition;

→ clarify concepts and terminology; unify representations and evaluation.

Parsing into Graph-Structured Representations
I Cottage industry of parsers with output structures beyond rooted trees;
I distinct techniques, e.g. based on transitions, composition, ‘translation’;

→ evaluate across frameworks; learning from complementary knowledge.

Two Distinct Tracks in MRP 2020
I Cross-Framework Perspective: Seek commonality and complementarity.
I Cross-Lingual Perspective: In-framework transfer to another language.
                                                                                6
Graph Structure vs. Node (or Edge) Decorations
Zero-Arity Predicates vs. Constants
I Nodes and edges can be labeled (e.g. by relation and role identifiers);

                                                                            7
Graph Structure vs. Node (or Edge) Decorations
Zero-Arity Predicates vs. Constants
I Nodes and edges can be labeled (e.g. by relation and role identifiers);
I labels can be internally structured: node properties and edge attributes;
I properties (and attributes) are non-recursive attribute–value matrices;
I node (and edge) label is merely a distinguished property (or attribute);

                                                                              7
Graph Structure vs. Node (or Edge) Decorations
Zero-Arity Predicates vs. Constants
I Nodes and edges can be labeled (e.g. by relation and role identifiers);
I labels can be internally structured: node properties and edge attributes;
I properties (and attributes) are non-recursive attribute–value matrices;
I node (and edge) label is merely a distinguished property (or attribute);
I distinction is not commonly discussed, but used by many frameworks.

                 person
                                                          person
            name          age
                                                        name       age

       name               temporal-quantity
    OP1 Pierre                                     name            temporal-quantity
    OP2 Vinken                QUANT 61
                                                  OP1   OP2
                                                                     QUANT    unit
                                 unit
                                              Pierre    Vinken     61                year
                                year

                            Pierre Vinken is 61 years old.
                                                                                            7
Anchoring in the Surface String
Relating Pieces of Meaning to the Linguistic Signal
I Intuitively, sub-structures of meaning relate to sub-parts of the input;
I semantic frameworks vary in how much weight to put on this relation;

                                                                             8
Anchoring in the Surface String
Relating Pieces of Meaning to the Linguistic Signal
I Intuitively, sub-structures of meaning relate to sub-parts of the input;
I semantic frameworks vary in how much weight to put on this relation;
I anchoring of graph elements in sub-strings of the underlying utterance;
I can be part of semantic annotations or not; can take different forms;

                                                                             8
Anchoring in the Surface String
Relating Pieces of Meaning to the Linguistic Signal
I Intuitively, sub-structures of meaning relate to sub-parts of the input;
I semantic frameworks vary in how much weight to put on this relation;
I anchoring of graph elements in sub-strings of the underlying utterance;
I can be part of semantic annotations or not; can take different forms;
I hierarchy of anchoring types: Flavor (0)–(2); bilexical graphs strictest;

       Name           Example                 Type of Anchoring
(0) bilexical   DM,
                hh
                ( (h(PSD
                      h(
                      (
                       h     nodes are sub-set of surface tokens
(1) anchored EDS, PTG, UCCA free node–sub-string correspondences
(2) unanchored AMR, DRG     no explicit sub-string correspondences
                                                                              8
Anchoring in the Surface String
Relating Pieces of Meaning to the Linguistic Signal
I Intuitively, sub-structures of meaning relate to sub-parts of the input;
I semantic frameworks vary in how much weight to put on this relation;
I anchoring of graph elements in sub-strings of the underlying utterance;
I can be part of semantic annotations or not; can take different forms;
I hierarchy of anchoring types: Flavor (0)–(2); bilexical graphs strictest;
I anchoring central in parsing, explicit or latent; aka ‘alignment’ for AMR;
I relevant to at least some downstream tasks; should impact evaluation.

       Name           Example                 Type of Anchoring
(0) bilexical   DM,
                hh
                ( (h(PSD
                      h(
                      (
                       h     nodes are sub-set of surface tokens
(1) anchored EDS, PTG, UCCA free node–sub-string correspondences
(2) unanchored AMR, DRG     no explicit sub-string correspondences
                                                                               8
A Selection of Semantic Graphbanks
Selection Criteria
I ‘Full-sentence’ semantics: all content-bearing units receive annotations;
I natively graph-based: meaning representation through (directed) graphs;
I large-scale, gold-standard annotations and parsers are publicly available;

                                                                               9
A Selection of Semantic Graphbanks
Selection Criteria
I ‘Full-sentence’ semantics: all content-bearing units receive annotations;
I natively graph-based: meaning representation through (directed) graphs;
I large-scale, gold-standard annotations and parsers are publicly available;

→ five distinct frameworks, bi-lexical to unanchored; sadly, English only;

                                                                               9
A Selection of Semantic Graphbanks
Selection Criteria
I ‘Full-sentence’ semantics: all content-bearing units receive annotations;
I natively graph-based: meaning representation through (directed) graphs;
I large-scale, gold-standard annotations and parsers are publicly available;

→ five distinct frameworks, bi-lexical to unanchored; sadly, English only;

                                                                               9
A Selection of Semantic Graphbanks
Selection Criteria
I ‘Full-sentence’ semantics: all content-bearing units receive annotations;
I natively graph-based: meaning representation through (directed) graphs;
I large-scale, gold-standard annotations and parsers are publicly available;

→ five distinct frameworks, bi-lexical to unanchored; sadly, English only;
→ new in MRP 2020: one additional language for four of the frameworks.

                                                                               9
A Selection of Semantic Graphbanks
Selection Criteria
I ‘Full-sentence’ semantics: all content-bearing units receive annotations;
I natively graph-based: meaning representation through (directed) graphs;
I large-scale, gold-standard annotations and parsers are publicly available;

→ five distinct frameworks, bi-lexical to unanchored; sadly, English only;
→ new in MRP 2020: one additional language for four of the frameworks.

(With Apologies to) Non-Graph or Non-Meaning Banks
 I PropBank (Palmer et al., 2005), Framenet (Baker et al., 1998), . . . ;
I Universal Decompositional Semantics (White et al., 2016);
I Enhanced Universal Dependencies (Schuster & Manning, 2016);
I ...
                                                                               9
(1) Elementary Dependency Structures (EDS)
Simplification of Underspecified Logical Forms (Oepen & Lønning, 2006)
I Converted from LinGO Redwoods Treebank (Flickinger et al., 2017);
I decomposition or construction meaning; anchors: arbitrary sub-strings.

                                           _almost_a_1
                                              ⟨23:29⟩

                                                 ARG1

           comp                          _impossible_a_for
            ⟨2:9⟩                             ⟨30:40⟩

              ARG1                               ARG1

        _similar_a_to          _a_q        _apply_v_to           udef_q      _other_a_1
            ⟨2:9⟩              ⟨0:1⟩         ⟨44:49⟩            ⟨53:100⟩       ⟨53:58⟩

                        ARG1     BV      ARG2                ARG3   BV     ARG1

                        _technique_n_1                         _crop_n_1
                            ⟨10:19⟩                              ⟨59:65⟩
                                                                                          10
(1) Prague Tectogrammatical Graphs (PTG)
Simplification of FGD Tectogrammatical ‘Trees’ (Zeman & Hajič, 2020)
I Prague (Czech–English) Dependency Treebanks (Hajič et al., 2012);
I unanchored nodes for unexpressed arguments, e.g. #Benef and #Gen.

                                                         PRED

                                                        be
                                                      ⟨20:22⟩

                                                  ACT           PAT

                                         apply                     possible
                                    ⟨41:43⟩ ⟨44:49⟩                ⟨30:40⟩

                                ADDR        PAT                        BEN      EXT

                   crop            technique                                          almost
                                                          ACT      #Benef
              ⟨50:52⟩ ⟨59:64⟩     ⟨0:1⟩ ⟨10:19⟩                                       ⟨23:29⟩

                     RSTR                 RSTR                     coref.gram

                  other                similar
                                                            #Gen
                 ⟨53:58⟩                ⟨2:9⟩
                                                                                                11
(1) Universal Conceptual Cognitive Annotation (UCCA)
Multi-Layered Design (Abend & Rappoport, 2013); Foundational Layer
I Tree backbone: semantic ‘constituents’ are scenes (‘clauses’) and units;

                                         A   F   D     F   P         A        U

                               ⟨20:22⟩           ⟨41:43⟩       ⟨44:49⟩                 ⟨65:66⟩

                 F   E                   E   C                           R     E        C

     ⟨0:1⟩               C     ⟨23:29⟩       ⟨30:40⟩       ⟨50:52⟩           ⟨53:58⟩             ⟨59:65⟩

             S       A

     ⟨2:9⟩           ⟨10:19⟩

     A similar technique is almost impossible to apply to other crops.
                                                                                                           12
(1) Universal Conceptual Cognitive Annotation (UCCA)
Multi-Layered Design (Abend & Rappoport, 2013); Foundational Layer
I Tree backbone: semantic ‘constituents’ are scenes (‘clauses’) and units;
I scenes (Process or State): pArticipants and aDverbials (plus F and U);

                                         A   F   D     F   P         A        U

                               ⟨20:22⟩           ⟨41:43⟩       ⟨44:49⟩                 ⟨65:66⟩

                 F   E                   E   C                           R     E        C

     ⟨0:1⟩               C     ⟨23:29⟩       ⟨30:40⟩       ⟨50:52⟩           ⟨53:58⟩             ⟨59:65⟩

             S       A

     ⟨2:9⟩           ⟨10:19⟩

     A similar technique is almost impossible to apply to other crops.
                                                                                                           12
(1) Universal Conceptual Cognitive Annotation (UCCA)
Multi-Layered Design (Abend & Rappoport, 2013); Foundational Layer
I Tree backbone: semantic ‘constituents’ are scenes (‘clauses’) and units;
I scenes (Process or State): pArticipants and aDverbials (plus F and U);
I complex units distinguish Center and Elaborator(s); allow remote edges.

                                         A   F   D     F   P         A        U

                               ⟨20:22⟩           ⟨41:43⟩       ⟨44:49⟩                 ⟨65:66⟩

                 F   E                   E   C                           R     E        C

     ⟨0:1⟩               C     ⟨23:29⟩       ⟨30:40⟩       ⟨50:52⟩           ⟨53:58⟩             ⟨59:65⟩

             S       A

     ⟨2:9⟩           ⟨10:19⟩

     A similar technique is almost impossible to apply to other crops.
                                                                                                           12
(2) Abstract Meaning Representation (AMR)

                                                Banarescu et al. (2013)
       possible-01
       polarity -
                                                I Abstractly (if not linguistically)
                                                  similar to EDS, but unanchored;
     mod (domain) ARG1
                                                I verbal senses from PropBank++ ;
  almost           apply-02
                                                I negation as node-local property;
                       ARG1   ARG2
                                                I tree-like annotation: inversed
           technique           crop               edges normalized for evaluation;
               (ARG1)-of         mod (domain)   I originally designed for (S)MT;
                                                  various NLU applications to date.
      resemble-01              other

     A similar technique is almost impossible to apply to other crops.

                                                                                       13
(2) Discourse Representation Graphs (DRG)
 Graph Encoding of DRS ‘Nested Boxes’ (Kamp & Reyle, 1993)
 I From Groningen Parallel Meaning Bank (Abzianidze et al., 2017);
 I explicit encoding of scope (boxes and in edges), using reified roles.

                                                      in                                          ATTRIBUTION        PRESUPPOSITION

                                             impossible.a.01

                       in          in                                              in                                in

                            Time          Topic                        Degree               in                  apply.v.01

                                                           Theme                                                                      in

        in           time.n.08     proposition.n.01                       almost.r.01            Theme              Goal                   in

                       EQU                                                      technique.n.01                            crop.n.01

                      "now"                                Attribute                                                         NEQ

      similar.a.01                                                                                                                 crop.n.01
                                                                                                                                                14
Cross-Framework Training and Evaluation Data
                            EDS         PTG       UCCA       AMR        DRG

               Flavor        1           1          1          2          2

              Text Type   newspaper   newspaper    mixed      mixed     mixed
   train

              Sentences     37,192      42,024     6,872      57,885     6,606
               Tokens      861,831    1,026,033   171,838   1,049,083   44,692

              Text Type    mixed       mixed      mixed      mixed      mixed
   validate

              Sentences    3,302       1,664       1,585     3,560       885
               Tokens      65,564      40,770     25,982     61,722     5,541

              Text Type    mixed      newspaper   mixed      mixed      mixed
   test

              Sentences    4,040        2,507      600       2,457       898
               Tokens      68,280       59,191    18,633     49,760     5,991

                                                                                 15
Cross-Framework Training and Evaluation Data
                            EDS         PTG       UCCA       AMR        DRG

               Flavor        1           1          1          2          2

              Text Type   newspaper   newspaper    mixed      mixed     mixed
   train

              Sentences     37,192      42,024     6,872      57,885     6,606
               Tokens      861,831    1,026,033   171,838   1,049,083   44,692

              Text Type    mixed       mixed      mixed      mixed      mixed
   validate

              Sentences    3,302       1,664       1,585     3,560       885
               Tokens      65,564      40,770     25,982     61,722     5,541

              Text Type    mixed      newspaper   mixed      mixed      mixed
   test

              Sentences    4,040        2,507      600       2,457       898
               Tokens      68,280       59,191    18,633     49,760     5,991

                                                                                 15
Cross-Framework Training and Evaluation Data
                            EDS         PTG       UCCA       AMR        DRG

               Flavor        1           1          1          2          2

              Text Type   newspaper   newspaper    mixed      mixed     mixed
   train

              Sentences     37,192      42,024     6,872      57,885     6,606
               Tokens      861,831    1,026,033   171,838   1,049,083   44,692

              Text Type    mixed       mixed      mixed      mixed      mixed
   validate

              Sentences    3,302       1,664       1,585     3,560       885
               Tokens      65,564      40,770     25,982     61,722     5,541

              Text Type    mixed      newspaper   mixed      mixed      mixed
   test

              Sentences    4,040        2,507      600       2,457       898
               Tokens      68,280       59,191    18,633     49,760     5,991

 I Validation split is MRP 2019 evaluation data; allowed for fine-tuning;

                                                                                 15
Cross-Framework Training and Evaluation Data
                            EDS         PTG       UCCA       AMR        DRG

               Flavor        1           1          1          2          2

              Text Type   newspaper   newspaper    mixed      mixed     mixed
   train

              Sentences     37,192      42,024     6,872      57,885     6,606
               Tokens      861,831    1,026,033   171,838   1,049,083   44,692

              Text Type    mixed       mixed      mixed      mixed      mixed
   validate

              Sentences    3,302       1,664       1,585     3,560       885
               Tokens      65,564      40,770     25,982     61,722     5,541

              Text Type    mixed      newspaper   mixed      mixed      mixed
   test

              Sentences    4,040        2,507      600       2,457       898
               Tokens      68,280       59,191    18,633     49,760     5,991

 I Validation split is MRP 2019 evaluation data; allowed for fine-tuning;
 I linguistics: smallish WSJ sample in all frameworks publicly available;
 I evaluation: subset of 100 sentences from The Little Prince is public.
                                                                                 15
Cross-Lingual Training and Evaluation Data

                         PTG       UCCA     AMR       DRG
           Language     Czech      German   Chinese   German
            Flavor        1           1        1         2
           Text Type   newspaper   mixed     mixed    mixed
   train

           Sentences     43,955     4,125    18,365   1,575
            Tokens      740,466    95,634   428,054   9,088
           Text Type   newspaper   mixed    mixed     mixed
   test

           Sentences      5,476     444      1,713     403
            Tokens       92,643    10,585   39,228    2,384

                                                               16
Cross-Lingual Training and Evaluation Data

                         PTG       UCCA     AMR       DRG
           Language     Czech      German   Chinese   German
            Flavor        1           1        1         2
           Text Type   newspaper   mixed     mixed    mixed
   train

           Sentences     43,955     4,125    18,365   1,575
            Tokens      740,466    95,634   428,054   9,088
           Text Type   newspaper   mixed    mixed     mixed
   test

           Sentences      5,476     444      1,713     403
            Tokens       92,643    10,585   39,228    2,384

                                                               16
Cross-Lingual Training and Evaluation Data

                         PTG       UCCA     AMR       DRG
           Language     Czech      German   Chinese   German
            Flavor        1           1        1         2
           Text Type   newspaper   mixed     mixed    mixed
   train

           Sentences     43,955     4,125    18,365   1,575
            Tokens      740,466    95,634   428,054   9,088
           Text Type   newspaper   mixed    mixed     mixed
   test

           Sentences      5,476     444      1,713     403
            Tokens       92,643    10,585   39,228    2,384

                                                               16
Cross-Lingual Training and Evaluation Data

                          PTG         UCCA       AMR         DRG
            Language      Czech      German      Chinese    German
             Flavor         1           1           1          2
            Text Type   newspaper     mixed      mixed       mixed
    train

            Sentences     43,955       4,125     18,365      1,575
             Tokens      740,466      95,634    428,054      9,088
            Text Type   newspaper     mixed      mixed       mixed
    test

            Sentences      5,476       444        1,713       403
             Tokens       92,643      10,585     39,228      2,384

I Gold-standard graphs for one additional language in four frameworks;

                                                                         16
Cross-Lingual Training and Evaluation Data

                           PTG         UCCA         AMR         DRG
            Language       Czech       German      Chinese     German
             Flavor          1            1           1           2
            Text Type   newspaper       mixed       mixed       mixed
    train

            Sentences     43,955         4,125      18,365      1,575
             Tokens      740,466        95,634     428,054      9,088
            Text Type   newspaper       mixed       mixed       mixed
    test

            Sentences      5,476         444         1,713       403
             Tokens       92,643        10,585      39,228      2,384

I Gold-standard graphs for one additional language in four frameworks;
I ‘low-resource’ training setting for two frameworks: UCCA and DRG;
? explor opportunities for cross-lingual transfer learning (in-framework).
                                                                             16
Cross-Framework Evaluation: MRP Graph Similarity
I Break down graphs into types of information: per-type and overall F1 ;

                                    Different Types of Semantic Graph ‘Atoms’
   _retire_v_1           proper_q
      ⟨7:14⟩               ⟨0:6⟩                       EDS PTG UCCA AMR DRG

                                     Top Nodes         3    3     3    3    3
                 ARG1     BV
                                     Labeled Edges     3    3     3    3   (3)
                                     Node Labels       3    3     7    3    3
                   named
                                     Node Properties   3    3     7    3    7
                 CARG Pierre
                    ⟨0:6⟩            Node Anchoring    3   (3)   (3)   7    7
                                     Edge Attributes   7    3     3    7    7

        Pierre retired.
                                                                                 17
Cross-Framework Evaluation: MRP Graph Similarity
I Break down graphs into types of information: per-type and overall F1 ;
I tops

                                    Different Types of Semantic Graph ‘Atoms’
   _retire_v_1           proper_q
      ⟨7:14⟩               ⟨0:6⟩                       EDS PTG UCCA AMR DRG

                                     Top Nodes         3    3     3    3    3
                 ARG1     BV
                                     Labeled Edges     3    3     3    3   (3)
                                     Node Labels       3    3     7    3    3
                   named
                                     Node Properties   3    3     7    3    7
                 CARG Pierre
                    ⟨0:6⟩            Node Anchoring    3   (3)   (3)   7    7
                                     Edge Attributes   7    3     3    7    7

        Pierre retired.
                                                                                 17
Cross-Framework Evaluation: MRP Graph Similarity
I Break down graphs into types of information: per-type and overall F1 ;
I tops and (labeled) edges;

                                    Different Types of Semantic Graph ‘Atoms’
   _retire_v_1           proper_q
      ⟨7:14⟩               ⟨0:6⟩                       EDS PTG UCCA AMR DRG

                                     Top Nodes         3    3     3    3    3
                 ARG1     BV
                                     Labeled Edges     3    3     3    3   (3)
                                     Node Labels       3    3     7    3    3
                   named
                                     Node Properties   3    3     7    3    7
                 CARG Pierre
                    ⟨0:6⟩            Node Anchoring    3   (3)   (3)   7    7
                                     Edge Attributes   7    3     3    7    7

        Pierre retired.
                                                                                 17
Cross-Framework Evaluation: MRP Graph Similarity
I Break down graphs into types of information: per-type and overall F1 ;
I tops and (labeled) edges; labels,

                                    Different Types of Semantic Graph ‘Atoms’
   _retire_v_1           proper_q
      ⟨7:14⟩               ⟨0:6⟩                       EDS PTG UCCA AMR DRG

                                     Top Nodes         3    3     3    3    3
                 ARG1     BV
                                     Labeled Edges     3    3     3    3   (3)
                                     Node Labels       3    3     7    3    3
                   named
                                     Node Properties   3    3     7    3    7
                 CARG Pierre
                    ⟨0:6⟩            Node Anchoring    3   (3)   (3)   7    7
                                     Edge Attributes   7    3     3    7    7

        Pierre retired.
                                                                                 17
Cross-Framework Evaluation: MRP Graph Similarity
I Break down graphs into types of information: per-type and overall F1 ;
I tops and (labeled) edges; labels, properties,

                                    Different Types of Semantic Graph ‘Atoms’
   _retire_v_1           proper_q
      ⟨7:14⟩               ⟨0:6⟩                       EDS PTG UCCA AMR DRG

                                     Top Nodes         3    3     3    3    3
                 ARG1     BV
                                     Labeled Edges     3    3     3    3   (3)
                                     Node Labels       3    3     7    3    3
                   named
                                     Node Properties   3    3     7    3    7
                 CARG Pierre
                    ⟨0:6⟩            Node Anchoring    3   (3)   (3)   7    7
                                     Edge Attributes   7    3     3    7    7

        Pierre retired.
                                                                                 17
Cross-Framework Evaluation: MRP Graph Similarity
I Break down graphs into types of information: per-type and overall F1 ;
I tops and (labeled) edges; labels, properties, anchors,

                                    Different Types of Semantic Graph ‘Atoms’
   _retire_v_1           proper_q
      ⟨7:14⟩               ⟨0:6⟩                       EDS PTG UCCA AMR DRG

                                     Top Nodes         3    3     3    3    3
                 ARG1     BV
                                     Labeled Edges     3    3     3    3   (3)
                                     Node Labels       3    3     7    3    3
                   named
                                     Node Properties   3    3     7    3    7
                 CARG Pierre
                    ⟨0:6⟩            Node Anchoring    3   (3)   (3)   7    7
                                     Edge Attributes   7    3     3    7    7

        Pierre retired.
                                                                                 17
Cross-Framework Evaluation: MRP Graph Similarity
I Break down graphs into types of information: per-type and overall F1 ;
I tops and (labeled) edges; labels, properties, anchors, and attributes;

                                    Different Types of Semantic Graph ‘Atoms’
   _retire_v_1           proper_q
      ⟨7:14⟩               ⟨0:6⟩                       EDS PTG UCCA AMR DRG

                                     Top Nodes         3    3     3    3    3
                 ARG1     BV
                                     Labeled Edges     3    3     3    3   (3)
                                     Node Labels       3    3     7    3    3
                   named
                                     Node Properties   3    3     7    3    7
                 CARG Pierre
                    ⟨0:6⟩            Node Anchoring    3   (3)   (3)   7    7
                                     Edge Attributes   7    3     3    7    7

        Pierre retired.
                                                                                 17
Cross-Framework Evaluation: MRP Graph Similarity
I Break down graphs into types of information: per-type and overall F1 ;
I tops and (labeled) edges; labels, properties, anchors, and attributes;
I requires node–node correspondences; search for overall maximum score;
I maximum common edge subgraph isomorphism (MCES) is NP-hard;

                                    Different Types of Semantic Graph ‘Atoms’
   _retire_v_1           proper_q
      ⟨7:14⟩               ⟨0:6⟩                       EDS PTG UCCA AMR DRG

                                     Top Nodes         3    3     3    3    3
                 ARG1     BV
                                     Labeled Edges     3    3     3    3   (3)
                                     Node Labels       3    3     7    3    3
                   named
                                     Node Properties   3    3     7    3    7
                 CARG Pierre
                    ⟨0:6⟩            Node Anchoring    3   (3)   (3)   7    7
                                     Edge Attributes   7    3     3    7    7

        Pierre retired.
                                                                                 17
Cross-Framework Evaluation: MRP Graph Similarity
I Break down graphs into types of information: per-type and overall F1 ;
I tops and (labeled) edges; labels, properties, anchors, and attributes;
I requires node–node correspondences; search for overall maximum score;
I maximum common edge subgraph isomorphism (MCES) is NP-hard;
→ smart initialization, scheduling, and pruning yield strong approximation.

                                    Different Types of Semantic Graph ‘Atoms’
   _retire_v_1           proper_q
      ⟨7:14⟩               ⟨0:6⟩                       EDS PTG UCCA AMR DRG

                                     Top Nodes         3    3     3    3    3
                 ARG1     BV
                                     Labeled Edges     3    3     3    3   (3)
                                     Node Labels       3    3     7    3    3
                   named
                                     Node Properties   3    3     7    3    7
                 CARG Pierre
                    ⟨0:6⟩            Node Anchoring    3   (3)   (3)   7    7
                                     Edge Attributes   7    3     3    7    7

        Pierre retired.
                                                                                 17
Graphbank Statistics (Kuhlmann & Oepen, 2016)
                                                           EDS     PTG UCCA AMR−1 DRG

                (02) Average Tokens per Graph              22.17 24.42     25.01   18.12    6.77
                (03) Average Nodes per Token                1.26 0.74       1.33    0.64    2.09
                (04) Distinct Edge Labels                     10    72        15     101      16
                (05)   Percentage   of   top nodes          0.99    1.27    1.66    3.77 3.40
  proportions

                (06)   Percentage   of   node labels       29.02   21.61       –   43.91 39.81
                (07)   Percentage   of   node properties   12.54   26.22       –    7.63     –
                (08)   Percentage   of   node anchors      29.02   19.63   38.80       –     –
                (09)   Percentage   of   (labeled) edges   28.43   26.10   56.88   44.69 56.79
                (10)   Percentage   of   edge attributes       –    5.17    2.66       –     –
                (11)   %g Rooted Trees                      0.09   22.63   28.19   22.05    0.35
                (12)   %g Treewidth One                    68.60   22.67   34.17   49.91    0.35
                (13)   Average Treewidth                   1.317   2.067   1.691   1.561   2.131
  treeness

                (14)   Maximal Treewidth                       3       7       4       5       5
                (15)   Average Edge Density                1.015   1.177   1.055   1.092   1.265
                (16)   %n Reentrant                        32.77   16.23    4.90   19.89   25.92
                (17)   %g Cyclic                            0.27   33.97    0.00    0.38    0.27
                (18)   %g Not Connected                     1.90    0.00    0.00    0.00    0.00
                (19)   %g Multi-Rooted                     99.93    0.00    0.00   71.64   32.32
                                                                                                   18
Graphbank Statistics (Kuhlmann & Oepen, 2016)
                                                           EDS     PTG UCCA AMR−1 DRG

                (02) Average Tokens per Graph              22.17 24.42     25.01   18.12    6.77
                (03) Average Nodes per Token                1.26 0.74       1.33    0.64    2.09
                (04) Distinct Edge Labels                     10    72        15     101      16
                (05)   Percentage   of   top nodes          0.99    1.27    1.66    3.77 3.40
  proportions

                (06)   Percentage   of   node labels       29.02   21.61       –   43.91 39.81
                (07)   Percentage   of   node properties   12.54   26.22       –    7.63     –
                (08)   Percentage   of   node anchors      29.02   19.63   38.80       –     –
                (09)   Percentage   of   (labeled) edges   28.43   26.10   56.88   44.69 56.79
                (10)   Percentage   of   edge attributes       –    5.17    2.66       –     –
                (11)   %g Rooted Trees                      0.09   22.63   28.19   22.05    0.35
                (12)   %g Treewidth One                    68.60   22.67   34.17   49.91    0.35
                (13)   Average Treewidth                   1.317   2.067   1.691   1.561   2.131
  treeness

                (14)   Maximal Treewidth                       3       7       4       5       5
                (15)   Average Edge Density                1.015   1.177   1.055   1.092   1.265
                (16)   %n Reentrant                        32.77   16.23    4.90   19.89   25.92
                (17)   %g Cyclic                            0.27   33.97    0.00    0.38    0.27
                (18)   %g Not Connected                     1.90    0.00    0.00    0.00    0.00
                (19)   %g Multi-Rooted                     99.93    0.00    0.00   71.64   32.32
                                                                                                   18
Graphbank Statistics (Kuhlmann & Oepen, 2016)
                                                           EDS     PTG UCCA AMR−1 DRG

                (02) Average Tokens per Graph              22.17 24.42     25.01   18.12    6.77
                (03) Average Nodes per Token                1.26 0.74       1.33    0.64    2.09
                (04) Distinct Edge Labels                     10    72        15     101      16
                (05)   Percentage   of   top nodes          0.99    1.27    1.66    3.77 3.40
  proportions

                (06)   Percentage   of   node labels       29.02   21.61       –   43.91 39.81
                (07)   Percentage   of   node properties   12.54   26.22       –    7.63     –
                (08)   Percentage   of   node anchors      29.02   19.63   38.80       –     –
                (09)   Percentage   of   (labeled) edges   28.43   26.10   56.88   44.69 56.79
                (10)   Percentage   of   edge attributes       –    5.17    2.66       –     –
                (11)   %g Rooted Trees                      0.09   22.63   28.19   22.05    0.35
                (12)   %g Treewidth One                    68.60   22.67   34.17   49.91    0.35
                (13)   Average Treewidth                   1.317   2.067   1.691   1.561   2.131
  treeness

                (14)   Maximal Treewidth                       3       7       4       5       5
                (15)   Average Edge Density                1.015   1.177   1.055   1.092   1.265
                (16)   %n Reentrant                        32.77   16.23    4.90   19.89   25.92
                (17)   %g Cyclic                            0.27   33.97    0.00    0.38    0.27
                (18)   %g Not Connected                     1.90    0.00    0.00    0.00    0.00
                (19)   %g Multi-Rooted                     99.93    0.00    0.00   71.64   32.32
                                                                                                   18
Graphbank Statistics (Kuhlmann & Oepen, 2016)
                                                           EDS     PTG UCCA AMR−1 DRG

                (02) Average Tokens per Graph              22.17 24.42     25.01   18.12    6.77
                (03) Average Nodes per Token                1.26 0.74       1.33    0.64    2.09
                (04) Distinct Edge Labels                     10    72        15     101      16
                (05)   Percentage   of   top nodes          0.99    1.27    1.66    3.77 3.40
  proportions

                (06)   Percentage   of   node labels       29.02   21.61       –   43.91 39.81
                (07)   Percentage   of   node properties   12.54   26.22       –    7.63     –
                (08)   Percentage   of   node anchors      29.02   19.63   38.80       –     –
                (09)   Percentage   of   (labeled) edges   28.43   26.10   56.88   44.69 56.79
                (10)   Percentage   of   edge attributes       –    5.17    2.66       –     –
                (11)   %g Rooted Trees                      0.09   22.63   28.19   22.05    0.35
                (12)   %g Treewidth One                    68.60   22.67   34.17   49.91    0.35
                (13)   Average Treewidth                   1.317   2.067   1.691   1.561   2.131
  treeness

                (14)   Maximal Treewidth                       3       7       4       5       5
                (15)   Average Edge Density                1.015   1.177   1.055   1.092   1.265
                (16)   %n Reentrant                        32.77   16.23    4.90   19.89   25.92
                (17)   %g Cyclic                            0.27   33.97    0.00    0.38    0.27
                (18)   %g Not Connected                     1.90    0.00    0.00    0.00    0.00
                (19)   %g Multi-Rooted                     99.93    0.00    0.00   71.64   32.32
                                                                                                   18
High-Level Overview of Submissions
›Cross-Framework   mCross-Lingual

 Teams                      AMR     DRG   EDS   PTG   UCCA

 Hitachi           ›m       ›m      ›m    ›     ›m    ›m
 ÚFAL              ›m       ›m      ›m    ›     ›m    ›m
 HIT-SCIR          ›m       ›m      ›m    ›     ›m    ›m
 HUJI-KU                    ›m      ›m    ›     ›m    ›m
 ISCAS                      ›       ›     ›     ›     ›
 TJU-BLCU                   ›m      ›m    ›     ›m    ›
 JBNU                       ›
 ÚFAL                       ›m ›m ›             ›m ›m

 ERG                                      ›                  19
Score Distribution

               1    ›       Cross-Framework (Full)
                        m Cross-Lingual              ›m
                                                     ›       ›
                                                             ›
                                                     ›m            › m
                        ›m                                   ›
                                                             ›     ›m
                                          ›                        ›
              0.8       ›                 ›m                 ›               m
                                                                             m
                                                                             m
                                            m                        m     ›
                                                                           ›
                                                                           ›m
                            m             ›          ›m
                        ›                            ›m
 All MRP F1

              0.6           m             ›
                                                                       m
                                          ›                        ›
                        ›                     m              ›
                                              m
              0.4                                    ›
                                                         m
                        ›                 ›
                                              m
              0.2           m                                      ›m
                                                                   ›
                                                                           ›
                                                                           ›
               0
                    Overall               AMR        DRG     EDS   PTG     UCCA

                                                                                  20
Score Distribution: Zoom In

                1
                     ›     Cross-Framework (Full)

              0.95    m    Cross-Lingual

                                                    ›
                                                    ›m    ›
                                                          ›
                                                                 m
                                                    ›m
               0.9
                                                                ›
                                                                ›m
 All MRP F1

                     ›m                                   ›
                                                          ›
              0.85
                                                                ›
                     ›               ›                                 m
               0.8                   ›m                   ›            m
                                                                       m
                                      m                          m
              0.75                                                    ›
                                                                      ›m
                                                                      ›
                     Overall         AMR            DRG   EDS   PTG   UCCA

                                                                             21
HIT-SCIR        0.81

     Overall
                        Hitachi + ÚFAL         0.86

                     HIT-SCIR     0.70

                                ÚFAL      0.80

     AMR
                                Hitachi     0.82

                                 HIT-SCIR          0.89

                                         Hitachi      0.93

     DRG
                                           ÚFAL         0.94

                                HIT-SCIR         0.87

                                          ÚFAL        0.93

     EDS
                                          Hitachi       0.94
                                                               ›Cross-Framework Track: Full Evaluation

                             HIT-SCIR         0.84

                                     ÚFAL         0.88
     PTG

                                    Hitachi        0.89

               Hitachi + HIT-SCIR    0.75
     UCCA

                            ÚFAL       0.76
22
HIT-SCIR      0.69

     Overall
                                 Hitachi + ÚFAL        0.85

               HIT-SCIR   0.49

                                      ÚFAL      0.78

     AMR
                                      Hitachi    0.80

                          HIT-SCIR      0.68

                                                ÚFAL      0.90

     DRG
                                                Hitachi       0.93

     EDS
                                                                     mCross-Lingual Track: Full Evaluation

                                  HIT-SCIR      0.78

                                          Hitachi       0.87
     PTG

                                                ÚFAL      0.91

                                     Hitachi    0.79

                                   HIT-SCIR      0.80
     UCCA

                                        ÚFAL      0.81
23
Medal Ceremony!
›Cross-Framework   mCross-Lingual

 Teams                       AMR    DRG   EDS   PTG   UCCA

 Hitachi           ›m        ›m     ›m    ›     ›m    ›m
 ÚFAL              ›m        ›m     ›m    ›     ›m    ›m
 HIT-SCIR          ›m        ›m     ›m    ›     ›m    ›m
 HUJI-KU                     ›m     ›m    ›     ›m    ›m
 ISCAS                       ›      ›     ›     ›     ›
 TJU-BLCU                    ›m     ›m    ›     ›m    ›
 JBNU                        ›
 ÚFAL                        ›m ›m ›            ›m ›m

 ERG                                      ›                  24
HIT-SCIR          0.80

     Overall
                 Hitachi + ÚFAL            0.85

               HIT-SCIR     0.70

                      ÚFAL          0.78

     AMR
                     Hitachi        0.79

     DRG
                           HIT-SCIR          0.89

                                      ÚFAL        0.95

     EDS
                                     Hitachi        0.97
                                                           Cross-Framework Track: The Little Prince

                  HIT-SCIR          0.78

                          Hitachi     0.82
     PTG

                            ÚFAL       0.83

                     HIT-SCIR         0.82
     UCCA

                Hitachi + ÚFAL         0.83
25
HIT-SCIR    0.70                                    ÚFAL MRPipe     0.71

                   ÚFAL      0.78                                      Amazon     0.71

     AMR
                                                           AMR
                  Hitachi    0.79                                      Saarland    0.73

                                               MRP 2020:
                                                                                                         MRP 2019:

                                                                            Saarland             0.95

                                                                        SJTU–NICT                0.95

                                                           DM

     DRG
                                                                           HIT-SCIR               0.95

                  HIT-SCIR          0.89                               SJTU–NICT             0.91

                         ÚFAL           0.95                               Saarland             0.92

     EDS
                                                           EDS

                         Hitachi        0.97                          SUDA–Alibaba               0.93
                                                                                                                     State of the Art: The Little Prince

                HIT-SCIR     0.78                                         Saarland          0.88

                   Hitachi       0.82                                       Hitachi         0.88
                                                           PSD

     PTG

                     ÚFAL        0.83                                  SJTU–NICT            0.89

                 HIT-SCIR        0.82                                  Saarland     0.76

                                                                    SUDA–Alibaba         0.82
     UCCA
                                                           UCCA

            Hitachi + ÚFAL       0.83                                   HIT-SCIR          0.83
26
Interim Conclusions & Outlook
Lessons Learned (from Two Consecutive Shared Tasks)
I Good community interest: 180 subscribers; 19 data licenses (via LDC);
I technical barriers and ‘competitive selection’: 6 + 2 teams submitted;

                                                                           27
Interim Conclusions & Outlook
Lessons Learned (from Two Consecutive Shared Tasks)
I Good community interest: 180 subscribers; 19 data licenses (via LDC);
I technical barriers and ‘competitive selection’: 6 + 2 teams submitted;

→ advanced state of the art on four frameworks (but possibly not AMR);

                                                                           27
Interim Conclusions & Outlook
Lessons Learned (from Two Consecutive Shared Tasks)
I Good community interest: 180 subscribers; 19 data licenses (via LDC);
I technical barriers and ‘competitive selection’: 6 + 2 teams submitted;

→ advanced state of the art on four frameworks (but possibly not AMR);
→ greatly increased cross-framework uniformity; but limited (M)TL so far.

                                                                            27
Interim Conclusions & Outlook
Lessons Learned (from Two Consecutive Shared Tasks)
I Good community interest: 180 subscribers; 19 data licenses (via LDC);
I technical barriers and ‘competitive selection’: 6 + 2 teams submitted;

→ advanced state of the art on four frameworks (but possibly not AMR);
→ greatly increased cross-framework uniformity; but limited (M)TL so far.

Outlook: Beyond MRP 2020
I High-quality, robust meaning representation parsers generally available;
I MRP 2020 data, metrics, submissions, and scores as stable benchmark;

                                                                             27
Interim Conclusions & Outlook
Lessons Learned (from Two Consecutive Shared Tasks)
I Good community interest: 180 subscribers; 19 data licenses (via LDC);
I technical barriers and ‘competitive selection’: 6 + 2 teams submitted;

→ advanced state of the art on four frameworks (but possibly not AMR);
→ greatly increased cross-framework uniformity; but limited (M)TL so far.

Outlook: Beyond MRP 2020
I High-quality, robust meaning representation parsers generally available;
I MRP 2020 data, metrics, submissions, and scores as stable benchmark;

 ? post-mortem contrastive analysis of architectures (Buljan et al., 2020);

                                                                              27
Interim Conclusions & Outlook
Lessons Learned (from Two Consecutive Shared Tasks)
I Good community interest: 180 subscribers; 19 data licenses (via LDC);
I technical barriers and ‘competitive selection’: 6 + 2 teams submitted;

→ advanced state of the art on four frameworks (but possibly not AMR);
→ greatly increased cross-framework uniformity; but limited (M)TL so far.

Outlook: Beyond MRP 2020
I High-quality, robust meaning representation parsers generally available;
I MRP 2020 data, metrics, submissions, and scores as stable benchmark;

 ? post-mortem contrastive analysis of architectures (Buljan et al., 2020);
 ? increased focus on evaluation metrics: score ‘larger pieces’; SEMBLEU;

                                                                              27
Interim Conclusions & Outlook
Lessons Learned (from Two Consecutive Shared Tasks)
I Good community interest: 180 subscribers; 19 data licenses (via LDC);
I technical barriers and ‘competitive selection’: 6 + 2 teams submitted;

→ advanced state of the art on four frameworks (but possibly not AMR);
→ greatly increased cross-framework uniformity; but limited (M)TL so far.

Outlook: Beyond MRP 2020
I High-quality, robust meaning representation parsers generally available;
I MRP 2020 data, metrics, submissions, and scores as stable benchmark;

 ? post-mortem contrastive analysis of architectures (Buljan et al., 2020);
 ? increased focus on evaluation metrics: score ‘larger pieces’; SEMBLEU;
→ open discussion with 2020 participants towards the end of this session.
                                                                              27
Acknowledgments

         Jayeol Chun, Dotan Dvir, Dan Flickinger,
    Jiří Mírovský, Anna Nedoluzhko, Sebastian Schuster,
            Milan Straka, and Zdeňka Urešová

               Linguistic Data Consortium,
          Nordic Language Processing Laboratory

                                                          28
References I

Omri Abend & Ari Rappoport. 2013. Universal Conceptual Cognitive
  Annotation (UCCA). In Proceedings of the 51th Meeting of the
  Association for Computational Linguistics, pages 228 – 238, Sofia,
  Bulgaria.
Lasha Abzianidze, Johannes Bjerva, Kilian Evang, Hessel Haagsma, Rik
   van Noord, Pierre Ludmann, Duc-Duy Nguyen, & Johan Bos. 2017. The
   Parallel Meaning Bank. Towards a multilingual corpus of translations
   annotated with compositional meaning representations. In Proceedings
   of the 15th Meeting of the European Chapter of the Association for
   Computational Linguistics, pages 242–247, Valencia, Spain.
Collin F. Baker, Charles J. Fillmore, & John B. Lowe. 1998. The Berkeley
   FrameNet project. In Proceedings of the 17th International Conference
   on Computational Linguistics and the 36th Meeting of the Association
   for Computational Linguistics, pages 86 – 90, Stroudsburg, PA, USA.

                                                                           29
References II

Laura Banarescu, Claire Bonial, Shu Cai, Madalina Georgescu, Kira Griffitt,
   Ulf Hermjakob, Kevin Knight, Philipp Koehn, Martha Palmer, &
   Nathan Schneider. 2013. Abstract Meaning Representation for
   sembanking. In Proceedings of the 7th Linguistic Annotation Workshop
   and Interoperability with Discourse, pages 178 – 186, Sofia, Bulgaria.
Maja Buljan, Joakim Nivre, Stephan Oepen, & Lilja Øvrelid. 2020. A tale
  of three parsers. Towards diagnostic evaluation for meaning
  representation parsing. In Proceedings of the 12th Language Resources
  and Evaluation Conference, pages 1902–1909, Marseille, France.
  European Language Resources Association.
Dan Flickinger, Stephan Oepen, & Emily M. Bender. 2017. Sustainable
  development and refinement of complex linguistic annotations at scale.
  In Nacy Ide & James Pustejovsky, editors, Handbook of Linguistic
  Annotation, pages 353 – 377. Springer, Dordrecht, The Netherlands.

                                                                              30
References III
Jan Hajič, Eva Hajičová, Jarmila Panevová, Petr Sgall, Ondřej Bojar, Silvie
   Cinková, Eva Fučíková, Marie Mikulová, Petr Pajas, Jan Popelka, Jiří
   Semecký, Jana Šindlerová, Jan Štěpánek, Josef Toman, Zdeňka
   Urešová, & Zdeněk Žabokrtský. 2012. Announcing Prague
   Czech-English Dependency Treebank 2.0. In Proceedings of the 8th
   International Conference on Language Resources and Evaluation, pages
   3153 – 3160, Istanbul, Turkey.
Hardy & Andreas Vlachos. 2018. Guided neural language generation for
  abstractive summarization using abstract meaning representation. In
  Proceedings of the 2018 Conference on Empirical Methods in Natural
  Language Processing, Brussels, Belgium.
Hans Kamp & Uwe Reyle. 1993. From Discourse to Logic. An Introduction
  to Modeltheoretic Semantics of Natural Language, Formal Logic and
  DRT. Kluwer, Dordrecht, The Netherlands.
Marco Kuhlmann & Stephan Oepen. 2016. Towards a catalogue of
  linguistic graph banks. Computational Linguistics, 42(4):819 – 827.
                                                                              31
References IV

Stephan Oepen & Jan Tore Lønning. 2006. Discriminant-based MRS
   banking. In Proceedings of the 5th International Conference on
   Language Resources and Evaluation, pages 1250 – 1255, Genoa, Italy.
Martha Palmer, Dan Gildea, & Paul Kingsbury. 2005. The Proposition
  Bank. A corpus annotated with semantic roles. Computational
  Linguistics, 31(1):71 – 106.
Sebastian Schuster & Christopher D. Manning. 2016. Enhanced English
   Universal Dependencies. An improved representation for natural
   language understanding tasks. In Proceedings of the 10th International
   Conference on Language Resources and Evaluation, Portorož, Slovenia.
Aaron Steven White, Drew Reisinger, Keisuke Sakaguchi, Tim Vieira,
  Sheng Zhang, Rachel Rudinger, Kyle Rawlins, & Benjamin Van Durme.
  2016. Universal Decompositional Semantics on Universal Dependencies.
  In Proceedings of the 2016 Conference on Empirical Methods in Natural
  Language Processing, pages 1713 – 1723, Austin, TX, USA.

                                                                            32
References V

Daniel Zeman & Jan Hajič. 2020. FGD at MRP 2020: Prague
  Tectogrammatical Graphs. In Proceedings of the CoNLL 2020 Shared
  Task: Cross-Framework Meaning Representation Parsing, pages 33 – 39,
  Online.

                                                                         33
You can also read