Title: A Best-Fit Approach for Productive Analysis of Omitted Arguments
1A Best-Fit Approach for Productive Analysis of
Omitted Arguments
- Eva Mok John Bryant
- University of California, Berkeley
- International Computer Science Institute
2Simplify grammar by exploiting the language
understanding process
- Omission of arguments in Mandarin Chinese
- Construction grammar framework
- Model of language understanding
- Our best-fit approach
3Productive Argument Omission (in Mandarin)
- Mother (I) give you this (a toy).
ma1ma gei3 ni3 zhei4ge
mother give 2PS thisCLS
1
- You give auntie the peach.
2
ni3 gei3 yi2
2PS give auntie
- Oh (go on)! You give auntie that.
3
ao ni3 gei3 ya
EMP 2PS give EMP
4
gei3
give
CHILDES Beijing Corpus (Tardiff, 1993 Tardiff,
1996)
4Arguments are omitted with different probabilities
- All arguments omitted 30.6 No arguments
omitted 6.1
5Construction grammar approach
- Kay Fillmore 1999 Goldberg 1995
- Grammaticality form and function
- Basic unit of analysis construction, i.e. a
pairing of form and meaning constraints - Not purely lexically compositional
- Implies early use of semantics in processing
- Embodied Construction Grammar (ECG) (Bergen
Chang, 2005)
6Problem Proliferation of constructions
Subj Verb Obj1 Obj2
? ? ? ?
Giver Transfer Recipient Theme
Verb Obj1 Obj2
? ? ?
Transfer Recipient Theme
Subj Verb Obj2
? ? ?
Giver Transfer Theme
Subj Verb Obj1
? ? ?
Giver Transfer Recipient
7If the analysis process is smart, then...
Subj Verb Obj1 Obj2
? ? ? ?
Giver Transfer Recipient Theme
- The grammar needs only state one construction
- Omission of constituents is flexibly allowed
- The analysis process figures out what was omitted
8Best-fit analysis process takes burden off the
grammar representation
Constructions
Utterance
incremental, competition-based,
psycholinguistically plausible
Semantic Specification image schemas, frames,
action schemas
Simulation
9Competition-based analyzer finds the best analysis
- An analysis is made up of
- A constructional tree
- A set of resolutions
- A semantic specification
The best fit has the highest combined score
10Combined score that determines best-fit
- Syntactic Fit
- Constituency relations
- Combine with preferences on non-local elements
- Conditioned on syntactic context
- Antecedent Fit
- Ability to find referents in the context
- Conditioned on syntactic information, feature
agreement - Semantic Fit
- Semantic bindings for frame roles
- Frame roles fillers are scored
11Analyzing ni3 gei3 yi2 (You give auntie)
Two of the competing analyses
ni3 gei3 yi2 omitted
? ? ? ?
Giver Transfer Recipient Theme
ni3 gei3 omitted yi2
? ? ? ?
Giver Transfer Recipient Theme
- Syntactic Fit
- P(Theme omitted ditransitive cxn) 0.65
- P(Recipient omitted ditransitive cxn) 0.42
(1-0.78)(1-0.42)0.65 0.08
(1-0.78)(1-0.65)0.42 0.03
12Using frame and lexical information to restrict
type of reference
The Transfer Frame Giver Recipient Theme The Transfer Frame Giver Recipient Theme
Manner Means Place Purpose Reason Time
Lexical Unit gei3 Giver (DNI) Recipient (DNI) Theme (DNI)
13Can the omitted argument be recovered from
context?
ni3 gei3 yi2 omitted
? ? ? ?
Giver Transfer Recipient Theme
ni3 gei3 omitted yi2
? ? ? ?
Giver Transfer Recipient Theme
?
14How good of a theme is a peach? How about an
aunt?
ni3 gei3 yi2 omitted
? ? ? ?
Giver Transfer Recipient Theme
ni3 gei3 omitted yi2
? ? ? ?
Giver Transfer Recipient Theme
The Transfer Frame Giver (usually animate) Recipient (usually animate) Theme (usually inanimate)
15The argument omission patterns shown earlier can
be covered with just ONE construction
Subj Verb Obj1 Obj2
? ? ? ?
Giver Transfer Recipient Theme
P(omittedcxn)
0.78
0.42
0.65
- Each cxn is annotated with probabilities of
omission - Language-specific default probability can be set
16Research goal
- A computationally-precise modeling framework for
learning early constructions
Learning
New Construction
Linguistic Knowledge
17Frequent argument omission in pro-drop languages
- Mandarin example
- ni3 gei3 yi2 (you give auntie)
- Even in English, there are often no spoken
antecedents to pronouns in conversations
Learner must integrate cues from intentions,
gestures, prior discourse, etc
18A short dialogue
- bie2 mo3 wai4tou2 a 1_3 ! (?????)
- NEG-IMP apply forehead
- Dont apply lotion to your forehead
- mo3 wai4tou2 ke3 jiu4 bu4 hao3kan4 le a .
(??????????) - apply forehead LINKER LINKER NEG good looking CRS
SFP - If you apply lotion to your forehead then
you will not be pretty -
- ze ya a bie2 gei3 ma1ma wang3 lian3 shang4
mo3 e ! (??? ?????????) - INTERJ NEG-IMP BEN mother CV-DIR face on apply
- INTERJ Dont apply the lotion on your moms
face (for mom) - - low pitch motherese ma1ma bu4 mo3 you2 .
(?????) - mother NEG apply lotion
- Mom doesnt apply (use) lotion
19Goals, refined
- Demonstrate learning given
- embodied meaning representation
- structured representation of context
- Based on
- Usage-based learning
- Domain-general statistical learning mechanism
- Generalization / linguistic category formation
20Towards a precise computational model
- Modeling early grammar learning
- Context model Simulation
- Data annotation
- Finding the best analysis for learning
- Hypothesizing and reorganizing constructions
- Pilot results
21Embodied Construction Grammar
construction yi2-N subcase of Morpheme
form constraints self.f.orth lt--
"yi2" meaning _at_Aunt evokes RD as
rd constraints self.m lt--gt
rd.referent self.m lt--gt rd.ontological_category
22you specifies discourse role
construction ni3-N subcase of Morpheme
form constraints self.f.orth lt--
"ni3" meaning _at_Human evokes RD as
rd constraints self.m lt--gt
rd.referent self.m lt--gt rd.ontological_category
rd.discourse_participant_role lt--
_at_Addressee rd.set_size lt-- _at_Singleton
23The meaning of give is a schema with roles
schema Transfer subcase of Action roles giver
_at_Entity recipient _at_Entity theme
_at_Entity constraints giver lt--gt protagonist
construction gei3-V2 subcase of Morpheme
form constraints self.f.orth lt--
"gei3" meaning Give
schema Give subcase of Transfer constraints in
herent_aspect lt-- _at_Inherent_Achievement giver
lt-- _at_Animate recipient lt-- _at_Animate theme lt--
_at_Manipulable_Inanimate_Object
24Finally, you-give-aunt links up the roles
construction ni3-gei3-yi2 subcase of
Finite_Clause constructional constituents n
ni3-N g gei3-V2 y yi2-N form const
raints n.f meets g.f g.f meets y.f meaning
Give constraints self.m lt--gt
g.m self.m.giver lt--gt n.m self.m.recipient
lt--gt y.m
25The learning loop Hypothesize Reorganize
World Knowledge
Utterance
Linguistic Knowledge
reorganize
Analysis
reinforcement
Context Fitting
hypothesize
PartialSemSpec
26If the learner has a ditransitive cxn
Context
Meaning
Form
XIXI
MOT
addressee
speaker
giver
meets
Discourse Segment
meets
recipient
INV
theme
Peach
27Context fitting recovers more relations
Context
Meaning
Form
XIXI
MOT
addressee
speaker
giver
meets
giver
Discourse Segment
Give
recipient
meets
recipient
attentional-focus
INV
theme
theme
Peach
28But the learner does not yet have phrasal cxns
Context
Meaning
Form
XIXI
Give
INV
29Context bootstraps learning
Meaning
Form
construction ni3-gei3-yi2 subcase of
Finite_Clause constructional constituents n
ni3 g gei3 y yi2 form constraints
n.f meets g.f g.f meets y.f meaning
Give constraints self.m lt--gt
g.m self.m.giver lt--gt n.m self.m.recipient
lt--gt y.m
30A model of context is key to learning
- The context model makes it possible for the
learning model to - learn new constructions using contextually
available information - learn argument-structure constructions in
pro-drop languages
31Understanding an utterance in context
Transcripts
Schemas Constructions
Analysis Resolution
Context Fitting
Simulation
Semantic Specification
32Context model Events Utterances
Setting participants, entities, relations
Start
Event
Event
DS
Sub-Event
Sub-Event
33Entities and Relations are instantiated
Setting CHI, MOT (incl. body parts) livingroom (
incl. ground, ceiling, chair, etc), lotion
ds04 admonishing05 speaker MOT addressee
CHI forcefulness normal
caused_motion01 forceful_motion motion
Start
apply02 applier CHI substance lotion surface
face(CHI)
translational_motion03 mover lotion spg SPG
34The context model is updated dynamically
- Extended transcript annotation speech acts
events - Simulator inserts events into context model
updates it with the effects - Some relations persists over time some dont.
Simulation
35Competition-based analyzer finds the best analysis
- An analysis is made up of
- A constructional tree
- A semantic specification
- A set of resolutions
Bill gave Mary the book
36Combined score that determines best-fit
- Syntactic Fit
- Constituency relations
- Combine with preferences on non-local elements
- Conditioned on syntactic context
- Antecedent Fit
- Ability to find referents in the context
- Conditioned on syntactic information, feature
agreement - Semantic Fit
- Semantic bindings for frame roles
- Frame roles fillers are scored
37Context Fitting goes beyond resolution
Context
Meaning
Form
XIXI
MOT
addressee
speaker
giver
meets
giver
Discourse Segment
Give
recipient
meets
recipient
attentional-focus
INV
theme
theme
Peach
38Context Fitting, a.k.a. intention reading
- Context Fitting takes resolution a step further
- considers entire context model, ranked by recency
- considers relations amongst entities
- heuristically fits from top down, e.g.
- discourse-related entities
- complex processes
- simple processes
- other structured and unstructured entities
- more heuristics for future events (e.g. in cases
of commands or suggestions)
39Adult grammar size
- 615 constructions total
- 100 abstract cxns (26 to capture lexical
variants) - 70 phrasal/clausal cxns
- 440 lexical cxns (260 open class)
- 195 schemas (120 open class, 75 closed class)
40Starter learner grammar size
- No grammatical categories (except interjections)
- Lexical items only
- 440 lexical constructions
- 260 open class schema / ontology meanings
- 40 closed class pronouns, negation markers, etc
- 60 function words no meanings
- 195 schemas (120 open class, 75 closed class)
41The process hierarchy defined in schemas
Process
Proto_Transitive
State_Change
State
Action
Complex_Process
Intransitive_State
Two_Participant_State
Serial_Processes
Mental_State
Concurrent_Processes
Cause_Effect
Joint_Motion
Caused_Motion
42The process hierarchy defined in schemas
Action
Motion
Intransitive_Action
Translational_Motion
Expression
Self_Motion
Translational_Self_Motion
Force_Application
Continuous_Force_Application
Forceful_Motion
Agentive_Impact
43The process hierarchy defined in schemas
Action
Communication
Cause_Change
Obtainment
Transfer
Ingestion
Perception
Other_Transitive_Action
44Understanding an utterance in context
Transcripts
reorganize
Schemas Constructions
Analysis Resolution
Context Fitting
Simulation
reinforcement
hypothesize
Semantic Specification
45Hypothesize Reorganize
- Hypothesize
- utterance-driven
- relies on the analysis (SemSpec context)
- operations compose
- Reorganize
- grammar-driven
- can be triggered by usage (to be determined)
- operations generalize
46Composing new constructions
Context
MOT
XIXI
Peach
giver recipient theme
INV
- Compose operation If roles from different
constructions point to the same context element,
propose a new construction and set up a meaning
binding.
47Creating pivot constructions
- Pivot generalization Given a phrasal cxn, look
for another cxn that shares 1 constituents. Line
up roles and bindings. Create new cxn category
for the slot.
48Resulting constructions
construction ni3-gei3-cat01 constituents ni3,
gei3, cat01 meaning Give constraints self.
m.recipient lt--gt g.m
general construction cat01 subcase of
Morpheme meaning _at_Human
construction wo3 subcase of cat01 meaning
_at_Human
construction yi2 subcase of cat01 meaning _at_Aunt
49Pilot Results Sample constructions learned
chi1_fan4 ni3_chuan1_xie2 ni3_shuo1 bu4_na2 wo3_qu4 ni3_ping2zi_gei3_wo3 ni3_gei3_yi2 wo3_bu4_chi1 eat rice you wear shoe you say NEG take I go you bottle give me you give auntI NEG eat
ni3 shuo1, chuan1 ni3 shuo1, hua4 wo3 zhao3, qu4 bu4 na2, he1wo3, ma1 cheng2 you say, wear you say, draw I find, go NEG take, drink I, mom scoop
50Challenge 1 Non-compositional meaning
Context
Bake-Event
baker baked
MOT
CHI
Cake
- Non-compositional meaning Search for additional
meaning schemas (in context or in general) that
relate the meanings of the individual
constructions
51Challenge 2 Function words
Context
Bake-Event
baker baked
MOT
Cake
CHI
- Function words tend to indicate relations rather
than events or entities
52Challenge 3 How far up to generalize
- Eat rice
- Eat apple
- Eat watermelon
- Want rice
- Want apple
- Want chair
Inanimate Object
Manipulable Objects
Unmovable Objects
Food
Furniture
Fruit
Savory
Chair
Sofa
apple
watermelon
rice
53Challenge 4 Beyond pivot constructions
- Pivot constructions indexing on particular
constituent type - Eat rice Eat apple Eat watermelon
- Abstract constructions indexing on role-filler
relations between constituents
Schema Eat roles eater lt--gt agent food lt--gt
patient
Schema Want roles wanter lt--gt agent wanted
lt--gt patient
54Challenge 5 Omissible constituents
- Intuition
- Same context, two expressions that differ by one
constituent ? a general construction with the
constituent being omissible - May require verbatim memory traces of utterances
relevant context
55When does the learning stop?
Bayesian Learning Framework
- Most likely grammar given utterances and context
- The grammar prior is a preference for the kind
of grammar - In practice, take the log and minimize cost ?
Minimum Description Length (MDL)
56Intuition for MDL
- S -gt Give me NP
- NP -gt the book
- NP -gt a book
- S -gt Give me NP
- NP -gt DET book
- DET -gt the
- DET -gt a
Suppose that the prior is inversely proportional
to the size of the grammar (e.g. number of
rules) Its not worthwhile to make this
generalization
57Intuition for MDL
- S -gt Give me NP
- NP -gt the book
- NP -gt a book
- NP -gt the pen
- NP -gt a pen
- NP -gt the pencil
- NP -gt a pencil
- NP -gt the marker
- NP -gt a marker
- S -gt Give me NP
- NP -gt DET N
- DET -gt the
- DET -gt a
- N -gt book
- N -gt pen
- N -gt pencil
- N -gt marker
58How to calculate the prior of this grammar
- (Yet to be determined)
- There is evidence that the lexicalized
constructions do not completely go away - If the more lexicalized constructions are
retained, the size of grammar is a bad indication
of degree of generality