Capturing linguistic interaction in a grammar A method for empirically evaluating

advertisement
Capturing linguistic interaction
in a grammar
A method for empirically evaluating
the grammar of a parsed corpus
Sean Wallis
Survey of English Usage
University College London
s.wallis@ucl.ac.uk
Capturing linguistic interaction...
• Parsed corpus linguistics
• Empirical evaluation of grammar
• Experiments
– Attributive AJPs
– Preverbal AVPs
– Embedded postmodifying clauses
• Conclusions
– Comparing grammars or corpora
– Potential applications
Parsed corpus linguistics
• Several million-word parsed corpora exist
• Each sentence analysed in the form of a tree
– different languages have been analysed
– limited amount of spontaneous speech data
• Commitment to a particular grammar required
– different schemes have been applied
– problems: computational completeness + manual
consistency
• Tools support linguistic research in corpora
Parsed corpus linguistics
• An example tree from ICE-GB (spoken)
S1A-006 #23
Parsed corpus linguistics
• Three kinds of evidence may be obtained
from a parsed corpus
Frequency evidence of a particular known rule,
structure or linguistic event
Coverage evidence of new rules, etc.
Interaction evidence of the relationship
between rules, structures and events
• This evidence is necessarily framed within a
particular grammatical scheme
– So… how might we evaluate this grammar?
Empirical evaluation of grammar
• Many theories, frameworks and grammars
– no agreed evaluation method exists
– linguistics is divided into competing camps
– status of parsed corpora ‘suspect’
• Possible method: retrievability of events




circularity: you get out what you put in
redundancy: ‘improvement’ by mere addition
atomic: based on single events, not pattern
specificity: based on particular phenomena
• New method: retrievability of event sequences
Experiment 1: attributive AJPs
• Adjectives before a noun in English
• Simple idea: plot the frequency of NPs with
at least n = 0, 1, 2, 3… attributive AJPs
Experiment 1: attributive AJPs
• Adjectives before a noun in English
• Simple idea: plot the frequency of NPs with
at least n = 0, 1, 2, 3… attributive AJPs
6. 0000
200, 000
Raw frequency
180, 000
Log frequency
5. 0000
160, 000
140, 000
4. 0000
120, 000
3. 0000
100, 000
80, 000
2. 0000
60, 000
40, 000
1. 0000
20, 000
0. 0000
0
0
1
2
3
4
5
6
0
1
2
3
4
5
6
NB: not a straight line
Experiment 1: analysis of results
• If the log-frequency line is straight
– exponential fall in frequency (constant probability)
– no interaction between decisions (cf. coin tossing)
• Sequential probability analysis
– calculate probability of adding each AJP
– error bars (binomial)
– probability falls
• second < first
• third < second
• fourth < second
– decisions interact
Experiment 1: analysis of results
• If the log-frequency line is straight
– exponential fall in frequency (constant probability)
– no interaction between decisions (cf. coin tossing)
• Sequential probability analysis
– calculate probability of adding each AJP
– error bars (binomial)
probability
– probability falls
0. 25
0. 20
• second < first
• third < second
• fourth < second
– decisions interact
0. 1 5
0. 1 0
0. 05
0. 00
0
1
2
3
4
5
Experiment 1: analysis of results
• If the log-frequency line is straight
– exponential fall in frequency (constant probability)
– no interaction between decisions (cf. coin tossing)
• Sequential probability analysis
–
–
–
–
–
calculate probability of adding each AJP
error bars (binomial)
probability
probability falls
decisions interact
fit to a power law
• y = m.x k
• find m and x
-1.2793
0. 25
0. 20
0. 1 5
0. 1 0
0. 05
y = 0.1931x
0. 00
0
1
2
3
4
5
Experiment 1: explanations?
• Feedback loop: for each successive AJP,
it is more difficult to add a further AJP
– Explanation 1: semantic constraints
• tend to say tall green ship
• do not tend to say tall short ship or green tall ship
– Explanation 2: communicative economy
• once speaker said tall green ship, tends to only say ship
– Further investigation required
• General principle:
– significant change (usually, fall) in probability is
evidence of an interaction along grammatical axis
Experiments 2,3: variations
 Restrict head: common and proper nouns
– Common nouns: similar results
– Proper nouns and adjectives are often treated as
compounds (Northern England vs. lower Loire )
 Ignore grammar: adjective + noun strings
– Some misclassifications / miscounting (‘noise’)
• she was [beautiful, people] said; tall very [green ship]
– Similar results
• slightly weaker (third < second ns at p=0.01)
– Insufficient evidence for grammar
• null hypothesis: simple lexical adjacency
Experiment 4: preverbal AVPs
• Consider adverb phrases before a verb
– Results very different
• Probability does not fall significantly between first and
second AVP
• Probability does fall
between third and
second AVP
– Possible constraints
• (weak) communicative
• not (strong) semantic
– Further investigation
needed
Experiment 4: preverbal AVPs
• Consider adverb phrases before a verb
– Results very different
• Probability does not fall significantly between first and
second AVP
• Probability does fall
between third and
second AVP
0.06
0.05
– Possible constraints
• (weak) communicative
• not (strong) semantic
– Further investigation
needed
– Not power law: R2 < 0.24
0.04
0.03
0.02
0.01
probability
0.00
1
2
3
Experiment 5: embedded clauses
• Another way to specify nouns in English
– add clause after noun to explicate it
• the ship [that was tall and green]
• the ship [in the port]
– may be embedded
• the ship [in the port [with the ancient lighthouse]]
– or successively postmodified
• the ship [in the port][with a very old mast]
• Compare successive embedding and
sequential postmodifying clauses
– Axis = embedding depth / sequence length
Experiment 5: method
• Extract examples with FTFs
– at least n levels of embedded postmodification:
Experiment 5: method
• Extract examples with FTFs
– at least n levels of embedded postmodification:
0
1
2
(etc.)
Experiment 5: method
• Extract examples with FTFs
– at least n levels of embedded postmodification:
0
1
2
(etc.)
– problems:
• multiple matching cases (use ICECUP IV to classify)
• overlapping cases (subtract extra case)
• co-ordination of clauses or NPs (use alternative patterns)
Experiment 5: analysis of results
• Probability of adding a further embedded
clause falls with each level
– second < first
– sequential < embedding
• Embedding only:
– third < first
– insufficient data for
third < second
• Conclusion:
– Interaction along embedding and sequential axes
Experiment 5: analysis of results
• Probability of adding a further embedded
clause falls with each level
– second < first
– sequential < embedding
• Embedding only:
– third < first
– insufficient data for
third < second
• Conclusion:
0.06
embedded
0.05
0.04
0.03
sequential
0.02
0.01
probability
0.00
0
1
2
3
– Interaction along embedding and sequential axes
4
Experiment 5: analysis of results
• Probability of adding a further embedded
clause falls with each level
– second < first
– sequential < embedding
• Fitting to f = m.x k
– k < 0 = fall ( f = m/x |k|)
– |k| is high = steep
• Conclusion:
0.06
embedded
y = 0.0539x -1.2206
0.05
0.04
0.03
0.02
sequential
y = 0.0523x -1.6516
0.01
0.00
0
1
– Both match power law: R2 > 0.99
2
3
4
Experiment 5: explanations?
• Lexical adjacency?
– No: 87% of 2-level cases have at least one VP, NP
or clause between upper and lower heads
• Misclassified cases of embedding?
– No: very few (5%) semantically ambiguous cases
• Language production constraints?
– Possibly, could also be communicative economy
• contrast spontaneous speech with other modes
• Positive ‘proof’ of recursive tree grammar
– Established from parsed corpus
– cf. negative ‘proof’ (NLP parsing problems)
Conclusions
• A new method for evaluating interactions
along grammatical axes
– General purpose, robust, structural
– More abstract than ‘linguistic choice’ experiments
– Depends on a concept of grammatical distance
along an axis, based on the chosen grammar
• Method has philosophical implications
– Grammar viewed as structure of linguistic choices
– Linguistics as an evaluable observational science
• Signature (trace) of language production decisions
– A unification of theoretical and corpus linguistics?
Comparing grammars or corpora
• Can we reliably retrieve known interaction
patterns with different grammars?
– Do these patterns differ across corpora?
• Benefits over individual event retrieval
 non-circular: generalisation across local syntax
 not subject to redundancy: arbitrary terms
makes trends more difficult to retrieve
 not atomic: based on patterns of interaction
 general: patterns may have multiple explanations
• Supplements retrieval of events
Potential applications
• Corpus linguistics
– Optimising existing grammar
• e.g. co-ordination, compound nouns
• Theoretical linguistics
– Comparing different grammars, same language
– Comparing different languages or periods
• Psycholinguistics
– Search for evidence of language production
constraints in spontaneous speech corpora
• speech and language therapy
• language acquisition and development
Links and further reading
• Survey of English Usage
– www.ucl.ac.uk/english-usage
• Corpora and grammar
– .../projects/ice-gb
• Full paper
– .../staff/sean/resources/analysing-grammaticalinteraction.pdf
• Sequential analysis spreadsheet (Excel)
– .../staff/sean/resources/interaction-trends.xls
Download