Professional Documents
Culture Documents
DOI: 10.1007/s10726-006-9024-z
C Springer 2006
MOHAK SHAH
School of Information Technology and Engineering, University of Ottawa, Ottawa, Canada
(E-mail: mshah@site.uottawa.ca)
STAN SZPAKOWICZ
School of Information Technology and Engineering, University of Ottawa, Ottawa, Canada;
Institute of Computer Science, Polish Academy of Sciences, Warsaw, Poland
(E-mail: szpak@site.uottawa.ca)
Abstract
Various combination of Natural Language Processing and Machine Learning methods offer ample opportunities
wherever texts are an important element of an application or a research area. Such methods discover patterns and
regularities in the data, seek generalization and in effect learn new knowledge. We have employed such methods
in learning from a large amount of textual data. Our application is electronic negotiations. The genre of texts
found in electronic negotiations may seem limited. It is an important research question whether our methods and
findings apply equally well to texts that come from face-to-face negotiations. In order to confirm such more general
applicability, we have analyzed comparable collections of texts from electronic and face-to-face negotiations. We
present our findings on the extent of similarity between these two related but distinct genres. In this study we have
analyzed similarities in the text data of electronic and face-to-face negotiations. The results show that in certain
conditions vocabulary richness, language complexity and text predictability are similar.
Key words: Electronic negotiations, face-to-face negotiations, communication process, text data, corpus analysis
1. Introduction
Natural Language Processing (NLP) and Machine Learning (ML) techniques have been
applied in a wide variety of fields. They are, however, a new tool for the study of negotiation
or, more specifically, texts found in negotiation. NLP and ML methods work well with large
amounts of textual data. This makes them desirable research tools in the fields where the
volume of information to process grows rapidly. Bioinformatics and medicine are among
This is an expanded version of a paper published in the Proceedings of FINEXIN 2005 (Workshop on the Analysis
of Formal and Informal Information Exchange during Negotiations), 3142, Ottawa, Canada, May 2005.
128
the recent successful applications of NLP and ML that lead to the discovery of useful and
interesting patterns.
Learning from large amounts of negotiation-related information is becoming a necessity: we note a substantial increase of the volume of electronic negotiations (e-negotiations).
Traditionally, negotiation has been understood as interaction between two negotiating parties with at least oral contact (face-to-face, by telephone, and so on). The pervasiveness of
information technology, however, has made e-negotiations a new reality. They take advantage of electronic means, not necessarily limited to email or the Internet. They also include
other electronic assistance such as negotiation support systems or online analysis software
(Schoop, 2003).
E-negotiations restrict communication between the participants to the exchange of formal
or informal messages. Formal messages represent offers, positions and preferences unambiguously, but tend to mask the more interesting aspects of the negotiators personality.
Informal messages are unrestricted texts in which negotiators confirm decisions reflected
in a formal representation, and may also establish a personal relationship. When informal
messages are allowed, their language may affect the negotiation outcome, strategies that
participants use, and roles in negotiations. These results come from analyzing the data
gathered in e-negotiations (Sokolova et al. 2005b).
We want to keep our ML/NLP toolkit general and to apply it to problems other than
electronic negotiations. We study the data of face-to-face negotiations. We must identify
the differences and similarities between the two types of negotiation processes, and find out
how this affects the language in textual data. If the data of face-to-face negotiations exhibit
the characteristics similar to those of e-negotiations data, we can expect that the learning
methods applied to e-negotiation data will also work with face-to-face negotiation data.
We seek methods that capture similarities between the text data of face-to-face and electronic negotiations. We study the richness of the vocabulary, word distribution, predictability
of data and common sets of features. We have applied statistical analysis and modelling.
We show that these methods, if based on word frequencies, find common characteristics
in the language of negotiations. We apply the statistical analysis and modelling methods
to the language in face-to-face negotiations, and compare the results with those for the
corresponding e-negotiation data.
The face-to-face textual data that we analyze here come from the Cartoon negotiations
(Brett, 2001). The texts are part of manual transcripts with non-textual elements removed
(those are the elements that pertain to body language, gestures, pauses and so on). We
compare these data with the textual e-negotiation data collected by the Inspire system
(Kersten and Zhang, 2003). Both data sets arose in the context of the training of negotiators,
but they find good use in research. In addition to showing that the two domains resemble
each other closely with respect to word distribution, we look at whether process-specific
information can help characterize them, given their statistical similarity.
The remainder of the paper is organized as follows. Section 2 reviews elements of
negotiations in general and introduces the data from the Cartoon and Inspire negotiations.
Section 3 reports the results of the data analysis. Section 4 presents the role of the processspecific information in the two contexts and explores its distribution. In the concluding
Section 5 we emphasize the central aspects of our work and suggest possible future work.
129
2. Problem Statement
In this work, we focus on the data gathered in business negotiations. Rojot (1991) defines
negotiation as a process where two or more parties aim to settle what each shall give and
take in a transaction between them.
The Cartoon and Inspire negotiations are training negotiations. Both represent the class of
bilateral integrative business negotiations. One participant is a buyer and the other is a seller.
The goal of a negotiation is to make a virtual purchase. A negotiation is successful if the
virtual purchase has taken place. The language of communication during these negotiations
is English.
Despite the similarities between the Cartoon and Inspire negotiations, there are substantial differences in their setting. The Cartoon negotiations are conducted face-to-face (Brett,
2001). The issue is the virtual purchase of a TV program. The whole negotiation process
takes place during one uninterrupted meeting. The Inspire system supports e-negotiations
(Kersten, 2003). The issue is the virtual purchase of bicycle parts. The process lasts up
to three weeks, during which participants exchange freely written messages and formal
numerical offers. Table 1 compares the key elements of Cartoon and Inspire negotiations.
We have access to the tape record transcripts1 of 40 Cartoon negotiations, all of them
successful. 20 negotiations were held between a Japanese buyer and a Japanese seller (the
Cartoon Japanese data), the other 20 between a US buyer and a US seller (the Cartoon US
data). The text data contain 157,253 word tokens distributed among 4,244 word types, with
1686 rare word types, appearing only once in the data (Oakes, 1998). As to the Inspire data,
in the experiments described in this paper we have worked with all text messages exchanged
in 1427 successful negotiations. Negotiations are intercultural: the participants come from
50 countries. The text data contain 1,073,398 word tokens distributed among 25,085 word
types, with 12,897 rare word types. A more detailed analysis of the Inspire language data
appears in (Sokolova et al., 2005a).
This comparison highlights the differences between the Cartoon data and the successful Inspire negotiation data. One is a small corpus containing recorded, transcribed and
Cartoon
Inspire
Environment
Means
Topic
Goal
Rules
Communication
# of negotiators
Roles
Protocol
Business
Face-to-face meeting
Purchase of a TV program
To make a purchase
Business
Electronic system
Purchase of goods
To make a purchase
Synchronous
two
buyer and seller
imposed by Cartoon
Asynchronous
two
buyer and seller
imposed by Inspire
130
post-edited speech data, the other is a relatively large corpus of unedited written texts. We
seek the similarities between the Cartoon data and the successful Inspire negotiations data.
We are interested in similarities due to the common origin of the data communication
during a negotiation and to the process characteristics business, bilateral, integrative,
and successful.
Our research problem is to identify a class of methods that will detect this type of
similarities. In the following sections we present such methods of revealing similarities
between two data sets. It is worth noting that finding similarities between corpora is one of
the more complex tasks in corpus analysis (Kilgarriff, 2001).
3. Data Analysis
We first employ corpus-linguistic methods to analyze and compare corpora. Section 3.1 explores lexical richness by studying vocabulary growth, homogeneity and variety of words
and simple expressions in both corpora. It also investigates how two corpora are close
statistically. Section 3.2 models the Cartoon and the Inspire data to evaluate their lexical complexity and predictability. The reliability of word frequencies makes the results a
trustworthy measure of corpus comparison.
V (1, N )
;
N
(1)
V (2, N )
N
(2)
P(N ) is a common measure of the richness of the vocabulary of a text (Sichel, 1986). We
also calculate the ratio of bigrams bi(N )
bi(N ) =
# of bigrams
N
(3)
131
Inspire-successful
T T (N )
0.027
0.023
P(N )
S(N )
bi(N )
tri(N )
0.4
0.076
0.289
0.682
0.5
0.002
0.198
0.486
Statistics
# of trigrams
N
(4)
(5)
132
Cartoon
US
Cartoon
All
Insp-succ
Dialogues
Brown
WSJ
i
you
to
the
to
that
the
to
you
to
i
you
i
you
and
the
of
and
the
of
to
the
million
runs
a
you
i
and
we
i
that
and
we
the
we
and
a
the
to
ah
a
to
a
in
that
a
and
in
that
for
we
a
of
a
of
your
offer
it
in
is
was
for
one
and
is
is
for
know
he
is
Cartoon Japanese
Cartoon US
Insp-succ
Dialogue
Cartoon Japanese
Cartoon US
0.45
0.45
0.53
0.57
0.42
0.39
N is the number of words, and d is the difference in the ranks of the same word in two
corpora (Oakes, 1998). SRCC shows accurate results in comparing corpora of different sizes
(superior to other methods) and is a reliable indicator of the corpus similarity (Kilgarriff,
2001). We avoid SRCCs bias towards low-frequency words it tends to overemphasize
such words by calculating it on the most frequent words where its performance is balanced.
Spearmans coefficient, calculated on the most common words, has shown that the
Japanese data are closer to the Dialogue data than to the Cartoon US data. Same applies
to the US data: they are closer to the Dialogue data than to the Cartoon Japanese data; see
Table 4.
The results show that similarities between the Cartoon and the Inspire data are statistically
significant with the 0.80 significance level.
133
Inspire negotiation data instead of working with the whole set. That is because statistical
modelling results depend on the size of the data set (Chen and Goodman, 1996).
3.2.1. Sampling
We consider the size in tokens and the number of negotiations as the main parameters for
choosing samples. The average length of a successful Inspire negotiation is 750 tokens.
We have found a subclass of the Inspire negotiations where the average negotiation length
is 1,160 words. Both average lengths are closer to the Cartoon Japanese negotiations in
which the average length is 1,340 than to the Cartoon US negotiations with 6,250 tokens on
average. We construct two samples from the successful Inspire negotiation data. InspGen is
selected from all successful Inspire negotiations. InspEmail is selected from longer successful negotiations (as measured by the number of word tokens) where negotiators exchange
personal email addresses. For both samples we choose negotiations uniformly distributed
in the original population.
The results show that the number of tokens and the number of negotiations of the samples
correspond to those of the Cartoon Japanese data and differ from those of Cartoon US data,
while the word distribution preserves characteristics of the successful Inspire negotiation
data; see Table 5.
We calculate the sample statistics; see Table 6. The size dependencies of S(N ), bi(N ),
tri(N ) are clearly seen if we compare the new results with the results reported in Table 2.
Finally, we report 10 most frequent words from the samples; see Table 7.
The results confirm that the Cartoon corpus is heterogenous whereas the Inspiresuccessful corpus is homogenous.
The most frequent words show that the common type of texts the negotiation messages
makes the Cartoon and Inspire corpora similar.
Table 5. Word distribution in four collections.
Cartoon Japanese
Characteristics
# of negotiations
N
T
InspEmail
InspGen
Cartoon US
20
25
33
20
26795
1412
26987
2764
27218
2539
130458
3702
534
194
1333
390
1180
361
1357
519
V(1,N)
V(2,N)
Cartoon Japanese
InspEmail
InspGen
Cartoon US
T T (N )
P(N )
S(N )
0.053
0.020
0.007
0.102
0.049
0.014
0.093
0.043
0.013
0.028
0.011
0.003
bi(N )
tri(N )
0.377
0.714
0.510
0.827
0.337
0.529
0.303
0.699
134
Cartoon Japanese
InspEmail
InspGen
rank
Cartoon Japanese
InspEmail
InspGen
you
runs
and
2
3
4
5
you
to
the
million
the
to
you
and
to
i
the
a
7
8
9
10
a
for
we
and
we
is
for
your
we
for
your
offer
3.2.2. Modelling
The goal of statistical modelling is to find a way of approximating the data well. For textual
data, a model predicts which words, and with what probabilities, can appear after a given
sequence of words. The criterion of the models quality is the value of cross-entropy:
n
1
log(P(wi ))
n i=0
(6)
n is the number of words in the test set, P(wi ) is the probability (calculated by the model)
of the appearance of the word wi in the test data (Oakes, 1998). A models cross-entropy
is calculated when it is built from training data, or training and validation data, and tested
on previously unseen test data (Chen and Goodman, 1996). A model with the lower crossentropy models the data better. With respect to language, cross-entropy is one of the estimators of its complexity or predictability. Lower cross-entropy means that the data are
predictable, higher cross-entropy indicates more uncertainty in the data.
For text data, N -gram models are very widely used. An N -gram is a sequence of N
consecutive words w1 , . . . , w N . An N -gram model approximates the probability of the
appearance of word w N after the sequence of words w1 , . . . , w N 1 . This is expressed as
conditional probability P(w N |w1N 1 ). The approximation is as follows:
1
P w N |w1N 1 P w N |w NN k
The values of k are usually 1, 2 and 3. N -gram models for N > 4 almost do not appear
in practice, quadrigram models are rare, and trigram models tend to slightly outperform
bigram models. The probabilities of N -grams are calculated on training data and then the
same probability applies to each N -gram appearing in the test data. In our study we employ
the Kneser-Ney and Good-Turing trigram models; for details and formulae see (Chen and
Goodman, 1996; Katz, 1987). Here is why we use these models: their performance is high
compared with other models; they require only training and test data; they are easy to
implement and to apply, for example, have no adjustable parameters.
From the language point of view, the models explore different data properties. The
Kneser-Ney probability of a word depends on the number of preceding word types, that
is, the number of different words. A previously unseen trigram may contain a collocation
135
Cartoon Japan.
InspEmail
InspGen
Cartoon US
GTK
10.06
11.8
11.04
12.21
KN
6.67
7.15
7.27
7.79
(a noticeable arrangement of adjacent words). This often happens in negotiation data with
their restricted topic. The Kneser-Ney model treats such trigrams differently from trigrams
that do not contain collocations. On the other hand, the Good-Turing model calculates the
frequency of N-grams in the training set, so it heavily depends on the sizes of the training
and test sets.
An important concern in such modelling is smoothing that helps account in advance for
previously unseen N -grams. The Kneser-Ney model has smoothing built-in: the probability
of an unseen N -gram is substituted by the probability of the (N 1)-gram. The probability
of a previously unseen word is N1 , where N is the number of tokens. For the Good-Turing
model we use Katz smoothing that generally performs well on data with a high token-type
ratio (Katz, 1987).
We employ five-fold cross-validation to train and test statistical models. In K -fold cross
validation, the examples are randomly divided into K mutually disjoint sets of approximately
equal sizes. The concept is learned from the examples in K 1 sets, and is tested on the
examples from the remaining set. This is repeated K times, once for each set (that is, each set
is used once as a test set). The average error rate over all K sets is the cross-validated error
rate. Table 8 reports the average five-fold cross-validation results for all the experiments.
GTK stands for the Good-Turing model with Katz smoothing, KN for the Kneser-Ney
model.
A cross-entropy reduction of more than 0.16 is noteworthy (Rosenfeld, 2000). The results
show that the predictability of the Cartoon Japanese data is closer to the predictability of the
Inspire samples than to the predictability of the Cartoon US data. At the same time, for all
four data sets the KN model is significantly better than the GTK model. This corresponds
to the KN and GTK performance results that Chen and Goodman (1996) reported for the
Switchboard corpus. We cite the results here because the Switchboard corpus contains
transcripts of spoken dialogues between a telephone operator and a client. While it is not a
business negotiation, such conversation sometimes is referred to as collaborative negotiation
(Chu-Carroll and Carberry, 2000).
Because the models depend on the sizes of the data sets, we can explain this by fact that
the KN model performs better on small data than the GTK model. Another conclusion is
related to the smoothing used in the KN model: in our data the number of different words
preceding word wk is a better indicator of wk than the N -gram frequencies.
The following question arises naturally: how well can one data set predict a word from
another set? Put differently, how closely can the Cartoon and Inspire samples approximate
each other with respect to the models? To answer this question, we have built cross-models
of the data for Cartoon Japanese and US samples and InspEmail and InspGen samples. For
cross-modelling we use a training set from one data class and a test set from another data
136
Model
Cartoon Japan
vs
Cartoon US
InspEmail
vs
InspGen
InspGen
vs
InspEmail
Cartoon US
vs
Cartoon Japan
GTK
KN
9.84
7.35
11.33
7.14
11.49
7.39
12.30
7.62
class. Table 9 presents the results of cross-modelling. X vs Y means that X is a training set
and Y is a test set. To make a fair comparison we use the same training and test sets as in
the previous modelling experiments.
Prediction of a word becomes harder when a model is trained on InspGen and Cartoon
US data and tested on InspEmail and Cartoon Japanese data respectively. The same holds
for a KN model trained on Cartoon Japanese and tested on Cartoon US data. The GTK
model, however, improves its fitness for Cartoon US data when trained on Cartoon Japanese
data. These results are intriguing but inconclusive. Their confirmation and interpretation
are a matter of future work.
Other improvements of the model fitness are due to the sample data. Cross-entropy
reported in InspEmail vs InspGen is smaller than in InspEmail. Note that the same training
data were used in both cases. The difference in cross-entropy results signals that the InspGen
test set has less unseen data than the InspEmail test set. This supports the previously stated
hypothesis that people use diversified vocabulary when they exchange personal information.
Having shown a considerable similarity of the Cartoon and the Inspire data, we now
investigate whether process-specific information can be used to represent and characterize
them.
137
through electronic means. On the other hand, the largest available collection of negotiation
texts consists merely of some 2500 negotiations. The number of available examples is
small compared with the number of features word types. This makes it necessary to
look for a much smaller number of features that can still capture the most important data
characteristics. Process-specific information helps do just that. We show that when it is used,
our textual e-negotiation data can be represented by merely 124 word types/features. If this
can be generalized, we will indeed be able to deal successfully with negotiation data at the
process level. In order to obtain process-specific information, we apply a semi-automatic
procedure discussed (Shah et al., 2004). Very briefly, the procedure takes a set of texts and
a communication model (Hargie and Dickson, 2004); identifies process-specific elements
of the model; builds a lexicon and the bigram and trigram models of the data; performs
simple word-sense disambiguation; constructs a set of semantic categories, among them
process-specific categories, and tags all word types in the lexicon with these categories;
finally, represents a text using only word types tagged as process-specific.
We have found the distribution of the process-specific words in the Cartoon data. In order
to obtain process-specific information, we use the same procedure as described before. The
results support those produced from the Inspire data; see Table 10. Table 11 reports the most
frequent bigrams used in the process-specific context to support that the process-specific
words play important role in the negotiation data. Recall that usually the most frequent
bigrams are prepositions, determiners, and words such as pronouns or interjections.
Here are a few detailed findings after applying the procedure to our data. A noun appears
among 10 most frequent unigrams in the Cartoon Japanese data usually, all top unigrams
are function words. The noun million is specific to the Cartoon negotiations. The most
Table 10. Percentage of process-specific words in 100 most
frequent unigrams (excluding function words).
Cartoon Japan
Cartoon US
Cartoon All
Insp-succ
64.8
51.1
56.6
57.9
Cartoon US
Cartoon All
InspEmail
InspGen
Insp-succ
5 million
want to
ultra ranger
i see
down payment
8 runs
the program
ok
if you
how about
you know
i think
going to
per episode
of the
okay. . .
we can
if we
i mean
would be
you know
i think
per episode
going to
we can
of the
okay. . .
want to
if we
in the
i am
the price
i think
if you
for you
in the
we are
and i
i hope
that we
i am
your offer
for you
the price
we are
i have
thank you
if you
and i
i hope
i am
we are
your offer
the price
i hope
for your
we can
that we
if you
i have
138
frequent noun in the Cartoon US data appears among the top 40 words. The noun episode
is also specific to the Cartoon negotiations.
Inspire
hapax legomena
dis legomena
unigram token ratio
0.4
0.076
0.027
0.5
0.002
0.023
0.289
0.682
6.98
11.12
significantly
58.0
0.198
0.486
7.47
11.42
similar
57.3
Method
Corpus characteristics
Measure
Analysis
lexical richness
Modelling
lexical complexity
Comparison
Feature selection
139
This knowledge can in the future be useful in assisting negotiation support systems, since
with time more and more negotiations will take place online. In addition, the approach is
independent of the particular electronic means that are employed for electronic negotiations.
This allows the results to be generalized beyond some specific negotiation support system
or other electronic assistance. Finally, representations that use process-specific information
have the potential to be extended to other types of human communication.
Acknowledgments
This work has been partially supported by the Social Sciences and Humanities Research
Council of Canada and by the Natural Sciences and Engineering Research Council of
Canada. We thank Jeanne M. Brett and Wendi L. Adair for making the Cartoon data
available.
Notes
1. Some transcripts skip the introductory part of a negotiation.
2. The Modelling and Feature Selection sections average results from Table 8.
References
Allen, D. E. and Guy, R. F. (1974). Conversation Analysis: The Sociology of Talk, Mouton, The Hague.
Brett, J. M. (2001). Negotiating Globally, Jossey-Bass, San Francisco.
Francis, W. N. and Kucera, H. (1979). Brown Corpus Manual, http://helmer.aksis.uib.no/icame/brown/bcm.
html.
Cellich, C. and Jain, S. C. (2004). Global Business Negotiations: A Practical Guide, Thomson, South-Western.
Chen, S. F. and Goodman, J. (1996). An Empirical Study of Smoothing Techniques for Language Modeling,
Proceedings of the 34th Annual Meeting of the Association for Computational Linguistics, 310318.
Chu-Carroll, J. and Carberry, S. (2000). Conflict Resolution in Collaborative Planning Dialogues, International
Journal of Human-Computer Studies 53(6), 9691015.
Hargie, O. and Dickson, D. (2004). Skilled Interpersonal Communication: Research, Theory and Practice,
Routledge.
Katz, S. M. (1987). Estimation of Probabilities from Sparse Data for the Language Model Component of a Speech
Recognizer, IEEE Transactions on Acoustics, Speech and Signal Processing (ASSP) 35(3), 400401.
Kersten, G. E. (2003). The Science and Engineering of E-negotiation: An Introduction, Proceedings of 36th
Hawaii International Conference on System Sciences (HICSS 2003), 27 p.
Kersten, G. E. and Zhang, G. (2003). Mining Inspire Data for the Determinants of Successful Internet Negotiations, Central European Journal of Operational Research 11(3), 297316.
Kilgarriff, A. (2001). Comparing Corpora, International Journal of Corpus Linguistics 6(1), 97133.
Oakes, M. P. (1998). Statistics for Corpus Linguistics, Edinburg University Press.
Paul, D. B. and Baker, J. M. (1992). The Design for the Wall Street Journal-Based CSR Corpus, Proceedings of
ICSLP92, 357361.
Rojot, J. (1991). Negotiation: From Theory to Practice, Macmillan Academic and Professional.
Rosenfeld, R. (2000). Two Decades of Statistical Language Modeling: Where Do We Go from There? Proceedings of IEEE 88, 12701278.
140
Schoop, M. (2003). A Language-Action Approach to Electronic Negotiations, Proceedings of LAP 2003, 143
160.
Sebastiani, F. (2002). Machine Learning in Automated Text Categorization, ACM Computing Surveys, 147.
Sichel, H. (1986). Word Frequency Distributions and Type-Token Characteristics, Mathematical Scientist 11,
4572.
Shah, M., Sokolova, M., and Szpakowicz, S. (2004). The Role of Domain-Specific Knowledge in Classifying the
Language of E-negotiations, Proceedings of 3rd International Conference on Natural Language Processing
(ICON 2004), Allied Publishers, India, 99108.
Sokolova, M., Nastase, V., Szpakowicz, S., and Shah, M. (2005a). Analysis and Models of Language in Electronic
Negotiations, in M. Draminski, P. Grzegorzewski, K. Trojanowski and S. Zadrozny (Eds.) Issues in Intelligent
Systems, Models and Techniques, EXIT, Warszawa, 197211.
Sokolova, M., Nastase, V., Shah, M., and Szpakowicz, S. (2005b). Feature Selection in Electronic Negotiation
Texts, Proceedings of Recent Advances in Natural Language Processing (RANLP2005), 518524.
Strobel, M. (2000). Effects of Electronic Markets on Negotiation Processes, Proceedings of ECIS 2000, 445452.
Thomas, L., Pfister, H. P., and Peterson, P. (2004). Issues Related to the Construction of a Purpose-Built DomainSpecific Word Corpus, Australian Journal of Educational and Developmental Psychology 4, 1328.
Thompson, L. and Nadler, J. (2002). Negotiating Via Information Technology: Theory and Application, Journal
of Social Issues 58(1), 109124.
Tweedie, F. J. and Baayen, R. H. (1998). How Variable May a Constant be? Measures of Lexical Richness in
Perspective, Computers and the Humanities 32, 323352.