Abstract
Several frequency-based measures are influenced by corpus size (e.g. lexical diversity or text similarity measures). It is largely unquestioned, however, that normalised frequencies correct for the influence of corpus size – but it has not yet been systematically tested whether and how they might be influenced by corpus size themselves. The central question is whether the normalised frequency of an element in a smaller corpus can be meaningfully compared to the normalised frequency of the same element in a larger corpus. We are testing the association between lists of normalised frequencies derived from corpus samples of different sizes from six languages. Our results suggest that the size of the underlying corpora does not negatively influence comparisons of normalised frequency lists, i.e. different corpus sizes do not lead to normalised frequencies no longer being comparable. For lower-frequency types, these associations decrease rather quickly. These empirical findings converge with predictions from statistical theory.
1 Introduction
One of the most important measures in corpus linguistics is the frequency of occurrence of lexical elements. Sometimes, these are compiled into frequency lists, for example, to find “‘important’ words” (cf. Laurence 2021: 108), in whatever sense, in a corpus. Of course, this is not restricted to single words but can be expanded to parts of words or multi-word units.
When we compare the frequency of occurrence of several elements over several (sub-)corpora, the absolute frequency of occurrence might not be enough to state that “the element observed more often is more frequent because the observed frequencies are of course dependent on the sizes of the corpus parts that are compared” (Gries 2010: 271). In this case, we have to compute normalised (sometimes also called relative) frequencies. These are typically reported as occurrences in one thousand or one million tokens by multiplying the quotient between raw (or absolute) frequency fraw and corpus size s (number of tokens) by the normalisation quantity c: fnorm = fraw/s*c. Just like the compilation of lists of raw frequencies, the normalisation of frequencies can be considered standard corpus linguistic practice: “frequencies must be normalised before the lists can be compared directly” (Baron et al. 2009).
It is therefore not surprising that not only normalised frequencies per se are omnipresent in corpus linguistic research,[1] but that the comparison of frequency values, often of the same linguistic element in (sub-)corpora of differing size, is relevant in many studies. Gries (2010) uses this is as an introductory example when he compares the normalised frequencies of give and bring across the spoken and written parts of the International Corpus of English (British component, ICE-GB). One of the most prominent examples is what Michel et al. (2011) call “usage frequency” when they normalise the absolute number of n-grams by the total number of words in the Google Books corpus in that year. They go on to compare these usage frequencies for single words (e.g., slavery) over the years as well as comparing usage frequencies of alternative expressions (e.g., the Great War, World War I) diachronically. Other comparisons of normalised frequencies between (sub-)corpora of different sizes include, for example, stylometric analyses for two columnists of the Daily Telegraph (Grieve 2023) and the study of changes in the use of the epistemic stance markers of modal verbs in climate science over time (Poole and Hayes 2023). In a discourse study of media language in the Egyptian revolution of 2011, Attia and Romero-Trillo (2022) compare, i.a., the frequency of specific keywords (and their collocations) in three corpora of different size (Arabic and English versions of Al Jazeera and Al Arabiya as well as BBC and CNN). Probabilities of non-lexical items are also being compared by calculating normalised frequencies, e.g. laughter in a hospital setting (Macqueen et al. 2024). Here, the sub-corpora are defined by the participant’s role in the interactions (patient, nurse, researcher, doctor etc.), and because the different roles contribute differently to the overall corpus, the role-based sub-corpora also differ in size. The great diversity of these examples already suggests how widespread comparisons of normalised frequencies are. The logic is always the same: the probability of occurrence of the same or different elements is compared across sub-corpora of different sizes.
However, it is also a well-known fact in corpus linguistics that most, if not all, quantities vary systematically with corpus size. Measures that have been proposed to describe lexical richness of texts “change systematically with the text length” (Tweedie and Baayen 1998: 334). There are also some interesting differences between theoretical constancy, i.e. their mathematical properties given the assumption “that words are used randomly and independently in texts” (Tweedie and Baayen 1998: 332) and empirical constancy, i.e. their actual behaviour in coherent prose. Also, there is an effect of corpus size on the efficiency of a frequency threshold when extracting lexical bundles from corpora when these thresholds “are expressed in normalised frequency” (Bestgen 2018: 205). Furthermore, text similarity measures based on generalised entropies “depend heavily on the sample size” (Koplenig et al. 2019: 1). In the field of psycholinguistics, Burgess and Livesay (1998) show that the predictive power of word frequency measures grows with the corpus size when predicting reaction times for low- and medium-frequency words in a word recognition study.
These are just a few examples where the size of a corpus can have a decisive influence on quantitative measures that are meant to describe (some property of) a corpus or the linguistic material it contains. We therefore wondered whether corpus size might have an even more fundamental influence, namely on normalised frequencies themselves. In this article, we will not specifically evaluate the mathematical properties of normalised frequencies. Rather, we will use corpus data to evaluate the empirical effect(s) of corpus size on normalised frequencies. The basic approach we will use is the pairwise comparison of frequency lists. Each of the two lists originates from a different corpus sample, possibly with a different size (measured in number of included sentences). Overall, we are trying to create a situation that could be considered “optimal” for the calculation and comparison of normalised frequencies, both in terms of the corpus basis (as far as we were able to) and the sampling process. Among other things, we exclude any influences from the text/document level (but not from the sentence level) by using a corpus with scrambled sentences. This also means that we deliberately ignore the internal structure of corpora. As a result, we largely exclude phenomena such as “clumpiness” or “burstiness” (Altmann et al. 2009; Kilgarriff 2001: 241) from the analyses presented here. In addition, we keep the text type as constant as possible by only including generic web corpora for the languages under investigation. In this way, we can be sure that any effects we might find can be attributed solely to corpus size.
We base our analyses on six comparison measures and corpora for six typologically diverse languages and show that corpus size (measured as the number of included sentences) does not have a negative influence on the comparison of normalised frequencies. Rather, we can show that whenever a larger sample is involved in the comparisons, the association between the lists increases. We also show, though, that the associations systematically decrease for less frequent wordform types.
The remainder of this paper is structured as follows. In Section 2, we will introduce some basic concepts of statistical theory regarding our research question and will formulate two predictions. In Section 3.1, we describe the corpora under investigation and how we pre-processed the data. Section 3.2 describes the sampling process and how we arrived at the sets of wordform types we used for the frequency lists. In Section 3.3, we present the comparison measures we use for measuring the association between frequency lists and how we set up the comparison regimes for lower-frequency wordform types. In the following sections, we first present the results for full (Section 4.1) and truncated (Section 4.2) frequency lists. We discuss the results in Section 5 before wrapping up with the implications of our results (Section 6).
2 Statistical theory
As indicated above,[2] we are trying to create a situation that is “optimal” for calculating and comparing normalised frequencies. Using random samples of sentences is consistent with the random sampling assumption from inferential statistics, at least for the text type (generic web corpora) we have used in this study. Given this assumption, it is further assumed that normalised frequencies are unbiased estimates of the occurrence probability of type i, written as
In the current study, we manipulate sample size s directly via drawing varying amounts of sentences from the base corpora (see Section 3.2). Assuming a binomial sampling distribution, the standard deviation of the expected probability of occurrence of type i is given by the square root of the sampling variance.
Hence, we expect the sampling variation to decrease when the sample size increases. In other words, bigger samples should yield more precise estimates of the probability of occurrence of a specific wordform type.
Furthermore, we can calculate the relative sampling variation in analogy to the coefficient of variation (CV), i.e. the sampling variance relative to p i . For this, we divide the standard deviation given in eq. (1) by the expected probability of occurrence:
CV i increases when the expected probability of occurrence decreases and vice versa. We must therefore expect that normalised frequency values for low-frequency wordform types are more “unstable” than for high-frequency types. We therefore arrive at two predictions from statistical theory.
The more sentences we sample from the base corpora, the more precise the normalised frequency values should be, as indicated by Equation (1).
As normalised frequencies decrease, precision should also decrease. Or, in other words, normalised frequency values for low-frequency types are less stable than for high-frequency types, as indicated by Equation (2).
The remainder of the paper can also be seen as an empirical test of these theoretical predictions.
3 Methods
3.1 Data
We use language data from the Leipzig Corpora Collection (Goldhahn et al. 2012). The largest corpora freely available for download on the website[3] contain one million sentences. As we explain in Section 3.2, we need larger corpora for our analyses. We therefore contacted the researchers at the Wortschatz Leipzig (WSL) project directly, who provided us with corpora containing ten million sentences per language.[4] For Chinese (ISO code zho[5]), English (eng), Finnish (fin), French (fra), German (deu), and Vietnamese (vie), ten million randomly shuffled sentences were extracted from generic web corpora of the respective language. In selecting the languages, we were guided on the one hand by the availability of large corpora in the WSL project. English is still the predominant language in corpus linguistics and therefore seemed like the natural choice. While English is fundamentally a Germanic language, much of its vocabulary stems from the Romance languages. Therefore, we chose German and French as additional languages. Chinese and Finnish are seen as two extremes on the analytic-agglutinative continuum. To represent another language family, we chose Vietnamese.
We tokenised and part-of-speech-tagged each corpus with UDPipe, using the R (R Core Team 2024) package {udpipe} (Wijffels 2022) with ready-made models for each of the languages. We excluded all wordforms that span several tokens in the UDPipe tokenisation because the tokenisation process adds the non-contracted elements, e.g., “zu” (Engl. to) and “dem” (Engl. the) are being added for German “zum” and we exclude the “zum”. In French, for example “à” (Engl. at/in/to) and “le” (Engl. the) are being added for “au” and we exclude the “au”.[6] Since the number of sentences is kept constant and sentences differ in length, the number of overall tokens (including punctuation) in the base corpora differs between languages (Chinese: 369,434,986 tokens; English: 227,654,199 tokens; Finnish: 133,665,750 tokens; French: 231,398,746 tokens; German: 182,120,095 tokens; Vietnamese: 211,459,440 tokens).
3.2 Sampling process
For each language, we sampled 100,000 (henceforth “100k”), 500,000 (“500k”), and one million sentences (“1M”). Each size was sampled five times. Whenever a sentence was sampled from the overall corpus of ten million sentences, we excluded this sentence from further sampling. Hence, all 15 samples (five 100k samples, five 500k samples and five 1M samples) are mutually exclusive, i.e. none of the sentences appears twice in the final datasets. This is also the reason why we needed ten million sentences for each language (technically, an overall corpus of 5 × 100,000 + 5 × 500,000 + 5 × 1,000,000 = 8,000,000 sentences would have sufficed but we wanted to allow for some of the sentences not to be sampled at all).
Since we are mainly interested in whether the normalised frequency of a given word form type is affected by corpus size, we need to make sure that this type is actually present in all comparison corpora (= samples and sizes). So, for each language, we computed the set of wordform types that appeared in all 15 samples (henceforth “overlapping types”). The numbers of overlapping types are as follows for each of the six languages: Chinese: 5,086 types; English: 35,699 types; Finnish: 52,345 types; French: 39,410 types; German: 40,266 types; Vietnamese: 18,910 types.[7] This procedure implicates that all overlapping types must have at least an absolute frequency of 15 in the overall ten million sentence corpus.
3.3 Analyses
For each sample, frequencies were normalised to occurrences in one million words. After[8] normalisation, each of the frequency lists was restricted to the overlapping wordform types because we can only compare frequencies of wordform types that occur in all samples for this language.
To answer our main question, whether corpus size has an influence on normalised frequencies, we set up a comparison regime. The logic here is that two frequency lists based on two different corpora should show a high association of normalised frequency values. This is especially true for the corpus samples we use here because the underlying corpus from which we draw is homogeneous (constant text type) and the sentences in the corpus are shuffled. So, if the normalisation of frequencies really works as intended, a differing sample size should not result in lower associations. In fact, we would expect larger samples to work ‘better’ because the normalised frequency values are based on more data.
We compared each sample frequency list to all the other sample frequency lists for the same language. We used the following measures for these comparisons:[9]
Kendall’s rank correlation coefficient τ B , which measures the ordinal association between the two frequency lists by comparing the number of concordant and discordant pairs of observations while accounting for tied pairs. A pair is concordant if both differences between two data points in a bivariate series of measurements point in the same direction. Accordingly, a pair is discordant if the differences point in the opposite direction.[10]
We used the considerably faster (compared to cor(…, method = “kendall”) in {stats}) implementation in the {pcaPP} R package (Filzmoser et al. 2023). This algorithm goes back to Knight (1966) and is being described in more detail by Abrevaya (1999) and Christensen (2005). Higher values of τ B indicate a stronger association between two frequency lists and it ranges between −1 (perfect negative association) and 1 (perfect positive association).
The proportion of concordant pairs p C between the two frequency lists, which is one component of the calculation of τ B . Higher values of p C indicate a stronger association between the two frequency lists. p C ranges between 0 and 1. p C is defined as
where n C is the number of concordant, n T the number of tied and n D the number of discordant pairs.
The proportion of concordant and tied pairs pCT between the two frequency lists. We included this measure because tied pairs can also be indicative of an association between the two lists and together with p C , we get a better insight into the procedure for calculating τ B . Also, by comparing p C and pCT, we are able to track down potential effects of the number of tied pairs alone. Higher values of pCT indicate a stronger association and its maximum value is 1. Note that, by definition, pCT is always higher or equal to p C because it is defined as
The mean normed deviation of normalised frequencies Δ f between the two frequency lists A and B. This measure is defined as
where N is the number of wordform types on the two frequency lists, a i and b i are the normalised frequencies of type i on list A and list B, respectively, and m i is the arithmetic mean of a i and b i . Note that the absolute deviation of normalised frequencies (without the norming by m i ) would also be possible. However, this would underestimate the deviations for low-frequency types because the “potential” for deviations is much higher for higher-frequency types.[11] The optimal value of Δ f is 0 which would indicate a perfect match of the normalised frequencies of all types.
We have included this comparison measure because, unlike τ B , p C and pCT, it does not operate on the basis of pairs but implements the comparison of two frequency values for the same word form type in a more direct way.
The proportion of types with a higher normalised frequency on list A than on list B pHH (HH stands for “half-half”). The rationale behind this comparison measure is that, if normed frequencies, indeed, fluctuate randomly between two frequency lists, around 50 % of all wordform types on list A should have a higher normalised frequency than on list B (the same is the case for the other direction). pHH is defined as
where N A>B is the number of wordform types with a higher normalised frequency on list A than on list B. The optimal value of pHH is 0.5 which would indicate that exactly half of the types have a higher normalised frequency value on list A than on list B (and the other way around). There are no cases in our datasets where the normalised frequencies of any wordform type are exactly equal.
The mean normalised frequency ratio r f (cf. Gries 2010: 272) between the normalised frequencies of any given wordform type on list A and B. r f is defined as
The optimal value for r f is 1. Note, however, that a value of 1 does not necessarily mean that all ratios are exactly 1 (i.e. that all pairs of a i and b i are equal) but that the mean of these ratios is 1.
To investigate whether the frequency range of the wordform types (from highly frequent to very infrequent types) has an impact on the comparison measures, we subsequently exclude types from the top of the frequency lists. We used the following steps: i) all types are being used, ii) 80 % of types from the bottom of the frequency list are being used, i.e. 20 % of the top frequency types are being excluded, iii) bottom 60 %, iv) bottom 40 %, v) bottom 20 %, i.e. 80 % of the top frequency types are being excluded. Results for these truncated frequency lists are presented in Section 4.2.
All in all, there are 3 (sizes of sample 1) × 5 (sample 1 iterations) × 3 (sizes of sample 2) × 5 (sample 2 iterations) × 5 (exclusions) = 1,125 comparisons for each language. From these, 3 (sizes) × 5 (samples) × 5 (exclusions) = 75 comparisons compare the frequency list with itself, so we conduct 1,125 − 75 = 1,050 comparisons per language. For each of these comparisons, we compute the six measures outlined above.
4 Results
4.1 Full frequency lists
Figure 1 presents the results for Kendall’s rank correlation coefficient (y-axis) for each language (colour and shape) and all combinations between the size of the first sample (x-axis) and the size of the second sample (plot panel). Each data point stands for one comparison (20 per sample-sample combination) and the lines connect the mean values of each group of data points.[12] The dashed line indicates the optimal value of τ B = 1.

The influence of the size of sample 1 (x-axis) and sample 2 (plot panels) on Kendall’s τ B (y-axis) measuring the correspondence of ranks for all types on the two frequency lists. Each data point is one comparison. The mean values for each group of data points are connected by lines. Languages are distinguished by colour and point symbol. The maximum value of τ B = 1 is marked by the dashed line. Note: The lines are only inserted to make the distinction between languages easier, i.e. there are no data points between the tick marks on the x-axis. The languages are slightly dodged on the x-axis to allow for easier distinction, i.e. sample sizes are all 100k, 500k or 1M sentences.
The different comparison data points within each group are not dispersed over a wide range of τ B values indicating little dispersion between samples. For example, in the left-most yellow/triangle group (all comparisons for the English corpus comparing 100k sentence samples with all other 100k sentence samples), the values of τ B vary between 0.729 and 0.734. Also, there is no systematic difference of effect patterns between languages. The central question, however, is whether sample size has an effect on τ B . That is, indeed, the case. Whenever a larger sample is involved in the comparison, τ B increases. Consequently, the smallest value of τ B is obtained for 100k versus 100k comparisons and the largest value is obtained for 1M versus 1M comparisons. These maximal observed values (mean τ B for Chinese: 0.896, English: 0.896, Finnish: 0.852, French: 0.887, German: 0.871, Vietnamese: 0.885) are close to the theoretical maximum of 1.
Given this pattern for comparisons via τ B , we conclude that corpus size indeed has an effect on the comparability of normalised frequencies. But it is exactly the kind of influence we would expect from larger corpus samples given prediction 1 from Section 2. Even though the inclusion of a larger sample leads to very disparate sample sizes (in our case, the most extreme difference in sample size is 100k vs. 1M sentences), τ B systematically increases, and this holds for all languages under investigation.
Figures 2 and 3 basically show the same pattern as Figure 1: as soon as (at least) one of the two corpus samples in the comparison is larger, p C and pCT increase with maximum mean values of 0.942 (p C , Chinese), 0.944 (p C , English), 0.917 (p C , Finnish), 0.938 (p C , French), 0.929 (p C , German), 0.936 (p C , Vietnamese) and 0.950 (pCT, Chinese), 0.950 (pCT, English), 0.930 (pCT, Finnish), 0.946 (pCT, French), 0.938 (pCT, German), 0.945 (pCT, Vietnamese).

The influence of the size of sample 1 (x-axis) and sample 2 (plot panels) on the proportion of concordant pairs p C (y-axis). Plot organisation is equivalent to Figure 1.

The influence of the size of sample 1 (x-axis) and sample 2 (plot panels) on the proportion of concordant and tied pairs pCT (y-axis). Plot organisation is equivalent to Figure 1.
For Δ f , we observe effect patterns similar to the previous comparative measures (see Figure 4). This time, however, a better match between the frequency lists is indicated by a lower value (with the theoretical minimum being 0). Again, as soon as a frequency list based on a larger corpus sample is part of the comparison, the measure is closer to the optimum value. Since Δ f is the mean normed deviation between the normalised frequencies of each type, this result could be influenced by a few outlier types with very high absolute deviations. However, if we calculate the median normed deviation (see Supplementary Figure 1), this effect pattern does not change considerably. Overall, the median values are slightly lower (e.g., for English, the range for the mean normed deviation is 0.0247–0.0779, for the median normed deviation, it is 0.0132–0.0341). These findings also generalise over languages.

The influence of the size of sample 1 (x-axis) and sample 2 (plot panels) on the mean normed deviation of normalised frequencies Δ f (y-axis). Plot organisation is equivalent to Figure 1.
For pHH, we only plot symmetric comparisons, i.e. where samples 1 and 2 contain the same number of sentences. Thus, we can lose the distinction via plot panels in Figure 5.[13] With larger sample sizes, the data points for the different symmetric comparisons lie closer to the optimum value. Hence, the maximum distance to pHH = 0.5 is for a 100k versus 100k comparison (e.g., for French: 0.425 and 0.575 for the symmetric comparison). The minimum distance can be observed for a 1M versus 1M comparison (again, for French, 0.477 and 0.523). The comparisons for 500k versus 500k sentence samples lie in between, and this holds for all languages.

The influence of the sizes of samples 1 and 2 (x-axis) on the proportion of word types that is more frequent in sample 1 than in sample 2 pHH (y-axis). Each data point is one comparison. Languages are distinguished by colour and point symbol.
For the mean normalised frequency ratio (see Figure 6), we see that, again, as soon as one larger sample is involved in the comparison, the comparison measure approaches its optimum value (here r f = 1). For example, for German, we observe the lowest mean value for the 1M versus 1M comparison (mean r f = 1.035). The mean value for the 500k versus 1M comparison is very close though (mean r f = 1.037).

The influence of the size of sample 1 (x-axis) and sample 2 (plot panels) on the mean normalised frequency ratio r f (y-axis). Plot organisation is equivalent to Figure 1.
4.2 Truncated frequency lists
We proceed by replicating the analyses from the previous section and then restricting the analyses in a stepwise manner to lower-frequency types. We do this to investigate whether the size of the samples that enter the comparison has a different effect on types in different frequency ranges. We start with the full frequency lists, i.e. replicating the analyses from the previous section and exclude 20 % more in each step, yielding five steps where the last step only involves the 20 % of types from the bottom of the frequency lists. We report the raw number of included types in each step for all languages in Supplementary Table 1.
Figure 7 (plot A) shows the results for Kendall’s τ B . First, there is a clear and continuous influence of the number of included types, as given in prediction 2 from Section 2: the fewer types we include, the lower the correspondence of ranks as indicated by τ B . Secondly, we see a similar pattern distinguishing between the sample sizes as in the previous section: as soon as one larger sample is involved in the comparison, correspondence levels increase, this is both visible from the left to the right panels for each language as well as the relative positions of the lines connecting the means for each group of data points. By comparing the panel rows, we see that these effects are consistent over the six languages. Figure 7 (plot B) replicates this ‘behaviour’ for one component of τ B , the concordant pairs.

The influence of the number of included types from the bottom of the frequency lists (x-axis), size of sample 1 (point colour and symbol) and sample 2 (plot panels, column-wise) on Kendall’s τ B (y-axis, plot A) and the proportion of concordant pairs p C (plot B). The lines connect the mean values for each group of data points. The maximum values of τ B = 1 and p C = 1 are marked with dashed lines. Languages are distinguished by row-wise plot panels.
What is interesting, though, is the pattern in Figure 8 (plot A) where the proportion of concordant and tied pairs is visualised. While the overall pattern remains largely stable, there are slight deviations for the last step where we only include the bottom 20 % of types. This effect is especially pronounced for the comparisons between the smallest samples (100k vs. 100k sentences) in all investigated languages. Here, we see that pCT indicates a better correspondence than in previous steps and in the comparisons of larger samples. This contradicts all previous findings. It is thus interesting and important to track down the cause for this pattern. Since this picture could not be observed for p C , it must be related to the tied pairs, the number of which is apparently ‘irregularly’ higher for the smallest comparisons. This is solely due to the number (or proportion) of hapax legomena:[14] while this proportion is less than 2 % for the larger comparisons, an average of 19.2 % of all types can only be observed once in both samples for the smallest comparisons for English (30.5 % for Finnish, 20.7 % for French, 25.7 % for German, 26.4 % for Vietnamese).[15] All hapax legomena ‘produce’ tied pairs, therefore increasing pCT. Similar distortions can be observed for the mean normed deviation of normalised frequencies (Figure 8, plot B) and the mean relative frequency ratio (Figure 9, plot B).

The influence of the number of included types from the bottom of the frequency lists (x-axis), size of sample 1 (point colour and symbol) and sample 2 (plot panels, column-wise) on the proportion of concordant and tied pairs pCT (y-axis, plot A) and the mean normed deviation of normalised frequencies Δ f (plot B). The lines connect the mean values for each group of data points. The maximum values of pCT = 1 and Δ f = 0 are marked with dashed lines. Languages are distinguished by row-wise plot panels.

The influence of the number of included types from the bottom of the frequency lists (x-axis), size of sample 1 (point colour and symbol) and sample 2 (plot panels, column-wise) on proportion of word types that is more frequent in sample 1 than in sample 2 pHH (y-axis, plot A) and the mean normalised frequency ratio r f (y-axis, plot B). The optimal values of pHH = 0.5 and r f = 1 are marked with dashed lines. Languages are distinguished by row-wise plot panels.
For Δ f (see Figure 8, plot B), the optimum value is 0, hence lower values indicate higher correspondence between frequency lists. Apart from the effect of hapax legomena on the smallest comparisons (20 % of bottom frequency types, 100k vs. 100k sentence samples), its trajectory indicates higher correspondences for larger samples and lower association when fewer types are included in the comparisons. Again, this holds for all languages under investigation.
For pHH (see Figure 9, plot A), we again only plot symmetric comparisons, i.e. the comparisons between equally-sized samples. The development over the number of included types follows a funnel-shaped pattern: The more we restrict the analyses to rarer types, the further the values move away from the optimum value of pHH = 0.5. Again, the values for larger samples are closer to the optimum value. The only data point that does not follow this pattern is the smallest comparison for French (fourth panel row, 100k vs. 100k, 20 % included types) which is slightly closer to 0.5 than the data point for twice as many included types. We currently have no explanation for this effect, especially why it only holds for one of the languages under investigation.
The distortions for small sample comparisons (bottom 20 % of types, 100k vs. 100k) due to hapax legomena are quite pronounced for the mean relative frequency ratio r f (Figure 9, plot B). In Figure 6, we already saw that the differences for r f between 500k and 1M comparisons were rather small. This does not only hold for the full frequency lists. For each language, the lines for 500k and 1M converge if 60 % or more of the bottom frequency types are included.
5 Discussion
Several frequency-based corpus linguistic measures are strongly influenced by corpus size (e.g. measures of lexical diversity or text similarity metrics). Based on this observation, we empirically investigated whether the very basal measure of normalised frequencies is also influenced by corpus size. To our knowledge, it is largely unquestioned that normalised frequencies are supposed to correct for the influence of corpus size – but it has not yet been systematically tested empirically how they might be influenced by corpus size themselves. We approached this question by comparing frequency lists, or rather testing the association that exists between two lists of normalised frequencies. For generic web corpora of Chinese, English, Finnish, French, German, and Vietnamese, we kept all influencing factors as constant as possible and only systematically varied the size of the underlying corpus from which the frequency lists were derived.
The analyses for the complete frequency lists paint a consistent picture and confirm predictions from statistical theory, both over languages and comparison measures: whenever a larger corpus sample is involved in the comparison, the six comparison measures for the frequency lists derived from those samples are closer to the respective optimal value (e.g. 100k vs. 500k or 500k vs. 1M sentences). So, it is not the case that the association is always higher when the sample sizes are the same (e.g. 100k vs. 100k or 500k vs. 500k sentences) – an assumption that would logically follow if one would assume that differences in corpus size would systematically skew normalised frequencies.
At the same time, however, we see that the association strengths indicated by the measures decrease rapidly when we restrict the comparisons to lexical items that lie in lower frequency bands, again confirming the prediction based on statistical theory. This means that the comparison of normalised frequencies for these elements is less reliable. On the other hand, the main finding also applies here: as soon as a larger sample is involved in the comparisons, we observe higher associations of the frequency lists, just at a lower overall level. This is what one would expect based on the Zipfian distribution of language data (Zipf 1935), because the discriminatory power of frequencies decreases in lower frequency ranges.[16] An example with two low-frequency word form types: in the five English samples with 100k sentences, the type “glide” is more frequent than “stitch” in two samples, equally frequent in one sample and less frequent in two samples (an overview of raw and normalised frequencies for the example types is given in Supplementary Table 3). In contrast, the relationship is consistent for two high-frequency types: “on” is more frequent than “with” in every sample. Now, if we consider the largest samples in our data set, which are based on 1M sentences each, the ranking between “glide” and “stitch” is stable: “glide” is consistently more frequent than “stitch”. In this sense, larger corpora “help” because they are more sensitive to frequency differences in (previously) lower frequency ranges. However, it should be noted that these larger corpora have the same problem in their own lower frequency ranges. In other words, the problem “shifts” to even rarer words.
Regarding the comparative measures we have used, τ B , p C and pCT are redundant to a certain extent. However, we could see in the contrast between Figures 7 and 8 (plot A) that of these three measures, only pCT was able to draw attention to the high proportion of hapax legomena for the 20 % least frequent of overlapping types, which may give this measure some justification. r f shows certain weaknesses when comparing larger corpus samples in the sense that it can no longer cleanly distinguish the associations in frequency lists based on 500k sentences versus 1M sentences.
6 Implications
Our main result is reassuring: the corpus size of the underlying corpora does not negatively influence comparisons of normalised frequency lists, i.e. differences in corpus size do not lead to lower associations between the derived lists. Rather, the results suggest that larger corpora (even in combination with smaller corpora) always have a positive effect in the sense that the association of the compared lists increases. However, we can only draw this conclusion for the scenario we have chosen for the present study, i.e. if all influences except the corpus size (e.g., text type[17] or sampling processes) are kept as constant as possible. For smaller corpora in particular, however, the comparability for less frequent types can quickly decrease. It would be interesting to see whether these findings also hold for more heterogeneous base corpora, for example of varying modality or text type. As already mentioned in the Introduction, we wanted to create a setup here that can be considered “optimal” for the comparison of normalised frequencies.
So, does this mean that larger corpora are a cure-all when it comes to comparing normalised frequencies? Given that other potential sources of influence (e.g., corpus type or sampling processes) were kept as constant as possible, our results suggest that larger corpora always have a positive effect in the sense that the association of the derived frequency lists increases – just like statistical theory predicts. This is also true when larger corpora are combined with smaller ones. On the other hand, especially with smaller corpora, the comparability for less frequent types can quickly decrease. Against this background, our results thus indicate that, all other things being equal – taking into account important considerations such as corpus quality, its composition, and the balance among various text types (Biber 1993; Koplenig 2017; Koplenig 2019; Leech 2007) – Mercer’s claim that “more data is better data” (Church and Mercer 1993) holds validity.
We would also like to point out that some of the pre-processing steps that we have adopted here cannot equally be applied to all kinds of linguistic research questions. For example, we have excluded very low-frequency words that may be relevant for, e.g., productivity studies. In addition, we have deliberately destroyed the internal structure of the corpora, i.e. the sequence of sentences and their assignment to individual corpus documents, which may be highly relevant for other linguistic research endeavours. It therefore remains to be seen to what extent the results shown here can be transferred to these types of studies.
Acknowledgments
We would like to thank the team of the Leipzig Corpora Collection for providing us with the 10M sentence corpora.
-
Data availability: The Supplementary Material, R scripts and data used for the analyses in this manuscript are available via OSF at https://osf.io/64mpz/. Please see the Wiki page of the project for explanations concerning the availability of the underlying corpora.
References
Abrevaya, Jason. 1999. Computation of the maximum rank correlation estimator. Economics Letters 62(3). 279–285. https://doi.org/10.1016/S0165-1765(98)00255-9.Search in Google Scholar
Altmann, Eduardo G., Janet B. Pierrehumbert & Adilson E. Motter. 2009. Beyond word frequency: Bursts, lulls, and scaling in the temporal distributions of words. PLoS One 4(11). e7678. https://doi.org/10.1371/journal.pone.0007678.Search in Google Scholar
Attia, Safa & Jesus Romero-Trillo. 2022. A corpus-assisted discourse study of the media language in the Egyptian revolution. Language, Discourse & Society 10(2). 105–127.10.32714/ricl.10.01.05Search in Google Scholar
Baron, Alistair, Paul Rayson & Dawn Archer. 2009. Word frequency and key word statistics in historical corpus linguistics. Anglistik 20(1). 41–67.Search in Google Scholar
Bestgen, Yves. 2018. Evaluating the frequency threshold for selecting lexical bundles by means of an extension of the Fisher’s exact test. Corpora 13(2). 205–228. https://doi.org/10.3366/cor.2018.0144.Search in Google Scholar
Biber, Douglas. 1993. Representativeness in corpus design. Literary and Linguistic Computing 8(4). 243–257. https://doi.org/10.1093/llc/8.4.243.Search in Google Scholar
Burgess, Curt & Kay Livesay. 1998. The effect of corpus size in predicting reaction time in a basic word recognition task: Moving on from Kučera and Francis. Behavior Research Methods, Instruments, & Computers 30(2). 272–277. https://doi.org/10.3758/BF03200655.Search in Google Scholar
Christensen, David. 2005. Fast algorithms for the calculation of Kendall’s τ. Computational Statistics 20(1). 51–62. https://doi.org/10.1007/BF02736122.Search in Google Scholar
Church, Kenneth W. & Robert L. Mercer. 1993. Introduction to the special issue on computational linguistics using large corpora. Computational Linguistics 19(1). 1–24.Search in Google Scholar
Filzmoser, Peter, Heinrich Fritz & Klaudius Kalcher. 2023. pcaPP: Robust PCA by projection pursuit. Available at: https://CRAN.R-project.org/package=pcaPP.Search in Google Scholar
Goldhahn, Dirk, Thomas Eckart & Uwe Quasthoff. 2012. Building large monolingual dictionaries at the leipzig corpora collection: From 100 to 200 languages. In Proceedings of the eighth international conference on language resources and evaluation (LREC’12), 759–765. Istanbul, Turkey: European Language Resources Association (ELRA). Available at: http://www.lrec-conf.org/proceedings/lrec2012/pdf/327_Paper.pdf.Search in Google Scholar
Gries, Stefan Th. 2010. Useful statistics for corpus linguistics. In Aquilino Sánchez & Moisés Almela (eds.), A mosaic of corpus linguistics: Selected approaches, 269–291. Frankfurt am Main: Peter Lang.Search in Google Scholar
Grieve, Jack. 2023. Register variation explains stylometric authorship analysis. Corpus Linguistics and Linguistic Theory 19(1). 47–77. https://doi.org/10.1515/cllt-2022-0040.Search in Google Scholar
Kilgarriff, Adam. 2001. Comparing corpora. International Journal of Corpus Linguistics 6(1). 97–133. https://doi.org/10.1075/ijcl.6.1.05kil.Search in Google Scholar
Knight, William R. 1966. A computer method for calculating Kendall’s Tau with ungrouped data. Journal of the American Statistical Association 61(314). 436–439. https://doi.org/10.1080/01621459.1966.10480879.Search in Google Scholar
Koplenig, Alexander. 2017. The impact of lacking metadata for the measurement of cultural and linguistic change using the Google Ngram data sets – Reconstructing the composition of the German corpus in times of WWII. Digital Scholarship in the Humanities 32(1). 169–188. https://doi.org/10.1093/llc/fqv037.Search in Google Scholar
Koplenig, Alexander. 2019. Against statistical significance testing in corpus linguistics. Corpus Linguistics and Linguistic Theory 15(2). 321–346. https://doi.org/10.1515/cllt-2016-0036.Search in Google Scholar
Koplenig, Alexander, Sascha Wolfer & Carolin Müller-Spitzer. 2019. Studying lexical dynamics and language change via generalized entropies: The problem of sample size. Entropy 21(5). https://doi.org/10.3390/e21050464.Search in Google Scholar
Laurence, Anthony. 2021. What can corpus software do? In Anne O’Keeffe, Michael McCarthy (eds.), The Routledge handbook of corpus linguistics, 2nd edn., 103–125. Abingdon, Oxon; New York, NY: Routledge.10.4324/9780367076399-9Search in Google Scholar
Leech, Geoffrey. 2007. New resources, or just better old ones? The holy grail of representativeness. In Marianne Hundt, Nadja Nesselhauf & Carolin Biewer (eds.), Corpus linguistics and the web, 133–149. Amsterdam: Rodopi.10.1163/9789401203791_009Search in Google Scholar
Macqueen, Susy, Luke Collins, Gavin Brookes, Zsófia Demjén, Elena Semino & Diana Slade. 2024. Laughter in hospital emergency departments. Discourse Studies 26(3). 358–380. https://doi.org/10.1177/14614456231194845.Search in Google Scholar
Michel, Jean-Baptiste, Yuan Kui Shen, Aviva P. Aiden, Adrian Veres, Matthew K. Gray, Joseph P. Pickett, Dale Hoiberg, Dan Clancy, Peter Norvig, Jon Orwant, Steven Pinker, Martina A. Nowak & Erez Lieberman Aiden & The Google Books Team. 2011. Quantitative analysis of culture using millions of digitized books. Science 331(6014). 176–182.10.1126/science.1199644Search in Google Scholar
Müller-Spitzer, Carolin, Sascha Wolfer & Alexander Koplenig. 2015. Observing online dictionary users: Studies using wiktionary log files. International Journal of Lexicography 28(1). 1–26. https://doi.org/10.1093/ijl/ecu029.Search in Google Scholar
Poole, Robert & Nicholas Hayes. 2023. Stance in climate science: A diachronic analysis of epistemic stance features in IPCC physical science reports. Journal of Corpora and Discourse Studies 5. 37–60. https://doi.org/10.18573/jcads.100.Search in Google Scholar
R Core Team. 2024. R: A language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing. Available at: https://www.R-project.org.Search in Google Scholar
Tweedie, Fiona J. & R. Harald Baayen. 1998. How variable may a constant be? Measures of lexical richness in perspective. Computers and the Humanities 32(5). 323–352.10.1023/A:1001749303137Search in Google Scholar
Wijffels, Jan. 2022. udpipe: Tokenization, parts of speech tagging, lemmatization and dependency parsing with the “UDPipe” “NLP” toolkit. Available at: https://CRAN.R-project.org/package=udpipe.Search in Google Scholar
Zipf, George Kingsley. 1935. The psycho-biology of language. Mifflin: Oxford: Houghton.Search in Google Scholar
Supplementary Material
This article contains supplementary material (https://doi.org/10.1515/cllt-2024-0040).
© 2025 the author(s), published by De Gruyter, Berlin/Boston
This work is licensed under the Creative Commons Attribution 4.0 International License.