Digital Humanities 2010 – ABSTRACT Maciej EDER

Does Size Matter? Authorship Attribution, Small Samples, Big Problem

The aim of this study is to find such a minimal size of text samples for authorship attribution that would provide stable results independent of a random noise. A few controlled tests for different sample lengths, languages and genres are discussed and compared. Although I focus on Delta methodology, the results should be valid for most – if not all – multidimensional methods relying on word frequencies and “nearest neighbor” classifications. Introduction In the field of stylometry, and especially in authorship attribution, the reliability of the obtained results becomes even more essential than the results themselves: failed attribution is much better than false attribution (cf. Love, 2002). However, while dozens of outstanding papers deal with increasing the effectiveness of current stylometric methods, the problem of their reliability remains somehow underestimated. Especially, the simple yet fundamental question of the shortest acceptable sample length for reliable attribution has not been discussed convincingly. It is true that the problem is not new. In his investigation of style variation in Golding’s The Inheritors, Hoover noticed that truncating all the samples to the size of the shortest chapter spoils the results, probably due to the short sample effect (Hoover, 2003). In another instance, Rybicki discovered that his own results of remarkable similarities in the patterns of distance between idiolects in two different translations of a single novel were due to the gap between talkative and non-talkative characters, the latter simply not saying enough to produce a reliable sample (Rybicki, 2006, 2008). A few scholars have proposed an intuitive solution of this problem, e.g., that an analyzed text should be “long” (Craig, 2004), that “for stylometric reliability the minimum sample size allowed is 1000 words” (Holmes et al., 2001), that “with texts of 1500 words or more, the Delta procedure is effective enough to serve as a direct guide to likely authorship” (Burrows, 2002), etc. Those statements, however, have not been followed by thorough empirical 1

investigation. Additionally, numerous attribution studies do not obey even the limit of 1000 words per sample (cf. Juola & Baayen, 2005; Burrows, 2002; Jockers et al., 2008, etc.) In those – and many other – attribution studies based on short samples, despite their well-established hypotheses, convincing choice of style-markers, advanced statistics applied and brilliant results presented, one cannot avoid a very simple yet uneasy question whether those impressive results could be obtained by chance, or at least positively affected by randomness? This question can be also formulated in a different way: if a cross-checking experiment with numerous short samples were available, would the results be just as satisfying?

Hypothesis It is commonly known that word frequencies in a corpus are random variables; the same can be said about any written authorial text, like novel or poem. Being a probabilistic phenomenon, word frequency strongly depends on the size of the population (i.e., the size of the text used in the study). Now, if the observed frequency of a single word exhibits too much variation for establishing an index of vocabulary richness resistant to sample length (cf. Tweedie & Baayen, 1998), a multidimensional approach – based on several probabilistic word frequencies – should be even more questionable. On theoretical grounds, we can intuitively assume that the smallest acceptable sample length would be hundreds rather than dozens of words. Next, we can expect that, in a series of controlled authorship experiments with longer and longer samples tested, the probability of attribution success would at first increase very quickly, indicating a strong correlation with the current text size; but then, above a certain value, further increase of input samples would not affect the effectiveness of the attribution. In any attempt to find this critical point in terms of statistical investigation, one should be aware, however, that this point might depend – to some extent – on the language, genre, or even the text analyzed.

Experiment I: Words A few corpora of known authorship were prepared for different languages and genres: for English, Polish, German, Hungarian, and French novels, for English epic poetry, Latin poetry (Ancient and Modern), Latin prose (non-fiction), and for Ancient Greek epic poetry. The research procedure was as follows. For each text in a given corpus, 500 randomly chosen single words

2

100

Figure 1: English novels, attribution effectiveness. Single-word (red points) and text-passage (green points) samples

80 60 40 20

percentage of correct attributions

●●

●●● ●● ●

●● ●●● ● ● ●●

● ●● ● ● ● ●●●

● ● ●● ● ●

● ● ● ●● ● ● ● ● ●● ● ● ● ●● ●●●● ● ● ●●●●●● ●● ● ●●● ● ● ●● ● ● ● ●● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ●● ●●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ●●● ●● ● ●● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ●● ●● ● ● ● ●● ● ● ●● ● ●● ● ● ●● ● ● ● ●● ● ● ● ●● ● ● ●● ●● ●● ●● ● ● ● ●● ● ●●● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ●● ● ● ●● ● ● ● ●● ● ●● ● ● ●● ●● ●● ● ●● ● ● ● ●● ● ●● ●● ●● ● ● ● ● ● ● ● ●● ●●● ●● ● ● ● ● ●●● ● ● ●●●● ● ● ●● ● ● ●● ●● ● ●● ● ● ●● ●● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ●●● ● ●● ● ● ● ● ● ● ● ● ● ●

0

5000

10000

15000

20000

length of samples (in words)



were concatenated into a new sample. These new samples were analyzed using the classical Delta method as developed by Burrows (2002); the percentage of attributive success was regarded as a measure of effectiveness of the current sample length. The same steps of excerpting new samples from the original texts, followed by the stage of “guessing” the correct authors, were repeated for the length of 600, 700, 800, ..., 20000 words per sample. The results for a corpus of 63 English novels are shown on Fig. 1. The observed scores (red points on the graph) clearly indicate the existence of a trend (solid line): the curve, climbing up very quickly, tends to stabilize at a certain point, which indicates the minimal sample size for the best attributing rate. Although it is difficult to find the precise position of that point, it becomes quite obvious that samples shorter than 5000 words provide a poor “guessing”, because they can be immensely affected by random noise. Below the size of 3000 words, the obtained results are simply disastrous (more than 60% of false attributions for 1000-word samples may serve as a convincing caveat). Other analyzed corpora showed similarly disappointing 3

100

Figure 2: Latin prose, attribution effectiveness. Single-word (red points) and text-passage (green points) samples

●● ● ● ● ●● ● ● ● ●● ● ●● ●● ● ● ●●● ● ●●● ● ● ● ● ●●●●● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ●●● ● ●● ● ●● ● ● ● ● ● ● ●● ● ● ●● ●● ● ●● ●● ● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ●

80 60

●●

● ●

●● ●●







●● ● ●



● ●● ● ● ● ● ●



●● ●

40

percentage of correct attributions



20

● ●●

● ● ● ●

0●●

2000

4000

6000

8000

10000

length of samples (in words)

results: the critical point of attributive success could be found between 5000 and 10000 words per sample (amazingly, there was no significant difference between flexive and non-flexive languages). Better scores were obtained for the two poetic corpora: English and Latin (3500 words per sample were enough for good results), and, surprisingly, the corpus of Latin prose (its minimal effective sample size was of some 2500 words; cf. Fig. 2, red points). For each corpus analyzed, the “guessing” scores also seem to show that effectiveness would not increase in samples exceeding 15000 words; this is also a valuable observation, suggesting that there are limits to Hoover’s statement that “for statistical analysis, the longer the text the better” (Hoover, 2001).

Experiment II: Passages The way of preparing samples by extracting a mass of single words from the original texts seems to be an obvious solution for the problem of statistical representativeness. In most attribution studies, however, shorter or longer

4

80 ● ●













● ●



60



40



20

percentage of correct attributions

Figure 3: Polish novels, attribution effectiveness. Samples of 8192 words, concatenated of chunks of 2, 4, 8, 16, 32, ..., 8192 words

2

4

8

16

32

64

128

256

512

2048

8192

length of chunks within a 8192−word sample

passages of disputed works are usually analyzed (either randomly chosen from the entire text, or simply truncated to the desired size). The purpose of the current experiment was to test the attribution effectiveness of this typical sampling. The whole procedure was repeated step by step as in the previous test, but now, instead of collecting individual words, sequences of 500 words (then 600, 700, ..., 20000) were excerpted randomly from the original texts. Three main observations could be made here: 1. For each corpus analyzed, the effectiveness of such samples (excerpted passages) was always worse than the scores described in the former experiment, relying on the “bag-of-words” type of sample (cf. Fig. 1 and 2, green points). 2. The more flexive the language, the smaller the difference in correct attribution between both types of samples, the “passages” and the “words”: the greatest in the English novels (cf. Fig. 1, green points vs. red), the smallest in the Hungarian corpus. 3. For “passages”, the dispersion of the observed scores was always wider than for “words”, indicating the possible significance of the influence of random noise in attribution studies relying on excerpted passages. This 5

effect might be due to the obvious differences in word distribution between narrative and dialogue parts in novels (cf. Hoover, 2001); however, the same effect was equally strong for poetry (Latin and English) and non-literary prose (Latin).

Experiment III: Chunks At times we encounter an attribution problem where extant works by a disputed author are doubtlessly too short for being analyzed in separate samples. The question is, then, if a concatenated collection of short poems, epigrams, sonnets, etc. in one sample (cf. Eder & Rybicki, 2009) would reach the effectiveness comparable to that presented above? And, if concatenated samples are suitable for attribution tests, do we need to worry about the size of the original texts constituting the joint sample? The third experiment, then, was designed as follows. In 12 iterations, several word-chunks were randomly selected from each text into 8192-word samples: 4096 bi-grams, 2048 tetra-grams, 1024 chunks of 8 words in length, 512 of 16 words, and so on, up to 2 chunks of 4096 words. Thus, all the samples in question were 8192 words long. The obtained results were very similar for all the languages and genres tested. As shown in Fig. 3 (for the corpus of Polish novels), the effectiveness of “guessing” depends to some extent on the word-chunk size used. Although the attributive scores are slightly worse for long chunks within a sample (4096 words or so) than for bi-grams, 4-word chunks etc., every chunk size could be acceptable to constitute a concatenated sample. However, although this seems to be an optimistic result, we should remember that this test would not be feasible on real short poems. Epigrams, sonnets etc. are often masterpieces of concise language, with a domination of verbs over adjectives, particles and so on, and with a strong tendency to compression of content. For that reason, further investigation is needed here.

Conclusions The scores presented in this study, as obtained with classical Delta procedure, would be slightly better when solved with Delta Prime, and worse if either Cluster Analysis or Multidimensional Scaling is used. However, the shape of all the curves, as well as the point where the attributive success rate becomes stable, are quite identical for each of these methods. The same refers to different combinations of style-markers’ settings, like “culling”, the number of the Most Frequent Words analyzed, deleting/non-deleting

6

pronouns, etc. – although different settings provide different “guessing” (up to 100% for the most efficient), they never affect the shape of the curves. Thus, since the obtained results are method-independent, this leads us to a very important conclusion about the smallest acceptable sample size for future attribution experiments and other investigations in the field of stylometry. Unfortunately, it also means that some of the recent attribution studies should be at least re-considered. Until we develop style-markers more precise than word frequencies, we should be aware of some limits in our current approaches, the most troublesome of these being the limits of the sample length. As I tried to show, using 2500-word samples will hardly provide a reliable result, to say nothing of shorter texts.

References Burrows, J. F. (2002). ‘Delta’: a Measure of Stylistic Difference and a Guide to Likely Authorship. Literary and Linguistic Computing 3(17): 267-287. Craig, H. (2004). Stylistic Analysis and Authorship Studies. In: S. Schreibman, R. Siemens and J. Unsworth (eds.), A Companion to Digital Humanities. Blackwell Publishing, p. 273-288. Eder, M., Rybicki, J. (2009). PCA, Delta, JGAAP and Polish Poetry of the 16th and the 17th Centuries: Who Wrote the Dirty Stuff? In: Digital Humanities 2009: Conference Abstracts. University of Maryland, College Park, p. 242-244. Holmes, D., Gordon, L. J., Wilson, Ch. (2001). A Widow and Her Soldier: Stylometry and the American Civil War. Literary and Linguistic Computing 4(16): 403-420. Hoover, D. L. (2001). Statistical Stylistic and Authorship Attribution: an Empirical Investigation. Literary and Linguistic Computing 4(16): 421-444. Hoover, D. L. (2003). Multivariate Analysis and the Study of Style Variation. Literary and Linguistic Computing 4(18): 341-360. Jockers, M. L., Witten, D. M., Criddle, C. S. (2008). Reassessing Authorship of the ‘Book of Mormon’ Using Delta and Nearest Shrunken Centroid Classification. Literary and Linguistic Computing 4(23): 465491. Juola, P., Baayen R. H. (2005). A Controlled-corpus Experiment in Authorship Identification by Cross-entropy. Literary and Linguistic 7

Computing Suppl. Issue (20): 59-67. Love, H. (2002). Attributing Authorship: An Introduction. Cambridge: Cambridge University Press. Rudman, J. (1998). The State of Authorship Attribution Studies: Some Problems and Solutions. Computers and the Humanities 31: 351-365. Rybicki, J. (2006). Burrowing into Translation: Character Idiolects in Henryk Sienkiewicz’s Trilogy and Its Two English Translations. Literary and Linguistic Computing 1(21): 91-103. Rybicki, J. (2008). Does Size Matter? A Re-examination of a Time-proven Method. In: Digital Humanities 2008: Book of Abstracts. University of Oulu, p. 184. Tweedie, J. F.; Baayen, R. H. (1998). How Variable May a Constant be? Measures of Lexical Richness in Perspective. Computers and the Humanities 32: 323-352.

8

Does Size Matter? Authorship Attribution, Small ...

The aim of this study is to find such a minimal size of text samples for authorship attribution that would provide stable results independent of a random noise. A few controlled tests for different sample lengths, languages and genres are discussed and compared. Although I focus on Delta method- ology, the results should be ...

124KB Sizes 1 Downloads 221 Views

Recommend Documents

Does Culture Matter? - Semantic Scholar
game or public goods game (e.g. Henrich, Boyd, Bowles, Camerer, Fehr, Gintis, and McElreath ..... (Data Sources: World Values Survey, General Social Survey.).

Does Culture Matter? - Semantic Scholar
provisions to differences in aggregate outcomes such as average savings rates, fer% tility rates, or .... extent to which executives believe that labor relations are good is correlated with .... that more directly reflects the cultural attitudes of i

When does size not matter? Effects of stimulus ... - Wiley Online Library
Similarly, recent studies showed degrees of emotional modulation of autonomic responses and ... distance and retinal size are strictly related (Loftus & Harley,.

Does size matter? A comment on "The Perils of Picky ...
contribution in studies where both %F and %V data were given. An insect taxon, which is consistently overestimated in %V studies compared to %F datasets, ...

Chapter 11 - Does Culture Matter?
your institution's administrator. All other .... This chapter will primarily focus on the epidemiological approach to culture .... 15 years, religion, ethnicity, subjective health status, income, and years of schooling, as well as a dummy for whether

Does Money Really Matter
of children adopted into low-, middle- and high-SES families (Duyme, Dumaret, Annick- ... Other studies yield a mixed verdict on whether family income influences healthy child ..... cluster adjustment to account for the fact that in some cases there

Does Money Really Matter
Achievement with Data from Random-Assignment Experiments ... our analytic approach, while Section IV describes the data and measures used in our analysis.

When Does Measurement Invariance Matter?
can theoretically assume every value of the real line. Mea- surement ... empirical one, the second cannot be so construed. Whether ..... App Psychol Meas. 2002 ...

Does Direct Democracy Reduce the Size of Government?
Jun 9, 2009 - two questions: does direct democracy reduce government spending? .... expenditures exceed a certain monetary threshold (which is defined.

How Does Colonial Origin Matter for Economic ...
Mar 16, 2010 - broad range of variables that are important for economic growth.2 A key feature .... colonial origin affects economic outcomes, using only the SSA dataset. ..... In the first stage, we test the hypothesis that colonial origin matters f

How Much does Health Insurance Matter for Young ...
including entry-level wages, jobs without employer sponsored insurance, and high health premiums that ..... have low-wage, entry-level, and temporary jobs that do not offer employer-sponsored insurance. (Schwartz and ...... variables/2SLS and GMM est

Does Inflation Targeting Matter for Output Growth ...
Tel: +1-202-473-0360. Abstract: We examine in this paper the effects of inflation targeting (IT) on industrial and emerging economies' output growth over the ...

How Does Colonial Origin Matter for Economic Performance in sub ...
Mar 16, 2010 - forth, SSA) dataset, the internationally observed growth differential .... the main source of influence of colonial legacy on the post ... to international trade during 1960(2000 (OPEN) and export share in GDP during 1960(.

does management matter? evidence from india - Semantic Scholar
May 5, 2010 - Chicago, Columbia, the EBRD, Harvard Business School, IESE, the ... important character in the early years of the economics discipline as .... modern management practices is a type of skill-biased technical change (SBTC) since ... Our e

1 Does It Really Matter Whether Students ...
the students might have lost confidence in the computer tutor due to its inability ...... should support typed input because all available data generally support the ..... The application is able to write the document to unused space on the hard driv

1 Does It Really Matter Whether Students' Contributions ...
We thank our research colleagues in the Emotive Computing Group at the University of. Memphis ..... Window 1 (top of screen) is ...... .500. I11. Confidence can learn from computer tutor .706. I7. Enjoy learning new computer software .657. I4.

How Does Colonial Origin Matter for Economic ... - CiteSeerX
16 Mar 2010 - dence and the subsequent post colonial changes in explaining income differences amongst former SSA countries. ..... The objective is to capture possible changes in the colonial legacy of market distortion, brought about by ...... VPopul