k-NN Aggregation with a Stacked Email Representation Amandine Orecchioni, Nirmalie Wiratunga, Stewart Massie, and Susan Craw School of Computing, The Robert Gordon University, Aberdeen AB25 1HG, Scotland, UK {ao|nw|sm|smc}@comp.rgu.ac.uk

Abstract. The variety in email related tasks, as well as the increase in daily email load, has created a need for automated email management tools. In this paper, we provide an empirical evaluation of representational schemes and retrieval strategies for email. In particular, we study the impact of both textual and nontextual email content for case representation applied to Email task management. Our first contribution is S TACK, an email representation based on stacking. Multiple casebases are created, each using a different case representation related with attributes corresponding to semi-structured email content. A k-NN classifier is applied to each casebase and the output is used to form a new case representation. Our second contribution is a new evaluation method allowing the creation of random chronological stratified train-test trials that respect both temporal and class distribution aspects, crucial for the email domain. The Enron corpus was used to create a dataset for the email deletion prediction task. Evaluation results show significant improvements with S TACK over single casebase retrieval and multiple casebases retrieval combined using majority vote.

1 Introduction Over time, email has evolved from a simple medium of communication to one involving complex management tasks. Nowadays, it is not enough to simply read and reply to emails. One must also prioritise reading order, filter spam and phish emails, avoid viruses, organise and maintain information repositories and manage social networks and diaries. This increase in email related tasks coupled with the increase in email handling load has created the need for automated email management tools. Research in email management recognises five key areas [19, 9]: information management, task management, time management, contact management and security protection. Machine learning research applied to email has focused on individual prediction tasks within each area such as delete [8], reply [23, 18], attach [12], forward [22], filter [20, 15, 10] and classify [2, 4]. However, since each area typically involves several chronologically organised email tasks, these tasks can be seen as email workflows. Table 1 presents four workflows by decomposing each management activity into a set of chronologically organised tasks. The Contact Workflow combines contact management and email classification tasks, while the Task Workflow combines automated email response, reply prediction and email classification tasks. The advantage of the email workflow view is that it provides a template that can be used to guide prediction. For

instance, consider an incoming email, it would be useful to predict what action the user might take. Is the email going to be forwarded, if so to whom? Is the user likely to reply, if so can the reply be semi-authored? Should the email be deleted, if so when?

Task Workflow Helpdesk Query

Information Workflow Paper Recommendation

Time Workflow Meeting Announcement

Contact Workflow New Contact Detail

1. Read Email

1. Read Email

1. Read Email

1. Read Email

2. Reply

2. Open Attachment

2. Open calendar

2. Open Address Book

3. Keep as ”open queries”

3. Print Attachment

3. Add reminder

3. Add contact

4. Receive follow-up question

4. Save Attachment

4. Keep email in inbox

4. Save contact

5. Reply

5. Delete Email

5. Delete email after meeting

5. Delete email

6. File in ”closed queries”

Table 1. Email Workflow Examples

An email prediction task requires a learner to handle both local interactions between emails and potentially evolving email concepts. CBR’s ability to handle these challenges were demonstrated on an email filtering task [10]. An email is commonly represented as a bag-of-words of its textual content. However, there is also separate evidence to suggest the utility of non-textual email attributes for email classification [11, 17]. In this paper, we present a systematic evaluation of email attribute extraction from both textual and non-textual content. Our work differs from existing feature selection and extraction research in that we establish the importance of email sub-sections in case representation instead of that of features. Since every workflow in Table 1 terminates in the delete or filing tasks, we choose to focus our study on deletion prediction using the Enron corpus. The rest of this paper is organised as follows. Section 2 discusses related work. Multiple email representations are defined in Section 3 and how they can be used in multiple casebases in Section 4. Section 5 defines a new evaluation methodology allowing the creation of random chronological stratified train-test trials that respect both temporal and class distribution aspects. The evaluation and results are presented in Section 6. Finally, our conclusions and future work are highlighted in Section 7.

2 Related Work This section presents three different aspects of related work. We will start by discussing email representation and possible email attributes. We will then report the benefits of ensemble of classifiers documented in the literature. Finally, we will highlight the drawbacks of current evaluation methods for the email domain. Research in email categorization is commonly focused by classification into topical folders [2] or into speech-acts [4], email prioritisation [16] and Spam filtering [10]. A

standard information retrieval approach for document representation is to use a bag-ofwords representation. However, research focused on the acquisition of indexing vocabulary for email identifies three types of features: – structured : features extracted from the header such as date, from, to; – textual : keywords from the free text sections such as subject and body; and – handcrafted : features created from preprocessing the emails such as email length and number of special characters [11]. It was shown that handcrafted features typically do not improve prediction accuracy and so are not considered further in this paper. Email features can be incorporated in a single feature vector. However, previous work has highlighted the benefits of using an ensemble of classifiers, based on different feature subsets, over a single feature vector [7]. Performance improvements with ensembles are due to the aggregation of base-learners which are essentially local specialists. This gets round the problem of feature weight optimisation otherwise needed with a single feature vector [5]. Feature subsets can be generated randomly [1] or using feature selection [6]. One possible ensemble aggregation method is stacking, which is typically used to combine different types of base-learners into a meta-learner [25]. The idea is to use the prediction of each base-learner as input for the meta-learner. This requires each case in the meta-learner casebase to be represented with values corresponding to predictions of the base-learners. Stacking has been successfully applied to spam filtering by combining a memory-based classifier and a Naive Bayes classifier [21]. The temporal aspect of email is an important issue to consider when generating train-test splits for evaluation. Indeed, test set emails must be more recent than train set emails. In a real life situation, it would be impossible to make a decision about an incoming email based on emails not yet received. This aspect is not taken into account by standard evaluation methodologies such as cross-validation, leave one out or hold out. A possible approach is to order emails chronologically and use the earlier half for training and the later half for testing [17]. However, a single split of the dataset is problematic to evaluate statistical significance as it only permits one trial. Another approach is to create multiple splits from a chronologically ordered dataset. The classifier is trained on the first N messages and tested on the following N , then trained on the first 2N messages and tested on the following N , then trained on the first 3N messages and tested on the following N and so on [2]. This approach is similar to the previous in that it ensures a chronological ordering of the data. But, it also allows statistical significance testing as it creates N -1 trials instead of a single trial. However, class distribution should also be respected, and these approaches are only suitable when emails from each class are evenly distributed over time.

3 Case Representation for Emails The decomposition of a semi-structured document into constituents, such as from, to, subject and body in emails, allows case retrieval to focus on each of them separately.

Multi-Attribute

Single-Attribute

This is particularly useful when they have their own indexing vocabulary as they are further decomposable into feature vectors. The top-half of Table 2 presents nine attributes identified in relation to email sections. Date and From Address are nominal attributes while the others are textual attributes represented as binary feature vector. For instance, all the email addresses from the To field of the emails in the casebase form the indexing vocabulary of To Address. This attribute is represented for each email as a feature vector of the indexing vocabulary, where the value for each feature is 1 if the email’s To field contains the feature or 0 otherwise. From Name, To Name, CC Address, CC Name, Subject and Body are represented similarly. Accordingly, an email can be represented with a single-attribute representation or a multi-attribute representation using alternative combinations of attributes (see bottom-part of Table 2). For instance, From combines feature attributes From Address and From Name and Text represents all textual content by including attributes Subject and Body.

Descriptor

Email Representation

Date From Address From Name To Address To Name CC Address CC Name Subject Body From To CC Recipients Text All

Date = date f rom@ = (Addressf rom1 , ...Addressf romi ) f romN = (N amef rom1 , ...N amef romj ) to@ = (Addressto1 , ...Addresstok ) toN = (N ameto1 , ...N ametol ) Cc@ = (Addresscc1 , ...Addressccm ) ccN = (N amecc1 , ...N ameccn ) subject = (keywordsubj1 , ...keywordsubjn ) body = (keywordbody1 , ...keywordbodyn ) f rom = (f rom@ , f romN ) to = (to@ , toN ) cc = (Cc@ , ccN ) recipients = (to@ , toN , cc@ , ccN ) text = (subject, body) all = (Date, f rom@ , f romN , to@ , toN , Cc@ , ccN , subject, body) Table 2. Case representations for email

The advantage of a representation that preserves document structure is that similarity computations can be confined to corresponding email sections. Figure 1 illustrates the aggregation of local similarities Si into a global similarity S(E1 , E2 ) between emails E1 and E2 . Local similarities are computed for each corresponding attribute and aggregated using average. An obvious approach to represent email is to include all attributes in a single feature vector representation which we call All. The similarity between two emails with such a representation is computed as above. However, since seven out of the nine attributes are binary feature vectors, we considered multiple casebases where every casebase contains

E2

Date

From Address

To Address

To Name

CC Address

CC Name

Subject

Subject

Body

Similarity S9

CC Name

Similarity S8

Similarity S4

From Name

CC Address

Similarity S7

To Name

Similarity S6

To Address

Similarity S5

From Name

Similarity S3

Similarity S1

From Address

Similarity S2

Date S(E1,E2)=Average(Si)

E1

Body

Fig. 1. Similarity computation with semi-structured document representations.

the same set of emails but uses a different case representation. Each casebase uses one of the single-attribute representations listed in the top-half of Table 2. The prediction task involves aggregating retrieval results from multiple casebases. Essentially, our interest is to study how best to combine similarities from separate email sections.

4 Retrieval with Multiple Casebases

New Email

Body

Subject

CC Name

CC Address

To Name

From Name

3NN 3NN 3NN

To Address

From Address

Date

Previous work has shown that retrieval over multiple casebases can be achieved with ensembles of k-NN classifiers [7]. Each k-NN classifier constitutes a CBR system which we refer to as a base-learner. The final prediction is obtained by combining the predictions of each base-learner. Figure 2 illustrates majority voting, a common aggregation strategy. Each base-learner uses the same casebase, but with a different case representation, and the final prediction of the new email is the majority vote of the base-learners predictions.

3NN 3NN 3NN 3NN 3NN 3NN

base-learners predictions Final prediction = Majority Vote

Fig. 2. Aggregation with majority voting.

3NN

3NN 3NN 3NN

New Email

3NN 3NN 3NN Body

CC Name

3NN Subject

To Name

3NN CC Address

From Name

3NN To Address

Date

From Address

3NN

Meta Casebase

3NN

base-learners predictions

9 Casebases

However, voting only makes sense if the classifiers perform comparably well. If, for instance, six of the nine classifiers make incorrect predictions, the majority vote will be incorrect. Majority vote is therefore unsuitable when the relevances of base-learners vary. This is because it is unclear which classifier to trust and a dynamic weighting would have to be applied to capture the importance of each classifier in difference circumstances. For instance, when a case is represented using Date, the classifier is typically unable to make a prediction. However, when a prediction is made, it is highly likely to be correct. Therefore, when this classifier is able to make a prediction, its weight should be high, but otherwise low. In this work, since each classifier learns from the same set of emails represented differently, if each classifier performed comparably well, one classifier would suffice. The goal in using multiple classifiers is to exploit the complementarity of email attributes. Stacking combines multiple models differently by introducing the concept of meta-learner. It tries to learn which classifiers are reliable using another learning algorithm to discover how best to combine the predictions of the base-learners. It is suited to situations where base-learners are reliable in different circumstances [24]. Stacking is generally used to combine different types of base-learners. Here, we use stacking to combine the same type of base-learners, k-NN classifiers, but where each classifier uses a different case representation. Each base-learner provides a prediction, like in majority voting, but the meta-learner combines these predictions into a new case. Therefore, cases in the casebase used by the meta-learner, or meta-casebase, have as many attributes as there are base-learners.

3NN 3NN 3NN 3NN 3NN 3NN

3NN

base-learners predictions = new email representation

Final prediction

Fig. 3. S TACK representation

In this work, a new case is classified by nine base-learners, one for each singleattribute representation. The predictions are used to create a new case representation for the meta-learner called S TACK. The top-part of Figure 3 illustrates how the metacasebase is created by using the predictions of each base-learners to create a new case

representation, or S TACK representation. The bottom-part illustrates the classification of a new case. First, the new email is classified by the nine base-learners to obtain its S TACK representation. It is then classified by the meta-learner which uses the metacasebase where cases are represented with S TACK. The hypothesis supporting the S TACK representation is that the base-learners predictions of similar emails follow the same pattern. For instance, if two emails are classified similarly by the base-learners, their representation in the meta-casebase will be similar.

5 Creating Random Chronological Stratified Trials We introduce a new evaluation method, n-R CST, in order to create random chronological stratified trials that maintain the temporal aspect of email and respect the overall class distribution. Given a set of n emails belonging to m classes, n-R CST creates chronologically ordered email subsets Ei corresponding to each class. Each subset Ei is further split into k even splits Eij . Typically k=2, where one split is used for training and the other for testing. When using stacking, k=3: the new representation for the meta-learner is obtained for Ei1 and Ei2 using Ei2 and Ei3 respectively for training. Next, k-1 suitable  for each class ci , dates dj are identified in order to create k chronological splits Eij  . This enforces such that all emails in class ci received between dj−1 and dj form Eij th the temporal aspect on our trials because emails in any j split are more recent than emails in any j+1th split. Multiple stratified trials can then be generated by randomly  for testing and selecting a number of emails according to ci ’s distribution from Eij  Eij+1 for training. This algorithm is detailed in Figure 4 to illustrate its generality. Figure 5 illustrates n-R CST applied to a DELETE/KEEP classification task. First DELETE emails are separated from KEEP emails and put into chronologically ordered subsets. Assuming 2 splits are required, one for testing and one for training, each subset is further decomposed into 2 even splits (e.g. D1 and K1 ). The date at which the oldest email of each split has been received is identified (e.g. dD1 and dK1 ). Finally, a date d is chosen between dD1 and dK1 . Splits can be recreated so that delete1 and keep1 are all emails, from each class respectively, received after d, whilst delete2 and keep2 are those received before d. Trials are now created by randomly selecting emails from delete1 and keep1 for testing and delete2 and keep2 for training. We are now ensured that emails in the test set are more recent than those in the training set. Let assume we have a dataset of 1500 emails, including 900 emails labeled as DELETE and 600 emails labeled as KEEP . Imagine we want to create N trials, respecting the overall class distribution and the chronological order of the dataset, where each trial represents 10% of the dataset and contains 1/3 for testing and 2/3 for training. A trial must therefore contain 90 emails labeled as DELETE and 60 emails labeled as KEEP. Out of the 90 DELETE, 30 are used for testing and 60 for training. Similarly, out of the 60 KEEP, 20 are used for testing and 40 for training. After n-R CST is applied to the dataset (as in Figure 5), a training set is formed by randomly selecting 60 emails from delete2 and 40 from keep2 and the test set formed with 30 emails from delete1 and 20 from keep1 . Multiple trials can now be generated similarly, ensuring each is stratified

E = {e1 , ..., en }, set of emails C = {c1 , ..., cm }, set of classes D = {d1 , ..., dk }, set of dates k the number of splits required d(ei ) date at which email ei was received c(ei ) class of email ei Create a subset Ei per class ∀ci ∈ C, ∃Ei ⊂ E where, ∀e ∈ Ei , c(e) = ci Order subset chronologically ∀Ei , order so that ∀ej ∈ Ei , d(ej ) < d(ej+1 ) Create k even splits Eij for each subset Ei ∀Ei , Ei = Ei1 ∪ ... ∪ Eik and |Ei1 | ≈ ... ≈ |Eik | Get the date of the oldest email in each split Eij ∀Eij , dij = d(e∗ ), where e∗ ∈ Eij and ∀e ∈ Eij , d(e∗ ) < d(e) Create a set Dj with the dates of the oldest email in the j th split of all Eij subsets Dj = {dij } i Select a date dj between the most recent and the oldest date in Dj ∀Dj , min(Dj ) < dj < max(Dj )  Create k chronological splits Eij per class ci ,   ∀e ∈ E, Eij = Eij ∩ {e}, where dj−1 > d(e)  > dj , and c(e) = ci so that ∀ej ∈ Ej = Eji , d(ej ) > d(ej+1 )

DELETE (D)

KEEP (K)

DELETE (D)

KEEP (K)

delete1(D1)

keep1(K1)

dD1

dK1

delete2(D2) dD2

keep2(K2) dK2

delete’ 1 d

Personal User Emails

by selecting the correct number of emails from each class, and are chronological, as the emails from the test set have been received after the emails from the training set.

keep’ 1 d

delete’ 2

keep’ 2

i

= timeline

Fig. 4. n-R CST algorithm

Fig. 5. n-R CST for delete prediction

6 Evaluation and Results In this paper, we compare different email representations for the prediction of email deletion. An ideal dataset should contain the time at which an email has been deleted. Indeed, an email can be deleted for different reasons. If an email is irrelevant, it is likely to be deleted as soon as it is received. An email regarding a meeting or an event is likely to be kept until a certain date then deleted when it becomes obsolete. Therefore, prediction of email deletion should also include the time at which the email should be deleted. A dataset containing deletion time is not currently available and is hard to obtain for ethical reasons. Therefore, in this paper, we experiment on a binary classification task where an email is classified either as DELETE or KEEP. Our evaluation has three main objectives: – show that all email attributes contribute to the performance of a classifier – show that combining base-learners using a S TACK representation is a better approach than the combination of base-learners using majority voting or a multiattribute representation

– show that a meta-learner using a S TACK representation is more stable across users. In the remaining of this section, we will first present the Enron Corpus and how it was processed to create datasets for the purpose of our experiments. We will then define the experimental design. Finally, we will discuss the results for each objective. 6.1

Email Dataset

The raw version contains 619,446 emails belonging to 158 users [3]. The Czech Amphora Research Group (ARG) has processed the Enron corpus into 16 XML files [14]. We used the file which identifies duplicate emails to remove all duplicate emails and we also cleaned the corpus by removing folders such as all documents and discussion threads [13]. These folders are computer generated folders and do not reflect user behaviour. Our cleaned Enron corpus contains 288,195 emails belonging to 150 users. For each user, we assumed that every email in deleted items should be classified as DELETE and every other email should be classified as KEEP . We therefore do not take into consideration the time at which an email has been deleted. 6.2

Experimental Design

n-R CST could not be applied to some Enron users because the time period covered by emails labeled as DELETE and the time period covered by email labeled as KEEP do not overlap or do not have an overlapping period large enough to create train-test trials. We extracted 14 Enron users for which n-R CST was suitable and generated 25 trials for each. The class distribution and the number of emails for these users is detailed in Table 3. User % of Deleted Emails User Number of Emails Dean-C 0.205 Lucci-P 753 Watson-K 0.214 Bass-E 754 Heard-M 0.270 Giron-D 767 Quigley-D 0.376 Shively-H 769 White-S 0.426 Heard-M 833 Schoolcraft-D 0.447 Quigley-D 988 Giron-D 0.495 Mims-Thurston-P 1028 Zipper-A 0.511 Zipper-A 1056 Bass-E 0.533 Thomas-P 1099 Parks-J 0.612 Schoolcraft-D 1321 Thomas-P 0.621 Dean-C 1540 Mims-Thurston-P 0.661 Parks-G 1661 Lucci-P 0.754 Watson-K 1974 Shively-H 0.765 White-S 2296 Table 3. Class distribution and Number of Email for selected Enron Users

Three splits were generated for each user using n-R CST. In order to evaluate the email representations listed in Table 2, one test set and one train set are required. A trial is created by randomly selecting emails from the most recent split for testing and from the second most recent split for training. In order to evaluate the stack representation, one test set and one train set of emails represented with S TACK are required. Emails from the most recent split are classified by the base-learners using the second most recent split for training. Similarly, emails from the second most recent split are classified by the base-learners using the third split for training. This provides us with a stack representation for emails in the first and second splits, allowing us to use the first split for testing and the second for training. The creation of trials is illustrated in Figure 6 and Figure 7. We maintain consistency in that, for all representations, test sets and training sets are identical.

Delete’1

Test Set

Keep’1

Delete’2

Train Set

Keep’2

Fig. 6. Trial creation for single feature vector representation

9 base-learners Delete’1

set1

Keep’1

Delete’2a

set2

Keep’2a

Delete’2b

set3

Keep’2b

STACK Representation Test Set Train Set

Fig. 7. Trial creation for S TACK representation

Once the trials are generated, a casebase is created for each case representation for each trial. All base-learners and the meta-learner are k-Nearest-Neighbour classifiers

with k = 3. Further research on the optimal number of neighbours to consider for each base-learner would be beneficial but is not considered further in this paper. For binary feature vectors attributes, the similarity is computed using the Euclidean distance. For nominal attributes, the similarity is 1 if the values are identical, 0 otherwise. During retrieval, any ties between multiple candidates are resolved randomly. Alternative representations are compared using classification accuracy. Since our data is not normally distributed, significance results are based on a 95% confidence level when applying the Kruskal Wallis test to three or more datasets and 99% confidence when applying the Wilcoxen signed-rank test to two data sets. 6.3

Is using All attributes best?

We compared the classification accuracy using the 12 email representations listed in Table 2. Our results show that using only the textual attributes (Subject, Body and Text) or using all the attributes (All) result in significantly better results compared to using any of the other attributes. Additionally, using All attributes gives significantly better results across users than Subject, Body and Text. This suggests that representing emails only using the textual attributes is not enough and that useful information for retrieval can be extracted from non-textual attributes. However, it is important to note that the accuracy achieved with the best single-attribute representation (All) is outperformed by simply predicting the majority class for 8 users out of 14. This is clearly illustrated in Figure 8, where the accuracy achieved with the four best representations is compared to the accuracy achieved by systematically predicting the majority class. We can note that All tends to perform well when classes are evenly distributed but struggles on highly skewed data.

Fig. 8. Accuracy achieved with different email representations

6.4

What is the best way to combine base-learners?

The All representation, where all email attributes are included in a feature vector, is the simplest way to combine email attributes. Since a classifier using such a representation

does not perform well on highly skewed data, we evaluated an alternative approach. A casebase is created for each single-attribute representation. The performance of k-NN using All representation is compared to an ensemble of 9 k-NN classifiers combined using majority vote (M AJORITY), each using a different casebase. Unlike All, M AJORITY implicitly captures the importance of each attribute by giving a weight to each classifier based on the similarity between the retrieved neighbours and the new case. This is because the prediction of each base-learner is a numeric value between 0 and 1 based on the similarity of the new case to the 3 nearest neighbours. For instance, for a given base-learner, let the similarities between the new case and the 3 nearest neighbours 0.2, 0.6 and 0.8 and their class KEEP, DELETE and DELETE respectively. The prediction for this base learner is (0.2∗0+0.6∗1+0.8∗1)/(0.2+0.6+0.8) = 0.875. If the prediction is smaller than 0.5, the email is classified as KEEP, otherwise it is classified as DELETE. The closer the prediction is to 0 or 1, the more the base-learner is confident that the class for the new case should be KEEP or DELETE respectively. The majority vote is calculated by averaging the predictions of all the base-learners. We therefore expect M AJORITY to perform better than All. However, significance test show that both approaches perform comparably. This is clearly illustrated with a scatter-plot in Figure 9. This suggests that the global similarity, or similarity across all attributes, is equivalent to the combination of local similarities, or similarities of individual attributes.

Fig. 9. Accuracy and Delete precision for 15 Enron users using M AJORITY and All

We then compared S TACK to both previous approaches. Results show that k-NN using S TACK representation performs significantly better than M AJORITY or k-NN using All representation. The scatter-plot in Figure 10 provides a closer look at S TACK and M AJORITY results. The performance achieved using the S TACK representation can be explained by its ability to generate predictions based on similarity values computed over the predictions of multiple base-learners. When combining email attributes using M A JORITY , a good performance is expected only if base-learners tend to agree on a prediction. However, S TACK further exploits the fact that if the ensemble of base-learners agree and disagree in a similar way then the emails are also similar. For instance, if

k-NN using Date predicts DELETE and k-NN using From predicts KEEP, M AJORITY would struggle to make a judgment whilst S TACK will decide based on other emails classified similarly.

Fig. 10. Accuracy and Delete precision for 15 Enron users using M AJORITY and S TACK

It is interesting to note that S TACK is also more robust to highly skewed data. Figure 11 shows how S TACK significantly outperforms a classifier consistently predicting the majority class.

Fig. 11. Accuracy using different aggregation methods

6.5

Is a meta-learner more consistent across users than a base-learner?

Email management is challenging because every individual deals with emails in a different way. A classifier can perform very well for one user and very poorly for another.

In this work, the availability of emails from 14 different users permits us to compare the consistency of the 3 approaches. The accuracy, for each user, achieved with the 4 best email representations from Table 2 are illustrated in Figure 8. It is clear that all approaches achieve inconsistent results across users. The accuracy with Subject varies from 0.49 to 0.75; Body from 0.46 to 0.74; Text from 0.37 to 0.70 and All from 0.48 to 0.83. Even if All significantly performs better, it is clearly inconsistent. It may perform extremely well for some users such as Schoolcraft but extremely poorly for others such as Bass. Such an approach is therefore unsuitable in a real life system. The classification accuracy using M AJORITY, All and S TACK appear in Figure 11. M AJORITY seems to be comparable to S TACK in terms of consistency, but S TACK still significantly outperforms M AJORITY in terms of overall accuracy.

7 Conclusions and Future Work An email representation including both structured and non-structured content results in significantly better retrieval when compared to a bag-of-word representation of just the textual content. Semi-structured content can be dealt with at the representation stage by incorporating it into a single feature vector or alternatively at the retrieval stage by use of multiple casebases, each using a different representation. Multiple casebases, when combined with a stacked representation, perform significantly better than when combined using majority voting. S TACK, M AJORITY and other alternative representation strategies are evaluated using n-R CST; a novel evaluation technique to create stratified train-test trials respecting the temporal aspect of emails. This methodology is applicable to any classification task dealing with temporal data. Future work will investigate the impact of feature selection techniques to optimise case representation in each casebase and the allocation of weights to the prediction of each classifier. It is also important to evaluate the generality of the S TACK representation for other email management tasks (e.g Reply, Forward, File). Finally, it would be interesting to compare this approach with other machine learning methods such as Support Vector Machines and Naive Bayes.

References 1. Stephen D. Bay. Combining nearest neighbor classifiers through multiple feature subsets. In Proceedings of the International Conference on Machine Learning (ICML’98), pages 37–45. 1998. 2. Ron Bekkerman, Andrew McCallum, and Gary Huang. Automatic categorization of email into folders: Benchmark experiments on Enron and Sri Corpora. Technical report, UMass CIIR, 2004. 3. William W. Cohen. Enron email dataset. http://www.cs.cmu.edu/ enron/, April 2005. 4. William W. Cohen, Vitor R. Carvalho, and Tom M. Mitchell. Learning to classify email into “speech acts”. In Proceedings of the conference on Empirical Methods in Natural Language Processing (EMNLP’04), pages 309–316. 2004. 5. S. Craw, J. Jarmulak, and R. Rowe. Maintaining retrieval knowledge in a case-based reasoning system. Computational Intelligence, 17:346–363, 2001. 6. Padraig Cunningham and John Carney. Diversity versus quality in classification ensembles based on feature selection. In Proceedings of the European Conference on Machine Learning (ECML’00), pages 109–116. 2000.

7. Padraig Cunningham and Gabriele Zenobi. Case representation issues for case-based reasoning from ensemble research. In Proceedings of the International Conference on Case-Based Reasoning (ICCBR’01), 2001. 8. Laura Dabbish, Gina Venolia, and JJ Cadiz. Marked for deletion: an analysis of email data. In Proceedings of the Conference on Human Factors in Computing Systems (CHI’30), pages 924–925. 2003. 9. Laura A. Dabbish, Robert E. Kraut, Susan Fussel, and Sara Kiesler. Understanding email use: Predicting action on a message. In Proceedings of the Conference on Human Factors in Computing Systems (SIGCHI’05), pages 691–700. 2005. 10. S.J. Delany, P. Cunningham, and L. Coyle. An assessment of case-based reasoning for spam filtering. Volume 24, pages 359–378. Springer Science+Business Media B.V., 2005. 11. Yanlei Diao, Hongjun Lu, and Dekai Wu. A comparative study of classification based personal e-mail filtering. In Proceedings of the Fourth Pacific-Asia Conference on Knowledge Discovery and Data MiningPAKDD (PAKDD’–), 2000. 12. Mark Dredze, John Blitzer, and Fernando Pereira. ”Sorry, I forgot the attachment:” email attachment prediction. In Proceddings of the Conference on Email and Anti-Spam (CEAS’06), 2006. 13. Jiri Dvorsk, Petr Gajdos, Eliska Ochodkova, Jan Martinovic, and Vclav Snsel. Social network problem in enron corpus. In Proceedings of the East-European Conference on Advances in Databases and Information Systems (ADBIS’05). 2005 14. Amphora Research Group. http://arg.vsb.cz/arg/Enron Corpus/default.aspx. 15. A. Gupta and R. Sekar. An approach for detecting self-propagating email using anomaly detection. In Proceedings of the International Symposium on Recent Advances in Intrusion Detection. 2003. 16. Svetlana Kiritchenko and Stan Matwin. Email classification with co-training. In Proceedings of the Conference of the Centre for Advanced Studies on Collaborative research (CASCON’01). 2001. 17. Bryan Klimt and Yiming Yang. The Enron Corpus: A new dataset for email classification research. In Proceedings of the European Conference on Machine Learning (ECML’04). 2004. 18. Luc Lamontagne and Guy Lapalme. Textual reuse for email response. In Advances in CaseBased Reasoning, Lecture Notes in Computer Science, volume 3155, pages 242–256. 2004. 19. Wendy E. Mackay. Diversity in the use of electronic mail: a preliminary inquiry. ACM Transactions on Information Systems, 6(4):380–397, 1988. 20. Srikanth Palla and Ram Dantu. Detecting phishing in emails. Spam Conference 2006. 2006. 21. Georgios Sakkis, Ion Androutsopoulos, Georgios Paliouras, Vangelis Karkaletsis, Constantine D. Spyropoulos, and Panagiotis Stamatopoulos. Stacking classifiers for anti-spam filtering of e-mail. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 44–50, 2001. 22. Marc A. Smith, Jeff Ubois, and Benjamin M. Gross. Forward thinking. In Proceedings of the Conference on Email and Anti-Spam (CEAS’05). 2005. 23. Joshua R. Tyler and John C. Tang. ”When can I expect an email response?”: a study of rhythms in email usage. In Proceedings of the European Conference on Computer-Supported Cooperative Work. 2003. 24. I. H. Witten and Eibe. Frank. Data Mining: Practical Machine Learning Tools and Techniques. Morgan Kaufmann, 2nd ed. edition, 2005. 25. David H. Wolpert. Stacked generalization. Neural Networks, 5(2):241–259, 1992.

k-NN Aggregation with a Stacked Email Representation

Number of Emails. Dean-C. 0.205. Lucci-P. 753. Watson-K. 0.214. Bass-E. 754. Heard-M. 0.270 ..... Marc A. Smith, Jeff Ubois, and Benjamin M. Gross. Forward ...

452KB Sizes 0 Downloads 265 Views

Recommend Documents

Representation and aggregation of preferences ... - ScienceDirect.com
Available online 1 November 2007. Abstract. We axiomatize in the Anscombe–Aumann setting a wide class of preferences called rank-dependent additive ...

Harsanyi's Aggregation Theorem with Incomplete Preferences
... Investissements d'Ave- nir Program (ANR-10-LABX-93). .... Bi-utilitarianism aggregates two utility functions ui and vi for each individual i = 1, … , I, the former ...

Range Aggregation with Set Selection
“find the minimum price of 5-star hotels with free parking and a gym whose .... such pairs, such that the storage of all intersection counts ...... free, internet,. 3.

Harsanyi's Aggregation Theorem with Incomplete Preferences
rem to the case of incomplete preferences at the individual and social level. Individuals and society .... Say that the preference profile ( ≿ i) i=0. I satisfies Pareto ...

Chronologically stacked coupon mailer
Jun 9, 2011 - detachable redeemable coupons redeemable through a. 2,813,728 A ... Many businesses and cooperative advertisers mail sheets containing a ...

Using Sub-sequence Information with kNN for ...
in a computer system in order to detect signs of security problems [2]. ..... Rate of increase in false positive is less for Jaccard similarity measure (0.005-.

A utility representation theorem with weaker continuity ...
Sep 10, 2009 - We prove that a mixture continuous preference relation has a utility represen- tation if its domain is a convex subset of a finite dimensional ...

A utility representation theorem with weaker continuity ...
Sep 10, 2009 - ... of the Japan Society for the Promotion of Science (JSPS) and financial ... In conjunction with the independence axiom, Herstein and Milnor [7] ...

A utility representation theorem with weaker continuity condition
Sep 4, 2008 - The main step in the proof of our utility representation theorem is to show ...... If k ≥ 3, we can apply Lemma 1 again to D ∩ co({x0} ∪ Y ε/(k−1).

Fuzzy-KNN-Prediksii.pdf
of Secondary Structure of Proteins (DSSP) [17] where the secondary structure is. classified as eight states. By grouping these eights states into three classes Coil.

A utility representation theorem with weaker continuity condition
Sep 4, 2008 - http://www.wiwi.uni-bielefeld.de/˜imw/Papers/showpaper.php?401 .... We prove that for any linearly continuous preference relation, the set of ..... {si,si},i = 2,...,k − 1}), where ri(C) stands for the relative interior of the set C.

Black Stacked Xtra.pdf
Black Stacked Xtra.pdf. Black Stacked Xtra.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying Black Stacked Xtra.pdf.

Strategyproof and efficient preference aggregation with ...
intuitive as a technical continuity check, bounded response seems to lack a strong normative ... status-quo rules, though not K-efficient, are K-strategyproof on the entire profile domain. ..... manipulability comes at a significant cost to efficienc

Email and Email Servers - GitHub
Oct 19, 2017 - With a friend(s)… 1. Define Email. 2. Discuss what you think makes Email unique from other digital communication methods (e.g., IRC, Hangouts,. Facebook, Slack, etc.) Sorry this feels a bit like a lecture in a course… but hopefully

Brownian aggregation rate of colloid particles with ...
Aug 13, 2014 - fore to derive an analytical solution, which give a simple way of calculating the rate ... perimental data, as well as for the theoretical simulation of.

SVM-KNN: Discriminative Nearest Neighbor ...
Michael Maire. Jitendra Malik. Computer Science Division, EECS Department. Univ. of California, Berkeley, CA 94720 ... yields excellent results in practice. The basic idea is to find close neighbors to a query sample and ... rather by similarity to p

Journal of Functional Programming A representation ... - CiteSeerX
Sep 8, 2015 - programmers and computer scientists, providing and connecting ...... (n, bs, f)). Furthermore, given a traversal of T, a coalgebra for UR. ∗.

Brownian aggregation rate of colloid particles with ...
y = C0/C − 1,27 since they are linear in time for the partic- ular case of a .... dantseva, V. P. Maltsev, and A. V. Chernyshev, Colloids Surf., B 32, 245. (2003). 5E.

Journal of Functional Programming A representation ... - CiteSeerX
DOI: 10.1017/S0956796815000088, Published online: 08 September 2015 ... programmers and computer scientists, providing and connecting different views on ... over a class of functors, such as monads or applicative functors. ..... In order to make the

TATTING-JEWELRY-STACKED-FLOWER-PATTERN-DUSENBURY ...
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. TATTING-JEWELRY-STACKED-FLOWER-PATTERN-DUSENBURY-©1990.pdf. TATTING-JEWELRY-STACKED-FLOWER-PATTERN-DUSENBUR

A Unique Costly Contemplation Representation
printed and reproduced only for educational or research purposes, including use ...... which (heuristically) puts μ(u)/λ weight on each v = λu ∈ V. Then, by a sim-.

MultiVec: a Multilingual and Multilevel Representation Learning ...
of data. A number of contributions have extended this work to phrases [Mikolov et al., 2013b], text ... on a large corpus of text, the embeddings of a word Ci(w).

Online Rank Aggregation
We consider an online learning framework where the task is to predict a ... a loss defined as Kendall tau distance between the predicted permutation and the.