skip to main content
10.1145/1141277.1141536acmconferencesArticle/Chapter ViewAbstractPublication PagessacConference Proceedingsconference-collections
Article

Exploiting partial decision trees for feature subset selection in e-mail categorization

Published: 23 April 2006 Publication History

Abstract

In this paper we propose PARTfs which adopts a supervised machine learning algorithm, namely partial decision trees, as a method for feature subset selection. In particular, it is shown that an aggressive reduction of the feature space can be achieved with PARTfs while still allowing for comparable classification results with conventional feature selection metrics. The approach is empirically verified by employing two different document representations and four different text classification algorithms that are applied to a document collection consisting of personal e-mail messages. The results show that a reduction of the feature space in the magnitude of ten is achievable without loss of classification accuracy.

References

[1]
D. Aha, D. Kibler, and M. Albert. Instance-Based Learning Algorithms. Machine Learning, 6(1), 1991.
[2]
H. Berger, M. Köhle, and D. Merkl. On the Impact of Document Representation on Classifier Performance in eMail Categorization. In Proc. Int'l Conf. Information Systems Technology and its Applications, 2005.
[3]
H. Berger and D. Merkl. A Comparison of Text-Categorization Methods applied to N-Gram Frequency Statistics. In Proc. of the 17th Australian Joint Conf. on Artificial Intelligence, 2004.
[4]
W. B. Cavnar and J. M. Trenkle. N-gram-based text categorization. In Proc. Int'l Symp. on Document Analysis and Information Retrieval, 1994.
[5]
W. W. Cohen. Fast effective rule induction. In Proc. of the Int'l Conf. on Machine Learning, pages 115--123. Morgan Kaufmann, 1995.
[6]
E. Crawford, I. Koprinska, and J. Patrick. Phrases and feature selection in e-mail classification. In Proc. 9th Australasian Document Computing Symp., 2004.
[7]
G. Forman. An extensive empirical study of feature selection metrics for text classification. Journal of Machine Learning Research, 3:1289--1305, 2003.
[8]
E. Frank and I. H. Witten. Generating accurate rule sets without global optimization. In Proc. of the Int'l Conf. on Machine Learning, pages 144--151. Morgan Kaufmann Publishers Inc., 1998.
[9]
M. A. Hall and G. Holmes. Benchmarking attribute selection techniques for discrete class data mining. IEEE Transactions on Knowledge and Data Engineering, 15(6):1437--1447, 2003.
[10]
G. H. John, R. Kohavi, and K. Pfleger. Irrelevant features and the subset selection problem. In Proc. of the 11th Int'l Conf. on Machine Learning, pages 121--129, 1994.
[11]
A. McCallum and K. Nigam. A Comparison of Event Models for Naive Bayes Text Classification. In Proc. of AAAI-98 Workshop on Learning for Text Categorization, 1998.
[12]
T. Mitchell. Machine Learning. McGraw-Hill, 1997.
[13]
D. Mladenic. Feature subset selection in text-learning. In Proc. of the 10th European Conf. on Machine Learning, pages 95--100, UK, 1998.
[14]
J. Platt. Fast Training of Support Vector Machines using Sequential Minimal Optimization. In Advances in Kernel Methods - Support Vector Learning, pages 185--208. MIT Press, 1999.
[15]
J. R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers Inc., 1993.
[16]
I. H. Witten and E. Frank. Data Mining: Practical machine learning tools with Java implementations. Morgan Kaufmann, San Francisco, 2000.
[17]
Y. Yang and X. Liu. A re-examination of text categorization methods. In Proc. of the Int'l ACM SIGIR Conf. on R&D in Information Retrieval, 1999.

Cited By

View all

Index Terms

  1. Exploiting partial decision trees for feature subset selection in e-mail categorization

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    SAC '06: Proceedings of the 2006 ACM symposium on Applied computing
    April 2006
    1967 pages
    ISBN:1595931082
    DOI:10.1145/1141277
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 23 April 2006

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. feature selection
    2. indexing methods
    3. information filtering
    4. machine learning
    5. text categorization

    Qualifiers

    • Article

    Conference

    SAC06
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 1,650 of 6,669 submissions, 25%

    Upcoming Conference

    SAC '25
    The 40th ACM/SIGAPP Symposium on Applied Computing
    March 31 - April 4, 2025
    Catania , Italy

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)3
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 09 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media