skip to main content
10.1145/956750.956772acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
Article

Aggregation-based feature invention and relational concept classes

Published: 24 August 2003 Publication History

Abstract

Model induction from relational data requires aggregation of the values of attributes of related entities. This paper makes three contributions to the study of relational learning. (1) It presents a hierarchy of relational concepts of increasing complexity, using relational schema characteristics such as cardinality, and derives classes of aggregation operators that are needed to learn these concepts. (2) Expanding one level of the hierarchy, it introduces new aggregation operators that model the distributions of the values to be aggregated and (for classification problems) the differences in these distributions by class. (3) It demonstrates empirically on a noisy business domain that more-complex aggregation methods can increase generalization performance. Constructing features using target-dependent aggregations can transform relational prediction tasks so that well-understood feature-vector-based modeling algorithms can be applied successfully.

References

[1]
J. M. Aronis and F. J. Provost. Efficiently constructing relational features from background knowledge for inductive machine learning. In U. Fayyad and R. Uthurusamy, editors, In Working Notes of the AAAI-94 Workshop on Knowledge Discovery in Databases (KDD-94), pages 347--358, 1994.]]
[2]
Hendrik Blockeel and Luc De Raedt. Top-down induction of first-order logical decision trees. Artificial Intelligence, 101(1--2):285--297, 1998.]]
[3]
A. P. Bradley. The use of the area under the ROC curve in the evaluation of machine learning algorithms. In Pattern Recognition, volume 30(7), pages 1145--1159, 1997.]]
[4]
L. De Raedt, H. Blockeel, L. Dehaspe, and W. Van Laer. Three companions for data mining in first order logic. In Saso Dzeroski and Nada Lavrac, editors, Relational Data Mining, pages 105--139. Springer-Verlag, 2001.]]
[5]
J. Fürnkranz. Dimensionality reduction in ILP: A call to arms. In Raedt L. and Muggleton S., editors, Proceedings of the IJCAI-97 Workshop on Frontiers of Inductive Logic Programming, 1997.]]
[6]
D. Jensen and P. R. Cohen. Multiple comparisons in induction algorithms. In Machine Learning, volume 38, pages 309--338, 2000.]]
[7]
D. Jensen and J. Neville. Data mining in social networks. In Dynamic Social Networks Modeling and Analysis, 2002.]]
[8]
D. Jensen and J. Neville. Linkage and autocorrelation cause feature selection bias in relational learning. In 19th International Conference on Machine Learning, 2002.]]
[9]
M. Kirsten, S. Wrobel, and T. Horvath. Distance based approaches to relational learning and clustering. In Saso Dzeroski and Nada Lavrac, editors, Relational Data Mining, pages 213--232. Springer Verlag, 2001.]]
[10]
A. Knobbe, M. De Haas, and A. Siebes. Propositionalisation and aggregates. In LNAI, volume 2168, pages 277--288, 2001.]]
[11]
D. Koller and A. Pfeffer. Probabilistic frame-based systems. In AAAI/IAAI, pages 580--587, 1998.]]
[12]
S. Kramer, N. Lavrac, and P. Flach. Propositionalization approaches to relational data mining. In Saso Dzeroski and Nada Lavrac, editors, Relational Data Mining, pages 262--291. Springer-Verlag, 2001.]]
[13]
S. Kramer, B. Pfahringer, and C. Helma. Stochastic propositionalization of non-determinate background knowledge. In International Workshop on Inductive Logic Programming, pages 80--94, 1998.]]
[14]
P. C. Mahalanobis. On the generalized distance in statistics. In Proc. Natl. Institute of Science of India, volume 12, pages 49--55, 1936.]]
[15]
T. Masters. Practical Neural Network Recipes in C++. San Diego: Academic Press, 1993.]]
[16]
E. McCreath. Induction in first order logic from noisy training examples and fixed example set size. In PhD Thesis, 1999.]]
[17]
S. H. Muggleton. Cprogol4.4: a tutorial introduction. In Saso Dzeroski and Nada Lavrac, editors, Relational Data Mining, pages 105--139. Springer-Verlag, 2001.]]
[18]
S. H. Muggleton and L. DeRaedt. Inductive logic programming: Theory and methods. The Journal of Logic Programming, 19 & 20:629--680, May 1994.]]
[19]
A. Popescul, L. H. Ungar, S. Lawrence, and D. M. Pennock. Structural logistic regression: Combining relational and statistical learning. In Proceedings of the Workshop on Multi-Relational Data Mining (MRDM-2002), pages 130--141. University of Alberta, Edmonton, Canada, July 2002.]]
[20]
F. Provost, C. Perlich, and S. Macskassy. Relational learning problems and simple models. In Lise Getoor and David Jensen, editors, Working Notes of the IJCAI-2003 Workshop on Learning Statistical Models from Relational Data, pages 116--120, 2003.]]
[21]
J. R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers, Los Altos, California, 1993.]]
[22]
J. R. Quinlan and R. M. Cameron-Jones. Foil: A midterm report. In P. Brazdil, editor, Proceedings of the 6th European Conference on Machine Learning, volume 667, pages 3--20. Springer-Verlag, 1993.]]
[23]
M. Smith. Neural Networks for Statistical Modeling. Boston: International Thomson Computer Press, 1996.]]
[24]
A. Srinivasan and R. D. King. Feature construction with inductive logic programming: A study of quantitative predictions of biological activity aided by structural attributes. In S. Muggleton, editor, Proceedings of the 6th International Workshop on Inductive Logic Programming, pages 352--367. Stockholm University, Royal Institute of Technology, 1996.]]

Cited By

View all

Index Terms

  1. Aggregation-based feature invention and relational concept classes

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    KDD '03: Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining
    August 2003
    736 pages
    ISBN:1581137370
    DOI:10.1145/956750
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 24 August 2003

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. aggregation
    2. constructive induction
    3. feature construction
    4. propositionalization
    5. relational learning

    Qualifiers

    • Article

    Conference

    KDD03
    Sponsor:

    Acceptance Rates

    KDD '03 Paper Acceptance Rate 46 of 298 submissions, 15%;
    Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)5
    • Downloads (Last 6 weeks)1
    Reflects downloads up to 06 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media