skip to main content
10.1145/3097983.3098175acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

BDT: Gradient Boosted Decision Tables for High Accuracy and Scoring Efficiency

Published:13 August 2017Publication History

ABSTRACT

In this paper we present gradient boosted decision tables (BDTs). A d-dimensional decision table is essentially a mapping from a sequence of d boolean tests to a real value in {R}. We propose novel algorithms to fit decision tables. Our thorough empirical study suggests that decision tables are better weak learners in the gradient boosting framework and can improve the accuracy of the boosted ensemble. In addition, we develop an efficient data structure to represent decision tables and propose a novel fast algorithm to improve the scoring efficiency for boosted ensemble of decision tables. Experiments on public classification and regression datasets demonstrate that our method is able to achieve 1.5x to 6x speedups over the boosted regression trees baseline. We complement our experimental evaluation with a bias-variance analysis that explains how different weak models influence the predictive power of the boosted ensemble. Our experiments suggest gradient boosting with randomly backfitted decision tables distinguishes itself as the most accurate method on a number of classification and regression problems. We have deployed a BDT model to LinkedIn news feed system and achieved significant lift on key metrics.

References

  1. E. Bauer and R. Kohavi. An empirical comparison of voting classification algorithms: Bagging, boosting, and variants. Machine learning, 36(1):105--139, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. L. Breiman. Bagging predictors. Machine learning, 24(2):123--140, 1996. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. L. Breiman, J. Friedman, C. Stone, and R. Olshen. Classification and regression trees. CRC press, 1984.Google ScholarGoogle Scholar
  4. C. Burges. From ranknet to lambdarank to lambdamart: An overview. Learning, 11:23--581, 2010.Google ScholarGoogle Scholar
  5. G. Capannini, C. Lucchese, F. M. Nardini, S. Orlando, R. Perego, and N. Tonellotto. Quality versus efficiency in document scoring with learning-to-rank models. Information Processing & Management, 52(6):1161--1177, 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. R. Caruana and A. Niculescu-Mizil. An empirical comparison of supervised learning algorithms. In ICML, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. J. Friedman. Greedy function approximation: a gradient boosting machine. Annals of Statistics, 29:1189--1232, 2001. Google ScholarGoogle ScholarCross RefCross Ref
  8. J. Friedman. Stochastic gradient boosting. Computational Statistics and Data Analysis, 38:367--378, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical view of boosting (with discussion and a rejoinder by the authors). Annals of Statistics, 28:337--407, 2000. Google ScholarGoogle ScholarCross RefCross Ref
  10. Y. Ganjisaffar, R. Caruana, and C. V. Lopes. Bagging gradient-boosted trees for high precision, low variance ranking models. In SIGIR, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. S. Geman, E. Bienenstock, and R. Doursat. Neural networks and the bias/variance dilemma. Neural computation, 4(1):1--58, 1992. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. R. Kelley Pace and R. Barry. Sparse spatial autoregressions. Statistics & Probability Letters, 33(3):291--297, 1997. Google ScholarGoogle Scholar
  13. P. Li. Robust logitboost and adaptive base class (abc) logitboost. In UAI, 2010.Google ScholarGoogle Scholar
  14. P. Li, C. Burges, and Q. Wu. Mcrank: Learning to rank using multiple classification and gradient boosting. In NIPS, 2007.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Y. Lou, R. Caruana, J. Gehrke, and G. Hooker. Accurate intelligible models with pairwise interactions. In KDD, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. C. Lucchese, F. M. Nardini, S. Orlando, R. Perego, N. Tonellotto, and R. Venturini. Quickscorer: A fast algorithm to rank documents with additive ensembles of regression trees. In SIGIR, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. D. Y. Pavlov, A. Gorodilov, and C. A. Brunk. Bagboo: a scalable hybrid bagging-the-boosting model. In CIKM, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. S. Tyree, K. Q. Weinberger, K. Agrawal, and J. Paykin. Parallel boosted regression trees for web search ranking. In WWW, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. S. M. Weiss and N. Indurkhya. Rule-based machine learning methods for functional prediction. Journal of Artificial Intelligence Research, 3:383--403, 1995.yGoogle ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. BDT: Gradient Boosted Decision Tables for High Accuracy and Scoring Efficiency

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      KDD '17: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
      August 2017
      2240 pages
      ISBN:9781450348874
      DOI:10.1145/3097983

      Copyright © 2017 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 13 August 2017

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      KDD '17 Paper Acceptance Rate64of748submissions,9%Overall Acceptance Rate1,133of8,635submissions,13%

      Upcoming Conference

      KDD '24

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader