skip to main content
research-article
Free Access

Compressed linear algebra for declarative large-scale machine learning

Published:24 April 2019Publication History
Skip Abstract Section

Abstract

Large-scale Machine Learning (ML) algorithms are often iterative, using repeated read-only data access and I/O-bound matrix-vector multiplications. Hence, it is crucial for performance to fit the data into single-node or distributed main memory to enable fast matrix-vector operations. General-purpose compression struggles to achieve both good compression ratios and fast decompression for block-wise uncompressed operations. Therefore, we introduce Compressed Linear Algebra (CLA) for lossless matrix compression. CLA encodes matrices with lightweight, value-based compression techniques and executes linear algebra operations directly on the compressed representations. We contribute effective column compression schemes, cache-conscious operations, and an efficient sampling-based compression algorithm. Our experiments show good compression ratios and operations performance close to the uncompressed case, which enables fitting larger datasets into available memory. We thereby obtain significant end-to-end performance improvements.

References

  1. Abadi, M. et al. TensorFlow: A system for large-scale machine learning. In OSDI (2016). Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Alexandrov, A. et al. The stratosphere platform for big data analytics. VLDB J. 23, 6 (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. American Statistical Association (ASA). Airline on-time performance dataset. stat-computing.org/dataexpo/2009.Google ScholarGoogle Scholar
  4. Boehm, M., et al. SystemML: Declarative machine learning on spark. PVLDB 9, 13 (2016). Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Bottou, L. The infinite MNIST dataset. leon.bottou.org.Google ScholarGoogle Scholar
  6. Chitta, R. et al. Approximate Kernel k-means: Solution to large scale Kernel clustering. In KDD (2011). Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Cohen, J. et al. MAD skills: New analysis practices for big data. PVLDB 2, 2 (2009). Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Elgamal, T. et al. SPOOF: Sum-product optimization and operator fusion for large-scale machine learning. In CIDR (2017).Google ScholarGoogle Scholar
  9. Elgohary, A. et al. Compressed linear algebra for large-scale machine learning. PVLDB 9, 12 (2016). Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Elgohary, A. et al. Compressed linear algebra for large-scale machine learning. VLDB J. (2017a). Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Elgohary, A., Boehm, M., Haas, P.J., Reiss, F.R., and Reinwald, B. Scaling Machine Learning via Compressed Linear Algebra. SIGMOD Record 46, 1 (2017b). Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Good, I.J. The population frequencies of species and the estimation of population parameters. Biometrika (1953).Google ScholarGoogle Scholar
  13. Haas, P.J. and Stokes, L. Estimating the number of classes in a finite population. JASA 93, 444 (1998).Google ScholarGoogle Scholar
  14. Johnson, N.L. et al. Univariate Discrete Distributions, 2nd edn. Wiley, New York, 1992.Google ScholarGoogle Scholar
  15. Kourtis, K. et al. Optimizing sparse matrix-vector multiplication using index and value compression. In CF (2008). Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Lichman, M. UCI machine learning repository: Higgs, covertype, US census (1990). archive.ics.uci.edu/ml/.Google ScholarGoogle Scholar
  17. Schelter, S. et al. Samsara: Declarative machine learning on distributed dataflow systems. NIPS ML Systems (2016).Google ScholarGoogle Scholar
  18. Williams, S. et al. Roofline: An insightful visual performance model for multicore architectures. Commun. ACM 52, 4 (2009). Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Zadeh, R.B. et al. Matrix computations and optimization in apache spark. In KDD (2016).Google ScholarGoogle Scholar
  20. Zaharia, M. et al. Resilient distributed datasets: A fault-tolerant abstraction for in-memory cluster computing. In NSDI (2012). Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Compressed linear algebra for declarative large-scale machine learning

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image Communications of the ACM
        Communications of the ACM  Volume 62, Issue 5
        May 2019
        83 pages
        ISSN:0001-0782
        EISSN:1557-7317
        DOI:10.1145/3328504
        Issue’s Table of Contents

        Copyright © 2019 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 24 April 2019

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format