skip to main content
10.1145/1344671.1344702acmconferencesArticle/Chapter ViewAbstractPublication PagesfpgaConference Proceedingsconference-collections
research-article

Reconfigurable computing for learning Bayesian networks

Published:24 February 2008Publication History

ABSTRACT

Learning the structure of Bayesian networks(BNs) is known to be NP-complete and most of the recent work in the field is based on heuristics. Many recent approaches to the problem trade correctness and exactness for faster computation and are still computationally infeasible, except for networks with few variables. In this paper we present a software/hardware co-design approach to learning Bayesian networks from experimental data that is scalable to very large networks. Our implementation improves the performance of algorithms that are traditionally developed based on the Von Neumann computing paradigm by more than four orders of magnitude. Through parallel implementation and exploitation of the reconfigurability of Field Programmable Gate Array (FPGA) systems our design enables scientists to apply BN learning techniques to large problems such as studies in molecular biology where the number of variables in the system overwhelms any state of the art software implementations. We describe how we combine Markov Chain Monte Carlo (MCMC) sampling with Bayesian network learning techniques as well as supervised learning methods in a parallel and scalable design. We also present how our design is mapped and customized to run on the Berkeley Emulation Engine 2 (BEE2) multi-FPGA system. Experimental results are presented on synthetic data sets generated from standard Bayesian networks as well as a real life problem in the context of systems biology

References

  1. Bayesian Networks Repository: http://compbio.cs.huji.ac.il/Repository/.Google ScholarGoogle Scholar
  2. BEE2 Website: http://bee2.eecs.berkeley.edu/.Google ScholarGoogle Scholar
  3. L. Breiman, J. Friedman, R. Olshen, and C. Stone. Classification and Regression Trees. Wadsworth and Brooks, Monterey, CA, 1984.Google ScholarGoogle Scholar
  4. W.L. Buntine. Theory refinement of Bayesian networks. In Uncertainty in Artificial Intelligence, 1991. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. C. Chang, J. Wawrzynek, and R.W. Brodersen. Bee2:a high-end reconfigurable computing system. IEEE Design and Test of Computers, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. D.M. Chickering. Learning bayesian networks is np-complete. In Learning from Data: Artificial Intelligence and Statistics. V. Springer Verlag, 1996.Google ScholarGoogle Scholar
  7. G.F. Cooper and E. Herskovits. A bayesian method for the induction of probabilistic networks from data. Mach. Learn., 9(4):309--347, 1992. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. G.F. Cooper and C. Yoo. Causal discovery from a mixture of experimental and observational data. In UAI, pages 116--125, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. B. Ellis and W. Wong. Sampling bayesian network quickly. In Interface, 2006.Google ScholarGoogle Scholar
  10. J. Friedman. Greedy function approximation: a gradient boosting machine, 1999.Google ScholarGoogle Scholar
  11. J. Friedman. Stochastic gradient boosting, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. N. Friedman and D. Koller. Being Bayesian about Bayesian network structure: A Bayesian approach to structure discovery in Bayesian networks. Machine Learning, 50(1-2):95--125, 2003. Full version of UAI 2000 paper.Google ScholarGoogle ScholarCross RefCross Ref
  13. N. Friedman, I. Nachman, and D. Pe'er. Learning bayesian network structure from massive datasets: The "sparse candidate" algorithm. In Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence, pages 206--215. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. D. Geiger and D. Heckerman. Learning gaussian networks. Technical Report MSR-TR-94-10, Redmond, WA, 1994.Google ScholarGoogle ScholarCross RefCross Ref
  15. C.J. Geyer. Markov chain monte carlo maximum likelihood. In Computing Science and Statistics: Proceedings of 23rd Symposium on the Interface Foundation, pages 156--163. Fairfax Station, 1991.Google ScholarGoogle Scholar
  16. W.K. Hastings. Monte carlo sampling methods using markov chains and their applications. 1970.Google ScholarGoogle Scholar
  17. D. Heckerman, D. Geiger, and D.M. Chickering. Learning bayesian networks: The combination of knowledge and statistical data. Mach. Learn., 20(3):197--243, September 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. A. Moore and W.K. Wong. Optimal reinsertion: A new search operator for accelerated and more accurate bayesian network structure learning. In T. Fawcett and N. Mishra, editors, Proceedings of the 20th International Conference on Machine Learning (ICML'03), pages 552--559, Menlo Park, California, August 2003. AAAI Press.Google ScholarGoogle Scholar
  19. J. Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, September 1988. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. I. Pournara, C.S. Bouganis, and G.A. Constantinides. Fpga-accelerated bayesian learning for reconstruction of gene regulatory networks. pages 323--328, 2005.Google ScholarGoogle Scholar
  21. K. Sachs, O. Perez, D. Pe'er, D.A. Lauffenburger, and G.P. Nolan. Causal protein-signaling networks derived from multiparameter single-cell data. Science, 308(5721):523--529, 2005.Google ScholarGoogle ScholarCross RefCross Ref
  22. H.K.H. So. Borph: An operating system for fpga-based reconfigurable computers. PhD dissertaion, University of California, Berkeley. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. M. Teyssier and D. Koller. Ordering-based search: A simple and effective algorithm for learning bayesian networks. In Proceedings of the Twenty-first Conference on Uncertainty in AI (UAI), pages 584--590, Edinburgh, Scottland, UK, July 2005.Google ScholarGoogle Scholar

Index Terms

  1. Reconfigurable computing for learning Bayesian networks

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        FPGA '08: Proceedings of the 16th international ACM/SIGDA symposium on Field programmable gate arrays
        February 2008
        278 pages
        ISBN:9781595939340
        DOI:10.1145/1344671

        Copyright © 2008 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 24 February 2008

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate125of627submissions,20%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader