skip to main content
10.1145/3040718.3040722acmconferencesArticle/Chapter ViewAbstractPublication PagesfogaConference Proceedingsconference-collections
research-article

On the Statistical Learning Ability of Evolution Strategies

Authors Info & Claims
Published:12 January 2017Publication History

ABSTRACT

We explore the ability of Evolution Strategies (ESs) to statistically learn the local landscape. Specifically, we consider ESs operating only with isotropic Gaussian mutations near the optimum and investigate the covariance matrix when constructed out of selected individuals by truncation. Unlike previous studies, we do not assume a Derandomization adaptation scheme, nor do we use Information Geometric Optimization in our proofs. We prove that the statistically constructed covariance matrix over such selected decision vectors has the same eigenvectors as the Hessian matrix. We further prove that when the population size is increased, the covariance becomes proportional to the inverse of the Hessian. We also devise and corroborate an analytic approximation of this covariance matrix. In the framework we consider, this confirms the classical hypothesis that learning the landscape is an inherent property of standard ESs, and that this capability stems only from the usage of isotropic Gaussian mutations and rank-based selection.

References

  1. Y. Akimoto. Analysis of a natural gradient algorithm on monotonic convex-quadratic-composite functions. In Proceedings of the 14th Annual Conference on Genetic and Evolutionary Computation, GECCO '12, pages 1293--1300, New York, NY, USA, 2012. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. S.-I. Amari and H. Nagaoka. Methods of Information Geometry, volume 191 of Translations of Mathematical Monographs. American Mathematical Society, 2000.Google ScholarGoogle Scholar
  3. T. B\"ack, C. Foussette, and P. Krause. Contemporary Evolution Strategies. Natural Computing Series. Springer-Verlag Berlin Heidelberg, 2013.Google ScholarGoogle Scholar
  4. H.-G. Beyer. Convergence analysis of evolutionary algorithms that are based on the paradigm of information geometry. Evolutionary Computation, 22(4):679--709, Dec. 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. E. Castillo, A. S. Hadi, N. Balakrishnan, and J. M. Sarabia. Extreme Value and Related Models with Applications in Engineering and Science. John Wiley and Sons, 2004.Google ScholarGoogle Scholar
  6. P. Embrechts, C. Klüppelberg, and T. Mikosch. Modelling Extremal Events for Insurance and Finance. Springer-Verlag, 1997. Google ScholarGoogle ScholarCross RefCross Ref
  7. A. H. Feiveson and F. C. Delaney. The distribution and properties of a weighted sum of chi squares. Technical report, National Aeronautics and Space Administration, 1968.Google ScholarGoogle Scholar
  8. R. Fisher and L. Tippett. Limiting forms of the frequency distribution of the largest or smallest member of a sample. Proc. Cambridge Philos. Soc., 24:180--190, 1928. Google ScholarGoogle ScholarCross RefCross Ref
  9. S. S. Gupta. Order Statistics from the Gamma Distribution. Technometrics, 2, May 1960. Google ScholarGoogle ScholarCross RefCross Ref
  10. N. Hansen and A. Ostermeier. Completely Derandomized Self-Adaptation in Evolution Strategies. Evolutionary Computation, 9(2):159--195, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. P. Moschopoulos and W. Canada. The distribution function of a linear combination of chi-squares. Computers & Mathematics with Applications, 10(4):383--386, 1984. Google ScholarGoogle ScholarCross RefCross Ref
  12. Y. Ollivier, L. Arnold, A. Auger, and N. Hansen. Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles. arXiv:1106.3708v3, 2014.Google ScholarGoogle Scholar
  13. G. Rudolph. On Correlated Mutations in Evolution Strategies. In Parallel Problem Solving from Nature - PPSN II, pages 105--114, Amsterdam, 1992. Elsevier.Google ScholarGoogle Scholar
  14. G. Rudolph. Convergence rates of evolutionary algorithms for a class of convex objective functions. Control and Cybernetics, 26(3), 1997.Google ScholarGoogle Scholar
  15. G. Rudolph. Handbook of Natural Computing: Theory, Experiments, and Applications, chapter Evolutionary Strategies, pages 673--698. Springer-Verlag, Berlin-Heidelberg, Germany, 2012. Google ScholarGoogle ScholarCross RefCross Ref
  16. O. M. Shir, J. Roslund, D. Whitley, and H. Rabitz. Efficient retrieval of landscape hessian: Forced optimal covariance adaptive learning. Physical Review E, 89:063306, Jun 2014. Google ScholarGoogle ScholarCross RefCross Ref
  17. D. Wierstra, T. Schaul, T. Glasmachers, Y. Sun, J. Peters, and J. Schmidhuber. Natural evolution strategies. J. Mach. Learn. Res., 15(1):949--980, Jan. 2014.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. A. Zhigljavsky and A.vZilinskas. Stochastic Global Optimization. Springer Optimization and Its Applications. Springer US, 2007.Google ScholarGoogle Scholar

Index Terms

  1. On the Statistical Learning Ability of Evolution Strategies

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            FOGA '17: Proceedings of the 14th ACM/SIGEVO Conference on Foundations of Genetic Algorithms
            January 2017
            170 pages
            ISBN:9781450346511
            DOI:10.1145/3040718

            Copyright © 2017 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 12 January 2017

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            FOGA '17 Paper Acceptance Rate13of23submissions,57%Overall Acceptance Rate72of131submissions,55%

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader