skip to main content
10.1145/2464576.2482724acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
tutorial

An evolutionary approach for robust adaptation of robot behavior to sensor error

Published:06 July 2013Publication History

ABSTRACT

Evolutionary algorithms can adapt the behavior of individual agents to maximize the fitness of populations of agents. We use a genetic algorithm (GA) to optimize behavior in a team of simulated robots that mimic foraging ants. We introduce positional and resource detection error models into this simulation, emulating the sensor error characterized by our physical iAnt robot platform. Increased positional error and detection error both decrease resource collection rates. However, they have different effects on GA behavior. Positional error causes the GA to reduce time spent searching for local resources and to reduce the likelihood of returning to locations where resources were previously found. Detection error causes the GA to select for more thorough local searching and a higher likelihood of communicating the location of found resources to other agents via pheromones. Agents that live in a world with error and use parameters evolved specifically for those worlds perform significantly better than agents in the same error-prone world using parameters evolved for an error-free world. This work demonstrates the utility of employing evolutionary methods to adapt robot behaviors that are robust to sensor errors.

References

  1. B. D. Beverly, H. McLendon, S. Nacu, S. Holmes, and D. Gordon. How site fidelity leads to individual differences in the foraging activity of harvester ants. Behavioral Ecology, 20(3):633--638, 2009.Google ScholarGoogle ScholarCross RefCross Ref
  2. R. A. Brooks. New Approaches to Robotics. Science, 253(5025):1227--1232, 1991.Google ScholarGoogle ScholarCross RefCross Ref
  3. R. A. Brooks. Artifical Life and Real Robots. In Toward a Practice of Autonomous Systems: Proceedings of the First European Conference on Artificial Life, pages 3--10, 1992.Google ScholarGoogle Scholar
  4. S. A. Curtis, J. Mica, J. Nuth, G. Marr, M. Rilee, and M. Bhat. ANTS (Autonomous Nano-Technology Swarm): An artificial intelligence approach to asteroid belt resource exploration. Technical report, 2000.Google ScholarGoogle Scholar
  5. S. A. Curtis, W. Truszkowski, M. L. Rilee, and P. E. Clark. ANTS for Human Exploration and Development of Space. In IEEE Aerospace Conference, volume 1, pages 1--161, 2003.Google ScholarGoogle Scholar
  6. T. P. Flanagan, K. Letendre, W. R. Burnside, G. M. Fricke, and M. E. Moses. Quantifying the Effect of Colony Size and Food Distribution on Harvester Ant Foraging. PLoS ONE, 7(7):e39427, 2012.Google ScholarGoogle ScholarCross RefCross Ref
  7. D. W. Gage. Many-Robot MCM Search Systems. In Autonomous Vehicles in Mine Countermeasures Symposium, number April, pages 4--7, 1995.Google ScholarGoogle Scholar
  8. J. P. Hecker, K. Letendre, K. Stolleis, D. Washington, and M. E. Moses. Formica ex Machina: Ant Swarm Foraging From Physical to Virtual and Back Again. Swarm Intelligence, 7461:252--259, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. N. Jakobi. Half-baked, Ad-hoc and Noisy: Minimal Simulations for Evolutionary Robotics. In Fourth European Conference on Artificial Life, pages 348--357. MIT Press, 1997.Google ScholarGoogle Scholar
  10. N. Jakobi, P. Husbands, and I. Harvey. Noise and The Reality Gap: The Use of Simulation in Evolutionary Robotics. Advances in Artificial Intelligence, 704--720, 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. H. Kitano, S. Tadokoro, I. Noda, H. Matsubara, T. Takahashi, A. Shinjou, and S. Shimada. Robocup rescue: Search and rescue in large-scale disasters as a domain for autonomous agents research. In Systems, Man, and Cybernetics, 1999. IEEE SMC'99 Conference Proceedings. 1999 IEEE International Conference on, volume 6, pages 739--743. IEEE, 1999.Google ScholarGoogle Scholar
  12. C. S. Kong, N. A. Peng, and I. Rekleitis. Distributed coverage with multi-robot system. In IEEE International Conference on Robotics and Automation, number May, pages 2423--2429. Ieee, 2006.Google ScholarGoogle Scholar
  13. K. Letendre and M. E. Moses. Synergy in Ant Foraging Strategies: Memory and Communication Alone and In Combination. GECCO (in press), 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. E. B. Mallon and N. R. Franks. Ants estimate area using Buffon's needle. Proceedings of the Royal Society of London. Series B: Biological Sciences, 267(1445):765--770, 2000.Google ScholarGoogle ScholarCross RefCross Ref
  15. M. J. Matarić. Learning to Behave Socially. Third International Conference on Simulation of Adaptive Behavior, (617):453---462, 1994. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. M. J. Matarić. Behaviour-based control: examples from navigation, learning, group behavior. Journal of Experimental & Theoretical Artificial Intelligence, 9(2--3):323--336, 1997.Google ScholarGoogle ScholarCross RefCross Ref
  17. M. J. Matarić. Reinforcement learning in the multi-robot domain. Autonomous Robots, 4:73--83, 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. J. McLurkin. Using Cooperative Robots for Explosive Ordnance Disposal. Massachusetts Institute of Technology Artificial Intelligence Laboratory, pages 1--10, 1996.Google ScholarGoogle Scholar
  19. O. Miglino, H. H. Lund, and S. Nolfi. Evolving Mobile Robots in Simulated and Real Environments. Artificial Life, 2(4):417--434, Jan. 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. M. E. Moses. Metabolic scaling from individuals to societies. PhD thesis, University of New Mexico, 2005.Google ScholarGoogle Scholar
  21. A. Nelson, E. Grant, and T. Henderson. Evolution of neural controllers for competitive game playing with teams of mobile robots. Robotics and Autonomous Systems, 46(135--150), 2004.Google ScholarGoogle Scholar
  22. S. Nolfi, D. Floreano, O. Miglino, and F. Mondada. How to evolve autonomous robots: Different approaches in evolutionary robotics. Artificial Life, 4:190--197, 1993.Google ScholarGoogle Scholar
  23. L. Panait and S. Luke. Learning ant foraging behaviors. Proceedings of the Ninth International Conference on the Simulation and Synthesis of Living Systems (ALIFE9), 2004.Google ScholarGoogle Scholar
  24. L. E. Parker. ALLIANCE: An architecture for fault tolerant multirobot cooperation. Robotics and Automation, IEEE Transactions on, 14(2):220--240, 1998.Google ScholarGoogle Scholar
  25. T. Paz Flanagan, K. Letendre, W. Burnside, G. M. Fricke, and M. Moses. How Ants Turn Information into Food. IEEE Conference on Artificial Life, pages 178--185, 2011.Google ScholarGoogle ScholarCross RefCross Ref
  26. C. Pinciroli, V. Trianni, R. O. Grady, G. Pini, A. Brutschy, M. Brambilla, N. Mathews, E. Ferrante, G. D. Caro, F. Ducatelle, T. Stirling, A. Guti, L. M. Gambardella, and M. Dorigo. ARGoS: a Pluggable, Multi-Physics Engine Simulator for Heterogeneous Swarm Robotics. In IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), number December 2010, pages 5027--5034, 2011.Google ScholarGoogle Scholar
  27. C. Pinciroli, V. Trianni, R. O'Grady, G. Pini, A. Brutschy, M. Brambilla, N. Mathews, E. Ferrante, G. Di Caro, F. Ducatelle, M. Birattari, L. M. Gambardella, and M. Dorigo. ARGoS: a modular, parallel, multi-engine simulator for multi-robot systems. Swarm intelligence, 6:271--295, 2012.Google ScholarGoogle ScholarCross RefCross Ref
  28. B. Prabhakar, K. N. Dektar, and D. M. Gordon. The Regulation of Ant Colony Foraging Activity without Spatial Information. PLoS Computational Biology, 8(8):e1002670, 2012.Google ScholarGoogle ScholarCross RefCross Ref
  29. M. Singh and D. Parhi. Path optimisation of a mobile robot using an artificial neural network controller. International Journal of Systems Science, 42(1):107--120, Jan. 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. E. Tunstel, J. M. Dolan, T. Fong, and D. Schreckenghost. Mobile Robotic Surveying Performance for Planetary Surface Site Characterization. In Proceedings of the 8th Workshop on Performance Metrics for Intelligent Systems, pages 200--205, New York, NY, 2008. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. R. Vaughan. Massively multi-robot simulation in stage. Swarm Intelligence, 2008.Google ScholarGoogle Scholar
  32. R. A. Watson, S. G. Ficici, and J. B. Pollack. Embodied Evolution: Distributing an evolutionary algorithm in a population of robots. Robotics and Autonomous Systems, 39(1):1--18, Apr. 2002.Google ScholarGoogle ScholarCross RefCross Ref
  33. M. Wittlinger, R. Wehner, and H. Wolf. The Ant Odometer: Stepping on Stilts and Stumps. Science, 312(5782):1965--1967, 2006.Google ScholarGoogle Scholar

Index Terms

  1. An evolutionary approach for robust adaptation of robot behavior to sensor error

          Recommendations

          Reviews

          Jindong Liu

          When undertaking the basic task of mimicking ants as they collect food from remote sites to bring back to the nest, robot systems exhibit two types of errors: erroneous estimation of robot position and erroneous detection of a food resource. This paper proposes an evolutionary approach to managing these errors in a multiple robot system. The authors have developed a central-place foraging algorithm (CPFA) based on ant foraging behavior. The behavior of each robot is defined by a number of parameters used in the CPFA, such as the probability of stopping the search effort, and a genetic algorithm (GA) is used to optimize behavior. Two types of simulated worlds were used for the experiments: an error-free world (perfect world) and a world with error (imperfect world). Learned parameters from the perfect world were swapped with those in the imperfect world. The parameter swapping was also applied between two imperfect worlds with different grades of error. The results show that the GA can help the robot deal with sensor error, especially when the parameters are learned from an imperfect world and applied to a similar type of imperfect world. Overall, it is interesting to see how biological evidence can help solve problems in a robotic system. The CPFA looks to be a promising approach to solving this foraging task. The main drawback of the paper is that all of the experiments were conducted in simulated worlds. Although two types of errors were taken into account, it is still far from reality. For example, the authors do not consider the detection error generated when the robot recognizes a laid pheromone (the landmark trace to a food location). Training in an imperfect world and testing in another imperfect world is not really novel. Another drawback is that there are no comparisons to other algorithms, although this may be due to constraints on the length of the report. It would be more convincing if the authors could add real robot experiments and comparisons. Online Computing Reviews Service

          Access critical reviews of Computing literature here

          Become a reviewer for Computing Reviews.

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            GECCO '13 Companion: Proceedings of the 15th annual conference companion on Genetic and evolutionary computation
            July 2013
            1798 pages
            ISBN:9781450319645
            DOI:10.1145/2464576
            • Editor:
            • Christian Blum,
            • General Chair:
            • Enrique Alba

            Copyright © 2013 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 6 July 2013

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • tutorial

            Acceptance Rates

            Overall Acceptance Rate1,669of4,410submissions,38%

            Upcoming Conference

            GECCO '24
            Genetic and Evolutionary Computation Conference
            July 14 - 18, 2024
            Melbourne , VIC , Australia

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader