ABSTRACT
Creating system setups for controlled performance evaluation experiments of distributed systems is time-consuming and expensive. Re-creating experiment setups and reproducing experimental results that have been published by other researchers is even more challenging. In this paper, we present an experiment automation approach for evaluating distributed systems in compute cloud environments. We propose three concepts which should guide the design of experiment automation tools: (1) capture experiment plans in software modules, (2) run experiments in a publicly accessible cloud-based Elastic Lab, and (3) collaborate on experiments in an open, distributed collaboration system. We developed two tools which implement these basic concepts and discuss challenges and lessons learned during our implementation. An initial exemplary use case with Apache Cassandra on top of Amazon EC2 provides a first insight into the types of performance and scalability experiments enabled by our tools.
- A. Avetisyan, R. Campbell, I. Gupta, M. Heath, S. Ko, G. Ganger, M. Kozuch, D. O'Hallaron, M. Kunze, T. Kwan, K. Lai, M. Lyons, D. Milojicic, H. Y. Lee, Y. C. Soh, N. K. Ming, J.-Y. Luke, and H. Namgoong. Open Cirrus: A Global Cloud Computing Testbed. Computer, 43(4):35--43, 2010. Google ScholarDigital Library
- R. A. Bailey. Design of Comparative Experiments. Cambridge Series in Statistical and Probabilistic Mathematics. University of London, 2nd edition, 2008.Google Scholar
- C. Binnig, D. Kossmann, T. Kraska, and S. Loesing. How is the Weather tomorrow? Towards a Benchmark for the Cloud. In Proceedings of the Second International Workshop on Testing Database Systems, DBTest '09, pages 9:1--9:6, New York, NY, USA, 2009. ACM. Google ScholarDigital Library
- P. Bodik, A. Fox, M. J. Franklin, M. I. Jordan, and D. A. Patterson. Characterizing, modeling, and generating workload spikes for stateful services. In Proceedings of the 1st ACM symposium on Cloud computing, SoCC '10, pages 241--252, New York, NY, USA, 2010. ACM. Google ScholarDigital Library
- B. Chun, D. Culler, T. Roscoe, A. Bavier, L. Peterson, M. Wawrzoniak, and M. Bowman. PlanetLab: An Overlay Testbed for Broad-Coverage Services. SIGCOMM Comput. Commun. Rev., 33(3):3--12, 2003. Google ScholarDigital Library
- B. F. Cooper, A. Silberstein, E. Tam, R. Ramakrishnan, and R. Sears. Benchmarking Cloud Serving Systems with YCSB. In Proceedings of the 1st ACM symposium on Cloud Computing, SoCC '10, pages 143--154, New York, NY, USA, 2010. ACM. Google ScholarDigital Library
- C. Dumitrescu, I. Raicu, M. Ripeanu, and I. Foster. Diperf: An automated distributed performance testing framework. In Proceedings of the 5th IEEE/ACM International Workshop on Grid Computing, GRID '04, pages 289--296, Washington, DC, USA, 2004. IEEE Computer Society. Google ScholarDigital Library
- J. Elson and J. Howell. Handling ash crowds from your garage. In USENIX 2008 Annual Technical Conference on Annual Technical Conference, ATC'08, pages 171--184, Berkeley, CA, USA, 2008. USENIX Association. Google ScholarDigital Library
- B. A. Frigyik, A. Kapila, and M. R. Gupta. Introduction to the Dirichlet Distribution and Related Processes. Technical Report 206, Uni Washington, 2010.Google Scholar
- S. Gaisbauer, J. Kirschnick, N. Edwards, and J. Rolia. Vats: Virtualized-aware automated test service. In QEST, pages 93--102, 2008. Google ScholarDigital Library
- Z. Ganon and I. Zilbershtein. Cloud-based performance testing of network management systems. In Computer Aided Modeling and Design of Communication Links and Networks, 2009. CAMAD '09. IEEE 14th International Workshop on, pages 1--6, june 2009.Google ScholarCross Ref
- A. Iosup, N. Yigitbasi, and D. Epema. On the performance variability of production cloud services. In Cluster, Cloud and Grid Computing (CCGrid), 2011 11th IEEE/ACM International Symposium on, pages 104--113, may 2011. Google ScholarDigital Library
- M. Klems, D. Bermbach, and R. Weinert. A runtime quality measurement framework for cloud database service systems. In Proceedings of the 8th International Conference on the Quality of Information and Communications Technology, pages 38--46. IEEE, Conference Publishing Services (CPS), September 2012. Google ScholarDigital Library
- A. Lakshman and P. Malik. Cassandra: structured storage system on a P2P network. In Proceedings of the 28th ACM symposium on Principles of distributed computing, PODC '09, pages 5--5, New York, NY, USA, 2009. ACM. Google ScholarDigital Library
- S. Patil, M. Polte, K. Ren, W. Tantisiriroj, L. Xiao, J. López, G. Gibson, A. Fuchs, and B. Rinaldi. YCSB++: Benchmarking and Performance Debugging Advanced Features in Scalable Table Stores. In Proceedings of the 2nd ACM Symposium on Cloud Computing, SoCC '11, New York, NY, USA, 2011. Google ScholarDigital Library
- A. Pavlo, P. Couvares, R. Gietzel, A. Karp, I. D. Alderman, M. Livny, and C. Bacon. The nmi build & test laboratory: Continuous integration framework for distributed computing software. In LISA, pages 263--273, 2006. Google ScholarDigital Library
- Priyanka, I. Chana, and A. Rana. Empirical evaluation of cloud-based testing techniques: a systematic review. SIGSOFT Softw. Eng. Notes, 37(3):1--9, May 2012. Google ScholarDigital Library
- S. Ruby, D. Thomas, and D. H. Hansson. Agile Web Development with Rails. The Pragmatic Programmers. The Facets of Ruby Series. Pragmatic Bookshelf, Raleigh, NC, 3. ed. edition, 2009. Google ScholarDigital Library
- C. C. Ruiz Sanabria, O. Richard, B. Videau, and I. Oleg. Managing Large Scale Experiments in Distributed Testbeds. Research Report RR-8106, INRIA, Oct. 2012.Google Scholar
- C. Sapuntzakis, D. Brumley, R. Chandra, N. Zeldovich, J. Chow, M. S. Lam, and M. Rosenblum. Virtual Appliances for Deploying and Maintaining Software. In Proceedings of the Seventeenth Large Installation Systems Administration Conference, October 2003. Google ScholarDigital Library
- Y. Wang, M. J. Rutherford, A. Carzaniga, and A. L. Wolf. Automating experimentation on distributed testbeds. In Proceedings of the 20th IEEE/ACM international Conference on Automated software engineering, ASE '05, pages 164--173, New York, NY, USA, 2005. ACM. Google ScholarDigital Library
Index Terms
- Position paper: cloud system deployment and performance evaluation tools for distributed databases
Recommendations
Position paper: cloud-based performance testing: issues and challenges
HotTopiCS '13: Proceedings of the 2013 international workshop on Hot topics in cloud servicesConducting performance testing is essential to evaluate system performance. With the emergence of cloud computing, applying cloud resources for large-scale performance testing become very attractive. Many organizations have applied cloud-based ...
Integrating Big Data technologies in a dynamic environment EIAH dedicated to e-learning systems based on cloud infrastructure
ICCWCS'17: Proceedings of the 2nd International Conference on Computing and Wireless Communication SystemsOnline learning has experienced a lot of change, largely as a result of a number of technological Innovations. Newer e-learning (online learning) authoring tools have been developed, and more sophisticated Learning Management Systems (LMSs) have been ...
Pragmatic assessment of research intensive areas in cloud: a systematic review
Cloud computing is a name given to a set of systems for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly ...
Comments