Rally provides a framework for performance analysis and benchmarking of individual OpenStack components as well as full production OpenStack cloud deployments
Go to file
Andrey Kurilin d25bdcf76d [ci] Rework launching cli job
There are 2 critical blocking issues related to cli job:

 * bug related to db conflict between 2 tests (samples and lib check).
   We already spent ~ a month reserching the root of an issue and trying
   to fix it. Unfortunately, we did not succeded. Since we are planning
   to rework samples mechanism, the applied solution in this patch
   should not look like a dirty workaround - split launching main cli
   tests and samples check. To do this, we need to modify out pytest
   launcher a bit

 * Rally-CI has two cli jobs for mysql and postgres backends. While
   porting them to use `tox -ecli` command, passing the proper env
   variable was missed and sqlite is used now in both cases.
   Passing the right environment variable is not only required step,
   setting serial mode of pytest is important too, since we do not want
   to have conflicts in tests connecting to the one db.

Also, this patch removes unused option (--timeout) of pytest launcher.

Change-Id: I7fc109ee5785f426211fbff7eb9b4553a3671ba7
2017-08-03 01:17:58 +03:00
certification/openstack [CI] Fix image regexp and version 2017-02-16 14:29:06 +03:00
devstack [osclients] do not always strip auth_url 2017-04-19 20:47:20 +03:00
doc Merge "Fix step number in tutorial section in rally doc" 2017-07-28 21:10:32 +00:00
etc Extend tags feature of tasks 2017-06-26 15:11:20 +03:00
rally [cli] Add command to display db connection str 2017-08-02 19:08:12 +03:00
rally-jobs Ceilometer Samples Context should wait until samples are processed 2017-07-29 07:27:06 +00:00
samples Merge "Add nova.BootServerAttachVolumeAndListAttachemnt" 2017-07-21 14:08:04 +00:00
tests [ci] Rework launching cli job 2017-08-03 01:17:58 +03:00
.coveragerc [CI] Fix coverage job 2016-06-27 15:39:13 +03:00
.dockerignore Fix docker build command 2015-11-10 16:33:29 -08:00
.gitignore Update .gitignore 2017-07-07 04:28:58 +00:00
.gitreview Update .gitreview file to reflect repo rename 2015-04-18 00:37:36 +00:00
babel.cfg Add rally.sample.conf to project 2013-08-14 14:08:09 +04:00
bindep.txt [install] Add iputils to bindep 2016-12-30 18:37:41 +03:00
CONTRIBUTING.rst [docs][6] Re-design docs to cover all user-groups 2017-01-10 11:25:00 -08:00
Dockerfile [Bindep]Use bindep lib to install system packages 2016-12-24 11:13:20 +08:00
install_rally.sh [Install] Fix postgresql installation 2017-07-12 00:01:01 +00:00
LICENSE Initial commit 2013-08-03 09:17:25 -07:00
optional-requirements.txt Fixing docker-check jenkins job 2016-09-20 16:56:00 +03:00
README.rst Fixup Rally Tutorial Page Link in README 2017-06-25 19:37:54 -07:00
requirements.txt remove runtime dep on sphinx 2017-07-17 08:04:58 -04:00
setup.cfg Put db requirements to extra requirements 2017-07-18 21:46:01 +00:00
setup.py Updated from global requirements 2015-09-22 10:45:07 +00:00
test-requirements.txt [ci] Use not the latest xdist package 2017-07-21 11:47:51 +00:00
tox.ini [ci] Rework launching cli job 2017-08-03 01:17:58 +03:00

Rally

Team and repository tags

image

Latest Version

Gitter Chat

Trello Board

Apache License, Version 2.0

What is Rally

Rally is a Benchmark-as-a-Service project for OpenStack.

Rally is intended to provide the community with a benchmarking tool that is capable of performing specific, complicated and reproducible test cases on real deployment scenarios.

If you are here, you are probably familiar with OpenStack and you also know that it's a really huge ecosystem of cooperative services. When something fails, performs slowly or doesn't scale, it's really hard to answer different questions on "what", "why" and "where" has happened. Another reason why you could be here is that you would like to build an OpenStack CI/CD system that will allow you to improve SLA, performance and stability of OpenStack continuously.

The OpenStack QA team mostly works on CI/CD that ensures that new patches don't break some specific single node installation of OpenStack. On the other hand it's clear that such CI/CD is only an indication and does not cover all cases (e.g. if a cloud works well on a single node installation it doesn't mean that it will continue to do so on a 1k servers installation under high load as well). Rally aims to fix this and help us to answer the question "How does OpenStack work at scale?". To make it possible, we are going to automate and unify all steps that are required for benchmarking OpenStack at scale: multi-node OS deployment, verification, benchmarking & profiling.

Rally workflow can be visualized by the following diagram:

Rally Architecture

Who Is Using Rally

Who is Using Rally

Documentation

Rally documentation on ReadTheDocs is a perfect place to start learning about Rally. It provides you with an easy and illustrative guidance through this benchmarking tool. For example, check out the Rally step-by-step tutorial that explains, in a series of lessons, how to explore the power of Rally in benchmarking your OpenStack clouds.

Architecture

In terms of software architecture, Rally is built of 4 main components:

  1. Server Providers - provide servers (virtual servers), with ssh access, in one L3 network.
  2. Deploy Engines - deploy OpenStack cloud on servers that are presented by Server Providers
  3. Verification - component that runs tempest (or another specific set of tests) against a deployed cloud, collects results & presents them in human readable form.
  4. Benchmark engine - allows to write parameterized benchmark scenarios & run them against the cloud.

Use Cases

There are 3 major high level Rally Use Cases:

Rally Use Cases

Typical cases where Rally aims to help are:

  • Automate measuring & profiling focused on how new code changes affect the OS performance;
  • Using Rally profiler to detect scaling & performance issues;
  • Investigate how different deployments affect the OS performance:
    • Find the set of suitable OpenStack deployment architectures;
    • Create deployment specifications for different loads (amount of controllers, swift nodes, etc.);
  • Automate the search for hardware best suited for particular OpenStack cloud;
  • Automate the production cloud specification generation:
    • Determine terminal loads for basic cloud operations: VM start & stop, Block Device create/destroy & various OpenStack API methods;
    • Check performance of basic cloud operations in case of different loads.