Merge "Amended testing documentation"
This commit is contained in:
commit
a7821f45b8
@ -92,8 +92,8 @@ the following topic guides.
|
|||||||
topics/workflows
|
topics/workflows
|
||||||
topics/tables
|
topics/tables
|
||||||
topics/policy
|
topics/policy
|
||||||
topics/testing
|
|
||||||
topics/angularjs
|
topics/angularjs
|
||||||
|
topics/testing
|
||||||
topics/javascript_testing
|
topics/javascript_testing
|
||||||
topics/styling
|
topics/styling
|
||||||
topics/translation
|
topics/translation
|
||||||
|
@ -39,6 +39,16 @@ Tests can be run in two ways:
|
|||||||
The code linting job can be run with ``tox -e npm -- lint``, or
|
The code linting job can be run with ``tox -e npm -- lint``, or
|
||||||
``tox -e npm -- lintq`` to show errors, but not warnings.
|
``tox -e npm -- lintq`` to show errors, but not warnings.
|
||||||
|
|
||||||
|
To decipher where tests are failing it may be useful to use Jasmine in the
|
||||||
|
browser to run individual tests to see where the tests are specifically breaking.
|
||||||
|
To do this, navigate to your local horizon in the browser and add
|
||||||
|
'/jasmine' to the end of the url. e.g: 'http://localhost:8000/jasmine'. Once you
|
||||||
|
have the jasmine report you may click on the title of an individual test to
|
||||||
|
re-run just that test. From here, you can also use chrome dev tools or similar
|
||||||
|
to set breakpoints in the code by accessing the 'Sources' tab and clicking on
|
||||||
|
lines of code where you wish to break the code. This will then show you the exact
|
||||||
|
places where the code breaks.
|
||||||
|
|
||||||
Coverage Reports
|
Coverage Reports
|
||||||
----------------
|
----------------
|
||||||
|
|
||||||
@ -47,9 +57,8 @@ developing, be sure to check the coverage reports on the master branch and
|
|||||||
compare your development branch; this will help identify missing tests.
|
compare your development branch; this will help identify missing tests.
|
||||||
|
|
||||||
To generate coverage reports, run ``tox -e npm``. The coverage reports can be
|
To generate coverage reports, run ``tox -e npm``. The coverage reports can be
|
||||||
found at ``horizon/coverage-karma/`` (framework tests) and
|
found at ``cover/horizon/`` (framework tests) and ``cover/openstack_dashboard/``
|
||||||
``openstack_dashboard/coverage-karma/`` (dashboard tests). Load
|
(dashboard tests). Load ``<browser>/index.html`` in a browser to view the reports.
|
||||||
``<browser>/index.html`` in a browser to view the reports.
|
|
||||||
|
|
||||||
Writing Tests
|
Writing Tests
|
||||||
=============
|
=============
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
===================
|
================
|
||||||
Testing Topic Guide
|
Testing Overview
|
||||||
===================
|
================
|
||||||
|
|
||||||
Having good tests in place is absolutely critical for ensuring a stable,
|
Having good tests in place is absolutely critical for ensuring a stable,
|
||||||
maintainable codebase. Hopefully that doesn't need any more explanation.
|
maintainable codebase. Hopefully that doesn't need any more explanation.
|
||||||
@ -13,6 +13,11 @@ If you already know everything about testing but are fed up with trying to
|
|||||||
debug why a specific test failed, you can skip the intro and jump
|
debug why a specific test failed, you can skip the intro and jump
|
||||||
straight to :ref:`debugging_unit_tests`.
|
straight to :ref:`debugging_unit_tests`.
|
||||||
|
|
||||||
|
.. toctree::
|
||||||
|
:maxdepth: 1
|
||||||
|
|
||||||
|
Angular specific testing <javascript_testing>
|
||||||
|
|
||||||
An overview of testing
|
An overview of testing
|
||||||
======================
|
======================
|
||||||
|
|
||||||
@ -222,6 +227,23 @@ Tips and tricks
|
|||||||
+ },
|
+ },
|
||||||
'django.db.backends': {
|
'django.db.backends': {
|
||||||
|
|
||||||
|
Coverage reports
|
||||||
|
----------------
|
||||||
|
|
||||||
|
It is possible for tests to fail on your patch due to the npm-run-test not
|
||||||
|
passing the minimum threshold. This is not necessarily related directly to the
|
||||||
|
functions in the patch that have failed, but more that there are not enough tests
|
||||||
|
across horizon that are related to your patch.
|
||||||
|
|
||||||
|
The coverage reports may be found in the 'cover' directory. There's a subdirectory
|
||||||
|
for horizon and openstack_dashboard, and then under a directory for the browser
|
||||||
|
used to run the tests you should find an ``index.html``. This can then be viewed
|
||||||
|
to see the coverage details.
|
||||||
|
|
||||||
|
In this scenario you may need to submit a secondary patch to address test coverage
|
||||||
|
for another function within horizon to ensure tests rise above the coverage
|
||||||
|
threshold and your original patch can pass the necessary tests.
|
||||||
|
|
||||||
Common pitfalls
|
Common pitfalls
|
||||||
---------------
|
---------------
|
||||||
|
|
||||||
@ -292,421 +314,3 @@ result of an error in the conditions of the test. Using the
|
|||||||
:meth:`~horizon.test.helpers.TestCase.assertMessageCount` will make it readily
|
:meth:`~horizon.test.helpers.TestCase.assertMessageCount` will make it readily
|
||||||
apparent what the problem is in the majority of cases. If not, then use ``pdb``
|
apparent what the problem is in the majority of cases. If not, then use ``pdb``
|
||||||
and start interrupting the code flow to see where things are getting off track.
|
and start interrupting the code flow to see where things are getting off track.
|
||||||
|
|
||||||
Integration tests in Horizon
|
|
||||||
============================
|
|
||||||
|
|
||||||
The integration tests currently live in the Horizon repository, see `here`_,
|
|
||||||
which also contains instructions on how to run the tests. To make integration
|
|
||||||
tests more understandable and maintainable, the Page Object pattern is used
|
|
||||||
throughout them.
|
|
||||||
|
|
||||||
.. warning:: To enable integration tests support before running them, please
|
|
||||||
copy openstack_dashboard/local/local_settings.d/_20_integration_tests_scaffolds.py.example
|
|
||||||
to openstack_dashboard/local/local_settings.d/_20_integration_tests_scaffolds.py
|
|
||||||
and then run ./manage.py collectstatic --clear && ./manage.py compress.
|
|
||||||
|
|
||||||
Horizon repository also provides two shell `scripts`_, which are executed in
|
|
||||||
pre_test_hook and post_test_hook respectively. Pre hook is generally used for
|
|
||||||
modifying test environment, while post hook is used for running actual
|
|
||||||
integration tests with tox and collecting test artifacts. Thanks to the
|
|
||||||
incorporating all modifications to tests into Horizon repository, one can alter
|
|
||||||
both tests and test environment and see the immediate results in Jenkins job
|
|
||||||
output.
|
|
||||||
|
|
||||||
.. _here: https://github.com/openstack/horizon/tree/master/openstack_dashboard/test/integration_tests
|
|
||||||
.. _scripts: https://github.com/openstack/horizon/tree/master/tools/gate/integration
|
|
||||||
|
|
||||||
Page Object pattern
|
|
||||||
-------------------
|
|
||||||
|
|
||||||
Within any web application's user interface (UI) there are areas that the tests
|
|
||||||
interact with. A Page Object simply models these as objects within the test
|
|
||||||
code. This reduces the amount of duplicated code; if the UI changes, the fix
|
|
||||||
needs only be applied in one place.
|
|
||||||
|
|
||||||
Page Objects can be thought of as facing in two directions simultaneously.
|
|
||||||
Facing towards the developer of a test, they represent the services offered by
|
|
||||||
a particular page. Facing away from the developer, they should be the only
|
|
||||||
thing that has a deep knowledge of the structure of the HTML of a page (or
|
|
||||||
part of a page). It is simplest to think of the methods on a Page Object as
|
|
||||||
offering the "services" that a page offers rather than exposing the details
|
|
||||||
and mechanics of the page. As an example, think of the inbox of any web-based
|
|
||||||
email system. Amongst the services that it offers are typically the ability to
|
|
||||||
compose a new email, to choose to read a single email, and to list the subject
|
|
||||||
lines of the emails in the inbox. How these are implemented should not matter
|
|
||||||
to the test.
|
|
||||||
|
|
||||||
Writing reusable and maintainable Page Objects
|
|
||||||
----------------------------------------------
|
|
||||||
|
|
||||||
Because the main idea is to encourage the developer of a test to try and think
|
|
||||||
about the services that they are interacting with rather than the
|
|
||||||
implementation, Page Objects should seldom expose the underlying WebDriver
|
|
||||||
instance. To facilitate this, methods on the Page Object should return other
|
|
||||||
Page Objects. This means that we can effectively model the user's journey
|
|
||||||
through the application.
|
|
||||||
|
|
||||||
Another important thing to mention is that a Page Object need not represent an
|
|
||||||
entire page. It may represent a section that appears many times within a site
|
|
||||||
or page, such as site navigation. The essential principle is that there is
|
|
||||||
only one place in your test suite with knowledge of the structure of the HTML
|
|
||||||
of a particular (part of a) page. With this in mind, a test developer builds
|
|
||||||
up regions that become reusable components (`example of a base form`_). These
|
|
||||||
properties can then be redefined or overridden (e.g. selectors) in the actual
|
|
||||||
pages (subclasses) (`example of a tabbed form`_).
|
|
||||||
|
|
||||||
The page objects are read-only and define the read-only and clickable elements
|
|
||||||
of a page, which work to shield the tests. For instance, from the test
|
|
||||||
perspective, if "Logout" used to be a link but suddenly becomes an option in a
|
|
||||||
drop-down menu, there are no changes (in the test itself) because it still simply
|
|
||||||
calls the "click_on_logout" action method.
|
|
||||||
|
|
||||||
This approach has two main aspects:
|
|
||||||
|
|
||||||
* The classes with the actual tests should be as readable as possible
|
|
||||||
* The other parts of the testing framework should be as much about data as
|
|
||||||
possible, so that if the CSS etc. changes you only need to change that one
|
|
||||||
property. If the flow changes, only the action method should need to change.
|
|
||||||
|
|
||||||
There is little that is Selenium-specific in the Pages, except for the
|
|
||||||
properties. There is little coupling between the tests and the pages. Writing
|
|
||||||
the tests becomes like writing out a list of steps (by using the previously
|
|
||||||
mentioned action methods). One of the key points, particularly important for
|
|
||||||
this kind of UI driven testing is to isolate the tests from what is behind
|
|
||||||
them.
|
|
||||||
|
|
||||||
.. _example of a base form: https://github.com/openstack/horizon/blob/8.0.0/openstack_dashboard/test/integration_tests/regions/forms.py#L250
|
|
||||||
.. _example of a tabbed form: https://github.com/openstack/horizon/blob/8.0.0/openstack_dashboard/test/integration_tests/regions/forms.py#L322
|
|
||||||
|
|
||||||
List of references
|
|
||||||
------------------
|
|
||||||
|
|
||||||
* https://wiki.openstack.org/wiki/Horizon/Testing/UI#Page_Object_Pattern_.28Selected_Approach.29
|
|
||||||
* https://wiki.mozilla.org/QA/Execution/Web_Testing/Docs/Automation/StyleGuide#Page_Objects
|
|
||||||
* https://code.google.com/p/selenium/wiki/PageObjects
|
|
||||||
|
|
||||||
Debugging integration tests
|
|
||||||
===========================
|
|
||||||
|
|
||||||
Even perfectly designed Page Objects are not a guarantee that your integration
|
|
||||||
test will not ever fail. This can happen due to different causes:
|
|
||||||
|
|
||||||
The first and most anticipated kind of failure is the inability to perform a
|
|
||||||
testing scenario by a living person simply because some OpenStack service or
|
|
||||||
Horizon itself prevents them from doing so. This is exactly the kind that
|
|
||||||
integration tests are designed to catch. Let us call them "good" failures.
|
|
||||||
|
|
||||||
All other kinds of failures are unwanted and could be roughly split into the
|
|
||||||
two following categories:
|
|
||||||
|
|
||||||
#. The failures that occur due to changes in application's DOM. some CSS/ Xpath selectors no longer matching
|
|
||||||
Horizon app's DOM. The usual signature for that kind of failures is having
|
|
||||||
a DOM changing patch for which the test job fails with a message like
|
|
||||||
this `selenium.common.exceptions.NoSuchElementException: Message: Unable to
|
|
||||||
locate element: {"method":"css selector","selector":"div.modal-dialog"}`.
|
|
||||||
If you find yourself in such a situation, you should fix the Page Object
|
|
||||||
selectors according to the DOM changes you made.
|
|
||||||
|
|
||||||
#. Unfortunately it is still quite possible to get the above error for a patch
|
|
||||||
which didn't implement any DOM changes. Among the reasons of such behavior
|
|
||||||
observed in past were:
|
|
||||||
|
|
||||||
* Integration tests relying on relative ordering of form fields and table
|
|
||||||
actions that broke with the addition of a new field. This issue should
|
|
||||||
be fixed by now, but may reappear in future for different entities.
|
|
||||||
|
|
||||||
* Integration tests relying on popups disappearing by the time a specific
|
|
||||||
action needs to be taken (or not existing at all). This expectation
|
|
||||||
turned out to be very fragile, since the speed of tests execution by
|
|
||||||
Jenkins workers may change independently of integration test code (hence,
|
|
||||||
popups disappear too late to free the way for the next action). The
|
|
||||||
unexpected (both too long and too short) timeouts aren't limited to just
|
|
||||||
popups, but apply to every situation when the element state transition
|
|
||||||
is not instant (like opening an external link, going to another page in
|
|
||||||
Horizon, waiting for button to become active, waiting for a table row to
|
|
||||||
change its state). Luckily, most transitions of "element becomes visible/
|
|
||||||
emerge to existence from non-existence" kind are already bulletproofed
|
|
||||||
using `implicit_wait` parameter in `integration_tests/horizon.conf` file.
|
|
||||||
Selenium just waits for specified amount of seconds for an element to
|
|
||||||
become visible (if it's not already visible) giving up when it exceeds
|
|
||||||
(with the above error). Also it's worth mentioning `explicit_wait` parameter
|
|
||||||
which is considered when the selenium `wait_until` method is involved (and
|
|
||||||
it is used, e.g. in waiting for spinner and messages popups to disappear).
|
|
||||||
|
|
||||||
An inconvenient thing about reading test results in the `console.html` file
|
|
||||||
attached to every `gate-horizon-dsvm-integration` finished job is that the test
|
|
||||||
failure may appear either as failure (assertion failed), or as error (expected
|
|
||||||
element didn't show up). In both cases an inquirer should suspect a legitimate
|
|
||||||
failure first (i.e., treat errors as failures). Unfortunately, no clear method
|
|
||||||
exists for the separation of "good" from "bad" failures. Each case is
|
|
||||||
unique and full of mysteries.
|
|
||||||
|
|
||||||
The Horizon testing mechanism tries to alleviate this ambiguity by providing
|
|
||||||
several facilities to aid in failure investigation:
|
|
||||||
|
|
||||||
* First there comes a screenshot made for every failed test (in a separate
|
|
||||||
folder, on a same level as `console.html`) - almost instant snapshot of a
|
|
||||||
screen on the moment of failure (*almost* sometimes matters, especially in
|
|
||||||
a case of popups that hang on a screen for a limited time);
|
|
||||||
* Then the patient inquirer may skim through the vast innards of
|
|
||||||
`console.html`, looking at browser log first (all javascript and css errors
|
|
||||||
should come there),
|
|
||||||
* Then looking at a full textual snapshot of a page for which test failed
|
|
||||||
(sometimes it gives a more precise picture than a screenshot),
|
|
||||||
* And finally looking at test error stacktrace (most useful) and a lengthy
|
|
||||||
output of requests/ responses with a selenium server. The last log sometimes
|
|
||||||
might tell us how long a specific web element was polled before failing (in
|
|
||||||
case of `implicit_wait` there should be a series of requests to the same
|
|
||||||
element).
|
|
||||||
|
|
||||||
The best way to solve the cause of test failure is running and debugging the
|
|
||||||
troublesome test locally. You could use `pdb` or Python IDE of your choice to
|
|
||||||
stop test execution in arbitrary points and examining various Page Objects
|
|
||||||
attributes to understand what they missed. Looking at the real page structure
|
|
||||||
in browser developer tools also could explain why the test fails. Sometimes it
|
|
||||||
may be worth to place breakpoints in JavaScript code (provided that static is
|
|
||||||
served uncompressed) to examine the objects of interest. If it takes long, you
|
|
||||||
may also want to increase the webdriver's timeout so it will not close browser
|
|
||||||
windows forcefully. Finally, sometimes it may make sense to examine the
|
|
||||||
contents of `logs` directory, especially apache logs - but that is mostly the
|
|
||||||
case for the "good" failures.
|
|
||||||
|
|
||||||
Writing your first integration test
|
|
||||||
===================================
|
|
||||||
|
|
||||||
So, you are going to write your first integration test and looking for some
|
|
||||||
guidelines on how to do it. The first and the most comprehensive source of
|
|
||||||
knowledge is the existing codebase of integration tests. Look how other tests
|
|
||||||
are written, which Page Objects they use and learn by copying. Accurate imitation
|
|
||||||
will eventually lead to a solid understanding. Yet there are few things that may
|
|
||||||
save you some time when you know them in advance.
|
|
||||||
|
|
||||||
File and directory layout and go_to_*page() methods
|
|
||||||
---------------------------------------------------
|
|
||||||
Below is the filesystem structure that test helpers rely on.::
|
|
||||||
|
|
||||||
horizon/
|
|
||||||
└─ openstack_dashboard/
|
|
||||||
└─ test/
|
|
||||||
└─ integration_tests/
|
|
||||||
├─ pages/
|
|
||||||
│ ├─ admin/
|
|
||||||
│ │ ├─ __init__.py
|
|
||||||
│ │ └─ system/
|
|
||||||
│ │ ├─ __init__.py
|
|
||||||
│ │ └─ flavorspage.py
|
|
||||||
│ ├─ project/
|
|
||||||
│ │ └─ compute/
|
|
||||||
│ │ ├─ __init__.py
|
|
||||||
│ │ ├─ access_and_security/
|
|
||||||
│ │ │ ├─ __init__.py
|
|
||||||
│ │ │ └─ keypairspage.py
|
|
||||||
│ │ └─ imagespage.py
|
|
||||||
│ └─ navigation.py
|
|
||||||
├─ regions/
|
|
||||||
├─ tests/
|
|
||||||
├─ config.py
|
|
||||||
└─ horizon.conf
|
|
||||||
|
|
||||||
New tests are put into integration_tests/tests, where they are grouped
|
|
||||||
by the kind of entities being tested (test_instances.py, test_networks.py, etc).
|
|
||||||
All Page Objects to be used by tests are inside pages/directory, the nested
|
|
||||||
directory structure you see within it obeys the value of `Navigation.CORE_PAGE_STRUCTURE`
|
|
||||||
you can find at pages/navigation.py module. The contents of the `CORE_PAGE_STRUCTURE`
|
|
||||||
variable should in turn mirror the structure of standard dashboard sidebar menu.
|
|
||||||
If this condition is not met, the go_to_<pagename>page() methods which are generated
|
|
||||||
automatically at runtime will have problems matching the real sidebar items. How are
|
|
||||||
these go_to_*page() methods are generated? From the sidebar's point of view, dashboard
|
|
||||||
content could be at most four levels deep: Dashboard, Panel Group, Panel and Tab.
|
|
||||||
Given the mixture of these entities in existing dashboards, it was decided that:
|
|
||||||
|
|
||||||
* When panels need to be addressed with go_to_<pagename>page() methods, two components in
|
|
||||||
the method's name are enough for distinguishing the right path to go along, namely a Panel
|
|
||||||
name and a Panel Group name (or a Dashboard name, if no Panel Group exists above Panel).
|
|
||||||
For example,
|
|
||||||
|
|
||||||
* `go_to_system_flavorspage()` method to go to Admin->System->Flavors and
|
|
||||||
|
|
||||||
* `go_to_identity_projectspage()` method to go to Identity->Projects panel.
|
|
||||||
|
|
||||||
* When we need to go one level deeper, i.e. go to the specific TableTab on any panel that
|
|
||||||
has several tabs, three components are enough - Panel Group, Panel and Tab names. For
|
|
||||||
example, `go_to_compute_accessandsecurity_floatingipspage()` for navigating to
|
|
||||||
Project->Compute->Access & Security->Floating IPs tab. Note that one cannot navigate
|
|
||||||
to a Panel level if that Panel has several tabs (i.e., only terminal levels could be
|
|
||||||
navigated to).
|
|
||||||
|
|
||||||
As you might have noticed, method name components are chosen from normalized items of
|
|
||||||
the `CORE_PAGE_STRUCTURE` dictionary, where normalization means replacing spaces with `_`
|
|
||||||
symbol and `&` symbol with `and`, then downcasing all symbols.
|
|
||||||
|
|
||||||
Once the `go_to_*page()` method's name is parsed and the proper menu item is matched in
|
|
||||||
a dashboard, it should return the proper Page Object. For that to happen a properly
|
|
||||||
named class should reside in a properly named module located in the right place of the
|
|
||||||
filesystem. More specifically and top down:
|
|
||||||
|
|
||||||
#. Page Object class is located in:
|
|
||||||
|
|
||||||
* <dashboard>/<panel_group>/<panel>page.py file for non-tabbed pages
|
|
||||||
|
|
||||||
* <dashboard>/<panel_group>/<panel>/<tab>page.py file for tabbed pages
|
|
||||||
Values <dashboard>, <panel_group>, <panel> and <tab> are the normalized versions of
|
|
||||||
the items from the `CORE_PAGE_STRUCTURE` dictionary.
|
|
||||||
|
|
||||||
#. Within the above module a descendant of `basepage.BaseNavigationPage` should be
|
|
||||||
defined, its name should have the form <Panel>Page or <Tab>Page, where <Panel> and <Tab>
|
|
||||||
are capitalized versions of normalized <panel> and <tab> items respectively.
|
|
||||||
|
|
||||||
Reusable regions
|
|
||||||
----------------
|
|
||||||
|
|
||||||
* `TableRegion` binds to the HTML Horizon table using the `TableRegion`'s `name`
|
|
||||||
attribute. To bind to the proper table this attribute has to be the same as
|
|
||||||
the `name` attribute of a `Meta` subclass of a corresponding `tables.DataTable`
|
|
||||||
descendant in the Python code. `TableRegion` provides all the needed facilities for
|
|
||||||
solving the following table-related tasks.
|
|
||||||
|
|
||||||
* Getting a specific row from a table matched by the column name and a target
|
|
||||||
text within that column (use `get_row()` method) or taking all the existing
|
|
||||||
rows on a current table page with `rows` property.
|
|
||||||
* Once you have a reference to a specific row, it can either be marked with
|
|
||||||
`mark()` for further batch actions or split to cells (using `cells` property
|
|
||||||
which is dictionary representing column name as a key to cell wrapper as a
|
|
||||||
value).
|
|
||||||
|
|
||||||
* For interacting with actions `TableRegion` provides 2 decorators, namely
|
|
||||||
`@bind_table_action()` and `@bind_row_action()` which bind to the actual HTML
|
|
||||||
button widget and decorate the specific table methods. These methods in turn
|
|
||||||
should click a bound button (comes as these methods' second argument after `self`)
|
|
||||||
and usually return a new region which is most often bound to a modal form
|
|
||||||
being shown after clicking that button in real Horizon.
|
|
||||||
|
|
||||||
* Another important part of `TableRegion` are the facilities for checking the
|
|
||||||
properties of a paged table - `assert_definition()`, `is_next_link_available()`
|
|
||||||
and `is_prev_link_available()` helpers and `turn_next_page()` / `turn_prev_page()`
|
|
||||||
which obviously cause the next / prev table page to be shown.
|
|
||||||
|
|
||||||
* when interacting with modal and non-modal forms three flavors of form wrappers
|
|
||||||
can be used.
|
|
||||||
|
|
||||||
* `BaseFormRegion` is used for simplest forms which are usually 'Submit' /
|
|
||||||
'Cancel' dialogs with no fields to be filled.
|
|
||||||
|
|
||||||
* `FormRegion` is the most used wrapper which provides interaction with the
|
|
||||||
fields within that form. Every field is backed by its own wrapper class, while
|
|
||||||
the `FormRegion` acts as a container which initializes all the field wrappers in
|
|
||||||
its `__init__()` method. Field mappings passed to `__init__()` could be
|
|
||||||
|
|
||||||
* either a tuple of string labels, in that case the same label is used for
|
|
||||||
referencing the field in test code and for binding to the HTML input (should be
|
|
||||||
the same as `name` attribute of that widget, could be seen in Django code defining
|
|
||||||
that form in Horizon)
|
|
||||||
|
|
||||||
* or a dictionary, where the key will be used for referencing the test field
|
|
||||||
and the value will be used for binding to the HTML input. Also it is feasible
|
|
||||||
to provide values other than strings in that dictionary - in this case they are
|
|
||||||
meant to be a Python class. This Python class will be initialized as any
|
|
||||||
BaseRegion is usually initialized and then the value's key will be used for
|
|
||||||
referencing this object. This is useful when dealing with non-standard widgets
|
|
||||||
in forms (like Membership widget in Create/Edit Project form or Networks widget
|
|
||||||
in Launch Instance form).
|
|
||||||
|
|
||||||
* `TabbedFormRegion` is a slight variation of `FormRegion`, it has several tabs
|
|
||||||
and thus can accept a tuple of tuples / dictionaries of field mappings, where
|
|
||||||
every tuple corresponds to a tab of a real form, binding order is that first
|
|
||||||
tuple binds to leftmost tab, which has index 0. Passing `default_tab` other than
|
|
||||||
0 to `TabbedFormRegion.__init__` we can make the test form to be created with
|
|
||||||
the tab other than the leftmost being shown immediately. Finally the method `switch_to`
|
|
||||||
allows us to switch to any existing form's tab.
|
|
||||||
|
|
||||||
* `MessageRegion` is a small region, but is very important for asserting that
|
|
||||||
everything goes well in Horizon under test. Technically, the `find_message_and_dismiss`
|
|
||||||
method belongs to `BasePage` class, but whenever it is called, `regions.messages`
|
|
||||||
module is imported as well to pass a `messages.SUCCESS` / `messages.ERROR`
|
|
||||||
argument into. The method returns `True` / `False` depending on if the specified
|
|
||||||
message was found and dismissed (which could be then asserted for).
|
|
||||||
|
|
||||||
Customizing tests to a specific gate environment
|
|
||||||
------------------------------------------------
|
|
||||||
|
|
||||||
* Upstream gate environment is not the only possible environment where Horizon
|
|
||||||
integration tests can be run. Various downstream distributions may also
|
|
||||||
want to run them. To ease the adoption of upstream tests to possibly
|
|
||||||
different conditions of a downstream gate environment, integration tests use
|
|
||||||
a configuration machinery backed by oslo.config library. It includes the
|
|
||||||
following pieces of knowledge:
|
|
||||||
|
|
||||||
* integration_tests/config.py file where all possible setting groups and
|
|
||||||
settings are defined along with their descriptions and defaults. If you are
|
|
||||||
going to add a new setting to Horizon integration tests, you should add it
|
|
||||||
first to this file.
|
|
||||||
|
|
||||||
* integration_tests/horizon.conf file - where all the overrides are
|
|
||||||
actually located. For clarity its contents mirrors the default values
|
|
||||||
in config.py (although technically they could be completely commented out
|
|
||||||
with the same result).
|
|
||||||
|
|
||||||
* To make developers' lives easier a local-only (not tracked by git)
|
|
||||||
counterpart of horizon.conf could exist at the same directory, named
|
|
||||||
'local-horizon.conf'. It is meant solely for overriding values from
|
|
||||||
horizon.conf that a developer's environment might differ from the gate
|
|
||||||
environment (like Horizon url or admin user password).
|
|
||||||
|
|
||||||
* When integration tests are run by openstack-infra/devstack-gate scripts they
|
|
||||||
use 2 hooks to alter the devstack gate environment, namely pre_test_hook and
|
|
||||||
post_test_hook. Contents of both hooks are defined inside the corresponding
|
|
||||||
shell scripts located at 'tools/gate/integration' at the top-level of horizon
|
|
||||||
repo. If you find yourself asking which of the hooks you need to modify - pre
|
|
||||||
or post, keep the following things in mind.
|
|
||||||
|
|
||||||
* Pre hook is executed before the Devstack is deployed, that essentially
|
|
||||||
means that almost none of packages that are installed as OpenStack services
|
|
||||||
dependencies during Devstack deployment are going to be present in the
|
|
||||||
system. Yet all the repositories contained with `PROJECTS` variable defined
|
|
||||||
in `devstack-vm-gate-wrap.sh`_ script will be already cloned by the moment
|
|
||||||
pre hook is executed. So the natural use for it is to customize some Horizon
|
|
||||||
settings before they are used in operations like compressing statics etc.
|
|
||||||
That is how it is actually used now: it sets settings variable
|
|
||||||
`INTEGRATION_TESTS_SUPPORT` to `True`, so all the support code for integration
|
|
||||||
tests is included into Horizon static assets. If this variable was set to
|
|
||||||
`False`, we would not be able to run integration tests.
|
|
||||||
|
|
||||||
* Post hook is executed after Devstack is deployed, so integration tests
|
|
||||||
themselves are run inside that hook, as well as various test artifacts
|
|
||||||
collection. When you modify it, do not forget to save the exit code of
|
|
||||||
a tox integration tests run and emit at the end of the script - or you may
|
|
||||||
lose the SUCCESS/FAILURE status of the whole tests suite and tamper with the
|
|
||||||
job results!
|
|
||||||
|
|
||||||
.. _devstack-vm-gate-wrap.sh: https://github.com/openstack-infra/devstack-gate/blob/master/devstack-vm-gate-wrap.sh
|
|
||||||
|
|
||||||
|
|
||||||
Writing integration tests for Horizon plugins
|
|
||||||
---------------------------------------------
|
|
||||||
|
|
||||||
First, for more details on writing a Horizon plugin please refer to
|
|
||||||
:doc:`../tutorials/plugin`. Second, there are 2 possible setups when running
|
|
||||||
integration tests for Horizon plugins.
|
|
||||||
|
|
||||||
The first setup, which is suggested to be used in gate of \*-dashboard plugins
|
|
||||||
is to get horizon as a dependency of a plugin and then run integration tests
|
|
||||||
using horizon.conf config file inside the plugin repo. This way the plugin augments
|
|
||||||
the location of Horizon built-in Page Objects with the location of its own
|
|
||||||
Page Objects, contained within the `plugin_page_path` option and the Horizon
|
|
||||||
built-in nav structure with its own nav structure contained within
|
|
||||||
`plugin_page_structure`. Then the plugin integration tests are run against core
|
|
||||||
Horizon augmented with just this particular plugin content.
|
|
||||||
|
|
||||||
The second setup may be used when it is needed to run integration tests for
|
|
||||||
Horizon + several plugins. In other words, content from several plugins is
|
|
||||||
merged into core Horizon content, then the combined integration tests from core
|
|
||||||
Horizon and all the involved plugins are run against the resulting dashboards.
|
|
||||||
To make this possible both options `plugin_page_path` and
|
|
||||||
`plugin_page_structure` have MultiStrOpt type. This means that they may be
|
|
||||||
defined several times and all the specified values will be gathered in a list,
|
|
||||||
which is iterated over when running integration tests. In this setup it's easier to
|
|
||||||
run the tests from Horizon repo, using the horizon.conf file within it.
|
|
||||||
|
|
||||||
Also keep in mind that `plugin_page_structure` needs to be a strict JSON
|
|
||||||
string, w/o trailing commas etc.
|
|
||||||
|
Loading…
Reference in New Issue
Block a user