horizon/tools/unit_tests.sh
Akihiro Motoki ccaaf8cbb9 Fix regressions in unit_tests.sh during pytest mirgation
The pytest migration patch introduced a couple of regression in
tools/unit_tests.sh:

* We no longer cannot run a specific test explicitly
  (per python file, per test class, per specific test).
  Per module like openstack_dashboard, openstack_auth and horizon
  is the only possible way after the pytest migration.
* openstack_dashboard tests outside of openstack_dashboard/test/
  are not looked up when a subset is specified.

This commit restores the orginal behavior.

pytest uses a different way to specify tests.
We cannot use the style of python modules like
"openstack_auth.tests.unit.test_policy.PolicyLoaderTestCase.test_policy_file_load"
and we need to specify tests like file paths [1].

[1] https://docs.pytest.org/en/latest/usage.html#specifying-tests-selecting-tests

Closes-Bug: #1875123
Change-Id: Ic4b6c69d8a2ed774cce16521c4c4336f55ff737e
2020-04-26 20:39:10 +09:00

51 lines
2.5 KiB
Bash
Executable File

# Uses envpython and toxinidir from tox run to construct a test command
testcommand="pytest"
posargs="${@:2}"
tagarg="not selenium and not integration and not plugin_test"
# Attempt to identify if any of the arguments passed from tox is a test subset
if [ -n "$posargs" ]; then
for arg in "$posargs"
do
if [ ${arg:0:1} != "-" ]; then
subset=$arg
fi
done
fi
horizon_test_results="--junitxml=${1}/test_reports/horizon_test_results.xml --html=${1}/test_reports/horizon_test_results.html"
dashboard_test_results="--junitxml=${1}/test_reports/openstack_dashboard_test_results.xml --html=${1}/test_reports/openstack_dashboard_test_results.html"
auth_test_results="--junitxml=${1}/test_reports/openstack_auth_test_results.xml --html=${1}/test_reports/openstack_auth_test_results.html"
plugins_test_results="--junitxml=${1}/test_reports/plugin_test_results.xml --html=${1}/test_reports/plugin_test_results.html"
single_html="--self-contained-html"
# If we are running a test subset, supply the correct settings file.
# If not, simply run the entire test suite.
if [ -n "$subset" ]; then
project="${subset%%/*}"
if [ $project == "horizon" ]; then
$testcommand $posargs --ds=horizon.test.settings -v -m "$tagarg" $horizon_test_results $single_html
elif [ $project == "openstack_dashboard" ]; then
$testcommand $posargs --ds=openstack_dashboard.test.settings -v -m "$tagarg" $dashboard_test_results $single_html
elif [ $project == "openstack_auth" ]; then
$testcommand $posargs --ds=openstack_auth.tests.settings -v -m "$tagarg" $auth_test_results $single_html
elif [ $project == "plugin-test" ]; then
$testcommand ${1}/openstack_dashboard/test/test_plugins --ds=openstack_dashboard.test.settings -v -m plugin_test $plugins_test_results $single_html
fi
else
$testcommand ${1}/horizon/ --ds=horizon.test.settings -v -m "$tagarg" $horizon_test_results $single_html
horizon_tests=$?
$testcommand ${1}/openstack_dashboard/ --ds=openstack_dashboard.test.settings -v -m "$tagarg" $dashboard_test_results $single_html
openstack_dashboard_tests=$?
$testcommand ${1}/openstack_auth/tests/ --ds=openstack_auth.tests.settings -v -m "$tagarg" $auth_test_results $single_html
auth_tests=$?
$testcommand ${1}/openstack_dashboard/ --ds=openstack_dashboard.test.settings -v -m plugin_test $plugins_test_results $single_html
plugin_tests=$?
# we have to tell tox if either of these test runs failed
if [[ $horizon_tests != 0 || $openstack_dashboard_tests != 0 || \
$auth_tests != 0 || $plugin_tests != 0 ]]; then
exit 1;
fi
fi