2cb0e674f3
Change deferred_auth_method to trusts by default, and fixes unittests error. Closes-Bug: #1286157 Change-Id: Ifebbc1b5d71458d8502e08bcacbd94b7437128f9
208 lines
7.4 KiB
Python
208 lines
7.4 KiB
Python
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
|
|
import logging
|
|
import os
|
|
import sys
|
|
import time
|
|
|
|
import fixtures
|
|
import mox
|
|
from oslo_config import cfg
|
|
from oslotest import mockpatch
|
|
import testscenarios
|
|
import testtools
|
|
|
|
from heat.common import context
|
|
from heat.common import messaging
|
|
from heat.engine.clients.os import cinder
|
|
from heat.engine.clients.os import glance
|
|
from heat.engine.clients.os import keystone
|
|
from heat.engine.clients.os import neutron
|
|
from heat.engine.clients.os import nova
|
|
from heat.engine.clients.os import trove
|
|
from heat.engine import environment
|
|
from heat.engine import resources
|
|
from heat.engine import scheduler
|
|
from heat.tests import fakes
|
|
from heat.tests import utils
|
|
|
|
|
|
TEST_DEFAULT_LOGLEVELS = {'migrate': logging.WARN,
|
|
'sqlalchemy': logging.WARN}
|
|
_LOG_FORMAT = "%(levelname)8s [%(name)s] %(message)s"
|
|
_TRUE_VALUES = ('True', 'true', '1', 'yes')
|
|
|
|
|
|
class FakeLogMixin(object):
|
|
def setup_logging(self):
|
|
# Assign default logs to self.LOG so we can still
|
|
# assert on heat logs.
|
|
default_level = logging.INFO
|
|
if os.environ.get('OS_DEBUG') in _TRUE_VALUES:
|
|
default_level = logging.DEBUG
|
|
|
|
self.LOG = self.useFixture(
|
|
fixtures.FakeLogger(level=default_level, format=_LOG_FORMAT))
|
|
base_list = set([nlog.split('.')[0]
|
|
for nlog in logging.Logger.manager.loggerDict])
|
|
for base in base_list:
|
|
if base in TEST_DEFAULT_LOGLEVELS:
|
|
self.useFixture(fixtures.FakeLogger(
|
|
level=TEST_DEFAULT_LOGLEVELS[base],
|
|
name=base, format=_LOG_FORMAT))
|
|
elif base != 'heat':
|
|
self.useFixture(fixtures.FakeLogger(
|
|
name=base, format=_LOG_FORMAT))
|
|
|
|
|
|
class HeatTestCase(testscenarios.WithScenarios,
|
|
testtools.TestCase, FakeLogMixin):
|
|
|
|
TIME_STEP = 0.1
|
|
|
|
def setUp(self, mock_keystone=True):
|
|
super(HeatTestCase, self).setUp()
|
|
self.m = mox.Mox()
|
|
self.addCleanup(self.m.UnsetStubs)
|
|
self.setup_logging()
|
|
scheduler.ENABLE_SLEEP = False
|
|
self.useFixture(fixtures.MonkeyPatch(
|
|
'heat.common.exception._FATAL_EXCEPTION_FORMAT_ERRORS',
|
|
True))
|
|
|
|
def enable_sleep():
|
|
scheduler.ENABLE_SLEEP = True
|
|
|
|
self.addCleanup(enable_sleep)
|
|
|
|
mod_dir = os.path.dirname(sys.modules[__name__].__file__)
|
|
project_dir = os.path.abspath(os.path.join(mod_dir, '../../'))
|
|
env_dir = os.path.join(project_dir, 'etc', 'heat',
|
|
'environment.d')
|
|
|
|
cfg.CONF.set_default('environment_dir', env_dir)
|
|
cfg.CONF.set_override('error_wait_time', None)
|
|
self.addCleanup(cfg.CONF.reset)
|
|
|
|
messaging.setup("fake://", optional=True)
|
|
self.addCleanup(messaging.cleanup)
|
|
|
|
tri = resources.global_env().get_resource_info(
|
|
'AWS::RDS::DBInstance',
|
|
registry_type=environment.TemplateResourceInfo)
|
|
if tri is not None:
|
|
cur_path = tri.template_name
|
|
templ_path = os.path.join(project_dir, 'etc', 'heat', 'templates')
|
|
if templ_path not in cur_path:
|
|
tri.template_name = cur_path.replace('/etc/heat/templates',
|
|
templ_path)
|
|
|
|
# use CWLiteAlarm for testing.
|
|
resources.global_env().registry.load(
|
|
{"AWS::CloudWatch::Alarm": "OS::Heat::CWLiteAlarm"})
|
|
if mock_keystone:
|
|
self.stub_keystoneclient()
|
|
utils.setup_dummy_db()
|
|
self.addCleanup(utils.reset_dummy_db)
|
|
|
|
def stub_wallclock(self):
|
|
"""
|
|
Overrides scheduler wallclock to speed up tests expecting timeouts.
|
|
"""
|
|
self._wallclock = time.time()
|
|
|
|
def fake_wallclock():
|
|
self._wallclock += self.TIME_STEP
|
|
return self._wallclock
|
|
|
|
self.m.StubOutWithMock(scheduler, 'wallclock')
|
|
scheduler.wallclock = fake_wallclock
|
|
|
|
def patchobject(self, obj, attr, **kwargs):
|
|
mockfixture = self.useFixture(mockpatch.PatchObject(obj, attr,
|
|
**kwargs))
|
|
return mockfixture.mock
|
|
|
|
# NOTE(pshchelo): this overrides the testtools.TestCase.patch method
|
|
# that does simple monkey-patching in favor of mock's patching
|
|
def patch(self, target, **kwargs):
|
|
mockfixture = self.useFixture(mockpatch.Patch(target, **kwargs))
|
|
return mockfixture.mock
|
|
|
|
def stub_auth(self, ctx=None, **kwargs):
|
|
auth = self.patchobject(ctx or context.RequestContext,
|
|
"_create_auth_plugin")
|
|
fake_auth = fakes.FakeAuth(**kwargs)
|
|
auth.return_value = fake_auth
|
|
return auth
|
|
|
|
def stub_keystoneclient(self, fake_client=None, **kwargs):
|
|
client = self.patchobject(keystone.KeystoneClientPlugin, "_create")
|
|
fkc = fake_client or fakes.FakeKeystoneClient(**kwargs)
|
|
client.return_value = fkc
|
|
return fkc
|
|
|
|
def stub_KeypairConstraint_validate(self):
|
|
validate = self.patchobject(nova.KeypairConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_ImageConstraint_validate(self, num=None):
|
|
validate = self.patchobject(glance.ImageConstraint, 'validate')
|
|
if num is None:
|
|
validate.return_value = True
|
|
else:
|
|
validate.side_effect = [True for x in range(num)]
|
|
|
|
def stub_FlavorConstraint_validate(self):
|
|
validate = self.patchobject(nova.FlavorConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_VolumeConstraint_validate(self):
|
|
validate = self.patchobject(cinder.VolumeConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_SnapshotConstraint_validate(self):
|
|
validate = self.patchobject(
|
|
cinder.VolumeSnapshotConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_VolumeTypeConstraint_validate(self):
|
|
validate = self.patchobject(cinder.VolumeTypeConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_ServerConstraint_validate(self):
|
|
validate = self.patchobject(nova.ServerConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_NetworkConstraint_validate(self):
|
|
validate = self.patchobject(neutron.NetworkConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_PortConstraint_validate(self):
|
|
validate = self.patchobject(neutron.PortConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_TroveFlavorConstraint_validate(self):
|
|
validate = self.patchobject(trove.FlavorConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_SubnetConstraint_validate(self):
|
|
validate = self.patchobject(neutron.SubnetConstraint, 'validate')
|
|
validate.return_value = True
|
|
|
|
def stub_RouterConstraint_validate(self):
|
|
validate = self.patchobject(neutron.RouterConstraint, 'validate')
|
|
validate.return_value = True
|