
We are still having some issues in the gate where greenlets from previous tests continue to run while the next test starts, causing false negative failures in unit or functional test jobs. This adds a new fixture that will ensure GreenThreadPoolExecutor.shutdown() is called with wait=True, to wait for greenlets in the pool to finish running before moving on. In local testing, doing this does not appear to adversely affect test run times, which was my primary concern. As a baseline, I ran a subset of functional tests in a loop until failure without the patch and after 11 hours, I got a failure reproducing the bug. With the patch, running the same subset of functional tests in a loop has been running for 24 hours and has not failed yet. Based on this, I think it may be worth trying this out to see if it will help stability of our unit and functional test jobs. And if it ends up impacting test run times or causes other issues, we can revert it. Partial-Bug: #1946339 Change-Id: Ia916310522b007061660172fa4d63d0fde9a55ac
982 lines
38 KiB
Python
982 lines
38 KiB
Python
# Copyright 2010 United States Government as represented by the
|
|
# Administrator of the National Aeronautics and Space Administration.
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
"""Base classes for our unit tests.
|
|
|
|
Allows overriding of flags for use of fakes, and some black magic for
|
|
inline callbacks.
|
|
|
|
"""
|
|
|
|
import nova.monkey_patch # noqa
|
|
|
|
import abc
|
|
import builtins
|
|
import collections
|
|
import contextlib
|
|
import copy
|
|
import datetime
|
|
import inspect
|
|
import itertools
|
|
import os
|
|
import os.path
|
|
import pprint
|
|
import sys
|
|
from unittest import mock
|
|
|
|
import fixtures
|
|
from oslo_cache import core as cache
|
|
from oslo_concurrency import lockutils
|
|
from oslo_config import cfg
|
|
from oslo_config import fixture as config_fixture
|
|
from oslo_log.fixture import logging_error as log_fixture
|
|
from oslo_log import log as logging
|
|
from oslo_serialization import jsonutils
|
|
from oslo_utils.fixture import uuidsentinel as uuids
|
|
from oslo_utils import timeutils
|
|
from oslo_versionedobjects import fixture as ovo_fixture
|
|
from oslotest import base
|
|
from oslotest import mock_fixture
|
|
from sqlalchemy.dialects import sqlite
|
|
import testtools
|
|
|
|
from nova.api.openstack import wsgi_app
|
|
from nova.compute import rpcapi as compute_rpcapi
|
|
from nova import context
|
|
from nova.db.main import api as db_api
|
|
from nova import exception
|
|
from nova import objects
|
|
from nova.objects import base as objects_base
|
|
from nova import quota
|
|
from nova.scheduler.client import report
|
|
from nova.scheduler import utils as scheduler_utils
|
|
from nova.tests import fixtures as nova_fixtures
|
|
from nova.tests.unit import matchers
|
|
from nova import utils
|
|
from nova.virt import images
|
|
|
|
CONF = cfg.CONF
|
|
|
|
logging.register_options(CONF)
|
|
CONF.set_override('use_stderr', False)
|
|
logging.setup(CONF, 'nova')
|
|
cache.configure(CONF)
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
_TRUE_VALUES = ('True', 'true', '1', 'yes')
|
|
CELL1_NAME = 'cell1'
|
|
|
|
|
|
# For compatibility with the large number of tests which use test.nested
|
|
nested = utils.nested_contexts
|
|
|
|
|
|
class TestingException(Exception):
|
|
pass
|
|
|
|
|
|
# NOTE(claudiub): this needs to be called before any mock.patch calls are
|
|
# being done, and especially before any other test classes load. This fixes
|
|
# the mock.patch autospec issue:
|
|
# https://github.com/testing-cabal/mock/issues/396
|
|
mock_fixture.patch_mock_module()
|
|
|
|
|
|
def _poison_unfair_compute_resource_semaphore_locking():
|
|
"""Ensure that every locking on COMPUTE_RESOURCE_SEMAPHORE is called with
|
|
fair=True.
|
|
"""
|
|
orig_synchronized = utils.synchronized
|
|
|
|
def poisoned_synchronized(*args, **kwargs):
|
|
# Only check fairness if the decorator is used with
|
|
# COMPUTE_RESOURCE_SEMAPHORE. But the name of the semaphore can be
|
|
# passed as args or as kwargs.
|
|
# Note that we cannot import COMPUTE_RESOURCE_SEMAPHORE as that would
|
|
# apply the decorators we want to poison here.
|
|
if len(args) >= 1:
|
|
name = args[0]
|
|
else:
|
|
name = kwargs.get("name")
|
|
if name == "compute_resources" and not kwargs.get("fair", False):
|
|
raise AssertionError(
|
|
'Locking on COMPUTE_RESOURCE_SEMAPHORE should always be fair. '
|
|
'See bug 1864122.')
|
|
# go and act like the original decorator
|
|
return orig_synchronized(*args, **kwargs)
|
|
|
|
# replace the synchronized decorator factory with our own that checks the
|
|
# params passed in
|
|
utils.synchronized = poisoned_synchronized
|
|
|
|
|
|
# NOTE(gibi): This poisoning needs to be done in import time as decorators are
|
|
# applied in import time on the ResourceTracker
|
|
_poison_unfair_compute_resource_semaphore_locking()
|
|
|
|
|
|
class NovaExceptionReraiseFormatError(object):
|
|
real_log_exception = exception.NovaException._log_exception
|
|
|
|
@classmethod
|
|
def patch(cls):
|
|
exception.NovaException._log_exception = cls._wrap_log_exception
|
|
|
|
@staticmethod
|
|
def _wrap_log_exception(self):
|
|
exc_info = sys.exc_info()
|
|
NovaExceptionReraiseFormatError.real_log_exception(self)
|
|
raise exc_info[1]
|
|
|
|
|
|
# NOTE(melwitt) This needs to be done at import time in order to also catch
|
|
# NovaException format errors that are in mock decorators. In these cases, the
|
|
# errors will be raised during test listing, before tests actually run.
|
|
NovaExceptionReraiseFormatError.patch()
|
|
|
|
|
|
class TestCase(base.BaseTestCase):
|
|
"""Test case base class for all unit tests.
|
|
|
|
Due to the slowness of DB access, please consider deriving from
|
|
`NoDBTestCase` first.
|
|
"""
|
|
# USES_DB is set to False for tests that inherit from NoDBTestCase.
|
|
USES_DB = True
|
|
# USES_DB_SELF is set to True in tests that specifically want to use the
|
|
# database but need to configure it themselves, for example to setup the
|
|
# API DB but not the cell DB. In those cases the test will override
|
|
# USES_DB_SELF = True but inherit from the NoDBTestCase class so it does
|
|
# not get the default fixture setup when using a database (which is the
|
|
# API and cell DBs, and adding the default flavors).
|
|
USES_DB_SELF = False
|
|
REQUIRES_LOCKING = False
|
|
|
|
# Setting to True makes the test use the RPCFixture.
|
|
STUB_RPC = True
|
|
|
|
# The number of non-cell0 cells to create. This is only used in the
|
|
# base class when USES_DB is True.
|
|
NUMBER_OF_CELLS = 1
|
|
|
|
# The stable compute id stuff is intentionally singleton-ish, which makes
|
|
# it a nightmare for testing multiple host/node combinations in tests like
|
|
# we do. So, mock it out by default, unless the test is specifically
|
|
# designed to handle it.
|
|
STUB_COMPUTE_ID = True
|
|
|
|
def setUp(self):
|
|
"""Run before each test method to initialize test environment."""
|
|
# Ensure BaseTestCase's ConfigureLogging fixture is disabled since
|
|
# we're using our own (StandardLogging).
|
|
with fixtures.EnvironmentVariable('OS_LOG_CAPTURE', '0'):
|
|
super(TestCase, self).setUp()
|
|
|
|
self.useFixture(
|
|
nova_fixtures.PropagateTestCaseIdToChildEventlets(self.id()))
|
|
|
|
# How many of which service we've started. {$service-name: $count}
|
|
self._service_fixture_count = collections.defaultdict(int)
|
|
|
|
self.useFixture(nova_fixtures.OpenStackSDKFixture())
|
|
|
|
self.useFixture(log_fixture.get_logging_handle_error_fixture())
|
|
|
|
self.stdlog = self.useFixture(nova_fixtures.StandardLogging())
|
|
|
|
# NOTE(sdague): because of the way we were using the lock
|
|
# wrapper we ended up with a lot of tests that started
|
|
# relying on global external locking being set up for them. We
|
|
# consider all of these to be *bugs*. Tests should not require
|
|
# global external locking, or if they do, they should
|
|
# explicitly set it up themselves.
|
|
#
|
|
# The following REQUIRES_LOCKING class parameter is provided
|
|
# as a bridge to get us there. No new tests should be added
|
|
# that require it, and existing classes and tests should be
|
|
# fixed to not need it.
|
|
if self.REQUIRES_LOCKING:
|
|
lock_path = self.useFixture(fixtures.TempDir()).path
|
|
self.fixture = self.useFixture(
|
|
config_fixture.Config(lockutils.CONF))
|
|
self.fixture.config(lock_path=lock_path,
|
|
group='oslo_concurrency')
|
|
|
|
self.useFixture(nova_fixtures.ConfFixture(CONF))
|
|
|
|
if self.STUB_RPC:
|
|
self.useFixture(nova_fixtures.RPCFixture('nova.test'))
|
|
|
|
# we cannot set this in the ConfFixture as oslo only registers the
|
|
# notification opts at the first instantiation of a Notifier that
|
|
# happens only in the RPCFixture
|
|
CONF.set_default('driver', ['test'],
|
|
group='oslo_messaging_notifications')
|
|
|
|
# NOTE(danms): Make sure to reset us back to non-remote objects
|
|
# for each test to avoid interactions. Also, backup the object
|
|
# registry.
|
|
objects_base.NovaObject.indirection_api = None
|
|
self._base_test_obj_backup = copy.copy(
|
|
objects_base.NovaObjectRegistry._registry._obj_classes)
|
|
self.addCleanup(self._restore_obj_registry)
|
|
objects.Service.clear_min_version_cache()
|
|
|
|
# NOTE(danms): Reset the cached list of cells
|
|
from nova.compute import api
|
|
api.CELLS = []
|
|
context.CELL_CACHE = {}
|
|
context.CELLS = []
|
|
|
|
self.computes = {}
|
|
self.cell_mappings = {}
|
|
self.host_mappings = {}
|
|
# NOTE(danms): If the test claims to want to set up the database
|
|
# itself, then it is responsible for all the mapping stuff too.
|
|
if self.USES_DB:
|
|
# NOTE(danms): Full database setup involves a cell0, cell1,
|
|
# and the relevant mappings.
|
|
self.useFixture(nova_fixtures.Database(database='api'))
|
|
self._setup_cells()
|
|
self.useFixture(nova_fixtures.DefaultFlavorsFixture())
|
|
elif not self.USES_DB_SELF:
|
|
# NOTE(danms): If not using the database, we mock out the
|
|
# mapping stuff and effectively collapse everything to a
|
|
# single cell.
|
|
self.useFixture(nova_fixtures.SingleCellSimple())
|
|
self.useFixture(nova_fixtures.DatabasePoisonFixture())
|
|
|
|
# NOTE(blk-u): WarningsFixture must be after the Database fixture
|
|
# because sqlalchemy-migrate messes with the warnings filters.
|
|
self.useFixture(nova_fixtures.WarningsFixture())
|
|
|
|
self.useFixture(ovo_fixture.StableObjectJsonFixture())
|
|
|
|
# Reset the global QEMU version flag.
|
|
images.QEMU_VERSION = None
|
|
|
|
# Reset the compute RPC API globals (mostly the _ROUTER).
|
|
compute_rpcapi.reset_globals()
|
|
|
|
self.addCleanup(self._clear_attrs)
|
|
self.useFixture(fixtures.EnvironmentVariable('http_proxy'))
|
|
self.policy = self.useFixture(nova_fixtures.PolicyFixture())
|
|
|
|
self.useFixture(nova_fixtures.PoisonFunctions())
|
|
|
|
self.useFixture(nova_fixtures.ForbidNewLegacyNotificationFixture())
|
|
|
|
# NOTE(mikal): make sure we don't load a privsep helper accidentally
|
|
self.useFixture(nova_fixtures.PrivsepNoHelperFixture())
|
|
self.useFixture(mock_fixture.MockAutospecFixture())
|
|
|
|
# FIXME(danms): Disable this for all tests by default to avoid breaking
|
|
# any that depend on default/previous ordering
|
|
self.flags(build_failure_weight_multiplier=0.0,
|
|
group='filter_scheduler')
|
|
|
|
# NOTE(melwitt): Reset the cached set of projects
|
|
quota.UID_QFD_POPULATED_CACHE_BY_PROJECT = set()
|
|
quota.UID_QFD_POPULATED_CACHE_ALL = False
|
|
|
|
self.useFixture(nova_fixtures.GenericPoisonFixture())
|
|
self.useFixture(nova_fixtures.SysFsPoisonFixture())
|
|
|
|
# Additional module names can be added to this set if needed
|
|
self.useFixture(nova_fixtures.ImportModulePoisonFixture(
|
|
set(['guestfs', 'libvirt'])))
|
|
|
|
# make sure that the wsgi app is fully initialized for all testcase
|
|
# instead of only once initialized for test worker
|
|
wsgi_app.init_global_data.reset()
|
|
|
|
# Reset the placement client singleton
|
|
report.PLACEMENTCLIENT = None
|
|
|
|
# Reset our local node uuid cache (and avoid writing to the
|
|
# local filesystem when we generate a new one).
|
|
if self.STUB_COMPUTE_ID:
|
|
self.useFixture(nova_fixtures.ComputeNodeIdFixture())
|
|
|
|
# Reset globals indicating affinity filter support. Some tests may set
|
|
# self.flags(enabled_filters=...) which could make the affinity filter
|
|
# support globals get set to a non-default configuration which affects
|
|
# all other tests.
|
|
scheduler_utils.reset_globals()
|
|
|
|
# Wait for bare greenlets spawn_n()'ed from a GreenThreadPoolExecutor
|
|
# to finish before moving on from the test. When greenlets from a
|
|
# previous test remain running, they may attempt to access structures
|
|
# (like the database) that have already been torn down and can cause
|
|
# the currently running test to fail.
|
|
self.useFixture(nova_fixtures.GreenThreadPoolShutdownWait())
|
|
|
|
def _setup_cells(self):
|
|
"""Setup a normal cellsv2 environment.
|
|
|
|
This sets up the CellDatabase fixture with two cells, one cell0
|
|
and one normal cell. CellMappings are created for both so that
|
|
cells-aware code can find those two databases.
|
|
"""
|
|
celldbs = nova_fixtures.CellDatabases()
|
|
|
|
ctxt = context.get_context()
|
|
fake_transport = 'fake://nowhere/'
|
|
|
|
c0 = objects.CellMapping(
|
|
context=ctxt,
|
|
uuid=objects.CellMapping.CELL0_UUID,
|
|
name='cell0',
|
|
transport_url=fake_transport,
|
|
database_connection=objects.CellMapping.CELL0_UUID)
|
|
c0.create()
|
|
self.cell_mappings[c0.name] = c0
|
|
celldbs.add_cell_database(objects.CellMapping.CELL0_UUID)
|
|
|
|
for x in range(self.NUMBER_OF_CELLS):
|
|
name = 'cell%i' % (x + 1)
|
|
uuid = getattr(uuids, name)
|
|
cell = objects.CellMapping(
|
|
context=ctxt,
|
|
uuid=uuid,
|
|
name=name,
|
|
transport_url=fake_transport,
|
|
database_connection=uuid)
|
|
cell.create()
|
|
self.cell_mappings[name] = cell
|
|
# cell1 is the default cell
|
|
celldbs.add_cell_database(uuid, default=(x == 0))
|
|
|
|
self.useFixture(celldbs)
|
|
|
|
def _restore_obj_registry(self):
|
|
objects_base.NovaObjectRegistry._registry._obj_classes = \
|
|
self._base_test_obj_backup
|
|
|
|
def _clear_attrs(self):
|
|
# Delete attributes that don't start with _ so they don't pin
|
|
# memory around unnecessarily for the duration of the test
|
|
# suite
|
|
for key in [k for k in self.__dict__.keys() if k[0] != '_']:
|
|
# NOTE(gmann): Skip attribute 'id' because if tests are being
|
|
# generated using testscenarios then, 'id' attribute is being
|
|
# added during cloning the tests. And later that 'id' attribute
|
|
# is being used by test suite to generate the results for each
|
|
# newly generated tests by testscenarios.
|
|
if key != 'id':
|
|
del self.__dict__[key]
|
|
|
|
def stub_out(self, old, new):
|
|
"""Replace a function for the duration of the test.
|
|
|
|
Use the monkey patch fixture to replace a function for the
|
|
duration of a test. Useful when you want to provide fake
|
|
methods instead of mocks during testing.
|
|
"""
|
|
self.useFixture(fixtures.MonkeyPatch(old, new))
|
|
|
|
@staticmethod
|
|
def patch_exists(patched_path, result, other=None):
|
|
"""Provide a static method version of patch_exists(), which if you
|
|
haven't already imported nova.test can be slightly easier to
|
|
use as a context manager within a test method via:
|
|
|
|
def test_something(self):
|
|
with self.patch_exists(path, True):
|
|
...
|
|
"""
|
|
return patch_exists(patched_path, result, other)
|
|
|
|
@staticmethod
|
|
def patch_open(patched_path, read_data):
|
|
"""Provide a static method version of patch_open() which is easier to
|
|
use as a context manager within a test method via:
|
|
|
|
def test_something(self):
|
|
with self.patch_open(path, "fake contents of file"):
|
|
...
|
|
"""
|
|
return patch_open(patched_path, read_data)
|
|
|
|
def flags(self, **kw):
|
|
"""Override flag variables for a test."""
|
|
group = kw.pop('group', None)
|
|
for k, v in kw.items():
|
|
CONF.set_override(k, v, group)
|
|
|
|
def enforce_fk_constraints(self, engine=None):
|
|
if engine is None:
|
|
engine = db_api.get_engine()
|
|
dialect = engine.url.get_dialect()
|
|
if dialect == sqlite.dialect:
|
|
engine.connect().exec_driver_sql("PRAGMA foreign_keys = ON")
|
|
|
|
def start_service(self, name, host=None, cell_name=None, **kwargs):
|
|
# Disallow starting multiple scheduler services
|
|
if name == 'scheduler' and self._service_fixture_count[name]:
|
|
raise TestingException("Duplicate start_service(%s)!" % name)
|
|
|
|
cell = None
|
|
# if the host is None then the CONF.host remains defaulted to
|
|
# 'fake-mini' (originally done in ConfFixture)
|
|
if host is not None:
|
|
# Make sure that CONF.host is relevant to the right hostname
|
|
self.useFixture(nova_fixtures.ConfPatcher(host=host))
|
|
|
|
if name == 'compute' and self.USES_DB:
|
|
# NOTE(danms): We need to create the HostMapping first, because
|
|
# otherwise we'll fail to update the scheduler while running
|
|
# the compute node startup routines below.
|
|
ctxt = context.get_context()
|
|
cell_name = cell_name or CELL1_NAME
|
|
cell = self.cell_mappings[cell_name]
|
|
if (host or name) not in self.host_mappings:
|
|
# NOTE(gibi): If the HostMapping does not exists then this is
|
|
# the first start of the service so we create the mapping.
|
|
hm = objects.HostMapping(context=ctxt,
|
|
host=host or name,
|
|
cell_mapping=cell)
|
|
hm.create()
|
|
self.host_mappings[hm.host] = hm
|
|
svc = self.useFixture(
|
|
nova_fixtures.ServiceFixture(name, host, cell=cell, **kwargs))
|
|
|
|
# Keep track of how many instances of this service are running.
|
|
self._service_fixture_count[name] += 1
|
|
real_stop = svc.service.stop
|
|
|
|
# Make sure stopping the service decrements the active count, so that
|
|
# start,stop,start doesn't trigger the "Duplicate start_service"
|
|
# exception.
|
|
def patch_stop(*a, **k):
|
|
self._service_fixture_count[name] -= 1
|
|
return real_stop(*a, **k)
|
|
self.useFixture(fixtures.MockPatchObject(
|
|
svc.service, 'stop', patch_stop))
|
|
|
|
return svc.service
|
|
|
|
def _start_compute(self, host, cell_name=None):
|
|
"""Start a nova compute service on the given host
|
|
|
|
:param host: the name of the host that will be associated to the
|
|
compute service.
|
|
:param cell_name: optional name of the cell in which to start the
|
|
compute service
|
|
:return: the nova compute service object
|
|
"""
|
|
compute = self.start_service('compute', host=host, cell_name=cell_name)
|
|
self.computes[host] = compute
|
|
return compute
|
|
|
|
def _run_periodics(self):
|
|
"""Run the update_available_resource task on every compute manager
|
|
|
|
This runs periodics on the computes in an undefined order; some child
|
|
class redefine this function to force a specific order.
|
|
"""
|
|
|
|
ctx = context.get_admin_context()
|
|
for host, compute in self.computes.items():
|
|
LOG.info('Running periodic for compute (%s)', host)
|
|
# Make sure the context is targeted to the proper cell database
|
|
# for multi-cell tests.
|
|
with context.target_cell(
|
|
ctx, self.host_mappings[host].cell_mapping) as cctxt:
|
|
compute.manager.update_available_resource(cctxt)
|
|
LOG.info('Finished with periodics')
|
|
|
|
def restart_compute_service(self, compute, keep_hypervisor_state=True):
|
|
"""Stops the service and starts a new one to have realistic restart
|
|
|
|
:param:compute: the nova-compute service to be restarted
|
|
:param:keep_hypervisor_state: If true then already defined instances
|
|
will survive the compute service restart.
|
|
If false then the new service will see
|
|
an empty hypervisor
|
|
:returns: a new compute service instance serving the same host and
|
|
and node
|
|
"""
|
|
|
|
# NOTE(gibi): The service interface cannot be used to simulate a real
|
|
# service restart as the manager object will not be recreated after a
|
|
# service.stop() and service.start() therefore the manager state will
|
|
# survive. For example the resource tracker will not be recreated after
|
|
# a stop start. The service.kill() call cannot help as it deletes
|
|
# the service from the DB which is unrealistic and causes that some
|
|
# operation that refers to the killed host (e.g. evacuate) fails.
|
|
# So this helper method will stop the original service and then starts
|
|
# a brand new compute service for the same host and node. This way
|
|
# a new ComputeManager instance will be created and initialized during
|
|
# the service startup.
|
|
compute.stop()
|
|
|
|
# this service was running previously so we have to make sure that
|
|
# we restart it in the same cell
|
|
cell_name = self.host_mappings[compute.host].cell_mapping.name
|
|
|
|
if keep_hypervisor_state:
|
|
# NOTE(gibi): FakeDriver does not provide a meaningful way to
|
|
# define some servers that exists already on the hypervisor when
|
|
# the driver is (re)created during the service startup. This means
|
|
# that we cannot simulate that the definition of a server
|
|
# survives a nova-compute service restart on the hypervisor.
|
|
# Instead here we save the FakeDriver instance that knows about
|
|
# the defined servers and inject that driver into the new Manager
|
|
# class during the startup of the compute service.
|
|
old_driver = compute.manager.driver
|
|
with mock.patch(
|
|
'nova.virt.driver.load_compute_driver') as load_driver:
|
|
load_driver.return_value = old_driver
|
|
new_compute = self.start_service(
|
|
'compute', host=compute.host, cell_name=cell_name)
|
|
else:
|
|
new_compute = self.start_service(
|
|
'compute', host=compute.host, cell_name=cell_name)
|
|
|
|
return new_compute
|
|
|
|
def assertJsonEqual(self, expected, observed, message=''):
|
|
"""Asserts that 2 complex data structures are json equivalent.
|
|
|
|
We use data structures which serialize down to json throughout
|
|
the code, and often times we just need to know that these are
|
|
json equivalent. This means that list order is not important,
|
|
and should be sorted.
|
|
|
|
Because this is a recursive set of assertions, when failure
|
|
happens we want to expose both the local failure and the
|
|
global view of the 2 data structures being compared. So a
|
|
MismatchError which includes the inner failure as the
|
|
mismatch, and the passed in expected / observed as matchee /
|
|
matcher.
|
|
|
|
"""
|
|
if isinstance(expected, str):
|
|
expected = jsonutils.loads(expected)
|
|
if isinstance(observed, str):
|
|
observed = jsonutils.loads(observed)
|
|
|
|
def sort_key(x):
|
|
if isinstance(x, (set, list)) or isinstance(x, datetime.datetime):
|
|
return str(x)
|
|
if isinstance(x, dict):
|
|
items = ((sort_key(key), sort_key(value))
|
|
for key, value in x.items())
|
|
return sorted(items)
|
|
return x
|
|
|
|
def inner(expected, observed, path='root'):
|
|
if isinstance(expected, dict) and isinstance(observed, dict):
|
|
self.assertEqual(
|
|
len(expected), len(observed),
|
|
('path: %s. Different dict key sets\n'
|
|
'expected=%s\n'
|
|
'observed=%s\n'
|
|
'difference=%s') %
|
|
(path,
|
|
sorted(expected.keys()),
|
|
sorted(observed.keys()),
|
|
list(set(expected.keys()).symmetric_difference(
|
|
set(observed.keys())))))
|
|
expected_keys = sorted(expected)
|
|
observed_keys = sorted(observed)
|
|
self.assertEqual(
|
|
expected_keys, observed_keys,
|
|
'path: %s. Dict keys are not equal' % path)
|
|
for key in expected:
|
|
inner(expected[key], observed[key], path + '.%s' % key)
|
|
elif (isinstance(expected, (list, tuple, set)) and
|
|
isinstance(observed, (list, tuple, set))):
|
|
self.assertEqual(
|
|
len(expected), len(observed),
|
|
('path: %s. Different list items\n'
|
|
'expected=%s\n'
|
|
'observed=%s\n'
|
|
'difference=%s') %
|
|
(path,
|
|
sorted(expected, key=sort_key),
|
|
sorted(observed, key=sort_key),
|
|
[a for a in itertools.chain(expected, observed) if
|
|
(a not in expected) or (a not in observed)]))
|
|
|
|
expected_values_iter = iter(sorted(expected, key=sort_key))
|
|
observed_values_iter = iter(sorted(observed, key=sort_key))
|
|
|
|
for i in range(len(expected)):
|
|
inner(next(expected_values_iter),
|
|
next(observed_values_iter), path + '[%s]' % i)
|
|
else:
|
|
self.assertEqual(expected, observed, 'path: %s' % path)
|
|
|
|
try:
|
|
inner(expected, observed)
|
|
except testtools.matchers.MismatchError as e:
|
|
difference = e.mismatch.describe()
|
|
if message:
|
|
message = 'message: %s\n' % message
|
|
msg = "\nexpected:\n%s\nactual:\n%s\ndifference:\n%s\n%s" % (
|
|
pprint.pformat(expected),
|
|
pprint.pformat(observed),
|
|
difference,
|
|
message)
|
|
error = AssertionError(msg)
|
|
error.expected = expected
|
|
error.observed = observed
|
|
error.difference = difference
|
|
raise error
|
|
|
|
def assertXmlEqual(self, expected, observed, **options):
|
|
self.assertThat(observed, matchers.XMLMatches(expected, **options))
|
|
|
|
def assertPublicAPISignatures(self, baseinst, inst):
|
|
def get_public_apis(inst):
|
|
methods = {}
|
|
|
|
def findmethods(object):
|
|
return inspect.ismethod(object) or inspect.isfunction(object)
|
|
|
|
for (name, value) in inspect.getmembers(inst, findmethods):
|
|
if name.startswith("_"):
|
|
continue
|
|
methods[name] = value
|
|
return methods
|
|
|
|
baseclass = baseinst.__class__.__name__
|
|
basemethods = get_public_apis(baseinst)
|
|
implmethods = get_public_apis(inst)
|
|
|
|
extranames = []
|
|
for name in sorted(implmethods.keys()):
|
|
if name not in basemethods:
|
|
extranames.append(name)
|
|
|
|
self.assertEqual([], extranames,
|
|
"public APIs not listed in base class %s" %
|
|
baseclass)
|
|
|
|
for name in sorted(implmethods.keys()):
|
|
# NOTE(stephenfin): We ignore type annotations
|
|
baseargs = inspect.getfullargspec(basemethods[name])[:-1]
|
|
implargs = inspect.getfullargspec(implmethods[name])[:-1]
|
|
|
|
self.assertEqual(baseargs, implargs,
|
|
"%s args don't match base class %s" %
|
|
(name, baseclass))
|
|
|
|
|
|
class APICoverage(object):
|
|
|
|
cover_api = None
|
|
|
|
def test_api_methods(self):
|
|
self.assertIsNotNone(self.cover_api)
|
|
api_methods = [x for x in dir(self.cover_api)
|
|
if not x.startswith('_')]
|
|
test_methods = [x[5:] for x in dir(self)
|
|
if x.startswith('test_')]
|
|
self.assertThat(
|
|
test_methods,
|
|
testtools.matchers.ContainsAll(api_methods))
|
|
|
|
|
|
class SubclassSignatureTestCase(testtools.TestCase, metaclass=abc.ABCMeta):
|
|
"""Ensure all overridden methods of all subclasses of the class
|
|
under test exactly match the signature of the base class.
|
|
|
|
A subclass of SubclassSignatureTestCase should define a method
|
|
_get_base_class which:
|
|
|
|
* Returns a base class whose subclasses will all be checked
|
|
* Ensures that all subclasses to be tested have been imported
|
|
|
|
SubclassSignatureTestCase defines a single test, test_signatures,
|
|
which does a recursive, depth-first check of all subclasses, ensuring
|
|
that their method signatures are identical to those of the base class.
|
|
"""
|
|
@abc.abstractmethod
|
|
def _get_base_class(self):
|
|
raise NotImplementedError()
|
|
|
|
def setUp(self):
|
|
self.useFixture(nova_fixtures.ConfFixture(CONF))
|
|
self.base = self._get_base_class()
|
|
|
|
super(SubclassSignatureTestCase, self).setUp()
|
|
|
|
@staticmethod
|
|
def _get_argspecs(cls):
|
|
"""Return a dict of method_name->argspec for every method of cls."""
|
|
argspecs = {}
|
|
|
|
# getmembers returns all members, including members inherited from
|
|
# the base class. It's redundant for us to test these, but as
|
|
# they'll always pass it's not worth the complexity to filter them out.
|
|
for (name, method) in inspect.getmembers(cls, inspect.ismethod):
|
|
# Subclass __init__ methods can usually be legitimately different
|
|
if name == '__init__':
|
|
continue
|
|
|
|
while hasattr(method, '__wrapped__'):
|
|
# This is a wrapped function. The signature we're going to
|
|
# see here is that of the wrapper, which is almost certainly
|
|
# going to involve varargs and kwargs, and therefore is
|
|
# unlikely to be what we want. If the wrapper manupulates the
|
|
# arguments taken by the wrapped function, the wrapped function
|
|
# isn't what we want either. In that case we're just stumped:
|
|
# if it ever comes up, add more knobs here to work round it (or
|
|
# stop using a dynamic language).
|
|
#
|
|
# Here we assume the wrapper doesn't manipulate the arguments
|
|
# to the wrapped function and inspect the wrapped function
|
|
# instead.
|
|
method = getattr(method, '__wrapped__')
|
|
|
|
argspecs[name] = inspect.getfullargspec(method)
|
|
|
|
return argspecs
|
|
|
|
@staticmethod
|
|
def _clsname(cls):
|
|
"""Return the fully qualified name of cls."""
|
|
return "%s.%s" % (cls.__module__, cls.__name__)
|
|
|
|
def _test_signatures_recurse(self, base, base_argspecs):
|
|
for sub in base.__subclasses__():
|
|
sub_argspecs = self._get_argspecs(sub)
|
|
|
|
# Check that each subclass method matches the signature of the
|
|
# base class
|
|
for (method, sub_argspec) in sub_argspecs.items():
|
|
# Methods which don't override methods in the base class
|
|
# are good.
|
|
if method in base_argspecs:
|
|
self.assertEqual(base_argspecs[method], sub_argspec,
|
|
'Signature of %(sub)s.%(method)s '
|
|
'differs from superclass %(base)s' %
|
|
{'base': self._clsname(base),
|
|
'sub': self._clsname(sub),
|
|
'method': method})
|
|
|
|
# Recursively check this subclass
|
|
self._test_signatures_recurse(sub, sub_argspecs)
|
|
|
|
def test_signatures(self):
|
|
self._test_signatures_recurse(self.base, self._get_argspecs(self.base))
|
|
|
|
|
|
class TimeOverride(fixtures.Fixture):
|
|
"""Fixture to start and remove time override."""
|
|
|
|
def __init__(self, override_time=None):
|
|
self.override_time = override_time
|
|
|
|
def setUp(self):
|
|
super(TimeOverride, self).setUp()
|
|
timeutils.set_time_override(override_time=self.override_time)
|
|
self.addCleanup(timeutils.clear_time_override)
|
|
|
|
|
|
class NoDBTestCase(TestCase):
|
|
"""`NoDBTestCase` differs from TestCase in that DB access is not supported.
|
|
This makes tests run significantly faster. If possible, all new tests
|
|
should derive from this class.
|
|
"""
|
|
USES_DB = False
|
|
|
|
|
|
class MatchType(object):
|
|
"""Matches any instance of a specified type
|
|
|
|
The MatchType class is a helper for use with the mock.assert_called_with()
|
|
method that lets you assert that a particular parameter has a specific data
|
|
type. It enables stricter checking than the built in mock.ANY helper.
|
|
|
|
Example usage could be:
|
|
|
|
mock_some_method.assert_called_once_with(
|
|
"hello",
|
|
MatchType(objects.Instance),
|
|
mock.ANY,
|
|
"world",
|
|
MatchType(objects.KeyPair))
|
|
"""
|
|
|
|
def __init__(self, wanttype):
|
|
self.wanttype = wanttype
|
|
|
|
def __eq__(self, other):
|
|
return type(other) == self.wanttype
|
|
|
|
def __ne__(self, other):
|
|
return type(other) != self.wanttype
|
|
|
|
def __repr__(self):
|
|
return "<MatchType:" + str(self.wanttype) + ">"
|
|
|
|
|
|
class MatchObjPrims(object):
|
|
"""Matches objects with equal primitives."""
|
|
|
|
def __init__(self, want_obj):
|
|
self.want_obj = want_obj
|
|
|
|
def __eq__(self, other):
|
|
return objects_base.obj_equal_prims(other, self.want_obj)
|
|
|
|
def __ne__(self, other):
|
|
return not other == self.want_obj
|
|
|
|
def __repr__(self):
|
|
return '<MatchObjPrims:' + str(self.want_obj) + '>'
|
|
|
|
|
|
class ContainKeyValue(object):
|
|
"""Checks whether a key/value pair is in a dict parameter.
|
|
|
|
The ContainKeyValue class is a helper for use with the mock.assert_*()
|
|
method that lets you assert that a particular dict contain a key/value
|
|
pair. It enables stricter checking than the built in mock.ANY helper.
|
|
|
|
Example usage could be:
|
|
|
|
mock_some_method.assert_called_once_with(
|
|
"hello",
|
|
ContainKeyValue('foo', bar),
|
|
mock.ANY,
|
|
"world",
|
|
ContainKeyValue('hello', world))
|
|
"""
|
|
|
|
def __init__(self, wantkey, wantvalue):
|
|
self.wantkey = wantkey
|
|
self.wantvalue = wantvalue
|
|
|
|
def __eq__(self, other):
|
|
try:
|
|
return other[self.wantkey] == self.wantvalue
|
|
except (KeyError, TypeError):
|
|
return False
|
|
|
|
def __ne__(self, other):
|
|
try:
|
|
return other[self.wantkey] != self.wantvalue
|
|
except (KeyError, TypeError):
|
|
return True
|
|
|
|
def __repr__(self):
|
|
return "<ContainKeyValue: key " + str(self.wantkey) + \
|
|
" and value " + str(self.wantvalue) + ">"
|
|
|
|
|
|
@contextlib.contextmanager
|
|
def patch_exists(patched_path, result, other=None):
|
|
"""Selectively patch os.path.exists() so that if it's called with
|
|
patched_path, return result. Calls with any other path are passed
|
|
through to the real os.path.exists() function if other is not provided.
|
|
If other is provided then that will be the result of the call on paths
|
|
other than patched_path.
|
|
|
|
Either import and use as a decorator / context manager, or use the
|
|
nova.TestCase.patch_exists() static method as a context manager.
|
|
|
|
Currently it is *not* recommended to use this if any of the
|
|
following apply:
|
|
|
|
- You want to patch via decorator *and* make assertions about how the
|
|
mock is called (since using it in the decorator form will not make
|
|
the mock available to your code).
|
|
|
|
- You want the result of the patched exists() call to be determined
|
|
programmatically (e.g. by matching substrings of patched_path).
|
|
|
|
- You expect exists() to be called multiple times on the same path
|
|
and return different values each time.
|
|
|
|
Additionally within unit tests which only test a very limited code
|
|
path, it may be possible to ensure that the code path only invokes
|
|
exists() once, in which case it's slightly overkill to do
|
|
selective patching based on the path. In this case something like
|
|
like this may be more appropriate:
|
|
|
|
@mock.patch('os.path.exists', return_value=True)
|
|
def test_my_code(self, mock_exists):
|
|
...
|
|
mock_exists.assert_called_once_with(path)
|
|
"""
|
|
real_exists = os.path.exists
|
|
|
|
def fake_exists(path):
|
|
if path == patched_path:
|
|
return result
|
|
elif other is not None:
|
|
return other
|
|
else:
|
|
return real_exists(path)
|
|
|
|
with mock.patch.object(os.path, "exists") as mock_exists:
|
|
mock_exists.side_effect = fake_exists
|
|
yield mock_exists
|
|
|
|
|
|
@contextlib.contextmanager
|
|
def patch_open(patched_path, read_data):
|
|
"""Selectively patch open() so that if it's called with patched_path,
|
|
return a mock which makes it look like the file contains
|
|
read_data. Calls with any other path are passed through to the
|
|
real open() function.
|
|
|
|
Either import and use as a decorator, or use the
|
|
nova.TestCase.patch_open() static method as a context manager.
|
|
|
|
Currently it is *not* recommended to use this if any of the
|
|
following apply:
|
|
|
|
- The code under test will attempt to write to patched_path.
|
|
|
|
- You want to patch via decorator *and* make assertions about how the
|
|
mock is called (since using it in the decorator form will not make
|
|
the mock available to your code).
|
|
|
|
- You want the faked file contents to be determined
|
|
programmatically (e.g. by matching substrings of patched_path).
|
|
|
|
- You expect open() to be called multiple times on the same path
|
|
and return different file contents each time.
|
|
|
|
Additionally within unit tests which only test a very limited code
|
|
path, it may be possible to ensure that the code path only invokes
|
|
open() once, in which case it's slightly overkill to do
|
|
selective patching based on the path. In this case something like
|
|
like this may be more appropriate:
|
|
|
|
@mock.patch('builtins.open')
|
|
def test_my_code(self, mock_open):
|
|
...
|
|
mock_open.assert_called_once_with(path)
|
|
"""
|
|
real_open = builtins.open
|
|
m = mock.mock_open(read_data=read_data)
|
|
|
|
def selective_fake_open(path, *args, **kwargs):
|
|
if path == patched_path:
|
|
return m(patched_path)
|
|
return real_open(path, *args, **kwargs)
|
|
|
|
with mock.patch('builtins.open') as mock_open:
|
|
mock_open.side_effect = selective_fake_open
|
|
yield m
|