
This is to be consistent with add_nodes_to_cluster. This change also renames the 'node' parameter to 'nodes' and improves a debug message. Change-Id: I7054358d6a62248706fb2120e1aedb82b71541cd
566 lines
21 KiB
Python
566 lines
21 KiB
Python
# Copyright 2016 Mirantis, Inc.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
import os
|
|
import re
|
|
import time
|
|
import urllib2
|
|
|
|
from devops.helpers import helpers
|
|
from fuelweb_test.helpers import os_actions
|
|
from fuelweb_test import logger
|
|
from proboscis import asserts
|
|
|
|
from stacklight_tests.helpers import remote_ops
|
|
from stacklight_tests import settings
|
|
|
|
|
|
PLUGIN_PACKAGE_RE = re.compile(r'([^/]+)-(\d+\.\d+)-(\d+\.\d+\.\d+)')
|
|
|
|
|
|
class NotFound(Exception):
|
|
pass
|
|
|
|
|
|
class TimeoutException(Exception):
|
|
pass
|
|
|
|
|
|
def get_plugin_name(filename):
|
|
"""Extract the plugin name from the package filename.
|
|
|
|
:param filename: the plugin's filename.
|
|
:type filename: str
|
|
:returns: the plugin's name or None if not found
|
|
:rtype: str
|
|
"""
|
|
m = PLUGIN_PACKAGE_RE.search(filename or '')
|
|
if m:
|
|
return m.group(1)
|
|
else:
|
|
return None
|
|
|
|
|
|
def get_plugin_version(filename):
|
|
"""Extract the plugin version from the package filename.
|
|
|
|
:param filename: the plugin's filename.
|
|
:type filename: str
|
|
:returns: the plugin's version or None if not found
|
|
:rtype: str
|
|
"""
|
|
m = PLUGIN_PACKAGE_RE.search(filename or '')
|
|
if m:
|
|
return m.group(3)
|
|
else:
|
|
return None
|
|
|
|
|
|
def get_fixture(name):
|
|
"""Return the full path to a fixture."""
|
|
path = os.path.join(os.environ.get("WORKSPACE", "./"), "fixtures", name)
|
|
if not os.path.isfile(path):
|
|
raise NotFound("File {} not found".format(path))
|
|
return path
|
|
|
|
|
|
class PluginHelper(object):
|
|
"""Class for common help functions."""
|
|
|
|
def __init__(self, env):
|
|
self.env = env
|
|
self.fuel_web = self.env.fuel_web
|
|
self._cluster_id = None
|
|
self.nailgun_client = self.fuel_web.client
|
|
self._os_conn = None
|
|
|
|
@property
|
|
def cluster_id(self):
|
|
if self._cluster_id is None:
|
|
try:
|
|
self._cluster_id = self.fuel_web.get_last_created_cluster()
|
|
except urllib2.URLError:
|
|
raise EnvironmentError("No cluster was created.")
|
|
return self._cluster_id
|
|
|
|
@cluster_id.setter
|
|
def cluster_id(self, value):
|
|
self._cluster_id = value
|
|
|
|
@property
|
|
def os_conn(self):
|
|
if self._os_conn is None:
|
|
self._os_conn = os_actions.OpenStackActions(
|
|
self.fuel_web.get_public_vip(self.cluster_id))
|
|
return self._os_conn
|
|
|
|
def prepare_plugin(self, plugin_path):
|
|
"""Upload and install plugin by path."""
|
|
self.env.admin_actions.upload_plugin(plugin=plugin_path)
|
|
self.env.admin_actions.install_plugin(
|
|
plugin_file_name=os.path.basename(plugin_path))
|
|
|
|
def activate_plugin(self, name, version, options=None):
|
|
"""Enable and configure a plugin for the cluster.
|
|
|
|
:param name: name of the plugin.
|
|
:type name: str
|
|
:param version: version of the plugin.
|
|
:type name: str
|
|
:param options: configuration of the plugin (optional).
|
|
:type options: dict
|
|
:returns: None
|
|
"""
|
|
if options is None:
|
|
options = {}
|
|
msg = "Plugin {0} isn't found.".format(name)
|
|
asserts.assert_true(
|
|
self.fuel_web.check_plugin_exists(self.cluster_id, name),
|
|
msg)
|
|
|
|
logger.info("Updating settings for plugin {0} ({1}): {2}".format(
|
|
name, version, options))
|
|
attributes = self.nailgun_client.get_cluster_attributes(
|
|
self.cluster_id)
|
|
attributes = attributes['editable'][name]
|
|
|
|
plugin_data = None
|
|
for item in attributes['metadata']['versions']:
|
|
if item['metadata']['plugin_version'] == version:
|
|
plugin_data = item
|
|
break
|
|
asserts.assert_is_not_none(
|
|
plugin_data, "Plugin {0} ({1}) is not found".format(name, version))
|
|
|
|
attributes['metadata']['enabled'] = True
|
|
for option, value in options.items():
|
|
path = option.split("/")
|
|
for p in path[:-1]:
|
|
plugin_settings = plugin_data[p]
|
|
plugin_settings[path[-1]] = value
|
|
self.nailgun_client.update_cluster_attributes(self.cluster_id, {
|
|
"editable": {name: attributes}
|
|
})
|
|
|
|
def get_plugin_vip(self, vip_name):
|
|
"""Get plugin IP."""
|
|
networks = self.nailgun_client.get_networks(self.cluster_id)
|
|
vip = networks.get('vips').get(vip_name, {}).get('ipaddr', None)
|
|
asserts.assert_is_not_none(
|
|
vip, "Failed to get the IP of {} server".format(vip_name))
|
|
|
|
logger.debug("VIP '{0}': {1}".format(vip_name, vip))
|
|
return vip
|
|
|
|
def get_all_ready_nodes(self):
|
|
return [node for node in
|
|
self.nailgun_client.list_cluster_nodes(self.cluster_id)
|
|
if node["status"] == "ready"]
|
|
|
|
def create_cluster(self, name=None, settings=None):
|
|
"""Create a cluster.
|
|
|
|
:param name: name of the cluster.
|
|
:type name: str
|
|
:param settings: optional dict containing the cluster's configuration.
|
|
:type settings: dict
|
|
:returns: the cluster's id
|
|
:rtype: str
|
|
"""
|
|
if not name:
|
|
name = self.__class__.__name__
|
|
self._cluster_id = self.env.fuel_web.create_cluster(
|
|
name=name,
|
|
settings=settings,
|
|
mode='ha_compact')
|
|
return self._cluster_id
|
|
|
|
def deploy_cluster(self, nodes_roles, verify_network=False,
|
|
update_interfaces=True, check_services=True):
|
|
"""Assign roles to nodes and deploy the cluster.
|
|
|
|
:param nodes_roles: nodes to roles mapping.
|
|
:type nodes_roles: dict
|
|
:param verify_network: whether or not network verification should be
|
|
run before the deployment (default: False).
|
|
:type verify_network: boolean
|
|
:param update_interfaces: whether or not interfaces should be updated
|
|
before the deployment (default: True).
|
|
:type update_interfaces: boolean
|
|
:param check_services: whether or not OSTF tests should run after the
|
|
deployment (default: True).
|
|
:type check_services: boolean
|
|
:returns: None
|
|
"""
|
|
self.fuel_web.update_nodes(self.cluster_id, nodes_roles,
|
|
update_interfaces=update_interfaces)
|
|
if verify_network:
|
|
self.fuel_web.verify_network(self.cluster_id)
|
|
self.fuel_web.deploy_cluster_wait(self.cluster_id,
|
|
check_services=check_services)
|
|
|
|
def run_ostf(self, *args, **kwargs):
|
|
"""Run the OpenStack health checks."""
|
|
self.fuel_web.run_ostf(self.cluster_id, *args, **kwargs)
|
|
|
|
def run_single_ostf(self, test_sets, test_name, *args, **kwargs):
|
|
"""Run a subset of the OpenStack health checks."""
|
|
self.fuel_web.run_single_ostf_test(self.cluster_id, test_sets,
|
|
test_name, *args, **kwargs)
|
|
|
|
def add_nodes_to_cluster(self, nodes, redeploy=True, check_services=False):
|
|
"""Add nodes to the cluster.
|
|
|
|
:param nodes: list of nodes with their roles.
|
|
:type: nodes: dict
|
|
:param redeploy: whether to redeploy the cluster (default: True).
|
|
:type redeploy: boolean
|
|
:param check_services: run OSTF after redeploy (default: False).
|
|
:type check_services: boolean
|
|
"""
|
|
self.fuel_web.update_nodes(
|
|
self.cluster_id,
|
|
nodes,
|
|
)
|
|
if redeploy:
|
|
self.fuel_web.deploy_cluster_wait(self.cluster_id,
|
|
check_services=check_services)
|
|
|
|
def remove_nodes_from_cluster(self, nodes, redeploy=True,
|
|
check_services=False):
|
|
"""Remove nodes from the cluster.
|
|
|
|
:param nodes: list of nodes to remove from the cluster.
|
|
:type nodes: dict
|
|
:param redeploy: whether to redeploy the cluster (default: True).
|
|
:type redeploy: boolean
|
|
:param check_services: run OSTF after redeploy (default: False).
|
|
:type check_services: boolean
|
|
"""
|
|
self.fuel_web.update_nodes(
|
|
self.cluster_id,
|
|
nodes,
|
|
pending_addition=False, pending_deletion=True,
|
|
)
|
|
if redeploy:
|
|
self.fuel_web.deploy_cluster_wait(self.cluster_id,
|
|
check_services=check_services)
|
|
|
|
def get_master_node_by_role(self, role_name, excluded_nodes_fqdns=()):
|
|
"""Return the node running as the Designated Controller (DC).
|
|
"""
|
|
nodes = self.fuel_web.get_nailgun_cluster_nodes_by_roles(
|
|
self.cluster_id, role_name)
|
|
nodes = [node for node in nodes
|
|
if node['fqdn'] not in set(excluded_nodes_fqdns)]
|
|
with self.fuel_web.get_ssh_for_nailgun_node(nodes[0]) as remote:
|
|
stdout = remote.check_call(
|
|
'pcs status cluster | grep "Current DC:"')["stdout"][0]
|
|
for node in nodes:
|
|
if node['fqdn'] in stdout:
|
|
return node
|
|
|
|
@staticmethod
|
|
def full_vip_name(vip_name):
|
|
return "".join(["vip__", vip_name])
|
|
|
|
def get_node_with_vip(self, role_name, vip, exclude_node=None):
|
|
nailgun_nodes = self.fuel_web.get_nailgun_cluster_nodes_by_roles(
|
|
self.cluster_id, role_name)
|
|
lma_nodes = self.fuel_web.get_devops_nodes_by_nailgun_nodes(
|
|
nailgun_nodes)
|
|
lma_node = None
|
|
if exclude_node:
|
|
for node in lma_nodes:
|
|
if node.name != exclude_node.name:
|
|
lma_node = node
|
|
break
|
|
else:
|
|
lma_node = lma_nodes[0]
|
|
return self.fuel_web.get_pacemaker_resource_location(
|
|
lma_node.name, vip)[0]
|
|
|
|
def wait_for_vip_migration(self, old_master, role_name, vip,
|
|
timeout=5 * 60):
|
|
logger.info('Waiting for the migration of VIP {}'.format(vip))
|
|
msg = "VIP {0} has not been migrated away from {1}".format(
|
|
vip, old_master)
|
|
helpers.wait(
|
|
lambda: old_master != self.get_node_with_vip(
|
|
role_name, vip, exclude_node=old_master),
|
|
timeout=timeout, timeout_msg=msg)
|
|
|
|
def power_off_node(self, node):
|
|
"""Power off a node.
|
|
|
|
:param node: Devops node.
|
|
:type node: devops node instance
|
|
"""
|
|
msg = 'Node {0} has not become offline after hard shutdown'.format(
|
|
node.name)
|
|
logger.info('Power off node %s', node.name)
|
|
node.destroy()
|
|
logger.info('Wait a %s node offline status', node.name)
|
|
helpers.wait(lambda: not self.fuel_web.get_nailgun_node_by_devops_node(
|
|
node)['online'], timeout=60 * 5, timeout_msg=msg)
|
|
|
|
def emulate_whole_network_disaster(self, delay_before_recover=5 * 60,
|
|
wait_become_online=True):
|
|
"""Simulate a full network outage for all nodes.
|
|
|
|
:param delay_before_recover: outage interval in seconds (default: 300).
|
|
:type delay_before_recover: int
|
|
:param wait_become_online: whether to wait for nodes to be back online.
|
|
:type wait_become_online: bool
|
|
"""
|
|
nodes = [node for node in self.env.d_env.get_nodes()
|
|
if node.driver.node_active(node)]
|
|
|
|
networks_interfaces = nodes[1].interfaces
|
|
|
|
for interface in networks_interfaces:
|
|
interface.network.block()
|
|
|
|
time.sleep(delay_before_recover)
|
|
|
|
for interface in networks_interfaces:
|
|
interface.network.unblock()
|
|
|
|
if wait_become_online:
|
|
self.fuel_web.wait_nodes_get_online_state(nodes[1:])
|
|
|
|
def uninstall_plugin(self, plugin_name, plugin_version, exit_code=0,
|
|
msg=None):
|
|
"""Remove a plugin.
|
|
|
|
:param plugin_name: plugin's name.
|
|
:type plugin_name: str
|
|
:param plugin_version: plugin's version.
|
|
:type plugin_version: str
|
|
:param exit_code: expected exit code.
|
|
:type exit_code: int
|
|
:param msg: message in case of error.
|
|
:type msg: str
|
|
"""
|
|
logger.info("Trying to uninstall {name}({version}) plugin".format(
|
|
name=plugin_name,
|
|
version=plugin_version))
|
|
msg = msg or "Plugin {0} deletion failed: exit code is {1}"
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
exec_res = remote.execute("fuel plugins --remove"
|
|
" {0}=={1}".format(plugin_name,
|
|
plugin_version))
|
|
asserts.assert_equal(
|
|
exit_code, exec_res['exit_code'],
|
|
msg.format(plugin_name, exec_res['exit_code']))
|
|
|
|
def check_plugin_cannot_be_uninstalled(self, plugin_name, plugin_version):
|
|
"""Check that the plugin cannot be uninstalled.
|
|
|
|
:param plugin_name: plugin's name.
|
|
:type plugin_name: str
|
|
:param plugin_version: plugin's version.
|
|
:type plugin_version: str
|
|
"""
|
|
self.uninstall_plugin(
|
|
plugin_name=plugin_name, plugin_version=plugin_version,
|
|
exit_code=1,
|
|
msg='{name}({version}) plugin deletion must not be allowed '
|
|
'when it is deployed'.format(name=plugin_name,
|
|
version=plugin_version))
|
|
|
|
def get_hostname_by_node_name(self, changed_node):
|
|
node = self.fuel_web.get_nailgun_node_by_base_name(changed_node)
|
|
if node is None:
|
|
raise NotFound("Nailgun node with '{}' in name not found".format(
|
|
changed_node))
|
|
return node['hostname']
|
|
|
|
def fuel_createmirror(self, option="", exit_code=0):
|
|
cmd = "fuel-createmirror {0}".format(option)
|
|
logger.info("Executing '{}' command.".format(cmd))
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
exec_res = remote.execute(cmd)
|
|
asserts.assert_equal(
|
|
exit_code, exec_res['exit_code'],
|
|
'fuel-createmirror failed: {0}'.format(exec_res['stderr']))
|
|
|
|
def replace_ubuntu_mirror_with_mos(self):
|
|
cmds = ["fuel-mirror create -P ubuntu -G mos",
|
|
"fuel-mirror apply --replace -P ubuntu -G mos"]
|
|
logger.info("Executing '{}' commands.".format('\n'.join(cmds)))
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
for cmd in cmds:
|
|
remote.check_call(cmd)
|
|
|
|
def fuel_create_repositories(self, nodes):
|
|
"""Start task to setup repositories on provided nodes
|
|
|
|
:param nodes: list of nodes to run task on them
|
|
:type nodes: list
|
|
"""
|
|
nodes_ids = [str(node['id']) for node in nodes]
|
|
cmd = (
|
|
"fuel --env {env_id} "
|
|
"node --node-id {nodes_ids} "
|
|
"--tasks setup_repositories".format(
|
|
env_id=self.cluster_id,
|
|
nodes_ids=' '.join(nodes_ids))
|
|
)
|
|
logger.info(
|
|
"Executing {cmd} command.".format(cmd=cmd))
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
remote.check_call(cmd)
|
|
|
|
def run_tasks(self, nodes, tasks=None, start=None, end=None,
|
|
timeout=10 * 60):
|
|
"""Run a set of tasks on nodes and wait for completion.
|
|
|
|
The list of tasks is provided using the 'tasks' parameter and it can
|
|
also be specified using the 'start' and/or 'end' parameters. In the
|
|
latter case, the method will compute the exact set of tasks to be
|
|
executed.
|
|
|
|
:param nodes: list of nodes that should run the tasks
|
|
:type nodes: list
|
|
:param tasks: list of tasks to run.
|
|
:param tasks: list
|
|
:param start: the task from where to start the deployment.
|
|
:param start: str
|
|
:param end: the task where to end the deployment.
|
|
:param end: str
|
|
:param timeout: number of seconds to wait for the tasks completion
|
|
(default: 600).
|
|
:param timeout: int
|
|
"""
|
|
task_ids = []
|
|
if tasks is not None:
|
|
task_ids += tasks
|
|
if start is not None or end is not None:
|
|
task_ids += [
|
|
t["id"] for t in self.nailgun_client.get_end_deployment_tasks(
|
|
self.cluster_id, end=end or '', start=start or '')]
|
|
node_ids = ",".join([str(node["id"]) for node in nodes])
|
|
logger.info("Running tasks {0} for nodes {1}".format(
|
|
",".join(task_ids), node_ids))
|
|
result = self.nailgun_client.put_deployment_tasks_for_cluster(
|
|
self.cluster_id, data=task_ids, node_id=node_ids)
|
|
self.fuel_web.assert_task_success(result, timeout=timeout)
|
|
|
|
def apply_maintenance_update(self):
|
|
"""Method applies maintenance updates on whole cluster."""
|
|
logger.info("Applying maintenance updates on master node")
|
|
self.env.admin_install_updates()
|
|
|
|
logger.info("Applying maintenance updates on slaves")
|
|
slaves_mu_script_url = (
|
|
"https://github.com/Mirantis/tools-sustaining/"
|
|
"raw/master/scripts/mos_apply_mu.py")
|
|
|
|
path_to_mu_script = "/tmp/mos_apply_mu.py"
|
|
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
remote.check_call("wget {uri} -O {path}".format(
|
|
uri=slaves_mu_script_url,
|
|
path=path_to_mu_script)
|
|
)
|
|
|
|
remote.check_call(
|
|
"python {path} "
|
|
"--env-id={identifier} "
|
|
"--user={username} "
|
|
"--pass={password} "
|
|
"--tenant={tenant_name} --update".format(
|
|
path=path_to_mu_script,
|
|
identifier=self.cluster_id,
|
|
**settings.KEYSTONE_CREDS
|
|
)
|
|
)
|
|
|
|
controllers = self.fuel_web.get_nailgun_cluster_nodes_by_roles(
|
|
self.cluster_id, roles=['controller', ])
|
|
|
|
computes = self.fuel_web.get_nailgun_cluster_nodes_by_roles(
|
|
self.cluster_id, roles=['compute', ])
|
|
|
|
logger.info("Restarting all OpenStack services")
|
|
|
|
logger.info("Restarting services on controllers")
|
|
ha_services = (
|
|
"p_heat-engine",
|
|
"p_neutron-plugin-openvswitch-agent",
|
|
"p_neutron-dhcp-agent",
|
|
"p_neutron-metadata-agent",
|
|
"p_neutron-l3-agent")
|
|
non_ha_services = (
|
|
"heat-api-cloudwatch",
|
|
"heat-api-cfn",
|
|
"heat-api",
|
|
"cinder-api",
|
|
"cinder-scheduler",
|
|
"nova-objectstore",
|
|
"nova-cert",
|
|
"nova-api",
|
|
"nova-consoleauth",
|
|
"nova-conductor",
|
|
"nova-scheduler",
|
|
"nova-novncproxy",
|
|
"neutron-server",
|
|
)
|
|
for controller in controllers:
|
|
with self.fuel_web.get_ssh_for_nailgun_node(
|
|
controller) as remote:
|
|
for service in ha_services:
|
|
remote_ops.manage_pacemaker_service(remote, service)
|
|
for service in non_ha_services:
|
|
remote_ops.manage_initctl_service(remote, service)
|
|
|
|
logger.info("Restarting services on computes")
|
|
compute_services = (
|
|
"neutron-plugin-openvswitch-agent",
|
|
"nova-compute",
|
|
)
|
|
for compute in computes:
|
|
with self.fuel_web.get_ssh_for_nailgun_node(compute) as remote:
|
|
for service in compute_services:
|
|
remote_ops.manage_initctl_service(remote, service)
|
|
|
|
@staticmethod
|
|
def check_notifications(got_list, expected_list):
|
|
for event_type in expected_list:
|
|
asserts.assert_true(
|
|
event_type in got_list, "{} event type not found in {}".format(
|
|
event_type, got_list))
|
|
|
|
@staticmethod
|
|
def wait_for_resource_status(resource_client, resource, expected_status,
|
|
timeout=180, interval=30):
|
|
start = time.time()
|
|
finish = start + timeout
|
|
while start < finish:
|
|
curr_state = resource_client.get(resource).status
|
|
if curr_state == expected_status:
|
|
return
|
|
else:
|
|
logger.debug(
|
|
"Instance is not in {} status".format(expected_status))
|
|
time.sleep(interval)
|
|
start = time.time()
|
|
raise TimeoutException("Timed out waiting to become {}".format(
|
|
expected_status))
|
|
|
|
def get_fuel_release(self):
|
|
version = self.nailgun_client.get_api_version()
|
|
return version.get('release')
|