
This adds usage of the flake8-import-order extension to our flake8 checks to enforce consistency on our import ordering to follow the overall OpenStack code guidelines. Since we have now dropped Python 2, this also cleans up a few cases for things that were third party libs but became part of the standard library such as mock, which is now a standard part of unittest. Some questions, in order of importance: Q: Are you insane? A: Potentially. Q: Why should we touch all of these files? A: This adds consistency to our imports. The extension makes sure that all imports follow our published guidelines of having imports ordered by standard lib, third party, and local. This will be a one time churn, then we can ensure consistency over time. Q: Why bother. this doesn't really matter? A: I agree - but... We have the issue that we have less people actively involved and less time to perform thorough code reviews. This will make it objective and automated to catch these kinds of issues. But part of this, even though it maybe seems a little annoying, is for making it easier for contributors. Right now, we may or may not notice if something is following the guidelines or not. And we may or may not comment in a review to ask for a contributor to make adjustments to follow the guidelines. But then further along into the review process, someone decides to be thorough, and after the contributor feels like they've had to deal with other change requests and things are in really good shape, they get a -1 on something mostly meaningless as far as the functionality of their code. It can be a frustrating and disheartening thing. I believe this actually helps avoid that by making it an objective thing that they find out right away up front - either the code is following the guidelines and everything is happy, or it's not and running local jobs or the pep8 CI job will let them know right away and they can fix it. No guessing on whether or not someone is going to take a stand on following the guidelines or not. This will also make it easier on the code reviewers. The more we can automate, the more time we can spend in code reviews making sure the logic of the change is correct and less time looking at trivial coding and style things. Q: Should we use our hacking extensions for this? A: Hacking has had to keep back linter requirements for a long time now. Current versions of the linters actually don't work with the way we've been hooking into them for our hacking checks. We will likely need to do away with those at some point so we can move on to the current linter releases. This will help ensure we have something in place when that time comes to make sure some checks are automated. Q: Didn't you spend more time on this than the benefit we'll get from it? A: Yeah, probably. Change-Id: Ic13ba238a4a45c6219f4de131cfe0366219d722f Signed-off-by: Sean McGinnis <sean.mcginnis@gmail.com>
928 lines
39 KiB
Python
928 lines
39 KiB
Python
# Copyright (C) 2015 Tom Barron <tpb@dyncloud.net>
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
"""Tests for Backup NFS driver."""
|
|
|
|
import bz2
|
|
import filecmp
|
|
import hashlib
|
|
import os
|
|
import shutil
|
|
import stat
|
|
import tempfile
|
|
import threading
|
|
from unittest import mock
|
|
import zlib
|
|
|
|
import ddt
|
|
from eventlet import tpool
|
|
from os_brick import exception as brick_exception
|
|
from os_brick.remotefs import remotefs as remotefs_brick
|
|
from oslo_config import cfg
|
|
import six
|
|
|
|
from cinder.backup.drivers import nfs
|
|
from cinder import context
|
|
from cinder import db
|
|
from cinder import exception
|
|
from cinder.i18n import _
|
|
from cinder import objects
|
|
from cinder import test
|
|
from cinder.tests.unit import fake_constants as fake
|
|
|
|
CONF = cfg.CONF
|
|
|
|
FAKE_BACKUP_MOUNT_POINT_BASE = '/fake/mount-point-base'
|
|
FAKE_HOST = 'fake_host'
|
|
FAKE_EXPORT_PATH = 'fake/export/path'
|
|
FAKE_BACKUP_SHARE = '%s:/%s' % (FAKE_HOST, FAKE_EXPORT_PATH)
|
|
FAKE_BACKUP_PATH = os.path.join(FAKE_BACKUP_MOUNT_POINT_BASE,
|
|
FAKE_EXPORT_PATH)
|
|
FAKE_BACKUP_ID = fake.BACKUP_ID
|
|
FAKE_BACKUP_ID_PART1 = fake.BACKUP_ID[:2]
|
|
FAKE_BACKUP_ID_PART2 = fake.BACKUP_ID[2:4]
|
|
FAKE_BACKUP_ID_REST = fake.BACKUP_ID[4:]
|
|
UPDATED_CONTAINER_NAME = os.path.join(FAKE_BACKUP_ID_PART1,
|
|
FAKE_BACKUP_ID_PART2,
|
|
FAKE_BACKUP_ID)
|
|
FAKE_EGID = 1234
|
|
|
|
|
|
@ddt.ddt
|
|
class BackupNFSShareTestCase(test.TestCase):
|
|
|
|
def setUp(self):
|
|
super(BackupNFSShareTestCase, self).setUp()
|
|
self.ctxt = context.get_admin_context()
|
|
self.mock_object(nfs, 'LOG')
|
|
# Note(yikun): It mocks out the backup notifier to avoid to leak
|
|
# notifications into other test.
|
|
notify_patcher = mock.patch(
|
|
'cinder.volume.volume_utils.notify_about_backup_usage')
|
|
notify_patcher.start()
|
|
self.addCleanup(notify_patcher.stop)
|
|
|
|
def test_check_configuration_no_backup_share(self):
|
|
self.override_config('backup_share', None)
|
|
self.mock_object(nfs.NFSBackupDriver, '_init_backup_repo_path',
|
|
return_value=FAKE_BACKUP_PATH)
|
|
|
|
driver = nfs.NFSBackupDriver(self.ctxt)
|
|
self.assertRaises(exception.InvalidConfigurationValue,
|
|
driver.check_for_setup_error)
|
|
|
|
@mock.patch('os.getegid', return_value=FAKE_EGID)
|
|
@mock.patch('cinder.utils.get_file_gid')
|
|
@mock.patch('cinder.utils.get_file_mode')
|
|
@ddt.data((FAKE_EGID, 0),
|
|
(FAKE_EGID, stat.S_IWGRP),
|
|
(6666, 0),
|
|
(6666, stat.S_IWGRP))
|
|
@ddt.unpack
|
|
def test_init_backup_repo_path(self,
|
|
file_gid,
|
|
file_mode,
|
|
mock_get_file_mode,
|
|
mock_get_file_gid,
|
|
mock_getegid):
|
|
self.override_config('backup_share', FAKE_BACKUP_SHARE)
|
|
self.override_config('backup_mount_point_base',
|
|
FAKE_BACKUP_MOUNT_POINT_BASE)
|
|
mock_remotefsclient = mock.Mock()
|
|
mock_remotefsclient.get_mount_point = mock.Mock(
|
|
return_value=FAKE_BACKUP_PATH)
|
|
self.mock_object(nfs.NFSBackupDriver, 'check_for_setup_error')
|
|
self.mock_object(remotefs_brick, 'RemoteFsClient',
|
|
return_value=mock_remotefsclient)
|
|
|
|
with mock.patch.object(nfs.NFSBackupDriver, '_init_backup_repo_path'):
|
|
driver = nfs.NFSBackupDriver(self.ctxt)
|
|
|
|
mock_get_file_gid.return_value = file_gid
|
|
mock_get_file_mode.return_value = file_mode
|
|
mock_execute = self.mock_object(driver, '_execute')
|
|
|
|
path = driver._init_backup_repo_path()
|
|
|
|
self.assertEqual(FAKE_BACKUP_PATH, path)
|
|
mock_remotefsclient.mount.assert_called_once_with(FAKE_BACKUP_SHARE)
|
|
mock_remotefsclient.get_mount_point.assert_called_once_with(
|
|
FAKE_BACKUP_SHARE)
|
|
|
|
mock_execute_calls = []
|
|
if file_gid != FAKE_EGID:
|
|
mock_execute_calls.append(
|
|
mock.call('chgrp',
|
|
'-R',
|
|
FAKE_EGID,
|
|
path,
|
|
root_helper=driver._root_helper,
|
|
run_as_root=True))
|
|
|
|
if not (file_mode & stat.S_IWGRP):
|
|
mock_execute_calls.append(
|
|
mock.call('chmod',
|
|
'-R',
|
|
'g+w',
|
|
path,
|
|
root_helper=driver._root_helper,
|
|
run_as_root=True))
|
|
|
|
mock_execute.assert_has_calls(mock_execute_calls, any_order=True)
|
|
self.assertEqual(len(mock_execute_calls), mock_execute.call_count)
|
|
|
|
def test_init_backup_repo_path_unconfigured(self):
|
|
"""RemoteFsClient is not created if backup_share unset"""
|
|
self.override_config('backup_share', None)
|
|
|
|
mock_remotefsclient = mock.Mock()
|
|
self.mock_object(remotefs_brick, 'RemoteFsClient')
|
|
|
|
driver = nfs.NFSBackupDriver(self.ctxt)
|
|
driver._init_backup_repo_path()
|
|
|
|
self.assertEqual(0, mock_remotefsclient.call_count)
|
|
|
|
@mock.patch('time.sleep')
|
|
def test_init_backup_repo_path_mount_retry(self, mock_sleep):
|
|
self.override_config('backup_share', FAKE_BACKUP_SHARE)
|
|
self.override_config('backup_mount_attempts', 2)
|
|
|
|
mock_remotefsclient = mock.Mock()
|
|
self.mock_object(remotefs_brick, 'RemoteFsClient',
|
|
return_value=mock_remotefsclient)
|
|
|
|
mock_remotefsclient.mount.side_effect = [
|
|
brick_exception.BrickException] * 2
|
|
with mock.patch.object(nfs.NFSBackupDriver, '_init_backup_repo_path'):
|
|
driver = nfs.NFSBackupDriver(self.ctxt)
|
|
|
|
self.assertRaises(brick_exception.BrickException,
|
|
driver._init_backup_repo_path)
|
|
self.assertEqual([mock.call(FAKE_BACKUP_SHARE),
|
|
mock.call(FAKE_BACKUP_SHARE)],
|
|
mock_remotefsclient.mount.call_args_list)
|
|
|
|
|
|
def fake_md5(arg):
|
|
class result(object):
|
|
def hexdigest(self):
|
|
return 'fake-md5-sum'
|
|
|
|
ret = result()
|
|
return ret
|
|
|
|
|
|
class BackupNFSTestCase(test.TestCase):
|
|
"""Test Cases for NFS backup driver."""
|
|
|
|
_DEFAULT_VOLUME_ID = fake.VOLUME_ID
|
|
|
|
def _create_volume_db_entry(self, volume_id=_DEFAULT_VOLUME_ID):
|
|
vol = {'id': volume_id,
|
|
'size': 1,
|
|
'status': 'available',
|
|
'volume_type_id': self.vt['id']}
|
|
return db.volume_create(self.ctxt, vol)['id']
|
|
|
|
def _create_backup_db_entry(self,
|
|
volume_id=_DEFAULT_VOLUME_ID,
|
|
container='test-container',
|
|
backup_id=fake.BACKUP_ID,
|
|
parent_id=None,
|
|
status=None):
|
|
|
|
try:
|
|
db.volume_get(self.ctxt, volume_id)
|
|
except exception.NotFound:
|
|
self._create_volume_db_entry(volume_id=volume_id)
|
|
|
|
backup = {'id': backup_id,
|
|
'size': 1,
|
|
'container': container,
|
|
'volume_id': volume_id,
|
|
'parent_id': parent_id,
|
|
'user_id': fake.USER_ID,
|
|
'project_id': fake.PROJECT_ID,
|
|
'status': status,
|
|
}
|
|
return db.backup_create(self.ctxt, backup)['id']
|
|
|
|
def _write_effective_compression_file(self, data_size):
|
|
"""Ensure file contents can be effectively compressed."""
|
|
self.volume_file.seek(0)
|
|
self.volume_file.write(bytes([65] * data_size))
|
|
self.volume_file.seek(0)
|
|
|
|
def _store_thread(self, *args, **kwargs):
|
|
self.thread_dict['thread'] = threading.current_thread()
|
|
return self.thread_original_method(*args, **kwargs)
|
|
|
|
def setUp(self):
|
|
super(BackupNFSTestCase, self).setUp()
|
|
|
|
self.ctxt = context.get_admin_context()
|
|
self.mock_object(hashlib, 'md5', fake_md5)
|
|
self.volume_file = tempfile.NamedTemporaryFile()
|
|
self.temp_dir = tempfile.mkdtemp()
|
|
self.addCleanup(self.volume_file.close)
|
|
self.override_config('backup_share', FAKE_BACKUP_SHARE)
|
|
self.override_config('backup_mount_point_base',
|
|
FAKE_BACKUP_MOUNT_POINT_BASE)
|
|
self.override_config('backup_file_size', 52428800)
|
|
self.mock_object(nfs.NFSBackupDriver, '_init_backup_repo_path',
|
|
return_value=self.temp_dir)
|
|
# Remove tempdir.
|
|
self.addCleanup(shutil.rmtree, self.temp_dir)
|
|
self.size_volume_file = 0
|
|
for _i in range(0, 32):
|
|
self.volume_file.write(os.urandom(1024))
|
|
self.size_volume_file += 1024
|
|
|
|
# Use dictionary to share data between threads
|
|
self.thread_dict = {}
|
|
|
|
# Note(yikun): It mocks out the backup notifier to avoid to leak
|
|
# notifications into other test.
|
|
notify_patcher = mock.patch(
|
|
'cinder.volume.volume_utils.notify_about_backup_usage')
|
|
notify_patcher.start()
|
|
self.addCleanup(notify_patcher.stop)
|
|
|
|
def test_backup_uncompressed(self):
|
|
volume_id = fake.VOLUME_ID
|
|
self._create_backup_db_entry(volume_id=volume_id)
|
|
self.flags(backup_compression_algorithm='none')
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
|
|
def test_backup_bz2(self):
|
|
volume_id = fake.VOLUME_ID
|
|
self._create_backup_db_entry(volume_id=volume_id)
|
|
self.flags(backup_compression_algorithm='bz2')
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self._write_effective_compression_file(self.size_volume_file)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
|
|
def test_backup_zlib(self):
|
|
volume_id = fake.VOLUME_ID
|
|
self._create_backup_db_entry(volume_id=volume_id)
|
|
self.flags(backup_compression_algorithm='zlib')
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self._write_effective_compression_file(self.size_volume_file)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
|
|
def test_backup_default_container(self):
|
|
volume_id = fake.VOLUME_ID
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=None,
|
|
backup_id=FAKE_BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, FAKE_BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
backup = objects.Backup.get_by_id(self.ctxt, FAKE_BACKUP_ID)
|
|
self.assertEqual(backup['container'], UPDATED_CONTAINER_NAME)
|
|
|
|
def test_backup_cancel(self):
|
|
"""Test the backup abort mechanism when backup is force deleted."""
|
|
count = set()
|
|
|
|
def my_refresh():
|
|
# This refresh method will abort the backup after 1 chunk
|
|
count.add(len(count) + 1)
|
|
if len(count) == 2:
|
|
backup.destroy()
|
|
original_refresh()
|
|
|
|
volume_id = fake.VOLUME_ID
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=None,
|
|
backup_id=FAKE_BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, FAKE_BACKUP_ID)
|
|
original_refresh = backup.refresh
|
|
|
|
# We cannot mock refresh method in backup object directly because
|
|
# mock will raise AttributeError on context manager exit.
|
|
with mock.patch('cinder.objects.base.CinderPersistentObject.refresh',
|
|
side_effect=my_refresh), \
|
|
mock.patch.object(service, 'delete_object',
|
|
side_effect=service.delete_object) as delete:
|
|
# Driver shouldn't raise the NotFound exception
|
|
service.backup(backup, self.volume_file)
|
|
|
|
# Ensure we called the delete_backup method when abort is detected
|
|
self.assertEqual(1, delete.call_count)
|
|
|
|
@mock.patch('cinder.backup.drivers.posix.PosixBackupDriver.'
|
|
'update_container_name',
|
|
return_value='testcontainer1')
|
|
@mock.patch('cinder.backup.drivers.nfs.NFSBackupDriver.'
|
|
'_send_progress_end')
|
|
@mock.patch('cinder.backup.drivers.nfs.NFSBackupDriver.'
|
|
'_send_progress_notification')
|
|
def test_backup_container_notify_1(self, _send_progress,
|
|
_send_progress_end,
|
|
_mock_update_container_name):
|
|
# This unit test writes data to disk. It should be
|
|
# updated to not do that.
|
|
|
|
volume_id = fake.VOLUME_ID
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container='testcontainer1')
|
|
|
|
# If the backup_object_number_per_notification is set to 1,
|
|
# the _send_progress method will be called for sure.
|
|
_send_progress.reset_mock()
|
|
_send_progress_end.reset_mock()
|
|
CONF.set_override("backup_object_number_per_notification", 1)
|
|
CONF.set_override("backup_enable_progress_timer", False)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
self.assertTrue(_send_progress.called)
|
|
self.assertTrue(_send_progress_end.called)
|
|
|
|
@mock.patch('cinder.backup.drivers.posix.PosixBackupDriver.'
|
|
'update_container_name',
|
|
return_value='testcontainer2')
|
|
@mock.patch('cinder.backup.drivers.nfs.NFSBackupDriver.'
|
|
'_send_progress_end')
|
|
@mock.patch('cinder.backup.drivers.nfs.NFSBackupDriver.'
|
|
'_send_progress_notification')
|
|
def test_backup_container_notify_2(self, _send_progress,
|
|
_send_progress_end,
|
|
_mock_update_container_name):
|
|
# This unit test writes data to disk. It should be
|
|
# updated to not do that.
|
|
|
|
volume_id = fake.VOLUME_ID
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container='testcontainer2')
|
|
|
|
# If the backup_object_number_per_notification is increased to
|
|
# another value, the _send_progress method will not be called.
|
|
_send_progress.reset_mock()
|
|
_send_progress_end.reset_mock()
|
|
CONF.set_override("backup_object_number_per_notification", 10)
|
|
CONF.set_override("backup_enable_progress_timer", False)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
self.assertFalse(_send_progress.called)
|
|
self.assertTrue(_send_progress_end.called)
|
|
|
|
@mock.patch('cinder.backup.drivers.posix.PosixBackupDriver.'
|
|
'update_container_name',
|
|
return_value='testcontainer3')
|
|
@mock.patch('cinder.backup.drivers.nfs.NFSBackupDriver.'
|
|
'_send_progress_end')
|
|
@mock.patch('cinder.backup.drivers.nfs.NFSBackupDriver.'
|
|
'_send_progress_notification')
|
|
def test_backup_container_notify_3(self, _send_progress,
|
|
_send_progress_end,
|
|
_mock_update_container_name):
|
|
# This unit test writes data to disk. It should be
|
|
# updated to not do that.
|
|
|
|
volume_id = fake.VOLUME_ID
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container='testcontainer3')
|
|
|
|
# If the timer is enabled, the _send_progress will be called,
|
|
# since the timer can trigger the progress notification.
|
|
_send_progress.reset_mock()
|
|
_send_progress_end.reset_mock()
|
|
CONF.set_override("backup_object_number_per_notification", 10)
|
|
CONF.set_override("backup_enable_progress_timer", True)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
self.assertTrue(_send_progress.called)
|
|
self.assertTrue(_send_progress_end.called)
|
|
|
|
def test_backup_custom_container(self):
|
|
volume_id = fake.VOLUME_ID
|
|
container_name = 'fake99'
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
self.assertEqual(backup['container'], container_name)
|
|
|
|
def test_backup_shafile(self):
|
|
volume_id = fake.VOLUME_ID
|
|
|
|
def _fake_generate_object_name_prefix(self, backup):
|
|
az = 'az_fake'
|
|
backup_name = '%s_backup_%s' % (az, backup['id'])
|
|
volume = 'volume_%s' % (backup['volume_id'])
|
|
prefix = volume + '_' + backup_name
|
|
return prefix
|
|
|
|
self.mock_object(nfs.NFSBackupDriver,
|
|
'_generate_object_name_prefix',
|
|
_fake_generate_object_name_prefix)
|
|
|
|
container_name = self.temp_dir.replace(tempfile.gettempdir() + '/',
|
|
'', 1)
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
self.assertEqual(backup['container'], container_name)
|
|
|
|
# Verify sha contents
|
|
content1 = service._read_sha256file(backup)
|
|
self.assertEqual(32 * 1024 / content1['chunk_size'],
|
|
len(content1['sha256s']))
|
|
|
|
def test_backup_cmp_shafiles(self):
|
|
volume_id = fake.VOLUME_ID
|
|
|
|
def _fake_generate_object_name_prefix(self, backup):
|
|
az = 'az_fake'
|
|
backup_name = '%s_backup_%s' % (az, backup['id'])
|
|
volume = 'volume_%s' % (backup['volume_id'])
|
|
prefix = volume + '_' + backup_name
|
|
return prefix
|
|
|
|
self.mock_object(nfs.NFSBackupDriver,
|
|
'_generate_object_name_prefix',
|
|
_fake_generate_object_name_prefix)
|
|
|
|
container_name = self.temp_dir.replace(tempfile.gettempdir() + '/',
|
|
'', 1)
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
self.assertEqual(backup['container'], container_name)
|
|
|
|
# Create incremental backup with no change to contents
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP2_ID,
|
|
parent_id=fake.BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
service.backup(deltabackup, self.volume_file)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
self.assertEqual(deltabackup['container'], container_name)
|
|
|
|
# Compare shas from both files
|
|
content1 = service._read_sha256file(backup)
|
|
content2 = service._read_sha256file(deltabackup)
|
|
|
|
self.assertEqual(len(content1['sha256s']), len(content2['sha256s']))
|
|
self.assertEqual(set(content1['sha256s']), set(content2['sha256s']))
|
|
|
|
def test_backup_delta_two_objects_change(self):
|
|
volume_id = fake.VOLUME_ID
|
|
|
|
def _fake_generate_object_name_prefix(self, backup):
|
|
az = 'az_fake'
|
|
backup_name = '%s_backup_%s' % (az, backup['id'])
|
|
volume = 'volume_%s' % (backup['volume_id'])
|
|
prefix = volume + '_' + backup_name
|
|
return prefix
|
|
|
|
self.mock_object(nfs.NFSBackupDriver,
|
|
'_generate_object_name_prefix',
|
|
_fake_generate_object_name_prefix)
|
|
|
|
self.flags(backup_file_size=(8 * 1024))
|
|
self.flags(backup_sha_block_size_bytes=1024)
|
|
|
|
container_name = self.temp_dir.replace(tempfile.gettempdir() + '/',
|
|
'', 1)
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
self.assertEqual(backup['container'], container_name)
|
|
|
|
# Create incremental backup with no change to contents
|
|
self.volume_file.seek(16 * 1024)
|
|
self.volume_file.write(os.urandom(1024))
|
|
self.volume_file.seek(20 * 1024)
|
|
self.volume_file.write(os.urandom(1024))
|
|
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP2_ID,
|
|
parent_id=fake.BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
service.backup(deltabackup, self.volume_file)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
self.assertEqual(deltabackup['container'], container_name)
|
|
|
|
content1 = service._read_sha256file(backup)
|
|
content2 = service._read_sha256file(deltabackup)
|
|
|
|
# Verify that two shas are changed at index 16 and 20
|
|
self.assertNotEqual(content1['sha256s'][16], content2['sha256s'][16])
|
|
self.assertNotEqual(content1['sha256s'][20], content2['sha256s'][20])
|
|
|
|
def test_backup_delta_two_blocks_in_object_change(self):
|
|
volume_id = fake.VOLUME_ID
|
|
|
|
def _fake_generate_object_name_prefix(self, backup):
|
|
az = 'az_fake'
|
|
backup_name = '%s_backup_%s' % (az, backup['id'])
|
|
volume = 'volume_%s' % (backup['volume_id'])
|
|
prefix = volume + '_' + backup_name
|
|
return prefix
|
|
|
|
self.mock_object(nfs.NFSBackupDriver,
|
|
'_generate_object_name_prefix',
|
|
_fake_generate_object_name_prefix)
|
|
|
|
self.flags(backup_file_size=(8 * 1024))
|
|
self.flags(backup_sha_block_size_bytes=1024)
|
|
|
|
container_name = self.temp_dir.replace(tempfile.gettempdir() + '/',
|
|
'', 1)
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
self.assertEqual(backup['container'], container_name)
|
|
|
|
# Create incremental backup with no change to contents
|
|
self.volume_file.seek(16 * 1024)
|
|
self.volume_file.write(os.urandom(1024))
|
|
self.volume_file.seek(20 * 1024)
|
|
self.volume_file.write(os.urandom(1024))
|
|
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP2_ID,
|
|
parent_id=fake.BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
service.backup(deltabackup, self.volume_file)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
self.assertEqual(deltabackup['container'], container_name)
|
|
|
|
# Verify that two shas are changed at index 16 and 20
|
|
content1 = service._read_sha256file(backup)
|
|
content2 = service._read_sha256file(deltabackup)
|
|
self.assertNotEqual(content1['sha256s'][16], content2['sha256s'][16])
|
|
self.assertNotEqual(content1['sha256s'][20], content2['sha256s'][20])
|
|
|
|
def test_backup_backup_metadata_fail(self):
|
|
"""Test of when an exception occurs in backup().
|
|
|
|
In backup(), after an exception occurs in
|
|
self._backup_metadata(), we want to check the process of an
|
|
exception handler.
|
|
"""
|
|
volume_id = fake.VOLUME_ID
|
|
|
|
self._create_backup_db_entry(volume_id=volume_id)
|
|
self.flags(backup_compression_algorithm='none')
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
|
|
def fake_backup_metadata(self, backup, object_meta):
|
|
raise exception.BackupDriverException(reason=_('fake'))
|
|
|
|
# Raise a pseudo exception.BackupDriverException.
|
|
self.mock_object(nfs.NFSBackupDriver, '_backup_metadata',
|
|
fake_backup_metadata)
|
|
|
|
# We expect that an exception be notified directly.
|
|
self.assertRaises(exception.BackupDriverException,
|
|
service.backup,
|
|
backup, self.volume_file)
|
|
|
|
def test_backup_backup_metadata_fail2(self):
|
|
"""Test of when an exception occurs in an exception handler.
|
|
|
|
In backup(), after an exception occurs in
|
|
self._backup_metadata(), we want to check the process when the
|
|
second exception occurs in self.delete_backup().
|
|
"""
|
|
volume_id = fake.VOLUME_ID
|
|
|
|
self._create_backup_db_entry(volume_id=volume_id)
|
|
self.flags(backup_compression_algorithm='none')
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
|
|
def fake_backup_metadata(self, backup, object_meta):
|
|
raise exception.BackupDriverException(reason=_('fake'))
|
|
|
|
# Raise a pseudo exception.BackupDriverException.
|
|
self.mock_object(nfs.NFSBackupDriver, '_backup_metadata',
|
|
fake_backup_metadata)
|
|
|
|
def fake_delete(self, backup):
|
|
raise exception.BackupOperationError()
|
|
|
|
# Raise a pseudo exception.BackupOperationError.
|
|
self.mock_object(nfs.NFSBackupDriver, 'delete_backup', fake_delete)
|
|
|
|
# We expect that the second exception is notified.
|
|
self.assertRaises(exception.BackupOperationError,
|
|
service.backup,
|
|
backup, self.volume_file)
|
|
|
|
def test_restore_uncompressed(self):
|
|
volume_id = fake.VOLUME_ID
|
|
|
|
self._create_backup_db_entry(volume_id=volume_id)
|
|
self.flags(backup_compression_algorithm='none')
|
|
self.flags(backup_sha_block_size_bytes=32)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
|
|
service.backup(backup, self.volume_file)
|
|
|
|
with tempfile.NamedTemporaryFile() as restored_file:
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
backup.status = objects.fields.BackupStatus.RESTORING
|
|
backup.save()
|
|
service.restore(backup, volume_id, restored_file)
|
|
self.assertTrue(filecmp.cmp(self.volume_file.name,
|
|
restored_file.name))
|
|
|
|
def test_restore_bz2(self):
|
|
self.thread_original_method = bz2.decompress
|
|
volume_id = fake.VOLUME_ID
|
|
self.mock_object(bz2, 'decompress', side_effect=self._store_thread)
|
|
|
|
self._create_backup_db_entry(volume_id=volume_id)
|
|
self.flags(backup_compression_algorithm='bz2')
|
|
file_size = 1024 * 3
|
|
self.flags(backup_file_size=file_size)
|
|
self.flags(backup_sha_block_size_bytes=1024)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self._write_effective_compression_file(file_size)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
|
|
with tempfile.NamedTemporaryFile() as restored_file:
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
backup.status = objects.fields.BackupStatus.RESTORING
|
|
backup.save()
|
|
service.restore(backup, volume_id, restored_file)
|
|
self.assertTrue(filecmp.cmp(self.volume_file.name,
|
|
restored_file.name))
|
|
|
|
self.assertNotEqual(threading.current_thread(),
|
|
self.thread_dict['thread'])
|
|
|
|
def test_restore_zlib(self):
|
|
self.thread_original_method = zlib.decompress
|
|
self.mock_object(zlib, 'decompress', side_effect=self._store_thread)
|
|
volume_id = fake.VOLUME_ID
|
|
|
|
self._create_backup_db_entry(volume_id=volume_id)
|
|
self.flags(backup_compression_algorithm='zlib')
|
|
file_size = 1024 * 3
|
|
self.flags(backup_file_size=file_size)
|
|
self.flags(backup_sha_block_size_bytes=1024)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self._write_effective_compression_file(file_size)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
backup.status = objects.fields.BackupStatus.RESTORING
|
|
backup.save()
|
|
service.backup(backup, self.volume_file)
|
|
|
|
with tempfile.NamedTemporaryFile() as restored_file:
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.restore(backup, volume_id, restored_file)
|
|
self.assertTrue(filecmp.cmp(self.volume_file.name,
|
|
restored_file.name))
|
|
|
|
self.assertNotEqual(threading.current_thread(),
|
|
self.thread_dict['thread'])
|
|
|
|
def test_restore_abort_delta(self):
|
|
volume_id = fake.VOLUME_ID
|
|
count = set()
|
|
|
|
def _fake_generate_object_name_prefix(self, backup):
|
|
az = 'az_fake'
|
|
backup_name = '%s_backup_%s' % (az, backup['id'])
|
|
volume = 'volume_%s' % (backup['volume_id'])
|
|
prefix = volume + '_' + backup_name
|
|
return prefix
|
|
|
|
def my_refresh():
|
|
# This refresh method will abort the backup after 1 chunk
|
|
count.add(len(count) + 1)
|
|
if len(count) == 2:
|
|
backup.status = objects.fields.BackupStatus.AVAILABLE
|
|
backup.save()
|
|
original_refresh()
|
|
|
|
self.mock_object(nfs.NFSBackupDriver,
|
|
'_generate_object_name_prefix',
|
|
_fake_generate_object_name_prefix)
|
|
|
|
self.flags(backup_file_size=(1024 * 8))
|
|
self.flags(backup_sha_block_size_bytes=1024)
|
|
|
|
container_name = self.temp_dir.replace(tempfile.gettempdir() + '/',
|
|
'', 1)
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
|
|
# Create incremental backup with no change to contents
|
|
self.volume_file.seek(16 * 1024)
|
|
self.volume_file.write(os.urandom(1024))
|
|
self.volume_file.seek(20 * 1024)
|
|
self.volume_file.write(os.urandom(1024))
|
|
|
|
self._create_backup_db_entry(
|
|
volume_id=volume_id,
|
|
status=objects.fields.BackupStatus.RESTORING,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP2_ID,
|
|
parent_id=fake.BACKUP_ID)
|
|
self.volume_file.seek(0)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
service.backup(deltabackup, self.volume_file, True)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
original_refresh = backup.refresh
|
|
|
|
with tempfile.NamedTemporaryFile() as restored_file, \
|
|
mock.patch('cinder.objects.Backup.refresh',
|
|
side_effect=my_refresh):
|
|
|
|
self.assertRaises(exception.BackupRestoreCancel,
|
|
service.restore, backup, volume_id,
|
|
restored_file)
|
|
|
|
def test_restore_delta(self):
|
|
volume_id = fake.VOLUME_ID
|
|
|
|
def _fake_generate_object_name_prefix(self, backup):
|
|
az = 'az_fake'
|
|
backup_name = '%s_backup_%s' % (az, backup['id'])
|
|
volume = 'volume_%s' % (backup['volume_id'])
|
|
prefix = volume + '_' + backup_name
|
|
return prefix
|
|
|
|
self.mock_object(nfs.NFSBackupDriver,
|
|
'_generate_object_name_prefix',
|
|
_fake_generate_object_name_prefix)
|
|
|
|
self.flags(backup_file_size=(1024 * 8))
|
|
self.flags(backup_sha_block_size_bytes=1024)
|
|
|
|
container_name = self.temp_dir.replace(tempfile.gettempdir() + '/',
|
|
'', 1)
|
|
self._create_backup_db_entry(volume_id=volume_id,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP_ID)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
self.volume_file.seek(0)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.backup(backup, self.volume_file)
|
|
|
|
# Create incremental backup with no change to contents
|
|
self.volume_file.seek(16 * 1024)
|
|
self.volume_file.write(os.urandom(1024))
|
|
self.volume_file.seek(20 * 1024)
|
|
self.volume_file.write(os.urandom(1024))
|
|
|
|
self._create_backup_db_entry(
|
|
volume_id=volume_id,
|
|
status=objects.fields.BackupStatus.RESTORING,
|
|
container=container_name,
|
|
backup_id=fake.BACKUP2_ID,
|
|
parent_id=fake.BACKUP_ID)
|
|
self.volume_file.seek(0)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
service.backup(deltabackup, self.volume_file, True)
|
|
deltabackup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
|
|
with tempfile.NamedTemporaryFile() as restored_file:
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP2_ID)
|
|
service.restore(backup, volume_id,
|
|
restored_file)
|
|
self.assertTrue(filecmp.cmp(self.volume_file.name,
|
|
restored_file.name))
|
|
|
|
def test_delete(self):
|
|
volume_id = fake.VOLUME_ID
|
|
self._create_backup_db_entry(volume_id=volume_id)
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
backup = objects.Backup.get_by_id(self.ctxt, fake.BACKUP_ID)
|
|
service.delete_backup(backup)
|
|
|
|
def test_get_compressor(self):
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
compressor = service._get_compressor('None')
|
|
self.assertIsNone(compressor)
|
|
compressor = service._get_compressor('zlib')
|
|
self.assertEqual(compressor, zlib)
|
|
self.assertIsInstance(compressor, tpool.Proxy)
|
|
compressor = service._get_compressor('bz2')
|
|
self.assertEqual(compressor, bz2)
|
|
self.assertIsInstance(compressor, tpool.Proxy)
|
|
self.assertRaises(ValueError, service._get_compressor, 'fake')
|
|
|
|
def create_buffer(self, size):
|
|
# Set up buffer of zeroed bytes
|
|
fake_data = bytearray(size)
|
|
if six.PY2:
|
|
# On Python 2, zlib.compressor() accepts buffer, but not bytearray
|
|
# NOTE(jsbryant): Pep8 fails on py3 based installations as buffer()
|
|
# was removed. 'noqa' used here to avoid that failure.
|
|
fake_data = buffer(fake_data) # noqa
|
|
return fake_data
|
|
|
|
def test_prepare_output_data_effective_compression(self):
|
|
"""Test compression works on a native thread."""
|
|
self.thread_original_method = zlib.compress
|
|
self.mock_object(zlib, 'compress', side_effect=self._store_thread)
|
|
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
fake_data = self.create_buffer(128)
|
|
|
|
result = service._prepare_output_data(fake_data)
|
|
|
|
self.assertEqual('zlib', result[0])
|
|
self.assertGreater(len(fake_data), len(result[1]))
|
|
self.assertNotEqual(threading.current_thread(),
|
|
self.thread_dict['thread'])
|
|
|
|
def test_prepare_output_data_no_compresssion(self):
|
|
self.flags(backup_compression_algorithm='none')
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
fake_data = self.create_buffer(128)
|
|
|
|
result = service._prepare_output_data(fake_data)
|
|
|
|
self.assertEqual('none', result[0])
|
|
self.assertEqual(fake_data, result[1])
|
|
|
|
def test_prepare_output_data_ineffective_compression(self):
|
|
service = nfs.NFSBackupDriver(self.ctxt)
|
|
fake_data = self.create_buffer(128)
|
|
|
|
# Pre-compress so that compression in the driver will be ineffective.
|
|
already_compressed_data = service.compressor.compress(fake_data)
|
|
|
|
result = service._prepare_output_data(already_compressed_data)
|
|
|
|
self.assertEqual('none', result[0])
|
|
self.assertEqual(already_compressed_data, result[1])
|