from teuthology import contextutil
from teuthology import exceptions
from teuthology.orchestra import run
-import tasks.ceph_client as cclient
+from tasks import ceph_client as cclient
from teuthology.orchestra.daemon import DaemonGroup
from tasks.daemonwatchdog import DaemonWatchdog
restore the tweaked option at the /end/ of 'tasks' block.
"""
if config.get('mon-health-to-clog', 'true') == 'false':
- saved_options = {}
cluster = config.get('cluster', 'ceph')
manager = ctx.managers[cluster]
manager.raw_cluster_command(
time.sleep(4)
for i in range(3):
umount_dev = "{d}1".format(d=devs[i])
- r = remote.run(args=['sudo', 'umount', run.Raw(umount_dev)])
+ remote.run(args=['sudo', 'umount', run.Raw(umount_dev)])
cmd = 'purge ' + nodename
execute_cdeploy(admin, cmd, path)
cmd = 'purgedata ' + nodename
from teuthology.orchestra import run
from teuthology.exceptions import CommandFailedError
from tasks.thrasher import Thrasher
-from six import StringIO
try:
from subprocess import DEVNULL # py3k
# this is for cephadm clusters
def shell(ctx, cluster_name, remote, args, name=None, **kwargs):
- testdir = teuthology.get_testdir(ctx)
extra_args = []
if name:
extra_args = ['-n', name]
try:
info = remote.sh(cmd, wait=True)
except CommandFailedError as e:
- log.error("Failure of --op info command with {ret}".
- format(e.exitstatus))
+ log.error("Failure of --op info command with %s",
+ e.exitstatus)
ERRORS += 1
continue
if not str(pg) in info:
- log.error("Bad data from info: {info}".format(info=info))
+ log.error("Bad data from info: %s", info)
ERRORS += 1
log.info("Test pg logging")
def _shell(ctx, cluster_name, remote, args, extra_cephadm_args=[], **kwargs):
- testdir = teuthology.get_testdir(ctx)
+ teuthology.get_testdir(ctx)
return remote.run(
args=[
'sudo',
@contextlib.contextmanager
def ceph_clients(ctx, config):
cluster_name = config['cluster']
- testdir = teuthology.get_testdir(ctx)
log.info('Setting up client nodes...')
clients = ctx.cluster.only(teuthology.is_type('client', cluster_name))
- testdir = teuthology.get_testdir(ctx)
- coverage_dir = '{tdir}/archive/coverage'.format(tdir=testdir)
for remote, roles_for_host in clients.remotes.items():
for role in teuthology.cluster_roles_of_type(roles_for_host, 'client',
cluster_name):
teuthology.deep_merge(config, overrides.get('cephadm', {}))
log.info('Config: ' + str(config))
- testdir = teuthology.get_testdir(ctx)
-
# set up cluster context
if not hasattr(ctx, 'ceph'):
ctx.ceph = {}
import random
import time
-import tasks.ceph_manager
+from tasks import ceph_manager
from teuthology import misc as teuthology
log = logging.getLogger(__name__)
mon_role_name = 'mon.' + name
mds_role_name = 'mds.' + name
mgr_role_name = 'mgr.' + name
- m_osd = re.search('--id (\d+) --setuser ceph', r.stdout.getvalue())
+ m_osd = re.search('--id (\d+) --setuser ceph', ps_eaf)
if m_osd:
osd_service = 'ceph-osd@{m}.service'.format(m=m_osd.group(1))
remote.run(args=['sudo', 'systemctl', 'status',