From 745f7b48bdd67a799bd7f0ca86b6b7600b1b8594 Mon Sep 17 00:00:00 2001 From: Josh Durgin Date: Wed, 11 May 2016 12:06:01 -0700 Subject: [PATCH] tasks/resolve_stuck_peering: fix whitespace and indentation Signed-off-by: Josh Durgin --- tasks/resolve_stuck_peering.py | 143 ++++++++++++++++----------------- 1 file changed, 71 insertions(+), 72 deletions(-) diff --git a/tasks/resolve_stuck_peering.py b/tasks/resolve_stuck_peering.py index 9bef045e8ce8b..6a94468b80e4c 100644 --- a/tasks/resolve_stuck_peering.py +++ b/tasks/resolve_stuck_peering.py @@ -10,101 +10,100 @@ from util.rados import rados log = logging.getLogger(__name__) def task(ctx, config): - """ - Test handling resolve stuck peering + """ + Test handling resolve stuck peering - requires 3 osds on a single test node - """ - if config is None: - config = {} - assert isinstance(config, dict), \ - 'Resolve stuck peering only accepts a dict for config' + requires 3 osds on a single test node + """ + if config is None: + config = {} + assert isinstance(config, dict), \ + 'Resolve stuck peering only accepts a dict for config' - manager = ctx.managers['ceph'] + manager = ctx.managers['ceph'] - while len(manager.get_osd_status()['up']) < 3: - time.sleep(10) + while len(manager.get_osd_status()['up']) < 3: + time.sleep(10) - manager.wait_for_clean() + manager.wait_for_clean() - dummyfile = '/etc/fstab' - dummyfile1 = '/etc/resolv.conf' + dummyfile = '/etc/fstab' + dummyfile1 = '/etc/resolv.conf' - #create 1 PG pool - pool='foo' - log.info('creating pool foo') - manager.raw_cluster_cmd('osd', 'pool', 'create', '%s' % pool, '1') + #create 1 PG pool + pool='foo' + log.info('creating pool foo') + manager.raw_cluster_cmd('osd', 'pool', 'create', '%s' % pool, '1') - #set min_size of the pool to 1 - #so that we can continue with I/O - #when 2 osds are down - manager.set_pool_property(pool, "min_size", 1) + #set min_size of the pool to 1 + #so that we can continue with I/O + #when 2 osds are down + manager.set_pool_property(pool, "min_size", 1) - osds = [0, 1, 2] + osds = [0, 1, 2] - primary = manager.get_pg_primary('foo', 0) - log.info("primary osd is %d", primary) + primary = manager.get_pg_primary('foo', 0) + log.info("primary osd is %d", primary) - others = list(osds) - others.remove(primary) + others = list(osds) + others.remove(primary) - log.info('writing initial objects') - first_mon = teuthology.get_first_mon(ctx, config) - (mon,) = ctx.cluster.only(first_mon).remotes.iterkeys() - #create few objects - for i in range(100): - rados(ctx, mon, ['-p', 'foo', 'put', 'existing_%d' % i, dummyfile]) + log.info('writing initial objects') + first_mon = teuthology.get_first_mon(ctx, config) + (mon,) = ctx.cluster.only(first_mon).remotes.iterkeys() + #create few objects + for i in range(100): + rados(ctx, mon, ['-p', 'foo', 'put', 'existing_%d' % i, dummyfile]) - manager.wait_for_clean() + manager.wait_for_clean() - #kill other osds except primary - log.info('killing other osds except primary') - for i in others: - manager.kill_osd(i) - for i in others: - manager.mark_down_osd(i) + #kill other osds except primary + log.info('killing other osds except primary') + for i in others: + manager.kill_osd(i) + for i in others: + manager.mark_down_osd(i) - for i in range(100): - rados(ctx, mon, ['-p', 'foo', 'put', 'new_%d' % i, dummyfile1]) + for i in range(100): + rados(ctx, mon, ['-p', 'foo', 'put', 'new_%d' % i, dummyfile1]) - #kill primary osd - manager.kill_osd(primary) - manager.mark_down_osd(primary) + #kill primary osd + manager.kill_osd(primary) + manager.mark_down_osd(primary) - #revive other 2 osds - for i in others: - manager.revive_osd(i) + #revive other 2 osds + for i in others: + manager.revive_osd(i) - #make sure that pg is down - #Assuming pg number for single pg pool will start from 0 - pgnum=0 - pgstr = manager.get_pgid(pool, pgnum) - stats = manager.get_single_pg_stats(pgstr) - print stats['state'] + #make sure that pg is down + #Assuming pg number for single pg pool will start from 0 + pgnum=0 + pgstr = manager.get_pgid(pool, pgnum) + stats = manager.get_single_pg_stats(pgstr) + print stats['state'] - timeout=60 - start=time.time() + timeout=60 + start=time.time() - while 'down' not in stats['state']: - assert time.time() - start < timeout, \ - 'failed to reach down state before timeout expired' + while 'down' not in stats['state']: + assert time.time() - start < timeout, \ + 'failed to reach down state before timeout expired' - #mark primary as lost - manager.raw_cluster_cmd('osd', 'lost', '%d' % primary,\ - '--yes-i-really-mean-it') + #mark primary as lost + manager.raw_cluster_cmd('osd', 'lost', '%d' % primary,\ + '--yes-i-really-mean-it') - #expect the pg status to be active+undersized+degraded - #pg should recover and become active+clean within timeout - stats = manager.get_single_pg_stats(pgstr) - print stats['state'] + #expect the pg status to be active+undersized+degraded + #pg should recover and become active+clean within timeout + stats = manager.get_single_pg_stats(pgstr) + print stats['state'] - timeout=10 - start=time.time() - - while manager.get_num_down(): - assert time.time() - start < timeout, \ - 'failed to recover before timeout expired' + timeout=10 + start=time.time() + while manager.get_num_down(): + assert time.time() - start < timeout, \ + 'failed to recover before timeout expired' -- 2.39.5