rgw s3 auth use sts: true
rgw lc debug interval: 10
rgw_restore_debug_interval: 20
+ rgw_restore_processor_period: 10
rgw:
storage classes:
LUKEWARM:
extra_attrs: ["cloud_restore"]
lc_debug_interval: 10
rgw_restore_debug_interval: 20
+ rgw_restore_processor_period: 10
lifecycle_tests: True
cloudtier_tests: True
if rgw_restore_debug_interval:
s3tests_conf['s3 main']['rgw_restore_debug_interval'] = rgw_restore_debug_interval
+ rgw_restore_processor_period = properties.get('rgw_restore_processor_period')
+ if rgw_restore_processor_period:
+ s3tests_conf['s3 main']['rgw_restore_processor_period'] = rgw_restore_processor_period
+
if ctx.rgw_cloudtier is not None:
log.info(' ctx.rgw_cloudtier config is %s ...', ctx.rgw_cloudtier.config)
client_rgw_config = ctx.rgw_cloudtier.config.get(client)
ret = process_restore_entry(entry, y);
if (!ret && entry.status == rgw::sal::RGWRestoreStatus::RestoreAlreadyInProgress) {
- r_entries.push_back(entry);
+ r_entries.push_back(entry);
ldpp_dout(this, 20) << __PRETTY_FUNCTION__ << ": re-pushing entry: '" << entry
<< "' on shard:"
<< obj_names[index] << dendl;
}
- if (ret < 0)
+ // Skip the entry of object/bucket which no longer exists
+ if (ret < 0 && (ret != -ENOENT))
goto done;
///process all entries, trim and re-add