syncer = sync_cls(TestHandler.src, TestHandler.dest,
TestHandler.max_entries,
rgw_data_log_window=TestHandler.rgw_data_log_window,
- object_timeout=TestHandler.object_timeout)
+ object_sync_timeout=TestHandler.object_sync_timeout)
syncer.prepare()
syncer.sync(
TestHandler.num_workers,
meta_syncer = meta_cls(src, dest, args.max_entries)
data_syncer = data_cls(src, dest, args.max_entries,
rgw_data_log_window=args.rgw_data_log_window,
- object_timeout=args.object_sync_timeout)
+ object_sync_timeout=args.object_sync_timeout)
# fetch logs first since data logs need to wait before becoming usable
# due to rgw's window of data log updates during which the bucket index
self.worker_cls = None # filled in by subclass constructor
self.num_shards = None
self.max_entries = max_entries
- self.object_timeout = kwargs.get('object_timeout')
+ self.object_sync_timeout = kwargs.get('object_sync_timeout')
def init_num_shards(self):
if self.num_shards is not None:
self.dest,
daemon_id=self.daemon_id,
max_entries=max_entries,
- object_timeout=self.object_timeout,
+ object_sync_timeout=self.object_sync_timeout,
)
for i in xrange(num_workers)]
for process in processes:
super(DataWorker, self).__init__(*args, **kwargs)
self.type = 'data'
self.op_id = 0
- self.object_timeout = kwargs.get('object_timeout', 60 * 60 * 60)
+ self.object_sync_timeout = kwargs.get('object_sync_timeout', 60 * 60 * 60)
self.daemon_id = kwargs['daemon_id']
def sync_object(self, bucket, obj):
self.op_id += 1
local_op_id = self.local_lock_id + ':' + str(self.op_id)
try:
- until = time.time() + self.object_timeout
+ until = time.time() + self.object_sync_timeout
client.sync_object_intra_region(self.dest_conn, bucket, obj,
self.src.zone.name,
self.daemon_id,