Pass logger instances to AccountBroker/ContainerBroker

There are some processes which don't pass their logger instances
when creating AccountBrokers/ContainerBrokers, which causes some
error messages with a different setting from the other logs
generated by the processes.

This patch makes them pass logger instances, to make sure they
generate all logs according to their log settings.

Change-Id: I914c3a2811e1a2b7f19ad2bc9b3d042fcba63820
This commit is contained in:
Takashi Kajinami 2016-03-22 23:29:48 +09:00 committed by Takashi Kajinami
parent 89eced960c
commit 4b2ac606b1
10 changed files with 57 additions and 44 deletions

View File

@ -135,7 +135,7 @@ class AccountAuditor(Daemon):
""" """
start_time = time.time() start_time = time.time()
try: try:
broker = AccountBroker(path) broker = AccountBroker(path, logger=self.logger)
if not broker.is_deleted(): if not broker.is_deleted():
self.validate_per_policy_counts(broker) self.validate_per_policy_counts(broker)
self.logger.increment('passes') self.logger.increment('passes')

View File

@ -193,7 +193,8 @@ class AccountReaper(Daemon):
elif fname.endswith('.db'): elif fname.endswith('.db'):
self.start_time = time() self.start_time = time()
broker = \ broker = \
AccountBroker(os.path.join(hsh_path, fname)) AccountBroker(os.path.join(hsh_path, fname),
logger=self.logger)
if broker.is_status_deleted() and \ if broker.is_status_deleted() and \
not broker.empty(): not broker.empty():
self.reap_account( self.reap_account(

View File

@ -860,7 +860,8 @@ class ReplicatorRpc(object):
mkdirs(os.path.join(self.root, drive, 'tmp')) mkdirs(os.path.join(self.root, drive, 'tmp'))
if not self._db_file_exists(db_file): if not self._db_file_exists(db_file):
return HTTPNotFound() return HTTPNotFound()
return getattr(self, op)(self.broker_class(db_file), args) return getattr(self, op)(
self.broker_class(db_file, logger=self.logger), args)
@contextmanager @contextmanager
def debug_timing(self, name): def debug_timing(self, name):
@ -970,7 +971,7 @@ class ReplicatorRpc(object):
return HTTPNotFound() return HTTPNotFound()
if not os.path.exists(old_filename): if not os.path.exists(old_filename):
return HTTPNotFound() return HTTPNotFound()
broker = self.broker_class(old_filename) broker = self.broker_class(old_filename, logger=self.logger)
broker.newid(args[0]) broker.newid(args[0])
renamer(old_filename, db_file) renamer(old_filename, db_file)
return HTTPNoContent() return HTTPNoContent()
@ -987,8 +988,8 @@ class ReplicatorRpc(object):
tmp_filename = os.path.join(self.root, drive, 'tmp', args[0]) tmp_filename = os.path.join(self.root, drive, 'tmp', args[0])
if self._abort_rsync_then_merge(db_file, tmp_filename): if self._abort_rsync_then_merge(db_file, tmp_filename):
return HTTPNotFound() return HTTPNotFound()
new_broker = self.broker_class(tmp_filename) new_broker = self.broker_class(tmp_filename, logger=self.logger)
existing_broker = self.broker_class(db_file) existing_broker = self.broker_class(db_file, logger=self.logger)
db_file = existing_broker.db_file db_file = existing_broker.db_file
point = -1 point = -1
objects = existing_broker.get_items_since(point, 1000) objects = existing_broker.get_items_since(point, 1000)

View File

@ -111,7 +111,7 @@ class ContainerAuditor(Daemon):
""" """
start_time = time.time() start_time = time.time()
try: try:
broker = ContainerBroker(path) broker = ContainerBroker(path, logger=self.logger)
if not broker.is_deleted(): if not broker.is_deleted():
broker.get_info() broker.get_info()
self.logger.increment('passes') self.logger.increment('passes')

View File

@ -370,7 +370,7 @@ class ContainerReplicatorRpc(db_replicator.ReplicatorRpc):
# if the local db has started sharding since the original 'sync' # if the local db has started sharding since the original 'sync'
# request then abort object replication now; instantiate a fresh broker # request then abort object replication now; instantiate a fresh broker
# each time this check if performed so to get latest state # each time this check if performed so to get latest state
broker = ContainerBroker(db_file) broker = ContainerBroker(db_file, logger=self.logger)
return broker.sharding_initiated() return broker.sharding_initiated()
def _post_rsync_then_merge_hook(self, existing_broker, new_broker): def _post_rsync_then_merge_hook(self, existing_broker, new_broker):

View File

@ -334,7 +334,7 @@ class ContainerSync(Daemon):
""" """
broker = None broker = None
try: try:
broker = ContainerBroker(path) broker = ContainerBroker(path, logger=self.logger)
# The path we pass to the ContainerBroker is a real path of # The path we pass to the ContainerBroker is a real path of
# a container DB. If we get here, however, it means that this # a container DB. If we get here, however, it means that this
# path is linked from the sync_containers dir. In rare cases # path is linked from the sync_containers dir. In rare cases

View File

@ -33,10 +33,11 @@ from test.unit.account.test_backend import (
class FakeAccountBroker(object): class FakeAccountBroker(object):
def __init__(self, path): def __init__(self, path, logger):
self.path = path self.path = path
self.db_file = path self.db_file = path
self.file = os.path.basename(path) self.file = os.path.basename(path)
self.logger = logger
def is_deleted(self): def is_deleted(self):
return False return False

View File

@ -75,7 +75,7 @@ class FakeBroker(object):
class FakeAccountBroker(object): class FakeAccountBroker(object):
def __init__(self, containers): def __init__(self, containers, logger=None):
self.containers = containers self.containers = containers
self.containers_yielded = [] self.containers_yielded = []

View File

@ -29,7 +29,7 @@ from test.unit.container import test_backend
class FakeContainerBroker(object): class FakeContainerBroker(object):
def __init__(self, path): def __init__(self, path, logger):
self.path = path self.path = path
self.db_file = path self.db_file = path
self.file = os.path.basename(path) self.file = os.path.basename(path)

View File

@ -48,7 +48,7 @@ class FakeRing(object):
class FakeContainerBroker(object): class FakeContainerBroker(object):
def __init__(self, path, metadata=None, info=None, deleted=False, def __init__(self, path, metadata=None, info=None, deleted=False,
items_since=None): items_since=None, logger=None):
self.db_file = path self.db_file = path
self.db_dir = os.path.dirname(path) self.db_dir = os.path.dirname(path)
self.metadata = metadata if metadata else {} self.metadata = metadata if metadata else {}
@ -189,9 +189,9 @@ class TestContainerSync(unittest.TestCase):
mock.patch('swift.container.sync.sleep', fake_sleep), \ mock.patch('swift.container.sync.sleep', fake_sleep), \
mock.patch(gen_func) as fake_generator, \ mock.patch(gen_func) as fake_generator, \
mock.patch('swift.container.sync.ContainerBroker', mock.patch('swift.container.sync.ContainerBroker',
lambda p: FakeContainerBroker(p, info={ lambda p, logger: FakeContainerBroker(p, info={
'account': 'a', 'container': 'c', 'account': 'a', 'container': 'c',
'storage_policy_index': 0})): 'storage_policy_index': 0}, logger=logger)):
fake_generator.side_effect = [iter(['container.db']), fake_generator.side_effect = [iter(['container.db']),
iter(['container.db'])] iter(['container.db'])]
cs = sync.ContainerSync({}, container_ring=FakeRing()) cs = sync.ContainerSync({}, container_ring=FakeRing())
@ -235,9 +235,9 @@ class TestContainerSync(unittest.TestCase):
mock.patch('swift.container.sync.time', fake_time), \ mock.patch('swift.container.sync.time', fake_time), \
mock.patch(gen_func) as fake_generator, \ mock.patch(gen_func) as fake_generator, \
mock.patch('swift.container.sync.ContainerBroker', mock.patch('swift.container.sync.ContainerBroker',
lambda p: FakeContainerBroker(p, info={ lambda p, logger: FakeContainerBroker(p, info={
'account': 'a', 'container': 'c', 'account': 'a', 'container': 'c',
'storage_policy_index': 0})): 'storage_policy_index': 0}, logger=logger)):
fake_generator.side_effect = [iter(['container.db']), fake_generator.side_effect = [iter(['container.db']),
iter(['container.db'])] iter(['container.db'])]
cs = sync.ContainerSync({}, container_ring=FakeRing()) cs = sync.ContainerSync({}, container_ring=FakeRing())
@ -337,9 +337,10 @@ class TestContainerSync(unittest.TestCase):
self.assertEqual(['10.0.0.0'], cs._myips) self.assertEqual(['10.0.0.0'], cs._myips)
orig_ContainerBroker = sync.ContainerBroker orig_ContainerBroker = sync.ContainerBroker
try: try:
sync.ContainerBroker = lambda p: FakeContainerBroker( sync.ContainerBroker = lambda p, logger: FakeContainerBroker(
p, info={'account': 'a', 'container': 'c', p, info={'account': 'a', 'container': 'c',
'storage_policy_index': 0}) 'storage_policy_index': 0},
logger=logger)
cs._myips = ['127.0.0.1'] # No match cs._myips = ['127.0.0.1'] # No match
cs._myport = 1 # No match cs._myport = 1 # No match
cs.container_sync('isa.db') cs.container_sync('isa.db')
@ -370,9 +371,10 @@ class TestContainerSync(unittest.TestCase):
cs = sync.ContainerSync({}, container_ring=cring) cs = sync.ContainerSync({}, container_ring=cring)
orig_ContainerBroker = sync.ContainerBroker orig_ContainerBroker = sync.ContainerBroker
try: try:
sync.ContainerBroker = lambda p: FakeContainerBroker( sync.ContainerBroker = lambda p, logger: FakeContainerBroker(
p, info={'account': 'a', 'container': 'c', p, info={'account': 'a', 'container': 'c',
'storage_policy_index': 0}, deleted=False) 'storage_policy_index': 0}, deleted=False,
logger=logger)
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
# This complete match will cause the 1 container failure since the # This complete match will cause the 1 container failure since the
@ -380,9 +382,10 @@ class TestContainerSync(unittest.TestCase):
cs.container_sync('isa.db') cs.container_sync('isa.db')
self.assertEqual(cs.container_failures, 1) self.assertEqual(cs.container_failures, 1)
sync.ContainerBroker = lambda p: FakeContainerBroker( sync.ContainerBroker = lambda p, logger: FakeContainerBroker(
p, info={'account': 'a', 'container': 'c', p, info={'account': 'a', 'container': 'c',
'storage_policy_index': 0}, deleted=True) 'storage_policy_index': 0}, deleted=True,
logger=logger)
# This complete match will not cause any more container failures # This complete match will not cause any more container failures
# since the broker indicates deletion # since the broker indicates deletion
cs.container_sync('isa.db') cs.container_sync('isa.db')
@ -396,11 +399,12 @@ class TestContainerSync(unittest.TestCase):
cs = sync.ContainerSync({}, container_ring=cring) cs = sync.ContainerSync({}, container_ring=cring)
orig_ContainerBroker = sync.ContainerBroker orig_ContainerBroker = sync.ContainerBroker
try: try:
sync.ContainerBroker = lambda p: FakeContainerBroker( sync.ContainerBroker = lambda p, logger: FakeContainerBroker(
p, info={'account': 'a', 'container': 'c', p, info={'account': 'a', 'container': 'c',
'storage_policy_index': 0, 'storage_policy_index': 0,
'x_container_sync_point1': -1, 'x_container_sync_point1': -1,
'x_container_sync_point2': -1}) 'x_container_sync_point2': -1},
logger=logger)
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
# This complete match will be skipped since the broker's metadata # This complete match will be skipped since the broker's metadata
@ -409,12 +413,13 @@ class TestContainerSync(unittest.TestCase):
self.assertEqual(cs.container_failures, 0) self.assertEqual(cs.container_failures, 0)
self.assertEqual(cs.container_skips, 1) self.assertEqual(cs.container_skips, 1)
sync.ContainerBroker = lambda p: FakeContainerBroker( sync.ContainerBroker = lambda p, logger: FakeContainerBroker(
p, info={'account': 'a', 'container': 'c', p, info={'account': 'a', 'container': 'c',
'storage_policy_index': 0, 'storage_policy_index': 0,
'x_container_sync_point1': -1, 'x_container_sync_point1': -1,
'x_container_sync_point2': -1}, 'x_container_sync_point2': -1},
metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1)}) metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1)},
logger=logger)
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
# This complete match will be skipped since the broker's metadata # This complete match will be skipped since the broker's metadata
@ -423,12 +428,13 @@ class TestContainerSync(unittest.TestCase):
self.assertEqual(cs.container_failures, 0) self.assertEqual(cs.container_failures, 0)
self.assertEqual(cs.container_skips, 2) self.assertEqual(cs.container_skips, 2)
sync.ContainerBroker = lambda p: FakeContainerBroker( sync.ContainerBroker = lambda p, logger: FakeContainerBroker(
p, info={'account': 'a', 'container': 'c', p, info={'account': 'a', 'container': 'c',
'storage_policy_index': 0, 'storage_policy_index': 0,
'x_container_sync_point1': -1, 'x_container_sync_point1': -1,
'x_container_sync_point2': -1}, 'x_container_sync_point2': -1},
metadata={'x-container-sync-key': ('key', 1)}) metadata={'x-container-sync-key': ('key', 1)},
logger=logger)
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
# This complete match will be skipped since the broker's metadata # This complete match will be skipped since the broker's metadata
@ -437,13 +443,14 @@ class TestContainerSync(unittest.TestCase):
self.assertEqual(cs.container_failures, 0) self.assertEqual(cs.container_failures, 0)
self.assertEqual(cs.container_skips, 3) self.assertEqual(cs.container_skips, 3)
sync.ContainerBroker = lambda p: FakeContainerBroker( sync.ContainerBroker = lambda p, logger: FakeContainerBroker(
p, info={'account': 'a', 'container': 'c', p, info={'account': 'a', 'container': 'c',
'storage_policy_index': 0, 'storage_policy_index': 0,
'x_container_sync_point1': -1, 'x_container_sync_point1': -1,
'x_container_sync_point2': -1}, 'x_container_sync_point2': -1},
metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1), metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1),
'x-container-sync-key': ('key', 1)}) 'x-container-sync-key': ('key', 1)},
logger=logger)
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
cs.allowed_sync_hosts = [] cs.allowed_sync_hosts = []
@ -453,13 +460,14 @@ class TestContainerSync(unittest.TestCase):
self.assertEqual(cs.container_failures, 1) self.assertEqual(cs.container_failures, 1)
self.assertEqual(cs.container_skips, 3) self.assertEqual(cs.container_skips, 3)
sync.ContainerBroker = lambda p: FakeContainerBroker( sync.ContainerBroker = lambda p, logger: FakeContainerBroker(
p, info={'account': 'a', 'container': 'c', p, info={'account': 'a', 'container': 'c',
'storage_policy_index': 0, 'storage_policy_index': 0,
'x_container_sync_point1': -1, 'x_container_sync_point1': -1,
'x_container_sync_point2': -1}, 'x_container_sync_point2': -1},
metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1), metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1),
'x-container-sync-key': ('key', 1)}) 'x-container-sync-key': ('key', 1)},
logger=logger)
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
cs.allowed_sync_hosts = ['127.0.0.1'] cs.allowed_sync_hosts = ['127.0.0.1']
@ -478,14 +486,15 @@ class TestContainerSync(unittest.TestCase):
orig_ContainerBroker = sync.ContainerBroker orig_ContainerBroker = sync.ContainerBroker
orig_time = sync.time orig_time = sync.time
try: try:
sync.ContainerBroker = lambda p: FakeContainerBroker( sync.ContainerBroker = lambda p, logger: FakeContainerBroker(
p, info={'account': 'a', 'container': 'c', p, info={'account': 'a', 'container': 'c',
'storage_policy_index': 0, 'storage_policy_index': 0,
'x_container_sync_point1': -1, 'x_container_sync_point1': -1,
'x_container_sync_point2': -1}, 'x_container_sync_point2': -1},
metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1), metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1),
'x-container-sync-key': ('key', 1)}, 'x-container-sync-key': ('key', 1)},
items_since=['erroneous data']) items_since=['erroneous data'],
logger=logger)
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
cs.allowed_sync_hosts = ['127.0.0.1'] cs.allowed_sync_hosts = ['127.0.0.1']
@ -534,7 +543,7 @@ class TestContainerSync(unittest.TestCase):
'x-container-sync-key': ('key', 1)}, 'x-container-sync-key': ('key', 1)},
items_since=[{'ROWID': 1, 'name': 'o'}]) items_since=[{'ROWID': 1, 'name': 'o'}])
with mock.patch('swift.container.sync.ContainerBroker', with mock.patch('swift.container.sync.ContainerBroker',
lambda p: fcb), \ lambda p, logger: fcb), \
mock.patch('swift.container.sync.hash_path', fake_hash_path): mock.patch('swift.container.sync.hash_path', fake_hash_path):
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
@ -561,7 +570,7 @@ class TestContainerSync(unittest.TestCase):
('key', 1)}, ('key', 1)},
items_since=[{'ROWID': 1, 'name': 'o'}]) items_since=[{'ROWID': 1, 'name': 'o'}])
with mock.patch('swift.container.sync.ContainerBroker', with mock.patch('swift.container.sync.ContainerBroker',
lambda p: fcb), \ lambda p, logger: fcb), \
mock.patch('swift.container.sync.hash_path', fake_hash_path): mock.patch('swift.container.sync.hash_path', fake_hash_path):
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
@ -582,7 +591,8 @@ class TestContainerSync(unittest.TestCase):
metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1), metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1),
'x-container-sync-key': ('key', 1)}, 'x-container-sync-key': ('key', 1)},
items_since=[{'ROWID': 1, 'name': 'o'}]) items_since=[{'ROWID': 1, 'name': 'o'}])
with mock.patch('swift.container.sync.ContainerBroker', lambda p: fcb): with mock.patch('swift.container.sync.ContainerBroker',
lambda p, logger: fcb):
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
cs.allowed_sync_hosts = ['127.0.0.1'] cs.allowed_sync_hosts = ['127.0.0.1']
@ -607,7 +617,7 @@ class TestContainerSync(unittest.TestCase):
items_since=[{'ROWID': 1, 'name': 'o', 'created_at': '1.2', items_since=[{'ROWID': 1, 'name': 'o', 'created_at': '1.2',
'deleted': True}]) 'deleted': True}])
with mock.patch('swift.container.sync.ContainerBroker', with mock.patch('swift.container.sync.ContainerBroker',
lambda p: fcb), \ lambda p, logger: fcb), \
mock.patch('swift.container.sync.delete_object', mock.patch('swift.container.sync.delete_object',
fake_delete_object): fake_delete_object):
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
@ -631,7 +641,7 @@ class TestContainerSync(unittest.TestCase):
items_since=[{'ROWID': 1, 'name': 'o', 'created_at': '1.2', items_since=[{'ROWID': 1, 'name': 'o', 'created_at': '1.2',
'deleted': True}]) 'deleted': True}])
with mock.patch('swift.container.sync.ContainerBroker', with mock.patch('swift.container.sync.ContainerBroker',
lambda p: fcb), \ lambda p, logger: fcb), \
mock.patch('swift.container.sync.delete_object', mock.patch('swift.container.sync.delete_object',
lambda *x, **y: None): lambda *x, **y: None):
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
@ -671,7 +681,7 @@ class TestContainerSync(unittest.TestCase):
metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1), metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1),
'x-container-sync-key': ('key', 1)}, 'x-container-sync-key': ('key', 1)},
items_since=[{'ROWID': 1, 'name': 'o'}]) items_since=[{'ROWID': 1, 'name': 'o'}])
sync.ContainerBroker = lambda p: fcb sync.ContainerBroker = lambda p, logger: fcb
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
cs.allowed_sync_hosts = ['127.0.0.1'] cs.allowed_sync_hosts = ['127.0.0.1']
@ -701,7 +711,7 @@ class TestContainerSync(unittest.TestCase):
metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1), metadata={'x-container-sync-to': ('http://127.0.0.1/a/c', 1),
'x-container-sync-key': ('key', 1)}, 'x-container-sync-key': ('key', 1)},
items_since=[{'ROWID': 1, 'name': 'o'}]) items_since=[{'ROWID': 1, 'name': 'o'}])
sync.ContainerBroker = lambda p: fcb sync.ContainerBroker = lambda p, logger: fcb
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
cs.allowed_sync_hosts = ['127.0.0.1'] cs.allowed_sync_hosts = ['127.0.0.1']
@ -723,7 +733,7 @@ class TestContainerSync(unittest.TestCase):
'x-container-sync-key': ('key', 1)}, 'x-container-sync-key': ('key', 1)},
items_since=[{'ROWID': 1, 'name': 'o', 'created_at': '1.2', items_since=[{'ROWID': 1, 'name': 'o', 'created_at': '1.2',
'deleted': True}]) 'deleted': True}])
sync.ContainerBroker = lambda p: fcb sync.ContainerBroker = lambda p, logger: fcb
cs._myips = ['10.0.0.0'] # Match cs._myips = ['10.0.0.0'] # Match
cs._myport = 1000 # Match cs._myport = 1000 # Match
cs.allowed_sync_hosts = ['127.0.0.1'] cs.allowed_sync_hosts = ['127.0.0.1']
@ -779,7 +789,7 @@ class TestContainerSync(unittest.TestCase):
mock.patch('swift.container.sync.hash_path', mock.patch('swift.container.sync.hash_path',
fake_hash_path), \ fake_hash_path), \
mock.patch('swift.container.sync.ContainerBroker', mock.patch('swift.container.sync.ContainerBroker',
lambda p: fcb): lambda p, logger: fcb):
cring = FakeRing() cring = FakeRing()
cs = sync.ContainerSync({}, container_ring=cring, cs = sync.ContainerSync({}, container_ring=cring,
logger=self.logger) logger=self.logger)