Merge "Python 3: Fix usage of dict methods"
This commit is contained in:
commit
25cdd52c6a
@ -79,13 +79,11 @@ class TestGatekeeper(unittest.TestCase):
|
|||||||
'x-object-transient-sysmeta-foo': 'value'}
|
'x-object-transient-sysmeta-foo': 'value'}
|
||||||
x_timestamp_headers = {'X-Timestamp': '1455952805.719739'}
|
x_timestamp_headers = {'X-Timestamp': '1455952805.719739'}
|
||||||
|
|
||||||
forbidden_headers_out = dict(sysmeta_headers.items() +
|
forbidden_headers_out = dict(sysmeta_headers)
|
||||||
x_backend_headers.items() +
|
forbidden_headers_out.update(x_backend_headers)
|
||||||
object_transient_sysmeta_headers.items())
|
forbidden_headers_out.update(object_transient_sysmeta_headers)
|
||||||
forbidden_headers_in = dict(sysmeta_headers.items() +
|
forbidden_headers_in = dict(forbidden_headers_out)
|
||||||
x_backend_headers.items() +
|
shunted_headers_in = dict(x_timestamp_headers)
|
||||||
object_transient_sysmeta_headers.items())
|
|
||||||
shunted_headers_in = dict(x_timestamp_headers.items())
|
|
||||||
|
|
||||||
def _assertHeadersEqual(self, expected, actual):
|
def _assertHeadersEqual(self, expected, actual):
|
||||||
for key in expected:
|
for key in expected:
|
||||||
@ -126,8 +124,8 @@ class TestGatekeeper(unittest.TestCase):
|
|||||||
expected_headers.update({'X-Backend-Inbound-' + k: v
|
expected_headers.update({'X-Backend-Inbound-' + k: v
|
||||||
for k, v in self.shunted_headers_in.items()})
|
for k, v in self.shunted_headers_in.items()})
|
||||||
self._assertHeadersEqual(expected_headers, fake_app.req.headers)
|
self._assertHeadersEqual(expected_headers, fake_app.req.headers)
|
||||||
unexpected_headers = dict(self.forbidden_headers_in.items() +
|
unexpected_headers = dict(self.forbidden_headers_in)
|
||||||
self.shunted_headers_in.items())
|
unexpected_headers.update(self.shunted_headers_in)
|
||||||
self._assertHeadersAbsent(unexpected_headers, fake_app.req.headers)
|
self._assertHeadersAbsent(unexpected_headers, fake_app.req.headers)
|
||||||
|
|
||||||
def test_reserved_header_removed_inbound(self):
|
def test_reserved_header_removed_inbound(self):
|
||||||
@ -163,8 +161,8 @@ class TestGatekeeper(unittest.TestCase):
|
|||||||
app = self.get_app(fake_app, {}, shunt_inbound_x_timestamp='false')
|
app = self.get_app(fake_app, {}, shunt_inbound_x_timestamp='false')
|
||||||
resp = req.get_response(app)
|
resp = req.get_response(app)
|
||||||
self.assertEqual('200 OK', resp.status)
|
self.assertEqual('200 OK', resp.status)
|
||||||
expected_headers = dict(self.allowed_headers.items() +
|
expected_headers = dict(self.allowed_headers)
|
||||||
self.shunted_headers_in.items())
|
expected_headers.update(self.shunted_headers_in)
|
||||||
self._assertHeadersEqual(expected_headers, fake_app.req.headers)
|
self._assertHeadersEqual(expected_headers, fake_app.req.headers)
|
||||||
|
|
||||||
def test_reserved_header_shunt_bypassed_inbound(self):
|
def test_reserved_header_shunt_bypassed_inbound(self):
|
||||||
|
@ -346,7 +346,7 @@ class TestSsyncEC(TestBaseSsync):
|
|||||||
tx_tombstones['o5'][0].delete(t5)
|
tx_tombstones['o5'][0].delete(t5)
|
||||||
|
|
||||||
suffixes = set()
|
suffixes = set()
|
||||||
for diskfiles in (tx_objs.values() + tx_tombstones.values()):
|
for diskfiles in list(tx_objs.values()) + list(tx_tombstones.values()):
|
||||||
for df in diskfiles:
|
for df in diskfiles:
|
||||||
suffixes.add(os.path.basename(os.path.dirname(df._datadir)))
|
suffixes.add(os.path.basename(os.path.dirname(df._datadir)))
|
||||||
|
|
||||||
@ -536,7 +536,7 @@ class TestSsyncEC(TestBaseSsync):
|
|||||||
tx_tombstones['o5'][0].delete(t5)
|
tx_tombstones['o5'][0].delete(t5)
|
||||||
|
|
||||||
suffixes = set()
|
suffixes = set()
|
||||||
for diskfiles in (tx_objs.values() + tx_tombstones.values()):
|
for diskfiles in list(tx_objs.values()) + list(tx_tombstones.values()):
|
||||||
for df in diskfiles:
|
for df in diskfiles:
|
||||||
suffixes.add(os.path.basename(os.path.dirname(df._datadir)))
|
suffixes.add(os.path.basename(os.path.dirname(df._datadir)))
|
||||||
|
|
||||||
@ -706,7 +706,7 @@ class TestSsyncReplication(TestBaseSsync):
|
|||||||
tx_tombstones['o7'][0].delete(t7b)
|
tx_tombstones['o7'][0].delete(t7b)
|
||||||
|
|
||||||
suffixes = set()
|
suffixes = set()
|
||||||
for diskfiles in (tx_objs.values() + tx_tombstones.values()):
|
for diskfiles in list(tx_objs.values()) + list(tx_tombstones.values()):
|
||||||
for df in diskfiles:
|
for df in diskfiles:
|
||||||
suffixes.add(os.path.basename(os.path.dirname(df._datadir)))
|
suffixes.add(os.path.basename(os.path.dirname(df._datadir)))
|
||||||
|
|
||||||
@ -879,7 +879,7 @@ class TestSsyncReplication(TestBaseSsync):
|
|||||||
rx_tombstones['o7'][0].delete(next(self.ts_iter))
|
rx_tombstones['o7'][0].delete(next(self.ts_iter))
|
||||||
|
|
||||||
suffixes = set()
|
suffixes = set()
|
||||||
for diskfiles in (tx_objs.values() + tx_tombstones.values()):
|
for diskfiles in list(tx_objs.values()) + list(tx_tombstones.values()):
|
||||||
for df in diskfiles:
|
for df in diskfiles:
|
||||||
suffixes.add(os.path.basename(os.path.dirname(df._datadir)))
|
suffixes.add(os.path.basename(os.path.dirname(df._datadir)))
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user