2013-06-09 01:14:23 +08:00
|
|
|
#-*- coding:utf-8 -*-
|
2013-09-20 01:00:54 +08:00
|
|
|
# Copyright (c) 2010-2012 OpenStack Foundation
|
2010-07-12 17:03:45 -05:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
|
|
|
# implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2013-08-31 22:36:58 -04:00
|
|
|
"""Tests for swift.obj.server"""
|
2010-07-12 17:03:45 -05:00
|
|
|
|
|
|
|
import cPickle as pickle
|
Allow pre-1970 dates in If-[Un]Modified-Since
If I want to fetch an object only if it is newer than the first moon
landing, I send a GET request with header:
If-Modified-Since: Sun, 20 Jul 1969 20:18:00 UTC
Since that date is older than Swift, I expect a 2xx response. However,
I get a 412, which isn't even a valid thing to do for
If-Modified-Since; it should either be 2xx or 304. This is because of
two problems:
(a) Swift treats pre-1970 dates as invalid, and
(b) Swift returns 412 when a date is invalid instead of ignoring it.
This commit makes it so any time between datetime.datetime.min and
datetime.datetime.max is an acceptable value for If-Modified-Since and
If-Unmodified-Since. Dates outside that date range are treated as
invalid headers and thus are ignored, as RFC 2616 section 14.28
requires ("If the specified date is invalid, the header is ignored").
This only works for dates that the Python standard library can parse,
which on my machine is 01 Jan 1 to 31 Dec 9999. Eliminating those
restrictions would require implementing our own date parsing and
comparison, and that's almost certainly not worth it.
Change-Id: I4cb4903c4e5e3b6b3c9506c2cabbfbda62e82f35
2014-03-03 11:25:43 -08:00
|
|
|
import datetime
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
import operator
|
2010-07-12 17:03:45 -05:00
|
|
|
import os
|
2012-12-17 06:39:25 -05:00
|
|
|
import mock
|
2010-07-12 17:03:45 -05:00
|
|
|
import unittest
|
2013-11-07 04:45:27 +00:00
|
|
|
import math
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
import random
|
2010-07-12 17:03:45 -05:00
|
|
|
from shutil import rmtree
|
|
|
|
from StringIO import StringIO
|
Allow pre-1970 dates in If-[Un]Modified-Since
If I want to fetch an object only if it is newer than the first moon
landing, I send a GET request with header:
If-Modified-Since: Sun, 20 Jul 1969 20:18:00 UTC
Since that date is older than Swift, I expect a 2xx response. However,
I get a 412, which isn't even a valid thing to do for
If-Modified-Since; it should either be 2xx or 304. This is because of
two problems:
(a) Swift treats pre-1970 dates as invalid, and
(b) Swift returns 412 when a date is invalid instead of ignoring it.
This commit makes it so any time between datetime.datetime.min and
datetime.datetime.max is an acceptable value for If-Modified-Since and
If-Unmodified-Since. Dates outside that date range are treated as
invalid headers and thus are ignored, as RFC 2616 section 14.28
requires ("If the specified date is invalid, the header is ignored").
This only works for dates that the Python standard library can parse,
which on my machine is 01 Jan 1 to 31 Dec 9999. Eliminating those
restrictions would require implementing our own date parsing and
comparison, and that's almost certainly not worth it.
Change-Id: I4cb4903c4e5e3b6b3c9506c2cabbfbda62e82f35
2014-03-03 11:25:43 -08:00
|
|
|
from time import gmtime, strftime, time, struct_time
|
2011-01-19 14:18:37 -06:00
|
|
|
from tempfile import mkdtemp
|
2011-03-15 22:12:03 -07:00
|
|
|
from hashlib import md5
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
import itertools
|
2010-07-12 17:03:45 -05:00
|
|
|
|
2013-11-24 20:24:45 -06:00
|
|
|
from eventlet import sleep, spawn, wsgi, listen, Timeout, tpool
|
|
|
|
|
|
|
|
from nose import SkipTest
|
|
|
|
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
from test.unit import FakeLogger, debug_logger, mocked_http_conn
|
2014-05-27 01:17:13 -07:00
|
|
|
from test.unit import connect_tcp, readuntil2crlfs, patch_policies
|
2013-03-26 20:42:26 +00:00
|
|
|
from swift.obj import server as object_server
|
2013-07-17 16:32:35 -07:00
|
|
|
from swift.obj import diskfile
|
2014-05-27 01:17:13 -07:00
|
|
|
from swift.common import utils, storage_policy
|
2013-10-07 12:10:31 +00:00
|
|
|
from swift.common.utils import hash_path, mkdirs, normalize_timestamp, \
|
|
|
|
NullLogger, storage_directory, public, replication
|
2012-09-05 20:49:50 -07:00
|
|
|
from swift.common import constraints
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
from swift.common.swob import Request, HeaderKeyDict
|
2014-06-23 12:52:50 -07:00
|
|
|
from swift.common.storage_policy import POLICIES
|
2013-12-11 16:59:59 -05:00
|
|
|
from swift.common.exceptions import DiskFileDeviceUnavailable
|
2010-07-12 17:03:45 -05:00
|
|
|
|
2011-03-16 09:04:00 -07:00
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
def mock_time(*args, **kwargs):
|
|
|
|
return 5000.0
|
|
|
|
|
|
|
|
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
@patch_policies
|
2010-07-12 17:03:45 -05:00
|
|
|
class TestObjectController(unittest.TestCase):
|
2013-08-31 22:36:58 -04:00
|
|
|
"""Test swift.obj.server.ObjectController"""
|
2010-07-12 17:03:45 -05:00
|
|
|
|
|
|
|
def setUp(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
"""Set up for testing swift.object.server.ObjectController"""
|
2013-10-07 12:10:31 +00:00
|
|
|
utils.HASH_PATH_SUFFIX = 'endcap'
|
|
|
|
utils.HASH_PATH_PREFIX = 'startcap'
|
2011-01-19 14:18:37 -06:00
|
|
|
self.testdir = \
|
|
|
|
os.path.join(mkdtemp(), 'tmp_test_object_server_ObjectController')
|
2010-07-12 17:03:45 -05:00
|
|
|
conf = {'devices': self.testdir, 'mount_check': 'false'}
|
2013-11-18 11:41:58 -08:00
|
|
|
self.object_controller = object_server.ObjectController(
|
|
|
|
conf, logger=debug_logger())
|
2010-10-13 21:26:43 +00:00
|
|
|
self.object_controller.bytes_per_sync = 1
|
2013-07-17 16:32:35 -07:00
|
|
|
self._orig_tpool_exc = tpool.execute
|
|
|
|
tpool.execute = lambda f, *args, **kwargs: f(*args, **kwargs)
|
2013-12-11 16:59:59 -05:00
|
|
|
self.df_mgr = diskfile.DiskFileManager(conf,
|
|
|
|
self.object_controller.logger)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
|
|
|
def tearDown(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
"""Tear down for testing swift.object.server.ObjectController"""
|
2011-01-24 17:12:38 -08:00
|
|
|
rmtree(os.path.dirname(self.testdir))
|
2013-08-16 17:13:00 -04:00
|
|
|
tpool.execute = self._orig_tpool_exc
|
2010-07-12 17:03:45 -05:00
|
|
|
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
def _stage_tmp_dir(self, policy):
|
|
|
|
mkdirs(os.path.join(self.testdir, 'sda1',
|
|
|
|
diskfile.get_tmp_dir(int(policy))))
|
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
def check_all_api_methods(self, obj_name='o', alt_res=None):
|
|
|
|
path = '/sda1/p/a/c/%s' % obj_name
|
2013-06-09 01:14:23 +08:00
|
|
|
body = 'SPECIAL_STRING'
|
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
op_table = {
|
|
|
|
"PUT": (body, alt_res or 201, ''), # create one
|
|
|
|
"GET": ('', alt_res or 200, body), # check it
|
|
|
|
"POST": ('', alt_res or 202, ''), # update it
|
|
|
|
"HEAD": ('', alt_res or 200, ''), # head it
|
|
|
|
"DELETE": ('', alt_res or 204, '') # delete it
|
|
|
|
}
|
2013-06-09 01:14:23 +08:00
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
for method in ["PUT", "GET", "POST", "HEAD", "DELETE"]:
|
|
|
|
in_body, res, out_body = op_table[method]
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
path, environ={'REQUEST_METHOD': method},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test'})
|
|
|
|
req.body = in_body
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, res)
|
|
|
|
if out_body and (200 <= res < 300):
|
|
|
|
self.assertEqual(resp.body, out_body)
|
2013-06-09 01:14:23 +08:00
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
def test_REQUEST_SPECIAL_CHARS(self):
|
|
|
|
obj = 'special昆%20/%'
|
|
|
|
self.check_all_api_methods(obj)
|
2013-06-09 01:14:23 +08:00
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
def test_device_unavailable(self):
|
|
|
|
def raise_disk_unavail(*args, **kwargs):
|
|
|
|
raise DiskFileDeviceUnavailable()
|
2013-06-09 01:14:23 +08:00
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
self.object_controller.get_diskfile = raise_disk_unavail
|
|
|
|
self.check_all_api_methods(alt_res=507)
|
|
|
|
|
|
|
|
def test_allowed_headers(self):
|
|
|
|
dah = ['content-disposition', 'content-encoding', 'x-delete-at',
|
|
|
|
'x-object-manifest', 'x-static-large-object']
|
|
|
|
conf = {'devices': self.testdir, 'mount_check': 'false',
|
|
|
|
'allowed_headers': ','.join(['content-type'] + dah)}
|
|
|
|
self.object_controller = object_server.ObjectController(
|
|
|
|
conf, logger=debug_logger())
|
|
|
|
self.assertEqual(self.object_controller.allowed_headers, set(dah))
|
2013-06-09 01:14:23 +08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_POST_update_meta(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.POST
|
2011-03-24 13:03:49 -05:00
|
|
|
original_headers = self.object_controller.allowed_headers
|
2011-03-22 20:05:44 -05:00
|
|
|
test_headers = 'content-encoding foo bar'.split()
|
|
|
|
self.object_controller.allowed_headers = set(test_headers)
|
2010-07-12 17:03:45 -05:00
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test',
|
2011-03-22 20:05:44 -05:00
|
|
|
'Foo': 'fooheader',
|
|
|
|
'Baz': 'bazheader',
|
2010-07-12 17:03:45 -05:00
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-Two': 'Two'})
|
|
|
|
req.body = 'VERIFY'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
timestamp = normalize_timestamp(time())
|
2010-11-16 15:35:39 -08:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
2010-07-12 17:03:45 -05:00
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'X-Object-Meta-3': 'Three',
|
|
|
|
'X-Object-Meta-4': 'Four',
|
2011-03-22 18:17:47 -05:00
|
|
|
'Content-Encoding': 'gzip',
|
2011-03-22 20:05:44 -05:00
|
|
|
'Foo': 'fooheader',
|
|
|
|
'Bar': 'barheader',
|
2010-07-12 17:03:45 -05:00
|
|
|
'Content-Type': 'application/x-test'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-24 13:03:49 -05:00
|
|
|
self.assert_("X-Object-Meta-1" not in resp.headers and
|
|
|
|
"X-Object-Meta-Two" not in resp.headers and
|
|
|
|
"X-Object-Meta-3" in resp.headers and
|
|
|
|
"X-Object-Meta-4" in resp.headers and
|
|
|
|
"Foo" in resp.headers and
|
|
|
|
"Bar" in resp.headers and
|
|
|
|
"Baz" not in resp.headers and
|
2011-03-22 18:17:47 -05:00
|
|
|
"Content-Encoding" in resp.headers)
|
|
|
|
self.assertEquals(resp.headers['Content-Type'], 'application/x-test')
|
|
|
|
|
2011-04-19 15:57:44 -05:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-04-19 15:57:44 -05:00
|
|
|
self.assert_("X-Object-Meta-1" not in resp.headers and
|
|
|
|
"X-Object-Meta-Two" not in resp.headers and
|
|
|
|
"X-Object-Meta-3" in resp.headers and
|
|
|
|
"X-Object-Meta-4" in resp.headers and
|
|
|
|
"Foo" in resp.headers and
|
|
|
|
"Bar" in resp.headers and
|
|
|
|
"Baz" not in resp.headers and
|
|
|
|
"Content-Encoding" in resp.headers)
|
|
|
|
self.assertEquals(resp.headers['Content-Type'], 'application/x-test')
|
|
|
|
|
2011-03-22 18:17:47 -05:00
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-22 18:17:47 -05:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-24 13:03:49 -05:00
|
|
|
self.assert_("X-Object-Meta-3" not in resp.headers and
|
|
|
|
"X-Object-Meta-4" not in resp.headers and
|
|
|
|
"Foo" not in resp.headers and
|
|
|
|
"Bar" not in resp.headers and
|
2011-03-22 18:17:47 -05:00
|
|
|
"Content-Encoding" not in resp.headers)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.headers['Content-Type'], 'application/x-test')
|
|
|
|
|
2011-03-24 13:03:49 -05:00
|
|
|
# test defaults
|
|
|
|
self.object_controller.allowed_headers = original_headers
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test',
|
|
|
|
'Foo': 'fooheader',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Manifest': 'c/bar',
|
|
|
|
'Content-Encoding': 'gzip',
|
|
|
|
'Content-Disposition': 'bar',
|
|
|
|
})
|
|
|
|
req.body = 'VERIFY'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-24 13:03:49 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-24 13:03:49 -05:00
|
|
|
self.assert_("X-Object-Meta-1" in resp.headers and
|
|
|
|
"Foo" not in resp.headers and
|
|
|
|
"Content-Encoding" in resp.headers and
|
|
|
|
"X-Object-Manifest" in resp.headers and
|
|
|
|
"Content-Disposition" in resp.headers)
|
|
|
|
self.assertEquals(resp.headers['Content-Type'], 'application/x-test')
|
|
|
|
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'X-Object-Meta-3': 'Three',
|
|
|
|
'Foo': 'fooheader',
|
|
|
|
'Content-Type': 'application/x-test'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-24 13:03:49 -05:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-24 13:03:49 -05:00
|
|
|
self.assert_("X-Object-Meta-1" not in resp.headers and
|
|
|
|
"Foo" not in resp.headers and
|
|
|
|
"Content-Encoding" not in resp.headers and
|
|
|
|
"X-Object-Manifest" not in resp.headers and
|
|
|
|
"Content-Disposition" not in resp.headers and
|
2010-07-12 17:03:45 -05:00
|
|
|
"X-Object-Meta-3" in resp.headers)
|
|
|
|
self.assertEquals(resp.headers['Content-Type'], 'application/x-test')
|
|
|
|
|
2014-03-06 12:31:38 +05:30
|
|
|
# Test for empty metadata
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test',
|
|
|
|
'X-Object-Meta-3': ''})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 202)
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.headers["x-object-meta-3"], '')
|
|
|
|
|
2013-07-02 11:48:19 -04:00
|
|
|
def test_POST_old_timestamp(self):
|
|
|
|
ts = time()
|
|
|
|
timestamp = normalize_timestamp(ts)
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-Two': 'Two'})
|
|
|
|
req.body = 'VERIFY'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
# Same timestamp should result in 409
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'X-Object-Meta-3': 'Three',
|
|
|
|
'X-Object-Meta-4': 'Four',
|
|
|
|
'Content-Encoding': 'gzip',
|
|
|
|
'Content-Type': 'application/x-test'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 409)
|
|
|
|
|
|
|
|
# Earlier timestamp should result in 409
|
|
|
|
timestamp = normalize_timestamp(ts - 1)
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'X-Object-Meta-5': 'Five',
|
|
|
|
'X-Object-Meta-6': 'Six',
|
|
|
|
'Content-Encoding': 'gzip',
|
|
|
|
'Content-Type': 'application/x-test'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 409)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_POST_not_exist(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
2010-11-16 15:35:39 -08:00
|
|
|
req = Request.blank('/sda1/p/a/c/fail',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
2010-07-12 17:03:45 -05:00
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-2': 'Two',
|
|
|
|
'Content-Type': 'text/plain'})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
|
|
|
def test_POST_invalid_path(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c', environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-2': 'Two',
|
|
|
|
'Content-Type': 'text/plain'})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
def test_POST_no_timestamp(self):
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-2': 'Two',
|
|
|
|
'Content-Type': 'text/plain'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 400)
|
|
|
|
|
|
|
|
def test_POST_bad_timestamp(self):
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': 'bad',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-2': 'Two',
|
|
|
|
'Content-Type': 'text/plain'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 400)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_POST_container_connection(self):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def mock_http_connect(response, with_exc=False):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
class FakeConn(object):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def __init__(self, status, with_exc):
|
|
|
|
self.status = status
|
|
|
|
self.reason = 'Fake'
|
|
|
|
self.host = '1.2.3.4'
|
|
|
|
self.port = '1234'
|
|
|
|
self.with_exc = with_exc
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def getresponse(self):
|
|
|
|
if self.with_exc:
|
|
|
|
raise Exception('test')
|
|
|
|
return self
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def read(self, amt=None):
|
|
|
|
return ''
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
return lambda *args, **kwargs: FakeConn(response, with_exc)
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
old_http_connect = object_server.http_connect
|
|
|
|
try:
|
2013-07-02 11:48:19 -04:00
|
|
|
ts = time()
|
|
|
|
timestamp = normalize_timestamp(ts)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'Content-Length': '0'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(ts + 1),
|
|
|
|
'X-Container-Host': '1.2.3.4:0',
|
|
|
|
'X-Container-Partition': '3',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Timestamp': '1',
|
|
|
|
'Content-Type': 'application/new1'})
|
2010-07-12 17:03:45 -05:00
|
|
|
object_server.http_connect = mock_http_connect(202)
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(ts + 2),
|
|
|
|
'X-Container-Host': '1.2.3.4:0',
|
|
|
|
'X-Container-Partition': '3',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Timestamp': '1',
|
|
|
|
'Content-Type': 'application/new1'})
|
2010-07-12 17:03:45 -05:00
|
|
|
object_server.http_connect = mock_http_connect(202, with_exc=True)
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(ts + 3),
|
|
|
|
'X-Container-Host': '1.2.3.4:0',
|
|
|
|
'X-Container-Partition': '3',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Timestamp': '1',
|
|
|
|
'Content-Type': 'application/new2'})
|
2010-07-12 17:03:45 -05:00
|
|
|
object_server.http_connect = mock_http_connect(500)
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
|
|
|
finally:
|
|
|
|
object_server.http_connect = old_http_connect
|
|
|
|
|
2011-03-15 22:12:03 -07:00
|
|
|
def test_POST_quarantine_zbyte(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test'})
|
|
|
|
req.body = 'VERIFY'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-12-11 16:59:59 -05:00
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
objfile = self.df_mgr.get_diskfile('sda1', 'p', 'a', 'c', 'o')
|
2013-08-30 18:08:24 -07:00
|
|
|
objfile.open()
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
file_name = os.path.basename(objfile._data_file)
|
|
|
|
with open(objfile._data_file) as fp:
|
2013-07-17 16:32:35 -07:00
|
|
|
metadata = diskfile.read_metadata(fp)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.unlink(objfile._data_file)
|
|
|
|
with open(objfile._data_file, 'w') as fp:
|
2013-07-17 16:32:35 -07:00
|
|
|
diskfile.write_metadata(fp, metadata)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
self.assertEquals(os.listdir(objfile._datadir)[0], file_name)
|
2013-12-11 16:59:59 -05:00
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2013-12-11 16:59:59 -05:00
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
2013-09-01 15:10:39 -04:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(time())})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
quar_dir = os.path.join(
|
|
|
|
self.testdir, 'sda1', 'quarantined', 'objects',
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.path.basename(os.path.dirname(objfile._data_file)))
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(os.listdir(quar_dir)[0], file_name)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_PUT_invalid_path(self):
|
|
|
|
req = Request.blank('/sda1/p/a/c', environ={'REQUEST_METHOD': 'PUT'})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
|
|
|
|
def test_PUT_no_timestamp(self):
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT',
|
|
|
|
'CONTENT_LENGTH': '0'})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
|
|
|
|
def test_PUT_no_content_type(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '6'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'VERIFY'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
|
|
|
|
def test_PUT_invalid_content_type(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': '\xff\xff'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'VERIFY'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
self.assert_('Content-Type' in resp.body)
|
|
|
|
|
|
|
|
def test_PUT_no_content_length(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'VERIFY'
|
|
|
|
del req.headers['Content-Length']
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 411)
|
|
|
|
|
2013-04-19 00:10:38 -04:00
|
|
|
def test_PUT_zero_content_length(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
2013-04-19 00:10:38 -04:00
|
|
|
req.body = ''
|
|
|
|
self.assertEquals(req.headers['Content-Length'], '0')
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-04-19 00:10:38 -04:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
def test_PUT_bad_transfer_encoding(self):
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'VERIFY'
|
|
|
|
req.headers['Transfer-Encoding'] = 'bad'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 400)
|
|
|
|
|
2014-03-19 21:51:17 +00:00
|
|
|
def test_PUT_if_none_match_star(self):
|
|
|
|
# First PUT should succeed
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'If-None-Match': '*'})
|
|
|
|
req.body = 'VERIFY'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
# File should already exist so it should fail
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'If-None-Match': '*'})
|
|
|
|
req.body = 'VERIFY'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
|
|
|
def test_PUT_if_none_match(self):
|
|
|
|
# PUT with if-none-match set and nothing there should succede
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'If-None-Match': 'notthere'})
|
|
|
|
req.body = 'VERIFY'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
# PUT with if-none-match of the object etag should fail
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'If-None-Match': '0b4c12d7e0a73840c1c4f148fda3b037'})
|
|
|
|
req.body = 'VERIFY'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_PUT_common(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'VERIFY'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0),
|
|
|
|
'p', hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.data')
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assert_(os.path.isfile(objfile))
|
|
|
|
self.assertEquals(open(objfile).read(), 'VERIFY')
|
2013-07-17 16:32:35 -07:00
|
|
|
self.assertEquals(diskfile.read_metadata(objfile),
|
2014-06-10 22:17:47 -07:00
|
|
|
{'X-Timestamp': utils.Timestamp(timestamp).internal,
|
2010-07-12 17:03:45 -05:00
|
|
|
'Content-Length': '6',
|
|
|
|
'ETag': '0b4c12d7e0a73840c1c4f148fda3b037',
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'name': '/a/c/o'})
|
|
|
|
|
|
|
|
def test_PUT_overwrite(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'VERIFY'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp = normalize_timestamp(time())
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'Content-Encoding': 'gzip'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'VERIFY TWO'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2010-07-12 17:03:45 -05:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.data')
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assert_(os.path.isfile(objfile))
|
|
|
|
self.assertEquals(open(objfile).read(), 'VERIFY TWO')
|
2013-07-17 16:32:35 -07:00
|
|
|
self.assertEquals(diskfile.read_metadata(objfile),
|
2014-06-10 22:17:47 -07:00
|
|
|
{'X-Timestamp': utils.Timestamp(timestamp).internal,
|
2010-07-12 17:03:45 -05:00
|
|
|
'Content-Length': '10',
|
|
|
|
'ETag': 'b381a4c5dab1eaa1eb9711fa647cd039',
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'name': '/a/c/o',
|
|
|
|
'Content-Encoding': 'gzip'})
|
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
def test_PUT_overwrite_w_delete_at(self):
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'X-Delete-At': 9999999999,
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'VERIFY'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'Content-Encoding': 'gzip'})
|
|
|
|
req.body = 'VERIFY TWO'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2013-12-11 16:59:59 -05:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.data')
|
2013-12-11 16:59:59 -05:00
|
|
|
self.assertTrue(os.path.isfile(objfile))
|
|
|
|
self.assertEqual(open(objfile).read(), 'VERIFY TWO')
|
|
|
|
self.assertEqual(diskfile.read_metadata(objfile),
|
2014-06-10 22:17:47 -07:00
|
|
|
{'X-Timestamp': utils.Timestamp(timestamp).internal,
|
2013-12-11 16:59:59 -05:00
|
|
|
'Content-Length': '10',
|
|
|
|
'ETag': 'b381a4c5dab1eaa1eb9711fa647cd039',
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'name': '/a/c/o',
|
|
|
|
'Content-Encoding': 'gzip'})
|
|
|
|
|
2013-07-02 11:48:19 -04:00
|
|
|
def test_PUT_old_timestamp(self):
|
|
|
|
ts = time()
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(ts),
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
2013-07-02 11:48:19 -04:00
|
|
|
req.body = 'VERIFY'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(ts),
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'Content-Encoding': 'gzip'})
|
|
|
|
req.body = 'VERIFY TWO'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 409)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2013-09-01 15:10:39 -04:00
|
|
|
headers={
|
|
|
|
'X-Timestamp': normalize_timestamp(ts - 1),
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'Content-Encoding': 'gzip'})
|
2013-07-02 11:48:19 -04:00
|
|
|
req.body = 'VERIFY THREE'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 409)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_PUT_no_etag(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'text/plain'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'test'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
def test_PUT_invalid_etag(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'ETag': 'invalid'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'test'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 422)
|
|
|
|
|
|
|
|
def test_PUT_user_metadata(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'ETag': 'b114ab7b90d9ccac4bd5d99cc7ebb568',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-Two': 'Two'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'VERIFY THREE'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2010-07-12 17:03:45 -05:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.data')
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assert_(os.path.isfile(objfile))
|
|
|
|
self.assertEquals(open(objfile).read(), 'VERIFY THREE')
|
2013-07-17 16:32:35 -07:00
|
|
|
self.assertEquals(diskfile.read_metadata(objfile),
|
2014-06-10 22:17:47 -07:00
|
|
|
{'X-Timestamp': utils.Timestamp(timestamp).internal,
|
2010-07-12 17:03:45 -05:00
|
|
|
'Content-Length': '12',
|
|
|
|
'ETag': 'b114ab7b90d9ccac4bd5d99cc7ebb568',
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'name': '/a/c/o',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-Two': 'Two'})
|
|
|
|
|
2014-02-28 17:10:03 +00:00
|
|
|
def test_PUT_client_timeout(self):
|
|
|
|
class FakeTimeout(BaseException):
|
|
|
|
def __enter__(self):
|
|
|
|
raise self
|
|
|
|
|
|
|
|
def __exit__(self, typ, value, tb):
|
|
|
|
pass
|
|
|
|
# This is just so the test fails when run on older object server code
|
|
|
|
# instead of exploding.
|
|
|
|
if not hasattr(object_server, 'ChunkReadTimeout'):
|
|
|
|
object_server.ChunkReadTimeout = None
|
|
|
|
with mock.patch.object(object_server, 'ChunkReadTimeout', FakeTimeout):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'Content-Length': '6'})
|
|
|
|
req.environ['wsgi.input'] = StringIO('VERIFY')
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 408)
|
|
|
|
|
2014-03-10 11:46:58 +00:00
|
|
|
def test_PUT_system_metadata(self):
|
|
|
|
# check that sysmeta is stored in diskfile
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'ETag': '1000d172764c9dbc3a5798a67ec5bb76',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-Two': 'Two'})
|
|
|
|
req.body = 'VERIFY SYSMETA'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
|
|
|
hash_path('a', 'c', 'o')),
|
|
|
|
timestamp + '.data')
|
|
|
|
self.assert_(os.path.isfile(objfile))
|
|
|
|
self.assertEquals(open(objfile).read(), 'VERIFY SYSMETA')
|
|
|
|
self.assertEquals(diskfile.read_metadata(objfile),
|
|
|
|
{'X-Timestamp': timestamp,
|
|
|
|
'Content-Length': '14',
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'ETag': '1000d172764c9dbc3a5798a67ec5bb76',
|
|
|
|
'name': '/a/c/o',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-Two': 'Two'})
|
|
|
|
|
|
|
|
def test_POST_system_metadata(self):
|
|
|
|
# check that diskfile sysmeta is not changed by a POST
|
|
|
|
timestamp1 = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp1,
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'ETag': '1000d172764c9dbc3a5798a67ec5bb76',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-Two': 'Two'})
|
|
|
|
req.body = 'VERIFY SYSMETA'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
timestamp2 = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp2,
|
|
|
|
'X-Object-Meta-1': 'Not One',
|
|
|
|
'X-Object-Sysmeta-1': 'Not One',
|
|
|
|
'X-Object-Sysmeta-Two': 'Not Two'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 202)
|
|
|
|
|
|
|
|
# original .data file metadata should be unchanged
|
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
|
|
|
hash_path('a', 'c', 'o')),
|
|
|
|
timestamp1 + '.data')
|
|
|
|
self.assert_(os.path.isfile(objfile))
|
|
|
|
self.assertEquals(open(objfile).read(), 'VERIFY SYSMETA')
|
|
|
|
self.assertEquals(diskfile.read_metadata(objfile),
|
|
|
|
{'X-Timestamp': timestamp1,
|
|
|
|
'Content-Length': '14',
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'ETag': '1000d172764c9dbc3a5798a67ec5bb76',
|
|
|
|
'name': '/a/c/o',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-Two': 'Two'})
|
|
|
|
|
|
|
|
# .meta file metadata should have only user meta items
|
|
|
|
metafile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
|
|
|
hash_path('a', 'c', 'o')),
|
|
|
|
timestamp2 + '.meta')
|
|
|
|
self.assert_(os.path.isfile(metafile))
|
|
|
|
self.assertEquals(diskfile.read_metadata(metafile),
|
|
|
|
{'X-Timestamp': timestamp2,
|
|
|
|
'name': '/a/c/o',
|
|
|
|
'X-Object-Meta-1': 'Not One'})
|
|
|
|
|
|
|
|
def test_PUT_then_fetch_system_metadata(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'ETag': '1000d172764c9dbc3a5798a67ec5bb76',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-Two': 'Two'})
|
|
|
|
req.body = 'VERIFY SYSMETA'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
def check_response(resp):
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.content_length, 14)
|
|
|
|
self.assertEquals(resp.content_type, 'text/plain')
|
|
|
|
self.assertEquals(resp.headers['content-type'], 'text/plain')
|
|
|
|
self.assertEquals(
|
|
|
|
resp.headers['last-modified'],
|
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp)))))
|
|
|
|
self.assertEquals(resp.headers['etag'],
|
|
|
|
'"1000d172764c9dbc3a5798a67ec5bb76"')
|
|
|
|
self.assertEquals(resp.headers['x-object-meta-1'], 'One')
|
|
|
|
self.assertEquals(resp.headers['x-object-sysmeta-1'], 'One')
|
|
|
|
self.assertEquals(resp.headers['x-object-sysmeta-two'], 'Two')
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
check_response(resp)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'GET'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
check_response(resp)
|
|
|
|
|
|
|
|
def test_PUT_then_POST_then_fetch_system_metadata(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'ETag': '1000d172764c9dbc3a5798a67ec5bb76',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-1': 'One',
|
|
|
|
'X-Object-Sysmeta-Two': 'Two'})
|
|
|
|
req.body = 'VERIFY SYSMETA'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
timestamp2 = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp2,
|
|
|
|
'X-Object-Meta-1': 'Not One',
|
|
|
|
'X-Object-Sysmeta-1': 'Not One',
|
|
|
|
'X-Object-Sysmeta-Two': 'Not Two'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 202)
|
|
|
|
|
|
|
|
def check_response(resp):
|
|
|
|
# user meta should be updated but not sysmeta
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.content_length, 14)
|
|
|
|
self.assertEquals(resp.content_type, 'text/plain')
|
|
|
|
self.assertEquals(resp.headers['content-type'], 'text/plain')
|
|
|
|
self.assertEquals(
|
|
|
|
resp.headers['last-modified'],
|
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp2)))))
|
|
|
|
self.assertEquals(resp.headers['etag'],
|
|
|
|
'"1000d172764c9dbc3a5798a67ec5bb76"')
|
|
|
|
self.assertEquals(resp.headers['x-object-meta-1'], 'Not One')
|
|
|
|
self.assertEquals(resp.headers['x-object-sysmeta-1'], 'One')
|
|
|
|
self.assertEquals(resp.headers['x-object-sysmeta-two'], 'Two')
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
check_response(resp)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'GET'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
check_response(resp)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_PUT_container_connection(self):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def mock_http_connect(response, with_exc=False):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
class FakeConn(object):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def __init__(self, status, with_exc):
|
|
|
|
self.status = status
|
|
|
|
self.reason = 'Fake'
|
|
|
|
self.host = '1.2.3.4'
|
|
|
|
self.port = '1234'
|
|
|
|
self.with_exc = with_exc
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def getresponse(self):
|
|
|
|
if self.with_exc:
|
|
|
|
raise Exception('test')
|
|
|
|
return self
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def read(self, amt=None):
|
|
|
|
return ''
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
return lambda *args, **kwargs: FakeConn(response, with_exc)
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
old_http_connect = object_server.http_connect
|
|
|
|
try:
|
|
|
|
timestamp = normalize_timestamp(time())
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'X-Container-Host': '1.2.3.4:0',
|
|
|
|
'X-Container-Partition': '3',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Timestamp': '1',
|
|
|
|
'Content-Type': 'application/new1',
|
|
|
|
'Content-Length': '0'})
|
2010-07-12 17:03:45 -05:00
|
|
|
object_server.http_connect = mock_http_connect(201)
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
timestamp = normalize_timestamp(time())
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'X-Container-Host': '1.2.3.4:0',
|
|
|
|
'X-Container-Partition': '3',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Timestamp': '1',
|
|
|
|
'Content-Type': 'application/new1',
|
|
|
|
'Content-Length': '0'})
|
2010-07-12 17:03:45 -05:00
|
|
|
object_server.http_connect = mock_http_connect(500)
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
timestamp = normalize_timestamp(time())
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'X-Container-Host': '1.2.3.4:0',
|
|
|
|
'X-Container-Partition': '3',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Timestamp': '1',
|
|
|
|
'Content-Type': 'application/new1',
|
|
|
|
'Content-Length': '0'})
|
2010-07-12 17:03:45 -05:00
|
|
|
object_server.http_connect = mock_http_connect(500, with_exc=True)
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
finally:
|
|
|
|
object_server.http_connect = old_http_connect
|
|
|
|
|
|
|
|
def test_HEAD(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.HEAD
|
2013-07-24 12:55:25 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c', environ={'REQUEST_METHOD': 'HEAD'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
2014-03-23 10:24:21 -07:00
|
|
|
self.assertFalse('X-Backend-Timestamp' in resp.headers)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
|
|
|
timestamp = normalize_timestamp(time())
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-Two': 'Two'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'VERIFY'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.content_length, 6)
|
|
|
|
self.assertEquals(resp.content_type, 'application/x-test')
|
|
|
|
self.assertEquals(resp.headers['content-type'], 'application/x-test')
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
resp.headers['last-modified'],
|
2013-11-07 04:45:27 +00:00
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp)))))
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.headers['etag'],
|
|
|
|
'"0b4c12d7e0a73840c1c4f148fda3b037"')
|
|
|
|
self.assertEquals(resp.headers['x-object-meta-1'], 'One')
|
|
|
|
self.assertEquals(resp.headers['x-object-meta-two'], 'Two')
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2010-07-12 17:03:45 -05:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.data')
|
2010-07-12 17:03:45 -05:00
|
|
|
os.unlink(objfile)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-length': '6'})
|
|
|
|
req.body = 'VERIFY'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': timestamp})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 204)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
2014-06-10 22:17:47 -07:00
|
|
|
self.assertEquals(resp.headers['X-Backend-Timestamp'],
|
|
|
|
utils.Timestamp(timestamp).internal)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
2011-03-15 22:12:03 -07:00
|
|
|
def test_HEAD_quarantine_zbyte(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.GET
|
2011-03-15 22:12:03 -07:00
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test'})
|
|
|
|
req.body = 'VERIFY'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
disk_file = self.df_mgr.get_diskfile('sda1', 'p', 'a', 'c', 'o')
|
|
|
|
disk_file.open()
|
2011-03-15 22:12:03 -07:00
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
file_name = os.path.basename(disk_file._data_file)
|
|
|
|
with open(disk_file._data_file) as fp:
|
2013-07-17 16:32:35 -07:00
|
|
|
metadata = diskfile.read_metadata(fp)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.unlink(disk_file._data_file)
|
|
|
|
with open(disk_file._data_file, 'w') as fp:
|
2013-07-17 16:32:35 -07:00
|
|
|
diskfile.write_metadata(fp, metadata)
|
2011-03-15 22:12:03 -07:00
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
file_name = os.path.basename(disk_file._data_file)
|
|
|
|
self.assertEquals(os.listdir(disk_file._datadir)[0], file_name)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
quar_dir = os.path.join(
|
|
|
|
self.testdir, 'sda1', 'quarantined', 'objects',
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.path.basename(os.path.dirname(disk_file._data_file)))
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(os.listdir(quar_dir)[0], file_name)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_GET(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.GET
|
2013-07-24 12:55:25 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c', environ={'REQUEST_METHOD': 'GET'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
|
2013-07-24 12:55:25 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
2014-03-23 10:24:21 -07:00
|
|
|
self.assertFalse('X-Backend-Timestamp' in resp.headers)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-Two': 'Two'})
|
|
|
|
req.body = 'VERIFY'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-07-24 12:55:25 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.body, 'VERIFY')
|
|
|
|
self.assertEquals(resp.content_length, 6)
|
|
|
|
self.assertEquals(resp.content_type, 'application/x-test')
|
|
|
|
self.assertEquals(resp.headers['content-length'], '6')
|
|
|
|
self.assertEquals(resp.headers['content-type'], 'application/x-test')
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
resp.headers['last-modified'],
|
2013-11-07 04:45:27 +00:00
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp)))))
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.headers['etag'],
|
|
|
|
'"0b4c12d7e0a73840c1c4f148fda3b037"')
|
|
|
|
self.assertEquals(resp.headers['x-object-meta-1'], 'One')
|
|
|
|
self.assertEquals(resp.headers['x-object-meta-two'], 'Two')
|
|
|
|
|
2013-07-24 12:55:25 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.range = 'bytes=1-3'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 206)
|
|
|
|
self.assertEquals(resp.body, 'ERI')
|
|
|
|
self.assertEquals(resp.headers['content-length'], '3')
|
|
|
|
|
2013-07-24 12:55:25 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.range = 'bytes=1-'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 206)
|
|
|
|
self.assertEquals(resp.body, 'ERIFY')
|
|
|
|
self.assertEquals(resp.headers['content-length'], '5')
|
|
|
|
|
2013-07-24 12:55:25 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.range = 'bytes=-2'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 206)
|
|
|
|
self.assertEquals(resp.body, 'FY')
|
|
|
|
self.assertEquals(resp.headers['content-length'], '2')
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2010-07-12 17:03:45 -05:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.data')
|
2010-07-12 17:03:45 -05:00
|
|
|
os.unlink(objfile)
|
2013-07-24 12:55:25 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application:octet-stream',
|
|
|
|
'Content-Length': '6'})
|
|
|
|
req.body = 'VERIFY'
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': timestamp})
|
2013-07-24 12:55:25 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 204)
|
|
|
|
|
2013-07-24 12:55:25 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
2014-06-10 22:17:47 -07:00
|
|
|
self.assertEquals(resp.headers['X-Backend-Timestamp'],
|
|
|
|
utils.Timestamp(timestamp).internal)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
|
|
|
def test_GET_if_match(self):
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
etag = resp.etag
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Match': '*'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
2010-11-16 15:35:39 -08:00
|
|
|
req = Request.blank('/sda1/p/a/c/o2',
|
|
|
|
environ={'REQUEST_METHOD': 'GET'},
|
2010-07-12 17:03:45 -05:00
|
|
|
headers={'If-Match': '*'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Match': '"%s"' % etag})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Match': '"11111111111111111111111111111111"'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={
|
|
|
|
'If-Match': '"11111111111111111111111111111111", "%s"' % etag})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={
|
|
|
|
'If-Match':
|
|
|
|
'"11111111111111111111111111111111", '
|
|
|
|
'"22222222222222222222222222222222"'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
Support If-[None-]Match for object HEAD, SLO, and DLO
I moved the checking of If-Match and If-None-Match out of the object
server's GET method and into swob so that everyone can use it. The
interface is similar to the Range handling; make a response with
conditional_response=True, and you get handing of If-Match and
If-None-Match.
Since the only users of conditional_response are object GET, object
HEAD, SLO, and DLO, this has the effect of adding support for If-Match
and If-None-Match to just the latter three places and nowhere
else. This makes object GET and HEAD consistent for any kind of
object, large or small.
This also fixes a bug where various conditional headers (If-*) were
passed through to the object server on segment requests, which could
cause segment requests to fail with a 304 or 412 response. Now only
certain headers are copied to the segment requests, and that doesn't
include the conditional ones, so they can't goof up the segment
retrieval.
Note that I moved SegmentedIterable to swift.common.request_helpers
because it sprouted a transitive dependency on swob, and leaving it in
utils caused a circular import.
Bonus fix: unified the handling of DiskFileQuarantined and
DiskFileNotFound in object server GET and HEAD. Now in either case, a
412 will be returned if the client said "If-Match: *". If not, the
response is a 404, just like before.
Closes-Bug: 1279076
Closes-Bug: 1280022
Closes-Bug: 1280028
Change-Id: Id2ee78346244d516b980202e990aa38ce6812de5
2014-02-12 18:29:12 -08:00
|
|
|
def test_HEAD_if_match(self):
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
etag = resp.etag
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Match': '*'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o2',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Match': '*'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Match': '"%s"' % etag})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Match': '"11111111111111111111111111111111"'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={
|
|
|
|
'If-Match': '"11111111111111111111111111111111", "%s"' % etag})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={
|
|
|
|
'If-Match':
|
|
|
|
'"11111111111111111111111111111111", '
|
|
|
|
'"22222222222222222222222222222222"'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_GET_if_none_match(self):
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
etag = resp.etag
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-None-Match': '*'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o2',
|
|
|
|
environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-None-Match': '*'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-None-Match': '"%s"' % etag})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-None-Match': '"11111111111111111111111111111111"'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-None-Match':
|
|
|
|
'"11111111111111111111111111111111", '
|
|
|
|
'"%s"' % etag})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 304)
|
Support If-[None-]Match for object HEAD, SLO, and DLO
I moved the checking of If-Match and If-None-Match out of the object
server's GET method and into swob so that everyone can use it. The
interface is similar to the Range handling; make a response with
conditional_response=True, and you get handing of If-Match and
If-None-Match.
Since the only users of conditional_response are object GET, object
HEAD, SLO, and DLO, this has the effect of adding support for If-Match
and If-None-Match to just the latter three places and nowhere
else. This makes object GET and HEAD consistent for any kind of
object, large or small.
This also fixes a bug where various conditional headers (If-*) were
passed through to the object server on segment requests, which could
cause segment requests to fail with a 304 or 412 response. Now only
certain headers are copied to the segment requests, and that doesn't
include the conditional ones, so they can't goof up the segment
retrieval.
Note that I moved SegmentedIterable to swift.common.request_helpers
because it sprouted a transitive dependency on swob, and leaving it in
utils caused a circular import.
Bonus fix: unified the handling of DiskFileQuarantined and
DiskFileNotFound in object server GET and HEAD. Now in either case, a
412 will be returned if the client said "If-Match: *". If not, the
response is a 404, just like before.
Closes-Bug: 1279076
Closes-Bug: 1280022
Closes-Bug: 1280028
Change-Id: Id2ee78346244d516b980202e990aa38ce6812de5
2014-02-12 18:29:12 -08:00
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
def test_HEAD_if_none_match(self):
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
etag = resp.etag
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-None-Match': '*'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o2',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-None-Match': '*'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-None-Match': '"%s"' % etag})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-None-Match': '"11111111111111111111111111111111"'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-None-Match':
|
|
|
|
'"11111111111111111111111111111111", '
|
|
|
|
'"%s"' % etag})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 304)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.etag, etag)
|
|
|
|
|
|
|
|
def test_GET_if_modified_since(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
since = strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(float(timestamp) + 1))
|
2010-07-12 17:03:45 -05:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Modified-Since': since})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
|
|
|
|
since = \
|
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT', gmtime(float(timestamp) - 1))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Modified-Since': since})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
|
|
|
since = \
|
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT', gmtime(float(timestamp) + 1))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Modified-Since': since})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
|
2013-11-07 04:45:27 +00:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
since = resp.headers['Last-Modified']
|
|
|
|
self.assertEquals(since, strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp)))))
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Modified-Since': since})
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-11-07 04:45:27 +00:00
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
|
|
|
|
timestamp = normalize_timestamp(int(time()))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o2',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
since = strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(float(timestamp)))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o2',
|
|
|
|
environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Modified-Since': since})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
|
2014-03-04 11:52:48 -08:00
|
|
|
def test_HEAD_if_modified_since(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
|
|
|
since = strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(float(timestamp) + 1))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Modified-Since': since})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
|
|
|
|
since = \
|
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT', gmtime(float(timestamp) - 1))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Modified-Since': since})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
|
|
|
since = \
|
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT', gmtime(float(timestamp) + 1))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Modified-Since': since})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
since = resp.headers['Last-Modified']
|
|
|
|
self.assertEquals(since, strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp)))))
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Modified-Since': since})
|
|
|
|
resp = self.object_controller.GET(req)
|
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
|
|
|
|
timestamp = normalize_timestamp(int(time()))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o2',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
since = strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(float(timestamp)))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o2',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Modified-Since': since})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 304)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_GET_if_unmodified_since(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
|
|
|
since = strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(float(timestamp) + 1))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Unmodified-Since': since})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
|
|
|
since = \
|
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT', gmtime(float(timestamp) - 9))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Unmodified-Since': since})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
|
|
|
since = \
|
|
|
|
strftime('%a, %d %b %Y %H:%M:%S GMT', gmtime(float(timestamp) + 9))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Unmodified-Since': since})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
2013-11-07 04:45:27 +00:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
since = resp.headers['Last-Modified']
|
|
|
|
self.assertEquals(since, strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp)))))
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'If-Unmodified-Since': since})
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-11-07 04:45:27 +00:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
2014-03-04 11:52:48 -08:00
|
|
|
def test_HEAD_if_unmodified_since(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4'})
|
|
|
|
req.body = 'test'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
since = strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp)) + 1))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Unmodified-Since': since})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
|
|
|
since = strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp))))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Unmodified-Since': since})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
|
|
|
since = strftime('%a, %d %b %Y %H:%M:%S GMT',
|
|
|
|
gmtime(math.ceil(float(timestamp)) - 1))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'If-Unmodified-Since': since})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
2011-03-15 22:12:03 -07:00
|
|
|
def test_GET_quarantine(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.GET
|
2011-03-15 22:12:03 -07:00
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test'})
|
|
|
|
req.body = 'VERIFY'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
disk_file = self.df_mgr.get_diskfile('sda1', 'p', 'a', 'c', 'o')
|
|
|
|
disk_file.open()
|
|
|
|
file_name = os.path.basename(disk_file._data_file)
|
2011-03-15 22:12:03 -07:00
|
|
|
etag = md5()
|
|
|
|
etag.update('VERIF')
|
|
|
|
etag = etag.hexdigest()
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
metadata = {'X-Timestamp': timestamp, 'name': '/a/c/o',
|
2011-03-15 22:12:03 -07:00
|
|
|
'Content-Length': 6, 'ETag': etag}
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
diskfile.write_metadata(disk_file._fp, metadata)
|
|
|
|
self.assertEquals(os.listdir(disk_file._datadir)[0], file_name)
|
2011-03-15 22:12:03 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-09-01 15:10:39 -04:00
|
|
|
quar_dir = os.path.join(
|
|
|
|
self.testdir, 'sda1', 'quarantined', 'objects',
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.path.basename(os.path.dirname(disk_file._data_file)))
|
|
|
|
self.assertEquals(os.listdir(disk_file._datadir)[0], file_name)
|
2012-02-07 14:02:28 -08:00
|
|
|
body = resp.body # actually does quarantining
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(body, 'VERIFY')
|
|
|
|
self.assertEquals(os.listdir(quar_dir)[0], file_name)
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
|
|
|
def test_GET_quarantine_zbyte(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.GET
|
2011-03-15 22:12:03 -07:00
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test'})
|
|
|
|
req.body = 'VERIFY'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
disk_file = self.df_mgr.get_diskfile('sda1', 'p', 'a', 'c', 'o')
|
|
|
|
disk_file.open()
|
|
|
|
file_name = os.path.basename(disk_file._data_file)
|
|
|
|
with open(disk_file._data_file) as fp:
|
2013-07-17 16:32:35 -07:00
|
|
|
metadata = diskfile.read_metadata(fp)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.unlink(disk_file._data_file)
|
|
|
|
with open(disk_file._data_file, 'w') as fp:
|
2013-07-17 16:32:35 -07:00
|
|
|
diskfile.write_metadata(fp, metadata)
|
2011-03-15 22:12:03 -07:00
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
self.assertEquals(os.listdir(disk_file._datadir)[0], file_name)
|
2011-03-15 22:12:03 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
quar_dir = os.path.join(
|
|
|
|
self.testdir, 'sda1', 'quarantined', 'objects',
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.path.basename(os.path.dirname(disk_file._data_file)))
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(os.listdir(quar_dir)[0], file_name)
|
|
|
|
|
|
|
|
def test_GET_quarantine_range(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.GET
|
2011-03-15 22:12:03 -07:00
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test'})
|
|
|
|
req.body = 'VERIFY'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
disk_file = self.df_mgr.get_diskfile('sda1', 'p', 'a', 'c', 'o')
|
|
|
|
disk_file.open()
|
|
|
|
file_name = os.path.basename(disk_file._data_file)
|
2011-03-15 22:12:03 -07:00
|
|
|
etag = md5()
|
|
|
|
etag.update('VERIF')
|
|
|
|
etag = etag.hexdigest()
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
metadata = {'X-Timestamp': timestamp, 'name': '/a/c/o',
|
2011-03-15 22:12:03 -07:00
|
|
|
'Content-Length': 6, 'ETag': etag}
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
diskfile.write_metadata(disk_file._fp, metadata)
|
|
|
|
self.assertEquals(os.listdir(disk_file._datadir)[0], file_name)
|
2011-03-15 22:12:03 -07:00
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
|
|
|
req.range = 'bytes=0-4' # partial
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-09-01 15:10:39 -04:00
|
|
|
quar_dir = os.path.join(
|
|
|
|
self.testdir, 'sda1', 'quarantined', 'objects',
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.path.basename(os.path.dirname(disk_file._data_file)))
|
2013-07-23 16:41:45 -07:00
|
|
|
resp.body
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
self.assertEquals(os.listdir(disk_file._datadir)[0], file_name)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertFalse(os.path.isdir(quar_dir))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
|
|
|
req.range = 'bytes=1-6' # partial
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-09-01 15:10:39 -04:00
|
|
|
quar_dir = os.path.join(
|
|
|
|
self.testdir, 'sda1', 'quarantined', 'objects',
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.path.basename(os.path.dirname(disk_file._data_file)))
|
2013-07-23 16:41:45 -07:00
|
|
|
resp.body
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
self.assertEquals(os.listdir(disk_file._datadir)[0], file_name)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertFalse(os.path.isdir(quar_dir))
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
|
|
|
req.range = 'bytes=0-14' # full
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-09-01 15:10:39 -04:00
|
|
|
quar_dir = os.path.join(
|
|
|
|
self.testdir, 'sda1', 'quarantined', 'objects',
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
os.path.basename(os.path.dirname(disk_file._data_file)))
|
|
|
|
self.assertEquals(os.listdir(disk_file._datadir)[0], file_name)
|
2013-07-23 16:41:45 -07:00
|
|
|
resp.body
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertTrue(os.path.isdir(quar_dir))
|
|
|
|
req = Request.blank('/sda1/p/a/c/o')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-03-15 22:12:03 -07:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
@mock.patch("time.time", mock_time)
|
2010-07-12 17:03:45 -05:00
|
|
|
def test_DELETE(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.DELETE
|
2010-11-16 15:35:39 -08:00
|
|
|
req = Request.blank('/sda1/p/a/c',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
|
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
# self.assertRaises(KeyError, self.object_controller.DELETE, req)
|
|
|
|
|
2013-07-02 11:48:19 -04:00
|
|
|
# The following should have created a tombstone file
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
timestamp = normalize_timestamp(1000)
|
2010-07-12 17:03:45 -05:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': timestamp})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
ts_1000_file = os.path.join(
|
2013-09-01 15:10:39 -04:00
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2013-07-02 11:48:19 -04:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.ts')
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
self.assertTrue(os.path.isfile(ts_1000_file))
|
|
|
|
# There should now be a 1000 ts file.
|
|
|
|
self.assertEquals(len(os.listdir(os.path.dirname(ts_1000_file))), 1)
|
2013-07-02 11:48:19 -04:00
|
|
|
|
|
|
|
# The following should *not* have created a tombstone file.
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
timestamp = normalize_timestamp(999)
|
2013-07-02 11:48:19 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': timestamp})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
ts_999_file = os.path.join(
|
2013-09-01 15:10:39 -04:00
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2013-07-02 11:48:19 -04:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.ts')
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
self.assertFalse(os.path.isfile(ts_999_file))
|
|
|
|
self.assertTrue(os.path.isfile(ts_1000_file))
|
|
|
|
self.assertEquals(len(os.listdir(os.path.dirname(ts_1000_file))), 1)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
timestamp = normalize_timestamp(1002)
|
2010-07-12 17:03:45 -05:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
|
|
|
'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4',
|
2013-09-01 15:10:39 -04:00
|
|
|
})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'test'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
# There should now be 1000 ts and a 1001 data file.
|
|
|
|
data_1002_file = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.data')
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
self.assertTrue(os.path.isfile(data_1002_file))
|
|
|
|
self.assertEquals(len(os.listdir(os.path.dirname(data_1002_file))), 1)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
2013-07-02 11:48:19 -04:00
|
|
|
# The following should *not* have created a tombstone file.
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
timestamp = normalize_timestamp(1001)
|
2010-07-12 17:03:45 -05:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': timestamp})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 409)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
ts_1001_file = os.path.join(
|
2013-09-01 15:10:39 -04:00
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2010-07-12 17:03:45 -05:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.ts')
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
self.assertFalse(os.path.isfile(ts_1001_file))
|
|
|
|
self.assertTrue(os.path.isfile(data_1002_file))
|
|
|
|
self.assertEquals(len(os.listdir(os.path.dirname(ts_1001_file))), 1)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
timestamp = normalize_timestamp(1003)
|
2010-07-12 17:03:45 -05:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': timestamp})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 204)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
ts_1003_file = os.path.join(
|
2013-09-01 15:10:39 -04:00
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2010-07-12 17:03:45 -05:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.ts')
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
self.assertTrue(os.path.isfile(ts_1003_file))
|
|
|
|
self.assertEquals(len(os.listdir(os.path.dirname(ts_1003_file))), 1)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
2013-07-02 11:48:19 -04:00
|
|
|
def test_DELETE_container_updates(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.DELETE and container
|
2013-07-02 11:48:19 -04:00
|
|
|
# updates, making sure container update is called in the correct
|
|
|
|
# state.
|
2014-06-10 22:17:47 -07:00
|
|
|
start = time()
|
|
|
|
timestamp = utils.Timestamp(start)
|
2013-07-02 11:48:19 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={
|
2014-06-10 22:17:47 -07:00
|
|
|
'X-Timestamp': timestamp.internal,
|
2013-07-02 11:48:19 -04:00
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Content-Length': '4',
|
2013-09-01 15:10:39 -04:00
|
|
|
})
|
2013-07-02 11:48:19 -04:00
|
|
|
req.body = 'test'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
calls_made = [0]
|
|
|
|
|
|
|
|
def our_container_update(*args, **kwargs):
|
|
|
|
calls_made[0] += 1
|
|
|
|
|
|
|
|
orig_cu = self.object_controller.container_update
|
|
|
|
self.object_controller.container_update = our_container_update
|
|
|
|
try:
|
|
|
|
# The following request should return 409 (HTTP Conflict). A
|
|
|
|
# tombstone file should not have been created with this timestamp.
|
2014-06-10 22:17:47 -07:00
|
|
|
timestamp = utils.Timestamp(start - 0.00001)
|
2013-07-02 11:48:19 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
2014-06-10 22:17:47 -07:00
|
|
|
headers={'X-Timestamp': timestamp.internal})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 409)
|
2013-09-01 15:10:39 -04:00
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2013-07-02 11:48:19 -04:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.ts')
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertFalse(os.path.isfile(objfile))
|
|
|
|
self.assertEquals(len(os.listdir(os.path.dirname(objfile))), 1)
|
|
|
|
self.assertEquals(0, calls_made[0])
|
|
|
|
|
|
|
|
# The following request should return 204, and the object should
|
|
|
|
# be truly deleted (container update is performed) because this
|
|
|
|
# timestamp is newer. A tombstone file should have been created
|
|
|
|
# with this timestamp.
|
2014-06-10 22:17:47 -07:00
|
|
|
timestamp = utils.Timestamp(start + 0.00001)
|
2013-07-02 11:48:19 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
2014-06-10 22:17:47 -07:00
|
|
|
headers={'X-Timestamp': timestamp.internal})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 204)
|
2013-09-01 15:10:39 -04:00
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2013-07-02 11:48:19 -04:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.ts')
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assert_(os.path.isfile(objfile))
|
|
|
|
self.assertEquals(1, calls_made[0])
|
|
|
|
self.assertEquals(len(os.listdir(os.path.dirname(objfile))), 1)
|
|
|
|
|
|
|
|
# The following request should return a 404, as the object should
|
|
|
|
# already have been deleted, but it should have also performed a
|
|
|
|
# container update because the timestamp is newer, and a tombstone
|
|
|
|
# file should also exist with this timestamp.
|
2014-06-10 22:17:47 -07:00
|
|
|
timestamp = utils.Timestamp(start + 0.00002)
|
2013-07-02 11:48:19 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
2014-06-10 22:17:47 -07:00
|
|
|
headers={'X-Timestamp': timestamp.internal})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
2013-09-01 15:10:39 -04:00
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2013-07-02 11:48:19 -04:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.ts')
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assert_(os.path.isfile(objfile))
|
|
|
|
self.assertEquals(2, calls_made[0])
|
|
|
|
self.assertEquals(len(os.listdir(os.path.dirname(objfile))), 1)
|
|
|
|
|
|
|
|
# The following request should return a 404, as the object should
|
|
|
|
# already have been deleted, and it should not have performed a
|
|
|
|
# container update because the timestamp is older, or created a
|
|
|
|
# tombstone file with this timestamp.
|
2014-06-10 22:17:47 -07:00
|
|
|
timestamp = utils.Timestamp(start + 0.00001)
|
2013-07-02 11:48:19 -04:00
|
|
|
req = Request.blank('/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
2014-06-10 22:17:47 -07:00
|
|
|
headers={'X-Timestamp': timestamp.internal})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
2013-09-01 15:10:39 -04:00
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2013-07-02 11:48:19 -04:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(timestamp).internal + '.ts')
|
2013-07-02 11:48:19 -04:00
|
|
|
self.assertFalse(os.path.isfile(objfile))
|
|
|
|
self.assertEquals(2, calls_made[0])
|
|
|
|
self.assertEquals(len(os.listdir(os.path.dirname(objfile))), 1)
|
|
|
|
finally:
|
|
|
|
self.object_controller.container_update = orig_cu
|
|
|
|
|
2014-06-10 22:17:47 -07:00
|
|
|
def test_object_update_with_offset(self):
|
|
|
|
ts = (utils.Timestamp(t).internal for t in
|
|
|
|
itertools.count(int(time())))
|
|
|
|
container_updates = []
|
|
|
|
|
|
|
|
def capture_updates(ip, port, method, path, headers, *args, **kwargs):
|
|
|
|
container_updates.append((ip, port, method, path, headers))
|
|
|
|
# create a new object
|
|
|
|
create_timestamp = ts.next()
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='PUT', body='test1',
|
|
|
|
headers={'X-Timestamp': create_timestamp,
|
|
|
|
'X-Container-Host': '10.0.0.1:8080',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Partition': 'p',
|
|
|
|
'Content-Type': 'text/plain'})
|
|
|
|
with mocked_http_conn(
|
|
|
|
200, give_connect=capture_updates) as fake_conn:
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertRaises(StopIteration, fake_conn.code_iter.next)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
self.assertEquals(1, len(container_updates))
|
|
|
|
for update in container_updates:
|
|
|
|
ip, port, method, path, headers = update
|
|
|
|
self.assertEqual(ip, '10.0.0.1')
|
|
|
|
self.assertEqual(port, '8080')
|
|
|
|
self.assertEqual(method, 'PUT')
|
|
|
|
self.assertEqual(path, '/sda1/p/a/c/o')
|
|
|
|
expected = {
|
|
|
|
'X-Size': len('test1'),
|
|
|
|
'X-Etag': md5('test1').hexdigest(),
|
|
|
|
'X-Content-Type': 'text/plain',
|
|
|
|
'X-Timestamp': create_timestamp,
|
|
|
|
}
|
|
|
|
for key, value in expected.items():
|
|
|
|
self.assertEqual(headers[key], str(value))
|
|
|
|
container_updates = [] # reset
|
|
|
|
# read back object
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='GET')
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
self.assertEqual(resp.headers['X-Timestamp'],
|
|
|
|
utils.Timestamp(create_timestamp).normal)
|
|
|
|
self.assertEqual(resp.headers['X-Backend-Timestamp'],
|
|
|
|
create_timestamp)
|
|
|
|
self.assertEqual(resp.body, 'test1')
|
|
|
|
# send an update with an offset
|
|
|
|
offset_timestamp = utils.Timestamp(
|
|
|
|
create_timestamp, offset=1).internal
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='PUT', body='test2',
|
|
|
|
headers={'X-Timestamp': offset_timestamp,
|
|
|
|
'X-Container-Host': '10.0.0.1:8080',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Partition': 'p',
|
|
|
|
'Content-Type': 'text/html'})
|
|
|
|
with mocked_http_conn(
|
|
|
|
200, give_connect=capture_updates) as fake_conn:
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertRaises(StopIteration, fake_conn.code_iter.next)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
self.assertEquals(1, len(container_updates))
|
|
|
|
for update in container_updates:
|
|
|
|
ip, port, method, path, headers = update
|
|
|
|
self.assertEqual(ip, '10.0.0.1')
|
|
|
|
self.assertEqual(port, '8080')
|
|
|
|
self.assertEqual(method, 'PUT')
|
|
|
|
self.assertEqual(path, '/sda1/p/a/c/o')
|
|
|
|
expected = {
|
|
|
|
'X-Size': len('test2'),
|
|
|
|
'X-Etag': md5('test2').hexdigest(),
|
|
|
|
'X-Content-Type': 'text/html',
|
|
|
|
'X-Timestamp': offset_timestamp,
|
|
|
|
}
|
|
|
|
for key, value in expected.items():
|
|
|
|
self.assertEqual(headers[key], str(value))
|
|
|
|
container_updates = [] # reset
|
|
|
|
# read back new offset
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='GET')
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
self.assertEqual(resp.headers['X-Timestamp'],
|
|
|
|
utils.Timestamp(offset_timestamp).normal)
|
|
|
|
self.assertEqual(resp.headers['X-Backend-Timestamp'],
|
|
|
|
offset_timestamp)
|
|
|
|
self.assertEqual(resp.body, 'test2')
|
|
|
|
# now overwrite with a newer time
|
|
|
|
overwrite_timestamp = ts.next()
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='PUT', body='test3',
|
|
|
|
headers={'X-Timestamp': overwrite_timestamp,
|
|
|
|
'X-Container-Host': '10.0.0.1:8080',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Partition': 'p',
|
|
|
|
'Content-Type': 'text/enriched'})
|
|
|
|
with mocked_http_conn(
|
|
|
|
200, give_connect=capture_updates) as fake_conn:
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertRaises(StopIteration, fake_conn.code_iter.next)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
self.assertEquals(1, len(container_updates))
|
|
|
|
for update in container_updates:
|
|
|
|
ip, port, method, path, headers = update
|
|
|
|
self.assertEqual(ip, '10.0.0.1')
|
|
|
|
self.assertEqual(port, '8080')
|
|
|
|
self.assertEqual(method, 'PUT')
|
|
|
|
self.assertEqual(path, '/sda1/p/a/c/o')
|
|
|
|
expected = {
|
|
|
|
'X-Size': len('test3'),
|
|
|
|
'X-Etag': md5('test3').hexdigest(),
|
|
|
|
'X-Content-Type': 'text/enriched',
|
|
|
|
'X-Timestamp': overwrite_timestamp,
|
|
|
|
}
|
|
|
|
for key, value in expected.items():
|
|
|
|
self.assertEqual(headers[key], str(value))
|
|
|
|
container_updates = [] # reset
|
|
|
|
# read back overwrite
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='GET')
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
self.assertEqual(resp.headers['X-Timestamp'],
|
|
|
|
utils.Timestamp(overwrite_timestamp).normal)
|
|
|
|
self.assertEqual(resp.headers['X-Backend-Timestamp'],
|
|
|
|
overwrite_timestamp)
|
|
|
|
self.assertEqual(resp.body, 'test3')
|
|
|
|
# delete with an offset
|
|
|
|
offset_delete = utils.Timestamp(overwrite_timestamp,
|
|
|
|
offset=1).internal
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='DELETE',
|
|
|
|
headers={'X-Timestamp': offset_delete,
|
|
|
|
'X-Container-Host': '10.0.0.1:8080',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Partition': 'p'})
|
|
|
|
with mocked_http_conn(
|
|
|
|
200, give_connect=capture_updates) as fake_conn:
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertRaises(StopIteration, fake_conn.code_iter.next)
|
|
|
|
self.assertEqual(resp.status_int, 204)
|
|
|
|
self.assertEquals(1, len(container_updates))
|
|
|
|
for update in container_updates:
|
|
|
|
ip, port, method, path, headers = update
|
|
|
|
self.assertEqual(ip, '10.0.0.1')
|
|
|
|
self.assertEqual(port, '8080')
|
|
|
|
self.assertEqual(method, 'DELETE')
|
|
|
|
self.assertEqual(path, '/sda1/p/a/c/o')
|
|
|
|
expected = {
|
|
|
|
'X-Timestamp': offset_delete,
|
|
|
|
}
|
|
|
|
for key, value in expected.items():
|
|
|
|
self.assertEqual(headers[key], str(value))
|
|
|
|
container_updates = [] # reset
|
|
|
|
# read back offset delete
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='GET')
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
self.assertEqual(resp.headers['X-Timestamp'], None)
|
|
|
|
self.assertEqual(resp.headers['X-Backend-Timestamp'], offset_delete)
|
|
|
|
# and one more delete with a newer timestamp
|
|
|
|
delete_timestamp = ts.next()
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='DELETE',
|
|
|
|
headers={'X-Timestamp': delete_timestamp,
|
|
|
|
'X-Container-Host': '10.0.0.1:8080',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Partition': 'p'})
|
|
|
|
with mocked_http_conn(
|
|
|
|
200, give_connect=capture_updates) as fake_conn:
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertRaises(StopIteration, fake_conn.code_iter.next)
|
|
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
self.assertEquals(1, len(container_updates))
|
|
|
|
for update in container_updates:
|
|
|
|
ip, port, method, path, headers = update
|
|
|
|
self.assertEqual(ip, '10.0.0.1')
|
|
|
|
self.assertEqual(port, '8080')
|
|
|
|
self.assertEqual(method, 'DELETE')
|
|
|
|
self.assertEqual(path, '/sda1/p/a/c/o')
|
|
|
|
expected = {
|
|
|
|
'X-Timestamp': delete_timestamp,
|
|
|
|
}
|
|
|
|
for key, value in expected.items():
|
|
|
|
self.assertEqual(headers[key], str(value))
|
|
|
|
container_updates = [] # reset
|
|
|
|
# read back delete
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='GET')
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 404)
|
|
|
|
self.assertEqual(resp.headers['X-Timestamp'], None)
|
|
|
|
self.assertEqual(resp.headers['X-Backend-Timestamp'], delete_timestamp)
|
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
def test_call_bad_request(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test swift.obj.server.ObjectController.__call__
|
2010-07-12 17:03:45 -05:00
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
|
|
|
|
|
|
|
def start_response(*args):
|
2013-08-31 22:36:58 -04:00
|
|
|
"""Sends args to outbuf"""
|
2010-07-12 17:03:45 -05:00
|
|
|
outbuf.writelines(args)
|
|
|
|
|
|
|
|
self.object_controller.__call__({'REQUEST_METHOD': 'PUT',
|
|
|
|
'SCRIPT_NAME': '',
|
|
|
|
'PATH_INFO': '/sda1/p/a/c/o',
|
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False},
|
|
|
|
start_response)
|
|
|
|
self.assertEquals(errbuf.getvalue(), '')
|
|
|
|
self.assertEquals(outbuf.getvalue()[:4], '400 ')
|
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
def test_call_not_found(self):
|
2010-07-12 17:03:45 -05:00
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
|
|
|
|
def start_response(*args):
|
|
|
|
"""Sends args to outbuf"""
|
|
|
|
outbuf.writelines(args)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
self.object_controller.__call__({'REQUEST_METHOD': 'GET',
|
|
|
|
'SCRIPT_NAME': '',
|
|
|
|
'PATH_INFO': '/sda1/p/a/c/o',
|
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False},
|
|
|
|
start_response)
|
|
|
|
self.assertEquals(errbuf.getvalue(), '')
|
|
|
|
self.assertEquals(outbuf.getvalue()[:4], '404 ')
|
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
def test_call_bad_method(self):
|
2010-07-12 17:03:45 -05:00
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
|
|
|
|
def start_response(*args):
|
|
|
|
"""Sends args to outbuf"""
|
|
|
|
outbuf.writelines(args)
|
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
self.object_controller.__call__({'REQUEST_METHOD': 'INVALID',
|
|
|
|
'SCRIPT_NAME': '',
|
|
|
|
'PATH_INFO': '/sda1/p/a/c/o',
|
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False},
|
|
|
|
start_response)
|
|
|
|
self.assertEquals(errbuf.getvalue(), '')
|
|
|
|
self.assertEquals(outbuf.getvalue()[:4], '405 ')
|
2012-06-01 16:39:35 +02:00
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
def test_call_name_collision(self):
|
2013-03-22 09:00:40 +02:00
|
|
|
def my_check(*args):
|
|
|
|
return False
|
2013-07-23 16:41:45 -07:00
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
def my_hash_path(*args):
|
|
|
|
return md5('collide').hexdigest()
|
|
|
|
|
|
|
|
with mock.patch("swift.obj.diskfile.hash_path", my_hash_path):
|
|
|
|
with mock.patch("swift.obj.server.check_object_creation",
|
|
|
|
my_check):
|
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
|
|
|
|
|
|
|
def start_response(*args):
|
|
|
|
"""Sends args to outbuf"""
|
|
|
|
outbuf.writelines(args)
|
|
|
|
|
|
|
|
self.object_controller.__call__({
|
|
|
|
'REQUEST_METHOD': 'PUT',
|
|
|
|
'SCRIPT_NAME': '',
|
|
|
|
'PATH_INFO': '/sda1/p/a/c/o',
|
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'CONTENT_TYPE': 'text/html',
|
|
|
|
'HTTP_X_TIMESTAMP': normalize_timestamp(1.2),
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False},
|
|
|
|
start_response)
|
|
|
|
self.assertEquals(errbuf.getvalue(), '')
|
|
|
|
self.assertEquals(outbuf.getvalue()[:4], '201 ')
|
|
|
|
|
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
|
|
|
|
|
|
|
def start_response(*args):
|
|
|
|
"""Sends args to outbuf"""
|
|
|
|
outbuf.writelines(args)
|
|
|
|
|
|
|
|
self.object_controller.__call__({
|
|
|
|
'REQUEST_METHOD': 'PUT',
|
|
|
|
'SCRIPT_NAME': '',
|
|
|
|
'PATH_INFO': '/sda1/p/b/d/x',
|
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'CONTENT_TYPE': 'text/html',
|
|
|
|
'HTTP_X_TIMESTAMP': normalize_timestamp(1.3),
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False},
|
|
|
|
start_response)
|
|
|
|
self.assertEquals(errbuf.getvalue(), '')
|
|
|
|
self.assertEquals(outbuf.getvalue()[:4], '403 ')
|
2013-03-22 09:00:40 +02:00
|
|
|
|
2012-06-01 16:39:35 +02:00
|
|
|
def test_invalid_method_doesnt_exist(self):
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
2013-07-23 16:41:45 -07:00
|
|
|
|
2012-06-01 16:39:35 +02:00
|
|
|
def start_response(*args):
|
|
|
|
outbuf.writelines(args)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
|
|
|
|
self.object_controller.__call__({
|
|
|
|
'REQUEST_METHOD': 'method_doesnt_exist',
|
|
|
|
'PATH_INFO': '/sda1/p/a/c/o'},
|
2013-09-01 15:10:39 -04:00
|
|
|
start_response)
|
2012-06-01 16:39:35 +02:00
|
|
|
self.assertEquals(errbuf.getvalue(), '')
|
|
|
|
self.assertEquals(outbuf.getvalue()[:4], '405 ')
|
|
|
|
|
|
|
|
def test_invalid_method_is_not_public(self):
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
2013-07-23 16:41:45 -07:00
|
|
|
|
2012-06-01 16:39:35 +02:00
|
|
|
def start_response(*args):
|
|
|
|
outbuf.writelines(args)
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
|
2012-06-01 16:39:35 +02:00
|
|
|
self.object_controller.__call__({'REQUEST_METHOD': '__init__',
|
|
|
|
'PATH_INFO': '/sda1/p/a/c/o'},
|
|
|
|
start_response)
|
|
|
|
self.assertEquals(errbuf.getvalue(), '')
|
|
|
|
self.assertEquals(outbuf.getvalue()[:4], '405 ')
|
2010-07-12 17:03:45 -05:00
|
|
|
|
|
|
|
def test_chunked_put(self):
|
|
|
|
listener = listen(('localhost', 0))
|
|
|
|
port = listener.getsockname()[1]
|
|
|
|
killer = spawn(wsgi.server, listener, self.object_controller,
|
|
|
|
NullLogger())
|
|
|
|
sock = connect_tcp(('localhost', port))
|
|
|
|
fd = sock.makefile()
|
|
|
|
fd.write('PUT /sda1/p/a/c/o HTTP/1.1\r\nHost: localhost\r\n'
|
|
|
|
'Content-Type: text/plain\r\n'
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
'Connection: close\r\nX-Timestamp: %s\r\n'
|
2010-07-12 17:03:45 -05:00
|
|
|
'Transfer-Encoding: chunked\r\n\r\n'
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
'2\r\noh\r\n4\r\n hai\r\n0\r\n\r\n' % normalize_timestamp(
|
|
|
|
1.0))
|
2010-07-12 17:03:45 -05:00
|
|
|
fd.flush()
|
Rework to support RFC 2616 Sec 4.4 Message Length
RFC 2616 Sec 4.4 Message Length describes how the content-length and
transfer-encoding headers interact. Basically, if chunked transfer
encoding is used, the content-length header value is ignored and if
the content-length header is present, and the request is not using
chunked transfer-encoding, then the content-length must match the body
length.
The only Transfer-Coding value we support in the Transfer-Encoding
header (to date) is "chunked". RFC 2616 Sec 14.41 specifies that if
"multiple encodings have been applied to an entity, the
transfer-codings MUST be listed in the order in which they were
applied." Since we only supported "chunked". If the Transfer-Encoding
header value has multiple transfer-codings, we return a 501 (Not
Implemented) (see RFC 2616 Sec 3.6) without checking if chunked is the
last one specified. Finally, if transfer-encoding is anything but
"chunked", we return a 400 (Bad Request) to the client.
This patch adds a new method, message_length, to the swob request
object which will apply an algorithm based on RFC 2616 Sec 4.4
leveraging the existing content_length property.
In addition to these changes, the proxy server will now notice when
the message length specified by the content-length header is greater
than the configured object maximum size and fail the request with a
413, "Request Entity Too Large", before reading the entire body.
This work flows from https://review.openstack.org/27152.
Change-Id: I5d2a30b89092680dee9d946e1aafd017eaaef8c0
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-04-19 00:10:38 -04:00
|
|
|
headers = readuntil2crlfs(fd)
|
|
|
|
exp = 'HTTP/1.1 201'
|
|
|
|
self.assertEquals(headers[:len(exp)], exp)
|
|
|
|
sock = connect_tcp(('localhost', port))
|
|
|
|
fd = sock.makefile()
|
|
|
|
fd.write('GET /sda1/p/a/c/o HTTP/1.1\r\nHost: localhost\r\n'
|
|
|
|
'Connection: close\r\n\r\n')
|
|
|
|
fd.flush()
|
|
|
|
headers = readuntil2crlfs(fd)
|
|
|
|
exp = 'HTTP/1.1 200'
|
|
|
|
self.assertEquals(headers[:len(exp)], exp)
|
|
|
|
response = fd.read()
|
|
|
|
self.assertEquals(response, 'oh hai')
|
|
|
|
killer.kill()
|
|
|
|
|
|
|
|
def test_chunked_content_length_mismatch_zero(self):
|
|
|
|
listener = listen(('localhost', 0))
|
|
|
|
port = listener.getsockname()[1]
|
|
|
|
killer = spawn(wsgi.server, listener, self.object_controller,
|
|
|
|
NullLogger())
|
|
|
|
sock = connect_tcp(('localhost', port))
|
|
|
|
fd = sock.makefile()
|
|
|
|
fd.write('PUT /sda1/p/a/c/o HTTP/1.1\r\nHost: localhost\r\n'
|
|
|
|
'Content-Type: text/plain\r\n'
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
'Connection: close\r\nX-Timestamp: %s\r\n'
|
Rework to support RFC 2616 Sec 4.4 Message Length
RFC 2616 Sec 4.4 Message Length describes how the content-length and
transfer-encoding headers interact. Basically, if chunked transfer
encoding is used, the content-length header value is ignored and if
the content-length header is present, and the request is not using
chunked transfer-encoding, then the content-length must match the body
length.
The only Transfer-Coding value we support in the Transfer-Encoding
header (to date) is "chunked". RFC 2616 Sec 14.41 specifies that if
"multiple encodings have been applied to an entity, the
transfer-codings MUST be listed in the order in which they were
applied." Since we only supported "chunked". If the Transfer-Encoding
header value has multiple transfer-codings, we return a 501 (Not
Implemented) (see RFC 2616 Sec 3.6) without checking if chunked is the
last one specified. Finally, if transfer-encoding is anything but
"chunked", we return a 400 (Bad Request) to the client.
This patch adds a new method, message_length, to the swob request
object which will apply an algorithm based on RFC 2616 Sec 4.4
leveraging the existing content_length property.
In addition to these changes, the proxy server will now notice when
the message length specified by the content-length header is greater
than the configured object maximum size and fail the request with a
413, "Request Entity Too Large", before reading the entire body.
This work flows from https://review.openstack.org/27152.
Change-Id: I5d2a30b89092680dee9d946e1aafd017eaaef8c0
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-04-19 00:10:38 -04:00
|
|
|
'Content-Length: 0\r\n'
|
|
|
|
'Transfer-Encoding: chunked\r\n\r\n'
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
'2\r\noh\r\n4\r\n hai\r\n0\r\n\r\n' % normalize_timestamp(
|
|
|
|
1.0))
|
Rework to support RFC 2616 Sec 4.4 Message Length
RFC 2616 Sec 4.4 Message Length describes how the content-length and
transfer-encoding headers interact. Basically, if chunked transfer
encoding is used, the content-length header value is ignored and if
the content-length header is present, and the request is not using
chunked transfer-encoding, then the content-length must match the body
length.
The only Transfer-Coding value we support in the Transfer-Encoding
header (to date) is "chunked". RFC 2616 Sec 14.41 specifies that if
"multiple encodings have been applied to an entity, the
transfer-codings MUST be listed in the order in which they were
applied." Since we only supported "chunked". If the Transfer-Encoding
header value has multiple transfer-codings, we return a 501 (Not
Implemented) (see RFC 2616 Sec 3.6) without checking if chunked is the
last one specified. Finally, if transfer-encoding is anything but
"chunked", we return a 400 (Bad Request) to the client.
This patch adds a new method, message_length, to the swob request
object which will apply an algorithm based on RFC 2616 Sec 4.4
leveraging the existing content_length property.
In addition to these changes, the proxy server will now notice when
the message length specified by the content-length header is greater
than the configured object maximum size and fail the request with a
413, "Request Entity Too Large", before reading the entire body.
This work flows from https://review.openstack.org/27152.
Change-Id: I5d2a30b89092680dee9d946e1aafd017eaaef8c0
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-04-19 00:10:38 -04:00
|
|
|
fd.flush()
|
|
|
|
headers = readuntil2crlfs(fd)
|
|
|
|
exp = 'HTTP/1.1 201'
|
|
|
|
self.assertEquals(headers[:len(exp)], exp)
|
2010-07-12 17:03:45 -05:00
|
|
|
sock = connect_tcp(('localhost', port))
|
|
|
|
fd = sock.makefile()
|
|
|
|
fd.write('GET /sda1/p/a/c/o HTTP/1.1\r\nHost: localhost\r\n'
|
|
|
|
'Connection: close\r\n\r\n')
|
|
|
|
fd.flush()
|
Rework to support RFC 2616 Sec 4.4 Message Length
RFC 2616 Sec 4.4 Message Length describes how the content-length and
transfer-encoding headers interact. Basically, if chunked transfer
encoding is used, the content-length header value is ignored and if
the content-length header is present, and the request is not using
chunked transfer-encoding, then the content-length must match the body
length.
The only Transfer-Coding value we support in the Transfer-Encoding
header (to date) is "chunked". RFC 2616 Sec 14.41 specifies that if
"multiple encodings have been applied to an entity, the
transfer-codings MUST be listed in the order in which they were
applied." Since we only supported "chunked". If the Transfer-Encoding
header value has multiple transfer-codings, we return a 501 (Not
Implemented) (see RFC 2616 Sec 3.6) without checking if chunked is the
last one specified. Finally, if transfer-encoding is anything but
"chunked", we return a 400 (Bad Request) to the client.
This patch adds a new method, message_length, to the swob request
object which will apply an algorithm based on RFC 2616 Sec 4.4
leveraging the existing content_length property.
In addition to these changes, the proxy server will now notice when
the message length specified by the content-length header is greater
than the configured object maximum size and fail the request with a
413, "Request Entity Too Large", before reading the entire body.
This work flows from https://review.openstack.org/27152.
Change-Id: I5d2a30b89092680dee9d946e1aafd017eaaef8c0
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-04-19 00:10:38 -04:00
|
|
|
headers = readuntil2crlfs(fd)
|
|
|
|
exp = 'HTTP/1.1 200'
|
|
|
|
self.assertEquals(headers[:len(exp)], exp)
|
2010-07-12 17:03:45 -05:00
|
|
|
response = fd.read()
|
|
|
|
self.assertEquals(response, 'oh hai')
|
|
|
|
killer.kill()
|
|
|
|
|
|
|
|
def test_max_object_name_length(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
2012-09-05 20:49:50 -07:00
|
|
|
max_name_len = constraints.MAX_OBJECT_NAME_LENGTH
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/' + ('1' * max_name_len),
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'DATA'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/' + ('2' * (max_name_len + 1)),
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
2010-07-12 17:03:45 -05:00
|
|
|
req.body = 'DATA'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
|
|
|
|
def test_max_upload_time(self):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2014-03-03 17:28:48 -08:00
|
|
|
class SlowBody(object):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def __init__(self):
|
|
|
|
self.sent = 0
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def read(self, size=-1):
|
|
|
|
if self.sent < 4:
|
|
|
|
sleep(0.1)
|
|
|
|
self.sent += 1
|
|
|
|
return ' '
|
|
|
|
return ''
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2010-07-12 17:03:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'PUT', 'wsgi.input': SlowBody()},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '4', 'Content-Type': 'text/plain'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
self.object_controller.max_upload_time = 0.1
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2010-07-12 17:03:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'PUT', 'wsgi.input': SlowBody()},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '4', 'Content-Type': 'text/plain'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 408)
|
|
|
|
|
|
|
|
def test_short_body(self):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2014-03-03 17:28:48 -08:00
|
|
|
class ShortBody(object):
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def __init__(self):
|
|
|
|
self.sent = False
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
def read(self, size=-1):
|
|
|
|
if not self.sent:
|
|
|
|
self.sent = True
|
|
|
|
return ' '
|
|
|
|
return ''
|
2010-11-16 15:35:39 -08:00
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2010-07-12 17:03:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'PUT', 'wsgi.input': ShortBody()},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '4', 'Content-Type': 'text/plain'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 499)
|
|
|
|
|
|
|
|
def test_bad_sinces(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2010-07-12 17:03:45 -05:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '4', 'Content-Type': 'text/plain'},
|
|
|
|
body=' ')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
2010-07-12 17:03:45 -05:00
|
|
|
headers={'If-Unmodified-Since': 'Not a valid date'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
2010-07-12 17:03:45 -05:00
|
|
|
headers={'If-Modified-Since': 'Not a valid date'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
Allow pre-1970 dates in If-[Un]Modified-Since
If I want to fetch an object only if it is newer than the first moon
landing, I send a GET request with header:
If-Modified-Since: Sun, 20 Jul 1969 20:18:00 UTC
Since that date is older than Swift, I expect a 2xx response. However,
I get a 412, which isn't even a valid thing to do for
If-Modified-Since; it should either be 2xx or 304. This is because of
two problems:
(a) Swift treats pre-1970 dates as invalid, and
(b) Swift returns 412 when a date is invalid instead of ignoring it.
This commit makes it so any time between datetime.datetime.min and
datetime.datetime.max is an acceptable value for If-Modified-Since and
If-Unmodified-Since. Dates outside that date range are treated as
invalid headers and thus are ignored, as RFC 2616 section 14.28
requires ("If the specified date is invalid, the header is ignored").
This only works for dates that the Python standard library can parse,
which on my machine is 01 Jan 1 to 31 Dec 9999. Eliminating those
restrictions would require implementing our own date parsing and
comparison, and that's almost certainly not worth it.
Change-Id: I4cb4903c4e5e3b6b3c9506c2cabbfbda62e82f35
2014-03-03 11:25:43 -08:00
|
|
|
|
|
|
|
too_big_date_list = list(datetime.datetime.max.timetuple())
|
|
|
|
too_big_date_list[0] += 1 # bump up the year
|
|
|
|
too_big_date = strftime(
|
|
|
|
"%a, %d %b %Y %H:%M:%S UTC", struct_time(too_big_date_list))
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
Allow pre-1970 dates in If-[Un]Modified-Since
If I want to fetch an object only if it is newer than the first moon
landing, I send a GET request with header:
If-Modified-Since: Sun, 20 Jul 1969 20:18:00 UTC
Since that date is older than Swift, I expect a 2xx response. However,
I get a 412, which isn't even a valid thing to do for
If-Modified-Since; it should either be 2xx or 304. This is because of
two problems:
(a) Swift treats pre-1970 dates as invalid, and
(b) Swift returns 412 when a date is invalid instead of ignoring it.
This commit makes it so any time between datetime.datetime.min and
datetime.datetime.max is an acceptable value for If-Modified-Since and
If-Unmodified-Since. Dates outside that date range are treated as
invalid headers and thus are ignored, as RFC 2616 section 14.28
requires ("If the specified date is invalid, the header is ignored").
This only works for dates that the Python standard library can parse,
which on my machine is 01 Jan 1 to 31 Dec 9999. Eliminating those
restrictions would require implementing our own date parsing and
comparison, and that's almost certainly not worth it.
Change-Id: I4cb4903c4e5e3b6b3c9506c2cabbfbda62e82f35
2014-03-03 11:25:43 -08:00
|
|
|
headers={'If-Unmodified-Since': too_big_date})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
Allow pre-1970 dates in If-[Un]Modified-Since
If I want to fetch an object only if it is newer than the first moon
landing, I send a GET request with header:
If-Modified-Since: Sun, 20 Jul 1969 20:18:00 UTC
Since that date is older than Swift, I expect a 2xx response. However,
I get a 412, which isn't even a valid thing to do for
If-Modified-Since; it should either be 2xx or 304. This is because of
two problems:
(a) Swift treats pre-1970 dates as invalid, and
(b) Swift returns 412 when a date is invalid instead of ignoring it.
This commit makes it so any time between datetime.datetime.min and
datetime.datetime.max is an acceptable value for If-Modified-Since and
If-Unmodified-Since. Dates outside that date range are treated as
invalid headers and thus are ignored, as RFC 2616 section 14.28
requires ("If the specified date is invalid, the header is ignored").
This only works for dates that the Python standard library can parse,
which on my machine is 01 Jan 1 to 31 Dec 9999. Eliminating those
restrictions would require implementing our own date parsing and
comparison, and that's almost certainly not worth it.
Change-Id: I4cb4903c4e5e3b6b3c9506c2cabbfbda62e82f35
2014-03-03 11:25:43 -08:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
2010-07-12 17:03:45 -05:00
|
|
|
|
|
|
|
def test_content_encoding(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2010-07-12 17:03:45 -05:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '4', 'Content-Type': 'text/plain',
|
|
|
|
'Content-Encoding': 'gzip'},
|
|
|
|
body=' ')
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.headers['content-encoding'], 'gzip')
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'HEAD'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2010-07-12 17:03:45 -05:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.headers['content-encoding'], 'gzip')
|
|
|
|
|
2011-10-26 21:42:24 +00:00
|
|
|
def test_async_update_http_connect(self):
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
policy = random.choice(list(POLICIES))
|
|
|
|
self._stage_tmp_dir(policy)
|
2011-10-26 21:42:24 +00:00
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_http_connect(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
raise Exception('test')
|
|
|
|
|
|
|
|
orig_http_connect = object_server.http_connect
|
|
|
|
try:
|
|
|
|
object_server.http_connect = fake_http_connect
|
2013-09-01 15:10:39 -04:00
|
|
|
self.object_controller.async_update(
|
|
|
|
'PUT', 'a', 'c', 'o', '127.0.0.1:1234', 1, 'sdc1',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
{'x-timestamp': '1', 'x-out': 'set',
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': policy.idx}, 'sda1',
|
|
|
|
policy.idx)
|
2011-10-26 21:42:24 +00:00
|
|
|
finally:
|
|
|
|
object_server.http_connect = orig_http_connect
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
given_args,
|
|
|
|
['127.0.0.1', '1234', 'sdc1', 1, 'PUT', '/a/c/o', {
|
|
|
|
'x-timestamp': '1', 'x-out': 'set',
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'user-agent': 'object-server %s' % os.getpid(),
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': policy.idx}])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
2014-05-27 01:17:13 -07:00
|
|
|
@patch_policies([storage_policy.StoragePolicy(0, 'zero', True),
|
|
|
|
storage_policy.StoragePolicy(1, 'one'),
|
|
|
|
storage_policy.StoragePolicy(37, 'fantastico')])
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
def test_updating_multiple_delete_at_container_servers(self):
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
policy = random.choice(list(POLICIES))
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
self.object_controller.expiring_objects_account = 'exp'
|
|
|
|
self.object_controller.expiring_objects_container_divisor = 60
|
|
|
|
|
|
|
|
http_connect_args = []
|
2013-07-23 16:41:45 -07:00
|
|
|
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
def fake_http_connect(ipaddr, port, device, partition, method, path,
|
|
|
|
headers=None, query_string=None, ssl=False):
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
class SuccessfulFakeConn(object):
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
@property
|
|
|
|
def status(self):
|
|
|
|
return 200
|
|
|
|
|
|
|
|
def getresponse(self):
|
|
|
|
return self
|
|
|
|
|
|
|
|
def read(self):
|
|
|
|
return ''
|
|
|
|
|
|
|
|
captured_args = {'ipaddr': ipaddr, 'port': port,
|
|
|
|
'device': device, 'partition': partition,
|
|
|
|
'method': method, 'path': path, 'ssl': ssl,
|
|
|
|
'headers': headers, 'query_string': query_string}
|
|
|
|
|
|
|
|
http_connect_args.append(
|
2013-07-23 16:41:45 -07:00
|
|
|
dict((k, v) for k, v in captured_args.iteritems()
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
if v is not None))
|
|
|
|
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
return SuccessfulFakeConn()
|
|
|
|
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': '12345',
|
|
|
|
'Content-Type': 'application/burrito',
|
|
|
|
'Content-Length': '0',
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': policy.idx,
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
'X-Container-Partition': '20',
|
|
|
|
'X-Container-Host': '1.2.3.4:5',
|
|
|
|
'X-Container-Device': 'sdb1',
|
|
|
|
'X-Delete-At': 9999999999,
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At-Container': '9999999960',
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
'X-Delete-At-Host': "10.1.1.1:6001,10.2.2.2:6002",
|
|
|
|
'X-Delete-At-Partition': '6237',
|
|
|
|
'X-Delete-At-Device': 'sdp,sdq'})
|
|
|
|
|
2014-05-27 01:17:13 -07:00
|
|
|
with mock.patch.object(object_server, 'http_connect',
|
|
|
|
fake_http_connect):
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
|
|
|
|
http_connect_args.sort(key=operator.itemgetter('ipaddr'))
|
|
|
|
|
|
|
|
self.assertEquals(len(http_connect_args), 3)
|
|
|
|
self.assertEquals(
|
|
|
|
http_connect_args[0],
|
|
|
|
{'ipaddr': '1.2.3.4',
|
|
|
|
'port': '5',
|
|
|
|
'path': '/a/c/o',
|
|
|
|
'device': 'sdb1',
|
|
|
|
'partition': '20',
|
|
|
|
'method': 'PUT',
|
|
|
|
'ssl': False,
|
2013-09-01 15:10:39 -04:00
|
|
|
'headers': HeaderKeyDict({
|
|
|
|
'x-content-type': 'application/burrito',
|
|
|
|
'x-etag': 'd41d8cd98f00b204e9800998ecf8427e',
|
|
|
|
'x-size': '0',
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('12345').internal,
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': '37',
|
2013-09-01 15:10:39 -04:00
|
|
|
'referer': 'PUT http://localhost/sda1/p/a/c/o',
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'user-agent': 'object-server %d' % os.getpid(),
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': policy.idx,
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-trans-id': '-'})})
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
self.assertEquals(
|
|
|
|
http_connect_args[1],
|
|
|
|
{'ipaddr': '10.1.1.1',
|
|
|
|
'port': '6001',
|
|
|
|
'path': '/exp/9999999960/9999999999-a/c/o',
|
|
|
|
'device': 'sdp',
|
|
|
|
'partition': '6237',
|
|
|
|
'method': 'PUT',
|
|
|
|
'ssl': False,
|
2013-09-01 15:10:39 -04:00
|
|
|
'headers': HeaderKeyDict({
|
|
|
|
'x-content-type': 'text/plain',
|
|
|
|
'x-etag': 'd41d8cd98f00b204e9800998ecf8427e',
|
|
|
|
'x-size': '0',
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('12345').internal,
|
2013-09-01 15:10:39 -04:00
|
|
|
'referer': 'PUT http://localhost/sda1/p/a/c/o',
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'user-agent': 'object-server %d' % os.getpid(),
|
2014-06-23 12:52:50 -07:00
|
|
|
# system account storage policy is 0
|
|
|
|
'X-Backend-Storage-Policy-Index': 0,
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-trans-id': '-'})})
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
self.assertEquals(
|
|
|
|
http_connect_args[2],
|
|
|
|
{'ipaddr': '10.2.2.2',
|
|
|
|
'port': '6002',
|
|
|
|
'path': '/exp/9999999960/9999999999-a/c/o',
|
|
|
|
'device': 'sdq',
|
|
|
|
'partition': '6237',
|
|
|
|
'method': 'PUT',
|
|
|
|
'ssl': False,
|
2013-09-01 15:10:39 -04:00
|
|
|
'headers': HeaderKeyDict({
|
|
|
|
'x-content-type': 'text/plain',
|
|
|
|
'x-etag': 'd41d8cd98f00b204e9800998ecf8427e',
|
|
|
|
'x-size': '0',
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('12345').internal,
|
2013-09-01 15:10:39 -04:00
|
|
|
'referer': 'PUT http://localhost/sda1/p/a/c/o',
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'user-agent': 'object-server %d' % os.getpid(),
|
2014-06-23 12:52:50 -07:00
|
|
|
# system account storage policy is 0
|
|
|
|
'X-Backend-Storage-Policy-Index': 0,
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-trans-id': '-'})})
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
|
2014-05-27 01:17:13 -07:00
|
|
|
@patch_policies([storage_policy.StoragePolicy(0, 'zero', True),
|
|
|
|
storage_policy.StoragePolicy(1, 'one'),
|
|
|
|
storage_policy.StoragePolicy(26, 'twice-thirteen')])
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
def test_updating_multiple_container_servers(self):
|
|
|
|
http_connect_args = []
|
2013-07-23 16:41:45 -07:00
|
|
|
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
def fake_http_connect(ipaddr, port, device, partition, method, path,
|
|
|
|
headers=None, query_string=None, ssl=False):
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
class SuccessfulFakeConn(object):
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
@property
|
|
|
|
def status(self):
|
|
|
|
return 200
|
|
|
|
|
|
|
|
def getresponse(self):
|
|
|
|
return self
|
|
|
|
|
|
|
|
def read(self):
|
|
|
|
return ''
|
|
|
|
|
|
|
|
captured_args = {'ipaddr': ipaddr, 'port': port,
|
|
|
|
'device': device, 'partition': partition,
|
|
|
|
'method': method, 'path': path, 'ssl': ssl,
|
|
|
|
'headers': headers, 'query_string': query_string}
|
|
|
|
|
|
|
|
http_connect_args.append(
|
2013-07-23 16:41:45 -07:00
|
|
|
dict((k, v) for k, v in captured_args.iteritems()
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
if v is not None))
|
|
|
|
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
return SuccessfulFakeConn()
|
|
|
|
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': '12345',
|
|
|
|
'Content-Type': 'application/burrito',
|
|
|
|
'Content-Length': '0',
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': '26',
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
'X-Container-Partition': '20',
|
|
|
|
'X-Container-Host': '1.2.3.4:5, 6.7.8.9:10',
|
|
|
|
'X-Container-Device': 'sdb1, sdf1'})
|
|
|
|
|
2014-05-27 01:17:13 -07:00
|
|
|
with mock.patch.object(object_server, 'http_connect',
|
|
|
|
fake_http_connect):
|
|
|
|
req.get_response(self.object_controller)
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
|
|
|
|
http_connect_args.sort(key=operator.itemgetter('ipaddr'))
|
|
|
|
|
|
|
|
self.assertEquals(len(http_connect_args), 2)
|
|
|
|
self.assertEquals(
|
|
|
|
http_connect_args[0],
|
|
|
|
{'ipaddr': '1.2.3.4',
|
|
|
|
'port': '5',
|
|
|
|
'path': '/a/c/o',
|
|
|
|
'device': 'sdb1',
|
|
|
|
'partition': '20',
|
|
|
|
'method': 'PUT',
|
|
|
|
'ssl': False,
|
2013-09-01 15:10:39 -04:00
|
|
|
'headers': HeaderKeyDict({
|
|
|
|
'x-content-type': 'application/burrito',
|
|
|
|
'x-etag': 'd41d8cd98f00b204e9800998ecf8427e',
|
|
|
|
'x-size': '0',
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('12345').internal,
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': '26',
|
2013-09-01 15:10:39 -04:00
|
|
|
'referer': 'PUT http://localhost/sda1/p/a/c/o',
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'user-agent': 'object-server %d' % os.getpid(),
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-trans-id': '-'})})
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
self.assertEquals(
|
|
|
|
http_connect_args[1],
|
|
|
|
{'ipaddr': '6.7.8.9',
|
|
|
|
'port': '10',
|
|
|
|
'path': '/a/c/o',
|
|
|
|
'device': 'sdf1',
|
|
|
|
'partition': '20',
|
|
|
|
'method': 'PUT',
|
|
|
|
'ssl': False,
|
2013-09-01 15:10:39 -04:00
|
|
|
'headers': HeaderKeyDict({
|
|
|
|
'x-content-type': 'application/burrito',
|
|
|
|
'x-etag': 'd41d8cd98f00b204e9800998ecf8427e',
|
|
|
|
'x-size': '0',
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('12345').internal,
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': '26',
|
2013-09-01 15:10:39 -04:00
|
|
|
'referer': 'PUT http://localhost/sda1/p/a/c/o',
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'user-agent': 'object-server %d' % os.getpid(),
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-trans-id': '-'})})
|
Allow for multiple X-(Account|Container)-* headers.
When the number of account/container or container/object replicas are
different, Swift had a few misbehaviors. This commit fixes them.
* On an object PUT/POST/DELETE, if there were 3 object replicas and
only 2 container replicas, then only 2 requests would be made to
object servers. Now, 3 requests will be made, but the third won't
have any X-Container-* headers in it.
* On an object PUT/POST/DELETE, if there were 3 object replicas and 4
container replicas, then only 3/4 container servers would receive
immediate updates; the fourth would be ignored. Now one of the
object servers will receive multiple (comma-separated) values in the
X-Container-* headers and it will attempt to contact both of them.
One side effect is that multiple async_pendings may be written for
updates to the same object. They'll have differing timestamps,
though, so all but the newest will be deleted unread. To trigger
this behavior, you have to have more container replicas than object
replicas, 2 or more of the container servers must be down, and the
headers sent to one object server must reference 2 or more down
container servers; it's unlikely enough and the consequences are so
minor that it didn't seem worth fixing.
The situation with account/containers is analogous, only without the
async_pendings.
Change-Id: I98bc2de93fb6b2346d6de1d764213d7563653e8d
2012-12-12 17:47:04 -08:00
|
|
|
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
def test_object_delete_at_aysnc_update(self):
|
|
|
|
policy = random.choice(list(POLICIES))
|
2014-06-10 22:17:47 -07:00
|
|
|
ts = (utils.Timestamp(t) for t in
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
itertools.count(int(time())))
|
|
|
|
|
|
|
|
container_updates = []
|
|
|
|
|
|
|
|
def capture_updates(ip, port, method, path, headers, *args, **kwargs):
|
|
|
|
container_updates.append((ip, port, method, path, headers))
|
|
|
|
|
2014-06-10 22:17:47 -07:00
|
|
|
put_timestamp = ts.next().internal
|
|
|
|
delete_at_timestamp = utils.normalize_delete_at_timestamp(
|
|
|
|
ts.next().normal)
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
delete_at_container = (
|
|
|
|
int(delete_at_timestamp) /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', method='PUT', body='', headers={
|
|
|
|
'Content-Type': 'text/plain',
|
|
|
|
'X-Timestamp': put_timestamp,
|
|
|
|
'X-Container-Host': '10.0.0.1:6001',
|
|
|
|
'X-Container-Device': 'sda1',
|
|
|
|
'X-Container-Partition': 'p',
|
|
|
|
'X-Delete-At': delete_at_timestamp,
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
|
|
|
'X-Delete-At-Partition': 'p',
|
|
|
|
'X-Delete-At-Host': '10.0.0.2:6002',
|
|
|
|
'X-Delete-At-Device': 'sda1',
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': int(policy),
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
})
|
|
|
|
with mocked_http_conn(
|
|
|
|
500, 500, give_connect=capture_updates) as fake_conn:
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertRaises(StopIteration, fake_conn.code_iter.next)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
self.assertEquals(2, len(container_updates))
|
|
|
|
delete_at_update, container_update = container_updates
|
|
|
|
# delete_at_update
|
|
|
|
ip, port, method, path, headers = delete_at_update
|
|
|
|
self.assertEqual(ip, '10.0.0.2')
|
|
|
|
self.assertEqual(port, '6002')
|
|
|
|
self.assertEqual(method, 'PUT')
|
|
|
|
self.assertEqual(path, '/sda1/p/.expiring_objects/%s/%s-a/c/o' %
|
|
|
|
(delete_at_container, delete_at_timestamp))
|
|
|
|
expected = {
|
|
|
|
'X-Timestamp': put_timestamp,
|
2014-06-23 12:52:50 -07:00
|
|
|
# system account storage policy is 0
|
|
|
|
'X-Backend-Storage-Policy-Index': 0,
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
}
|
|
|
|
for key, value in expected.items():
|
|
|
|
self.assertEqual(headers[key], str(value))
|
|
|
|
# container_update
|
|
|
|
ip, port, method, path, headers = container_update
|
|
|
|
self.assertEqual(ip, '10.0.0.1')
|
|
|
|
self.assertEqual(port, '6001')
|
|
|
|
self.assertEqual(method, 'PUT')
|
|
|
|
self.assertEqual(path, '/sda1/p/a/c/o')
|
|
|
|
expected = {
|
|
|
|
'X-Timestamp': put_timestamp,
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': int(policy),
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
}
|
|
|
|
for key, value in expected.items():
|
|
|
|
self.assertEqual(headers[key], str(value))
|
|
|
|
# check async pendings
|
|
|
|
async_dir = os.path.join(self.testdir, 'sda1',
|
|
|
|
diskfile.get_async_dir(policy.idx))
|
|
|
|
found_files = []
|
|
|
|
for root, dirs, files in os.walk(async_dir):
|
|
|
|
for f in files:
|
|
|
|
async_file = os.path.join(root, f)
|
|
|
|
found_files.append(async_file)
|
|
|
|
data = pickle.load(open(async_file))
|
|
|
|
if data['account'] == 'a':
|
|
|
|
self.assertEquals(
|
2014-06-23 12:52:50 -07:00
|
|
|
int(data['headers']
|
|
|
|
['X-Backend-Storage-Policy-Index']), policy.idx)
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
elif data['account'] == '.expiring_objects':
|
|
|
|
self.assertEquals(
|
2014-06-23 12:52:50 -07:00
|
|
|
int(data['headers']
|
|
|
|
['X-Backend-Storage-Policy-Index']), 0)
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
else:
|
|
|
|
self.fail('unexpected async pending data')
|
|
|
|
self.assertEqual(2, len(found_files))
|
|
|
|
|
2011-10-26 21:42:24 +00:00
|
|
|
def test_async_update_saves_on_exception(self):
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
policy = random.choice(list(POLICIES))
|
|
|
|
self._stage_tmp_dir(policy)
|
2013-10-07 12:10:31 +00:00
|
|
|
_prefix = utils.HASH_PATH_PREFIX
|
|
|
|
utils.HASH_PATH_PREFIX = ''
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
def fake_http_connect(*args):
|
|
|
|
raise Exception('test')
|
|
|
|
|
|
|
|
orig_http_connect = object_server.http_connect
|
|
|
|
try:
|
|
|
|
object_server.http_connect = fake_http_connect
|
2013-09-01 15:10:39 -04:00
|
|
|
self.object_controller.async_update(
|
|
|
|
'PUT', 'a', 'c', 'o', '127.0.0.1:1234', 1, 'sdc1',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
{'x-timestamp': '1', 'x-out': 'set',
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': policy.idx}, 'sda1',
|
|
|
|
policy.idx)
|
2011-10-26 21:42:24 +00:00
|
|
|
finally:
|
|
|
|
object_server.http_connect = orig_http_connect
|
2013-10-07 12:10:31 +00:00
|
|
|
utils.HASH_PATH_PREFIX = _prefix
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
async_dir = diskfile.get_async_dir(policy.idx)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(
|
2013-09-01 15:10:39 -04:00
|
|
|
pickle.load(open(os.path.join(
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
self.testdir, 'sda1', async_dir, 'a83',
|
2014-06-10 22:17:47 -07:00
|
|
|
'06fbf0b514e5199dfc4e00f42eb5ea83-%s' %
|
|
|
|
utils.Timestamp(1).internal))),
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
{'headers': {'x-timestamp': '1', 'x-out': 'set',
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'user-agent': 'object-server %s' % os.getpid(),
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': policy.idx},
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
'account': 'a', 'container': 'c', 'obj': 'o', 'op': 'PUT'})
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
def test_async_update_saves_on_non_2xx(self):
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
policy = random.choice(list(POLICIES))
|
|
|
|
self._stage_tmp_dir(policy)
|
2013-10-07 12:10:31 +00:00
|
|
|
_prefix = utils.HASH_PATH_PREFIX
|
|
|
|
utils.HASH_PATH_PREFIX = ''
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
def fake_http_connect(status):
|
|
|
|
|
|
|
|
class FakeConn(object):
|
|
|
|
|
|
|
|
def __init__(self, status):
|
|
|
|
self.status = status
|
|
|
|
|
|
|
|
def getresponse(self):
|
|
|
|
return self
|
|
|
|
|
|
|
|
def read(self):
|
|
|
|
return ''
|
|
|
|
|
|
|
|
return lambda *args: FakeConn(status)
|
|
|
|
|
|
|
|
orig_http_connect = object_server.http_connect
|
|
|
|
try:
|
|
|
|
for status in (199, 300, 503):
|
|
|
|
object_server.http_connect = fake_http_connect(status)
|
2013-09-01 15:10:39 -04:00
|
|
|
self.object_controller.async_update(
|
|
|
|
'PUT', 'a', 'c', 'o', '127.0.0.1:1234', 1, 'sdc1',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
{'x-timestamp': '1', 'x-out': str(status),
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': policy.idx}, 'sda1',
|
|
|
|
policy.idx)
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
async_dir = diskfile.get_async_dir(policy.idx)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(
|
2013-09-01 15:10:39 -04:00
|
|
|
pickle.load(open(os.path.join(
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
self.testdir, 'sda1', async_dir, 'a83',
|
2014-06-10 22:17:47 -07:00
|
|
|
'06fbf0b514e5199dfc4e00f42eb5ea83-%s' %
|
|
|
|
utils.Timestamp(1).internal))),
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
{'headers': {'x-timestamp': '1', 'x-out': str(status),
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'user-agent':
|
|
|
|
'object-server %s' % os.getpid(),
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index':
|
|
|
|
policy.idx},
|
2011-10-26 21:42:24 +00:00
|
|
|
'account': 'a', 'container': 'c', 'obj': 'o',
|
|
|
|
'op': 'PUT'})
|
|
|
|
finally:
|
|
|
|
object_server.http_connect = orig_http_connect
|
2013-10-07 12:10:31 +00:00
|
|
|
utils.HASH_PATH_PREFIX = _prefix
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
def test_async_update_does_not_save_on_2xx(self):
|
2013-12-11 16:59:59 -05:00
|
|
|
_prefix = utils.HASH_PATH_PREFIX
|
|
|
|
utils.HASH_PATH_PREFIX = ''
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
def fake_http_connect(status):
|
|
|
|
|
|
|
|
class FakeConn(object):
|
|
|
|
|
|
|
|
def __init__(self, status):
|
|
|
|
self.status = status
|
|
|
|
|
|
|
|
def getresponse(self):
|
|
|
|
return self
|
|
|
|
|
|
|
|
def read(self):
|
|
|
|
return ''
|
|
|
|
|
|
|
|
return lambda *args: FakeConn(status)
|
|
|
|
|
|
|
|
orig_http_connect = object_server.http_connect
|
|
|
|
try:
|
|
|
|
for status in (200, 299):
|
|
|
|
object_server.http_connect = fake_http_connect(status)
|
2013-09-01 15:10:39 -04:00
|
|
|
self.object_controller.async_update(
|
|
|
|
'PUT', 'a', 'c', 'o', '127.0.0.1:1234', 1, 'sdc1',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
{'x-timestamp': '1', 'x-out': str(status)}, 'sda1', 0)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertFalse(
|
2013-09-01 15:10:39 -04:00
|
|
|
os.path.exists(os.path.join(
|
|
|
|
self.testdir, 'sda1', 'async_pending', 'a83',
|
2011-10-26 21:42:24 +00:00
|
|
|
'06fbf0b514e5199dfc4e00f42eb5ea83-0000000001.00000')))
|
|
|
|
finally:
|
|
|
|
object_server.http_connect = orig_http_connect
|
2013-12-11 16:59:59 -05:00
|
|
|
utils.HASH_PATH_PREFIX = _prefix
|
|
|
|
|
|
|
|
def test_async_update_saves_on_timeout(self):
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
policy = random.choice(list(POLICIES))
|
|
|
|
self._stage_tmp_dir(policy)
|
2013-12-11 16:59:59 -05:00
|
|
|
_prefix = utils.HASH_PATH_PREFIX
|
|
|
|
utils.HASH_PATH_PREFIX = ''
|
|
|
|
|
|
|
|
def fake_http_connect():
|
|
|
|
|
|
|
|
class FakeConn(object):
|
|
|
|
|
|
|
|
def getresponse(self):
|
|
|
|
return sleep(1)
|
|
|
|
|
|
|
|
return lambda *args: FakeConn()
|
|
|
|
|
|
|
|
orig_http_connect = object_server.http_connect
|
|
|
|
try:
|
|
|
|
for status in (200, 299):
|
|
|
|
object_server.http_connect = fake_http_connect()
|
|
|
|
self.object_controller.node_timeout = 0.001
|
|
|
|
self.object_controller.async_update(
|
|
|
|
'PUT', 'a', 'c', 'o', '127.0.0.1:1234', 1, 'sdc1',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
{'x-timestamp': '1', 'x-out': str(status)}, 'sda1',
|
|
|
|
policy.idx)
|
|
|
|
async_dir = diskfile.get_async_dir(int(policy))
|
2013-12-11 16:59:59 -05:00
|
|
|
self.assertTrue(
|
|
|
|
os.path.exists(os.path.join(
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
self.testdir, 'sda1', async_dir, 'a83',
|
2014-06-10 22:17:47 -07:00
|
|
|
'06fbf0b514e5199dfc4e00f42eb5ea83-%s' %
|
|
|
|
utils.Timestamp(1).internal)))
|
2013-12-11 16:59:59 -05:00
|
|
|
finally:
|
|
|
|
object_server.http_connect = orig_http_connect
|
|
|
|
utils.HASH_PATH_PREFIX = _prefix
|
2011-10-26 21:42:24 +00:00
|
|
|
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
def test_container_update_no_async_update(self):
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_async_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.async_update = fake_async_update
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/v1/a/c/o',
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '1234'})
|
2013-09-01 15:10:39 -04:00
|
|
|
self.object_controller.container_update(
|
|
|
|
'PUT', 'a', 'c', 'o', req, {
|
|
|
|
'x-size': '0', 'x-etag': 'd41d8cd98f00b204e9800998ecf8427e',
|
|
|
|
'x-content-type': 'text/plain', 'x-timestamp': '1'},
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'sda1', 0)
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
self.assertEquals(given_args, [])
|
|
|
|
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
def test_container_update_success(self):
|
|
|
|
container_updates = []
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
def capture_updates(ip, port, method, path, headers, *args, **kwargs):
|
|
|
|
container_updates.append((ip, port, method, path, headers))
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'/sda1/0/a/c/o',
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '123',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'X-Container-Host': 'chost:cport',
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
'X-Container-Partition': 'cpartition',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'X-Container-Device': 'cdevice',
|
|
|
|
'Content-Type': 'text/plain'}, body='')
|
|
|
|
with mocked_http_conn(
|
|
|
|
200, give_connect=capture_updates) as fake_conn:
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertRaises(StopIteration, fake_conn.code_iter.next)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
self.assertEqual(len(container_updates), 1)
|
|
|
|
ip, port, method, path, headers = container_updates[0]
|
|
|
|
self.assertEqual(ip, 'chost')
|
|
|
|
self.assertEqual(port, 'cport')
|
|
|
|
self.assertEqual(method, 'PUT')
|
|
|
|
self.assertEqual(path, '/cdevice/cpartition/a/c/o')
|
|
|
|
self.assertEqual(headers, HeaderKeyDict({
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'user-agent': 'object-server %s' % os.getpid(),
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'x-size': '0',
|
|
|
|
'x-etag': 'd41d8cd98f00b204e9800998ecf8427e',
|
|
|
|
'x-content-type': 'text/plain',
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp(1).internal,
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': '0', # default when not given
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'x-trans-id': '123',
|
|
|
|
'referer': 'PUT http://localhost/sda1/0/a/c/o'}))
|
|
|
|
|
|
|
|
def test_container_update_async(self):
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/0/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '123',
|
|
|
|
'X-Container-Host': 'chost:cport',
|
|
|
|
'X-Container-Partition': 'cpartition',
|
|
|
|
'X-Container-Device': 'cdevice',
|
|
|
|
'Content-Type': 'text/plain'}, body='')
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_pickle_async_update(*args):
|
|
|
|
given_args[:] = args
|
|
|
|
self.object_controller._diskfile_mgr.pickle_async_update = \
|
|
|
|
fake_pickle_async_update
|
|
|
|
with mocked_http_conn(500) as fake_conn:
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertRaises(StopIteration, fake_conn.code_iter.next)
|
|
|
|
self.assertEqual(resp.status_int, 201)
|
|
|
|
self.assertEqual(len(given_args), 7)
|
|
|
|
(objdevice, account, container, obj, data, timestamp,
|
|
|
|
policy_index) = given_args
|
|
|
|
self.assertEqual(objdevice, 'sda1')
|
|
|
|
self.assertEqual(account, 'a')
|
|
|
|
self.assertEqual(container, 'c')
|
|
|
|
self.assertEqual(obj, 'o')
|
2014-06-10 22:17:47 -07:00
|
|
|
self.assertEqual(timestamp, utils.Timestamp(1).internal)
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
self.assertEqual(policy_index, 0)
|
|
|
|
self.assertEqual(data, {
|
|
|
|
'headers': HeaderKeyDict({
|
|
|
|
'X-Size': '0',
|
Object services user-agent string uses full name
It does not appear that, aside from the user-agent string, the strings
"obj-server", "obj-updater", or "obj-replicator" (or "obj-<anything>"*)
appear in the swift code base, aside from the directory containing the
object services code being named "obj".
Furthermore, the container, account, and proxy services construct their
user-agent string, as reported in the logs, using their full name. In
addition, this full name also shows up as the name of the process via
"ps" or "top", etc., which can make it easier for admins to match log
entries with other tools.
For consistency, we update the object services to use an "object-"
prefix rather than "obj-" in its user agent string.
* obj-etag does appear in a unit test, but not part of the regular
code.
Change-Id: I914fc189514207df2535731eda10cb4b3d30cc6c
2014-06-23 12:59:24 -07:00
|
|
|
'User-Agent': 'object-server %s' % os.getpid(),
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'X-Content-Type': 'text/plain',
|
2014-06-10 22:17:47 -07:00
|
|
|
'X-Timestamp': utils.Timestamp(1).internal,
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'X-Trans-Id': '123',
|
|
|
|
'Referer': 'PUT http://localhost/sda1/0/a/c/o',
|
|
|
|
'X-Backend-Storage-Policy-Index': '0',
|
|
|
|
'X-Etag': 'd41d8cd98f00b204e9800998ecf8427e'}),
|
|
|
|
'obj': 'o',
|
|
|
|
'account': 'a',
|
|
|
|
'container': 'c',
|
|
|
|
'op': 'PUT'})
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
def test_container_update_bad_args(self):
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_async_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
req = Request.blank(
|
|
|
|
'/v1/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '123',
|
|
|
|
'X-Container-Host': 'chost,badhost',
|
|
|
|
'X-Container-Partition': 'cpartition',
|
|
|
|
'X-Container-Device': 'cdevice'})
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
with mock.patch.object(self.object_controller, 'async_update',
|
|
|
|
fake_async_update):
|
|
|
|
self.object_controller.container_update(
|
|
|
|
'PUT', 'a', 'c', 'o', req, {
|
|
|
|
'x-size': '0',
|
|
|
|
'x-etag': 'd41d8cd98f00b204e9800998ecf8427e',
|
|
|
|
'x-content-type': 'text/plain', 'x-timestamp': '1'},
|
|
|
|
'sda1', 0)
|
2013-12-11 16:59:59 -05:00
|
|
|
self.assertEqual(given_args, [])
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
errors = self.object_controller.logger.get_lines_for_level('error')
|
|
|
|
self.assertEqual(len(errors), 1)
|
|
|
|
msg = errors[0]
|
|
|
|
self.assertTrue('Container update failed' in msg)
|
|
|
|
self.assertTrue('different numbers of hosts and devices' in msg)
|
|
|
|
self.assertTrue('chost,badhost' in msg)
|
|
|
|
self.assertTrue('cdevice' in msg)
|
2013-12-11 16:59:59 -05:00
|
|
|
|
2013-06-03 23:50:05 +00:00
|
|
|
def test_delete_at_update_on_put(self):
|
|
|
|
# Test how delete_at_update works when issued a delete for old
|
|
|
|
# expiration info after a new put with no new expiration info.
|
2011-10-26 21:42:24 +00:00
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_async_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/v1/a/c/o',
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '123'})
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
with mock.patch.object(self.object_controller, 'async_update',
|
|
|
|
fake_async_update):
|
|
|
|
self.object_controller.delete_at_update(
|
|
|
|
'DELETE', 2, 'a', 'c', 'o', req, 'sda1', 0)
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
given_args, [
|
2014-01-13 17:45:56 +00:00
|
|
|
'DELETE', '.expiring_objects', '0000000000',
|
|
|
|
'0000000002-a/c/o', None, None, None,
|
2013-09-01 15:10:39 -04:00
|
|
|
HeaderKeyDict({
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': 0,
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('1').internal,
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-trans-id': '123',
|
|
|
|
'referer': 'PUT http://localhost/v1/a/c/o'}),
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'sda1', 0])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
2012-08-06 20:53:24 +00:00
|
|
|
def test_delete_at_negative(self):
|
2013-06-03 23:50:05 +00:00
|
|
|
# Test how delete_at_update works when issued a delete for old
|
|
|
|
# expiration info after a new put with no new expiration info.
|
2012-08-06 20:53:24 +00:00
|
|
|
# Test negative is reset to 0
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_async_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.async_update = fake_async_update
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/v1/a/c/o',
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '1234'})
|
2012-08-06 20:53:24 +00:00
|
|
|
self.object_controller.delete_at_update(
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'DELETE', -2, 'a', 'c', 'o', req, 'sda1', 0)
|
2012-08-06 20:53:24 +00:00
|
|
|
self.assertEquals(given_args, [
|
2014-01-13 17:45:56 +00:00
|
|
|
'DELETE', '.expiring_objects', '0000000000', '0000000000-a/c/o',
|
|
|
|
None, None, None,
|
2013-09-01 15:10:39 -04:00
|
|
|
HeaderKeyDict({
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': 0,
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('1').internal,
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-trans-id': '1234',
|
|
|
|
'referer': 'PUT http://localhost/v1/a/c/o'}),
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'sda1', 0])
|
2012-08-06 20:53:24 +00:00
|
|
|
|
|
|
|
def test_delete_at_cap(self):
|
2013-06-03 23:50:05 +00:00
|
|
|
# Test how delete_at_update works when issued a delete for old
|
|
|
|
# expiration info after a new put with no new expiration info.
|
2012-08-06 20:53:24 +00:00
|
|
|
# Test past cap is reset to cap
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_async_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.async_update = fake_async_update
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/v1/a/c/o',
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '1234'})
|
2012-08-06 20:53:24 +00:00
|
|
|
self.object_controller.delete_at_update(
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'DELETE', 12345678901, 'a', 'c', 'o', req, 'sda1', 0)
|
2012-08-06 20:53:24 +00:00
|
|
|
self.assertEquals(given_args, [
|
2013-06-03 23:50:05 +00:00
|
|
|
'DELETE', '.expiring_objects', '9999936000', '9999999999-a/c/o',
|
|
|
|
None, None, None,
|
2013-09-01 15:10:39 -04:00
|
|
|
HeaderKeyDict({
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': 0,
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('1').internal,
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-trans-id': '1234',
|
|
|
|
'referer': 'PUT http://localhost/v1/a/c/o'}),
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'sda1', 0])
|
2012-08-06 20:53:24 +00:00
|
|
|
|
2011-10-26 21:42:24 +00:00
|
|
|
def test_delete_at_update_put_with_info(self):
|
2013-06-03 23:50:05 +00:00
|
|
|
# Keep next test,
|
|
|
|
# test_delete_at_update_put_with_info_but_missing_container, in sync
|
|
|
|
# with this one but just missing the X-Delete-At-Container header.
|
2011-10-26 21:42:24 +00:00
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_async_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.async_update = fake_async_update
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/v1/a/c/o',
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '1234',
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At-Container': '0',
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
'X-Delete-At-Host': '127.0.0.1:1234',
|
|
|
|
'X-Delete-At-Partition': '3',
|
|
|
|
'X-Delete-At-Device': 'sdc1'})
|
2011-10-26 21:42:24 +00:00
|
|
|
self.object_controller.delete_at_update('PUT', 2, 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
req, 'sda1', 0)
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
given_args, [
|
2014-01-13 17:45:56 +00:00
|
|
|
'PUT', '.expiring_objects', '0000000000', '0000000002-a/c/o',
|
|
|
|
'127.0.0.1:1234',
|
2013-09-01 15:10:39 -04:00
|
|
|
'3', 'sdc1', HeaderKeyDict({
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': 0,
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-size': '0',
|
|
|
|
'x-etag': 'd41d8cd98f00b204e9800998ecf8427e',
|
|
|
|
'x-content-type': 'text/plain',
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('1').internal,
|
2013-09-01 15:10:39 -04:00
|
|
|
'x-trans-id': '1234',
|
|
|
|
'referer': 'PUT http://localhost/v1/a/c/o'}),
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'sda1', 0])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
2013-06-03 23:50:05 +00:00
|
|
|
def test_delete_at_update_put_with_info_but_missing_container(self):
|
|
|
|
# Same as previous test, test_delete_at_update_put_with_info, but just
|
|
|
|
# missing the X-Delete-At-Container header.
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_async_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.async_update = fake_async_update
|
|
|
|
self.object_controller.logger = FakeLogger()
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/v1/a/c/o',
|
2013-06-03 23:50:05 +00:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '1234',
|
|
|
|
'X-Delete-At-Host': '127.0.0.1:1234',
|
|
|
|
'X-Delete-At-Partition': '3',
|
|
|
|
'X-Delete-At-Device': 'sdc1'})
|
|
|
|
self.object_controller.delete_at_update('PUT', 2, 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
req, 'sda1', 0)
|
2013-06-03 23:50:05 +00:00
|
|
|
self.assertEquals(
|
|
|
|
self.object_controller.logger.log_dict['warning'],
|
|
|
|
[(('X-Delete-At-Container header must be specified for expiring '
|
|
|
|
'objects background PUT to work properly. Making best guess as '
|
|
|
|
'to the container name for now.',), {})])
|
|
|
|
|
2011-10-26 21:42:24 +00:00
|
|
|
def test_delete_at_update_delete(self):
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_async_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.async_update = fake_async_update
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/v1/a/c/o',
|
Enhance log msg to report referer and user-agent
Enhance internally logged messages to report referer and user-agent.
Pass the referering URL and METHOD between internal servers (when
known), and set the user-agent to be the server type (obj-server,
container-server, proxy-server, obj-updater, obj-replicator,
container-updater, direct-client, etc.) with the process PID. In
conjunction with the transaction ID, it helps to track down which PID
from a given system was responsible for initiating the request and
what that server was working on to make this request.
This has been helpful in tracking down interactions between object,
container and account servers.
We also take things a bit further performaing a bit of refactoring to
consolidate calls to transfer_headers() now that we have a helper
method for constructing them.
Finally we performed further changes to avoid header key duplication
due to string literal header key values and the various objects
representing headers for requests and responses. See below for more
details.
====
Header Keys
There seems to be a bit of a problem with the case of the various
string literals used for header keys and the interchangable way
standard Python dictionaries, HeaderKeyDict() and HeaderEnvironProxy()
objects are used.
If one is not careful, a header object of some sort (one that does not
normalize its keys, and that is not necessarily a dictionary) can be
constructed containing header keys which differ only by the case of
their string literals. E.g.:
{ 'x-trans-id': '1234', 'X-Trans-Id': '5678' }
Such an object, when passed to http_connect() will result in an
on-the-wire header where the key values are merged together, comma
separated, that looks something like:
HTTP_X_TRANS_ID: 1234,5678
For some headers in some contexts, this is behavior is desirable. For
example, one can also use a list of tuples which enumerate the multiple
values a single header should have.
However, in almost all of the contexts used in the code base, this is
not desirable.
This behavior arises from a combination of factors:
1. Header strings are not constants and different lower-case and
title-case header strings values are used interchangably in the
code at times
It might be worth the effort to make a pass through the code to
stop using string literals and use constants instead, but there
are plusses and minuses to doing that, so this was not attempted
in this effort
2. HeaderEnvironProxy() objects report their keys in ".title()"
case, but normalize all other key references to the form
expected by the Request class's environ field
swob.Request.headers fields are HeaderEnvironProxy() objects.
3. HeaderKeyDict() objects report their keys in ".lower()" case,
and normalize all other key references to ".lower()" case
swob.Response.headers fields are HeaderKeyDict() objects.
Depending on which object is used and how it is used, one can end up
with such a mismatch.
This commit takes the following steps as a (PROPOSED) solution:
1. Change HeaderKeyDict() to normalize using ".title()" case to
match HeaderEnvironProxy()
2. Replace standard python dictionary objects with HeaderKeyDict()
objects where possible
This gives us an object that normalizes key references to avoid
fixing the code to normalize the string literals.
3. Fix up a few places to use title case string literals to match
the new defaults
Change-Id: Ied56a1df83ffac793ee85e796424d7d20f18f469
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2012-11-15 16:34:45 -05:00
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '1234'})
|
2011-10-26 21:42:24 +00:00
|
|
|
self.object_controller.delete_at_update('DELETE', 2, 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
req, 'sda1', 0)
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
given_args, [
|
2014-01-13 17:45:56 +00:00
|
|
|
'DELETE', '.expiring_objects', '0000000000',
|
|
|
|
'0000000002-a/c/o', None, None,
|
2013-09-01 15:10:39 -04:00
|
|
|
None, HeaderKeyDict({
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': 0,
|
2014-06-10 22:17:47 -07:00
|
|
|
'x-timestamp': utils.Timestamp('1').internal,
|
|
|
|
'x-trans-id': '1234',
|
2013-09-01 15:10:39 -04:00
|
|
|
'referer': 'DELETE http://localhost/v1/a/c/o'}),
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'sda1', 0])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
2014-01-23 01:05:16 +00:00
|
|
|
def test_delete_backend_replication(self):
|
|
|
|
# If X-Backend-Replication: True delete_at_update should completely
|
|
|
|
# short-circuit.
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_async_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.async_update = fake_async_update
|
|
|
|
req = Request.blank(
|
|
|
|
'/v1/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': 1,
|
|
|
|
'X-Trans-Id': '1234',
|
|
|
|
'X-Backend-Replication': 'True'})
|
|
|
|
self.object_controller.delete_at_update(
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
'DELETE', -2, 'a', 'c', 'o', req, 'sda1', 0)
|
2014-01-23 01:05:16 +00:00
|
|
|
self.assertEquals(given_args, [])
|
|
|
|
|
2011-10-26 21:42:24 +00:00
|
|
|
def test_POST_calls_delete_at(self):
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_delete_at_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.delete_at_update = fake_delete_at_update
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
self.assertEquals(given_args, [])
|
|
|
|
|
|
|
|
sleep(.00001)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Type': 'application/x-test'})
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
|
|
|
self.assertEquals(given_args, [])
|
|
|
|
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp1 = normalize_timestamp(time())
|
|
|
|
delete_at_timestamp1 = str(int(time() + 1000))
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp1,
|
|
|
|
'Content-Type': 'application/x-test',
|
|
|
|
'X-Delete-At': delete_at_timestamp1})
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
given_args, [
|
|
|
|
'PUT', int(delete_at_timestamp1), 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
given_args[5], 'sda1', 0])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
while given_args:
|
|
|
|
given_args.pop()
|
|
|
|
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp2 = normalize_timestamp(time())
|
|
|
|
delete_at_timestamp2 = str(int(time() + 2000))
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': timestamp2,
|
|
|
|
'Content-Type': 'application/x-test',
|
|
|
|
'X-Delete-At': delete_at_timestamp2})
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
given_args, [
|
|
|
|
'PUT', int(delete_at_timestamp2), 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
given_args[5], 'sda1', 0,
|
2013-09-01 15:10:39 -04:00
|
|
|
'DELETE', int(delete_at_timestamp1), 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
given_args[5], 'sda1', 0])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
def test_PUT_calls_delete_at(self):
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_delete_at_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.delete_at_update = fake_delete_at_update
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
self.assertEquals(given_args, [])
|
|
|
|
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp1 = normalize_timestamp(time())
|
|
|
|
delete_at_timestamp1 = str(int(time() + 1000))
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': timestamp1,
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'X-Delete-At': delete_at_timestamp1})
|
|
|
|
req.body = 'TEST'
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
given_args, [
|
|
|
|
'PUT', int(delete_at_timestamp1), 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
given_args[5], 'sda1', 0])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
while given_args:
|
|
|
|
given_args.pop()
|
|
|
|
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp2 = normalize_timestamp(time())
|
|
|
|
delete_at_timestamp2 = str(int(time() + 2000))
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp2,
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'X-Delete-At': delete_at_timestamp2})
|
|
|
|
req.body = 'TEST'
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
self.assertEquals(
|
|
|
|
given_args, [
|
|
|
|
'PUT', int(delete_at_timestamp2), 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
given_args[5], 'sda1', 0,
|
2013-09-01 15:10:39 -04:00
|
|
|
'DELETE', int(delete_at_timestamp1), 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
given_args[5], 'sda1', 0])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
def test_GET_but_expired(self):
|
|
|
|
test_time = time() + 10000
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp = int(test_time + 100)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_timestamp /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 2000),
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp),
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
2011-10-26 21:42:24 +00:00
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time)})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
2011-11-01 23:16:11 +00:00
|
|
|
orig_time = object_server.time.time
|
|
|
|
try:
|
|
|
|
t = time()
|
|
|
|
object_server.time.time = lambda: t
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp = int(t + 1)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_timestamp /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2014-03-23 10:24:21 -07:00
|
|
|
put_timestamp = normalize_timestamp(test_time - 1000)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-11-01 23:16:11 +00:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
2014-03-23 10:24:21 -07:00
|
|
|
headers={'X-Timestamp': put_timestamp,
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp),
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
2011-11-01 23:16:11 +00:00
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-11-01 23:16:11 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-11-01 23:16:11 +00:00
|
|
|
environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time)})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-11-01 23:16:11 +00:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
finally:
|
|
|
|
object_server.time.time = orig_time
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
orig_time = object_server.time.time
|
|
|
|
try:
|
|
|
|
t = time() + 2
|
|
|
|
object_server.time.time = lambda: t
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(t)})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
2014-03-23 10:24:21 -07:00
|
|
|
self.assertEquals(resp.headers['X-Backend-Timestamp'],
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(put_timestamp))
|
2011-10-26 21:42:24 +00:00
|
|
|
finally:
|
|
|
|
object_server.time.time = orig_time
|
|
|
|
|
|
|
|
def test_HEAD_but_expired(self):
|
|
|
|
test_time = time() + 10000
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp = int(test_time + 100)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_timestamp /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 2000),
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp),
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
2011-10-26 21:42:24 +00:00
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time)})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
|
2011-11-01 23:16:11 +00:00
|
|
|
orig_time = object_server.time.time
|
|
|
|
try:
|
|
|
|
t = time()
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp = int(t + 1)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_timestamp /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2011-11-01 23:16:11 +00:00
|
|
|
object_server.time.time = lambda: t
|
2014-03-23 10:24:21 -07:00
|
|
|
put_timestamp = normalize_timestamp(test_time - 1000)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-11-01 23:16:11 +00:00
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
2014-03-23 10:24:21 -07:00
|
|
|
headers={'X-Timestamp': put_timestamp,
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp),
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
2011-11-01 23:16:11 +00:00
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-11-01 23:16:11 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-11-01 23:16:11 +00:00
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time)})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-11-01 23:16:11 +00:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
finally:
|
|
|
|
object_server.time.time = orig_time
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
orig_time = object_server.time.time
|
|
|
|
try:
|
|
|
|
t = time() + 2
|
|
|
|
object_server.time.time = lambda: t
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'HEAD'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time())})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
2014-03-23 10:24:21 -07:00
|
|
|
self.assertEquals(resp.headers['X-Backend-Timestamp'],
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(put_timestamp))
|
2011-10-26 21:42:24 +00:00
|
|
|
finally:
|
|
|
|
object_server.time.time = orig_time
|
|
|
|
|
|
|
|
def test_POST_but_expired(self):
|
|
|
|
test_time = time() + 10000
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp = int(test_time + 100)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_timestamp /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 2000),
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp),
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
2011-10-26 21:42:24 +00:00
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 1500)})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 202)
|
|
|
|
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp = int(time() + 1)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_timestamp /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 1000),
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp),
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
2011-10-26 21:42:24 +00:00
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
orig_time = object_server.time.time
|
|
|
|
try:
|
|
|
|
t = time() + 2
|
|
|
|
object_server.time.time = lambda: t
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time())})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
finally:
|
|
|
|
object_server.time.time = orig_time
|
|
|
|
|
2012-11-08 15:22:01 -08:00
|
|
|
def test_DELETE_but_expired(self):
|
|
|
|
test_time = time() + 10000
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp = int(test_time + 100)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_timestamp /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2012-11-08 15:22:01 -08:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 2000),
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp),
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
2012-11-08 15:22:01 -08:00
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2012-11-08 15:22:01 -08:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
orig_time = object_server.time.time
|
|
|
|
try:
|
|
|
|
t = test_time + 100
|
|
|
|
object_server.time.time = lambda: float(t)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2012-11-08 15:22:01 -08:00
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time())})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2012-11-08 15:22:01 -08:00
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
finally:
|
|
|
|
object_server.time.time = orig_time
|
|
|
|
|
2013-12-03 11:17:57 -08:00
|
|
|
def test_DELETE_if_delete_at_expired_still_deletes(self):
|
|
|
|
test_time = time() + 10
|
|
|
|
test_timestamp = normalize_timestamp(test_time)
|
|
|
|
delete_at_time = int(test_time + 10)
|
|
|
|
delete_at_timestamp = str(delete_at_time)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_time /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': test_timestamp,
|
|
|
|
'X-Delete-At': delete_at_timestamp,
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
|
|
|
# sanity
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'X-Timestamp': test_timestamp})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
self.assertEquals(resp.body, 'TEST')
|
|
|
|
objfile = os.path.join(
|
|
|
|
self.testdir, 'sda1',
|
2014-05-27 01:17:13 -07:00
|
|
|
storage_directory(diskfile.get_data_dir(0), 'p',
|
2013-12-03 11:17:57 -08:00
|
|
|
hash_path('a', 'c', 'o')),
|
2014-06-10 22:17:47 -07:00
|
|
|
utils.Timestamp(test_timestamp).internal + '.data')
|
2013-12-03 11:17:57 -08:00
|
|
|
self.assert_(os.path.isfile(objfile))
|
|
|
|
|
|
|
|
# move time past expirery
|
|
|
|
with mock.patch('swift.obj.diskfile.time') as mock_time:
|
|
|
|
mock_time.time.return_value = test_time + 100
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'GET'},
|
|
|
|
headers={'X-Timestamp': test_timestamp})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
# request will 404
|
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
# but file still exists
|
|
|
|
self.assert_(os.path.isfile(objfile))
|
|
|
|
|
2014-06-06 11:35:34 -07:00
|
|
|
# make the x-if-delete-at with some wrong bits
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': delete_at_timestamp,
|
|
|
|
'X-If-Delete-At': int(time() + 1)})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
self.assertTrue(os.path.isfile(objfile))
|
|
|
|
|
2013-12-03 11:17:57 -08:00
|
|
|
# make the x-if-delete-at with all the right bits
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': delete_at_timestamp,
|
|
|
|
'X-If-Delete-At': delete_at_timestamp})
|
|
|
|
resp = req.get_response(self.object_controller)
|
2014-06-06 11:35:34 -07:00
|
|
|
self.assertEquals(resp.status_int, 204)
|
2013-12-03 11:17:57 -08:00
|
|
|
self.assertFalse(os.path.isfile(objfile))
|
|
|
|
|
2014-06-06 11:35:34 -07:00
|
|
|
# make the x-if-delete-at with all the right bits (again)
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': delete_at_timestamp,
|
|
|
|
'X-If-Delete-At': delete_at_timestamp})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
self.assertFalse(os.path.isfile(objfile))
|
|
|
|
|
|
|
|
# make the x-if-delete-at for some not found
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o-not-found',
|
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': delete_at_timestamp,
|
|
|
|
'X-If-Delete-At': delete_at_timestamp})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 404)
|
|
|
|
|
2011-10-26 21:42:24 +00:00
|
|
|
def test_DELETE_if_delete_at(self):
|
|
|
|
test_time = time() + 10000
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 99),
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 98)})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 204)
|
|
|
|
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp = int(test_time - 1)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_timestamp /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 97),
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp),
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
2011-10-26 21:42:24 +00:00
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 95),
|
|
|
|
'X-If-Delete-At': str(int(test_time))})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 95)})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 204)
|
|
|
|
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp = int(test_time - 1)
|
|
|
|
delete_at_container = str(
|
|
|
|
delete_at_timestamp /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 94),
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp),
|
|
|
|
'X-Delete-At-Container': delete_at_container,
|
2011-10-26 21:42:24 +00:00
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'DELETE'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 92),
|
|
|
|
'X-If-Delete-At': str(int(test_time))})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 412)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'DELETE'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 92),
|
|
|
|
'X-If-Delete-At': delete_at_timestamp})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 204)
|
|
|
|
|
2013-12-03 13:01:15 -05:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(test_time - 92),
|
|
|
|
'X-If-Delete-At': 'abc'})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
|
2011-10-26 21:42:24 +00:00
|
|
|
def test_DELETE_calls_delete_at(self):
|
|
|
|
given_args = []
|
|
|
|
|
|
|
|
def fake_delete_at_update(*args):
|
|
|
|
given_args.extend(args)
|
|
|
|
|
|
|
|
self.object_controller.delete_at_update = fake_delete_at_update
|
|
|
|
timestamp1 = normalize_timestamp(time())
|
2013-06-03 23:50:05 +00:00
|
|
|
delete_at_timestamp1 = int(time() + 1000)
|
|
|
|
delete_at_container1 = str(
|
|
|
|
delete_at_timestamp1 /
|
|
|
|
self.object_controller.expiring_objects_container_divisor *
|
|
|
|
self.object_controller.expiring_objects_container_divisor)
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': timestamp1,
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream',
|
2013-06-03 23:50:05 +00:00
|
|
|
'X-Delete-At': str(delete_at_timestamp1),
|
|
|
|
'X-Delete-At-Container': delete_at_container1})
|
2011-10-26 21:42:24 +00:00
|
|
|
req.body = 'TEST'
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
self.assertEquals(given_args, [
|
|
|
|
'PUT', int(delete_at_timestamp1), 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
given_args[5], 'sda1', 0])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
while given_args:
|
|
|
|
given_args.pop()
|
|
|
|
|
|
|
|
sleep(.00001)
|
|
|
|
timestamp2 = normalize_timestamp(time())
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'DELETE'},
|
|
|
|
headers={'X-Timestamp': timestamp2,
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
2014-05-27 01:17:13 -07:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 204)
|
|
|
|
self.assertEquals(given_args, [
|
|
|
|
'DELETE', int(delete_at_timestamp1), 'a', 'c', 'o',
|
Add Storage Policy support to Object Updates
The object server will now send its storage policy index to the
container server synchronously and asynchronously (via async_pending).
Each storage policy gets its own async_pending directory under
/srv/node/$disk/objects-$N, so there's no need to change the on-disk
pickle format; the policy index comes from the async_pending's
filename. This avoids any hassle on upgrade. (Recall that policy 0's
objects live in /srv/node/$disk/objects, not objects-0.) Per-policy
tempdir as well.
Also clean up a couple little things in the object updater. Now it
won't abort processing when it encounters a file (not directory) named
"async_pending-\d+", and it won't process updates in a directory that
does not correspond to a storage policy.
That is, if you have policies 1, 2, and 3, but there's a directory on
your disk named "async_pending-5", the updater will now skip over that
entirely. It won't even bother doing directory listings at all. This
is a good idea, believe it or not, because there's nothing good that
the container server can do with an update from some unknown storage
policy. It can't update the listing, it can't move the object if it's
misplaced... all it can do is ignore the request, so it's better to
just not send it in the first place. Plus, if this is due to a
misconfiguration on one storage node, then the updates will get
processed once the configuration is fixed.
There's also a drive by fix to update some backend http mocks for container
update tests that we're not fully exercising their their request fakes.
Because the object server container update code is resilient to to all manor
of failure from backend requests the general intent of the tests was
unaffected but this change cleans up some confusing logging in the debug
logger output.
The object-server will send X-Storage-Policy-Index headers with all
requests to container severs, including X-Delete containers and all
object PUT/DELETE requests. This header value is persisted in the
pickle file for the update and sent along with async requests from the
object-updater as well.
The container server will extract the X-Storage-Policy-Index header from
incoming requests and apply it to container broker calls as appropriate
defaulting to the legacy storage policy 0 to support seemless migration.
DocImpact
Implements: blueprint storage-policies
Change-Id: I07c730bebaee068f75024fa9c2fa9e11e295d9bd
add to object updates
Change-Id: Ic97a422238a0d7bc2a411a71a7aba3f8b42fce4d
2014-03-17 17:54:42 -07:00
|
|
|
given_args[5], 'sda1', 0])
|
2011-10-26 21:42:24 +00:00
|
|
|
|
|
|
|
def test_PUT_delete_at_in_past(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'X-Delete-At': str(int(time() - 1)),
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
self.assertTrue('X-Delete-At in past' in resp.body)
|
|
|
|
|
|
|
|
def test_POST_delete_at_in_past(self):
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT'},
|
2011-10-26 21:42:24 +00:00
|
|
|
headers={'X-Timestamp': normalize_timestamp(time()),
|
|
|
|
'Content-Length': '4',
|
|
|
|
'Content-Type': 'application/octet-stream'})
|
|
|
|
req.body = 'TEST'
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
2011-10-26 21:42:24 +00:00
|
|
|
environ={'REQUEST_METHOD': 'POST'},
|
|
|
|
headers={'X-Timestamp': normalize_timestamp(time() + 1),
|
|
|
|
'X-Delete-At': str(int(time() - 1))})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2011-10-26 21:42:24 +00:00
|
|
|
self.assertEquals(resp.status_int, 400)
|
|
|
|
self.assertTrue('X-Delete-At in past' in resp.body)
|
|
|
|
|
2012-02-07 14:02:28 -08:00
|
|
|
def test_REPLICATE_works(self):
|
|
|
|
|
|
|
|
def fake_get_hashes(*args, **kwargs):
|
|
|
|
return 0, {1: 2}
|
|
|
|
|
2012-08-21 12:51:59 -07:00
|
|
|
def my_tpool_execute(func, *args, **kwargs):
|
2012-02-07 14:02:28 -08:00
|
|
|
return func(*args, **kwargs)
|
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
was_get_hashes = diskfile.get_hashes
|
2012-02-07 14:02:28 -08:00
|
|
|
was_tpool_exe = tpool.execute
|
|
|
|
try:
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
diskfile.get_hashes = fake_get_hashes
|
|
|
|
tpool.execute = my_tpool_execute
|
|
|
|
req = Request.blank('/sda1/p/suff',
|
|
|
|
environ={'REQUEST_METHOD': 'REPLICATE'},
|
|
|
|
headers={})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2012-02-07 14:02:28 -08:00
|
|
|
self.assertEquals(resp.status_int, 200)
|
|
|
|
p_data = pickle.loads(resp.body)
|
|
|
|
self.assertEquals(p_data, {1: 2})
|
|
|
|
finally:
|
|
|
|
tpool.execute = was_tpool_exe
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
diskfile.get_hashes = was_get_hashes
|
2012-02-07 14:02:28 -08:00
|
|
|
|
|
|
|
def test_REPLICATE_timeout(self):
|
|
|
|
|
|
|
|
def fake_get_hashes(*args, **kwargs):
|
|
|
|
raise Timeout()
|
|
|
|
|
2012-08-21 12:51:59 -07:00
|
|
|
def my_tpool_execute(func, *args, **kwargs):
|
2012-02-07 14:02:28 -08:00
|
|
|
return func(*args, **kwargs)
|
|
|
|
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
was_get_hashes = diskfile.get_hashes
|
2012-02-07 14:02:28 -08:00
|
|
|
was_tpool_exe = tpool.execute
|
|
|
|
try:
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
diskfile.get_hashes = fake_get_hashes
|
|
|
|
tpool.execute = my_tpool_execute
|
|
|
|
req = Request.blank('/sda1/p/suff',
|
|
|
|
environ={'REQUEST_METHOD': 'REPLICATE'},
|
|
|
|
headers={})
|
2012-02-07 14:02:28 -08:00
|
|
|
self.assertRaises(Timeout, self.object_controller.REPLICATE, req)
|
|
|
|
finally:
|
|
|
|
tpool.execute = was_tpool_exe
|
DiskFile API, with reference implementation
Refactor on-disk knowledge out of the object server by pushing the
async update pickle creation to the new DiskFileManager class (name is
not the best, so suggestions welcome), along with the REPLICATOR
method logic. We also move the mount checking and thread pool storage
to the new ondisk.Devices object, which then also becomes the new home
of the audit_location_generator method.
For the object server, a new setup() method is now called at the end
of the controller's construction, and the _diskfile() method has been
renamed to get_diskfile(), to allow implementation specific behavior.
We then hide the need for the REST API layer to know how and where
quarantining needs to be performed. There are now two places it is
checked internally, on open() where we verify the content-length,
name, and x-timestamp metadata, and in the reader on close where the
etag metadata is checked if the entire file was read.
We add a reader class to allow implementations to isolate the WSGI
handling code for that specific environment (it is used no-where else
in the REST APIs). This simplifies the caller's code to just use a
"with" statement once open to avoid multiple points where close needs
to be called.
For a full historical comparison, including the usage patterns see:
https://gist.github.com/portante/5488238
(as of master, 2b639f5, Merge
"Fix 500 from account-quota This Commit
middleware")
--------------------------------+------------------------------------
DiskFileManager(conf)
Methods:
.pickle_async_update()
.get_diskfile()
.get_hashes()
Attributes:
.devices
.logger
.disk_chunk_size
.keep_cache_size
.bytes_per_sync
DiskFile(a,c,o,keep_data_fp=) DiskFile(a,c,o)
Methods: Methods:
*.__iter__()
.close(verify_file=)
.is_deleted()
.is_expired()
.quarantine()
.get_data_file_size()
.open()
.read_metadata()
.create() .create()
.write_metadata()
.delete() .delete()
Attributes: Attributes:
.quarantined_dir
.keep_cache
.metadata
*DiskFileReader()
Methods:
.__iter__()
.close()
Attributes:
+.was_quarantined
DiskWriter() DiskFileWriter()
Methods: Methods:
.write() .write()
.put() .put()
* Note that the DiskFile class * Note that the DiskReader() object
implements all the methods returned by the
necessary for a WSGI app DiskFileOpened.reader() method
iterator implements all the methods
necessary for a WSGI app iterator
+ Note that if the auditor is
refactored to not use the DiskFile
class, see
https://review.openstack.org/44787
then we don't need the
was_quarantined attribute
A reference "in-memory" object server implementation of a backend
DiskFile class in swift/obj/mem_server.py and
swift/obj/mem_diskfile.py.
One can also reference
https://github.com/portante/gluster-swift/commits/diskfile for the
proposed integration with the gluster-swift code based on these
changes.
Change-Id: I44e153fdb405a5743e9c05349008f94136764916
Signed-off-by: Peter Portante <peter.portante@redhat.com>
2013-09-12 19:51:18 -04:00
|
|
|
diskfile.get_hashes = was_get_hashes
|
2010-07-12 17:03:45 -05:00
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
def test_REPLICATE_insufficient_storage(self):
|
|
|
|
conf = {'devices': self.testdir, 'mount_check': 'true'}
|
|
|
|
self.object_controller = object_server.ObjectController(
|
|
|
|
conf, logger=debug_logger())
|
|
|
|
self.object_controller.bytes_per_sync = 1
|
|
|
|
|
|
|
|
def fake_check_mount(*args, **kwargs):
|
|
|
|
return False
|
|
|
|
|
|
|
|
with mock.patch("swift.obj.diskfile.check_mount", fake_check_mount):
|
|
|
|
req = Request.blank('/sda1/p/suff',
|
|
|
|
environ={'REQUEST_METHOD': 'REPLICATE'},
|
|
|
|
headers={})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 507)
|
|
|
|
|
|
|
|
def test_REPLICATION_can_be_called(self):
|
|
|
|
req = Request.blank('/sda1/p/other/suff',
|
|
|
|
environ={'REQUEST_METHOD': 'REPLICATION'},
|
|
|
|
headers={})
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEqual(resp.status_int, 200)
|
|
|
|
|
2012-08-27 14:44:41 -07:00
|
|
|
def test_PUT_with_full_drive(self):
|
|
|
|
|
2014-03-03 17:28:48 -08:00
|
|
|
class IgnoredBody(object):
|
2012-08-27 14:44:41 -07:00
|
|
|
|
|
|
|
def __init__(self):
|
|
|
|
self.read_called = False
|
|
|
|
|
|
|
|
def read(self, size=-1):
|
|
|
|
if not self.read_called:
|
|
|
|
self.read_called = True
|
|
|
|
return 'VERIFY'
|
|
|
|
return ''
|
|
|
|
|
|
|
|
def fake_fallocate(fd, size):
|
|
|
|
raise OSError(42, 'Unable to fallocate(%d)' % size)
|
|
|
|
|
2013-07-17 16:32:35 -07:00
|
|
|
orig_fallocate = diskfile.fallocate
|
2012-08-27 14:44:41 -07:00
|
|
|
try:
|
2013-07-17 16:32:35 -07:00
|
|
|
diskfile.fallocate = fake_fallocate
|
2012-08-27 14:44:41 -07:00
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
body_reader = IgnoredBody()
|
2013-09-01 15:10:39 -04:00
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'PUT',
|
|
|
|
'wsgi.input': body_reader},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Length': '6',
|
|
|
|
'Content-Type': 'application/octet-stream',
|
|
|
|
'Expect': '100-continue'})
|
2013-08-16 17:13:00 -04:00
|
|
|
resp = req.get_response(self.object_controller)
|
2012-08-27 14:44:41 -07:00
|
|
|
self.assertEquals(resp.status_int, 507)
|
|
|
|
self.assertFalse(body_reader.read_called)
|
|
|
|
finally:
|
2013-07-17 16:32:35 -07:00
|
|
|
diskfile.fallocate = orig_fallocate
|
2012-08-27 14:44:41 -07:00
|
|
|
|
Object replication ssync (an rsync alternative)
For this commit, ssync is just a direct replacement for how
we use rsync. Assuming we switch over to ssync completely
someday and drop rsync, we will then be able to improve the
algorithms even further (removing local objects as we
successfully transfer each one rather than waiting for whole
partitions, using an index.db with hash-trees, etc., etc.)
For easier review, this commit can be thought of in distinct
parts:
1) New global_conf_callback functionality for allowing
services to perform setup code before workers, etc. are
launched. (This is then used by ssync in the object
server to create a cross-worker semaphore to restrict
concurrent incoming replication.)
2) A bit of shifting of items up from object server and
replicator to diskfile or DEFAULT conf sections for
better sharing of the same settings. conn_timeout,
node_timeout, client_timeout, network_chunk_size,
disk_chunk_size.
3) Modifications to the object server and replicator to
optionally use ssync in place of rsync. This is done in
a generic enough way that switching to FutureSync should
be easy someday.
4) The biggest part, and (at least for now) completely
optional part, are the new ssync_sender and
ssync_receiver files. Nice and isolated for easier
testing and visibility into test coverage, etc.
All the usual logging, statsd, recon, etc. instrumentation
is still there when using ssync, just as it is when using
rsync.
Beyond the essential error and exceptional condition
logging, I have not added any additional instrumentation at
this time. Unless there is something someone finds super
pressing to have added to the logging, I think such
additions would be better as separate change reviews.
FOR NOW, IT IS NOT RECOMMENDED TO USE SSYNC ON PRODUCTION
CLUSTERS. Some of us will be in a limited fashion to look
for any subtle issues, tuning, etc. but generally ssync is
an experimental feature. In its current implementation it is
probably going to be a bit slower than rsync, but if all
goes according to plan it will end up much faster.
There are no comparisions yet between ssync and rsync other
than some raw virtual machine testing I've done to show it
should compete well enough once we can put it in use in the
real world.
If you Tweet, Google+, or whatever, be sure to indicate it's
experimental. It'd be best to keep it out of deployment
guides, howtos, etc. until we all figure out if we like it,
find it to be stable, etc.
Change-Id: If003dcc6f4109e2d2a42f4873a0779110fff16d6
2013-08-28 16:10:43 +00:00
|
|
|
def test_global_conf_callback_does_nothing(self):
|
|
|
|
preloaded_app_conf = {}
|
|
|
|
global_conf = {}
|
|
|
|
object_server.global_conf_callback(preloaded_app_conf, global_conf)
|
|
|
|
self.assertEqual(preloaded_app_conf, {})
|
|
|
|
self.assertEqual(global_conf.keys(), ['replication_semaphore'])
|
2013-11-24 20:24:45 -06:00
|
|
|
try:
|
|
|
|
value = global_conf['replication_semaphore'][0].get_value()
|
|
|
|
except NotImplementedError:
|
|
|
|
# On some operating systems (at a minimum, OS X) it's not possible
|
|
|
|
# to introspect the value of a semaphore
|
|
|
|
raise SkipTest
|
|
|
|
else:
|
|
|
|
self.assertEqual(value, 4)
|
Object replication ssync (an rsync alternative)
For this commit, ssync is just a direct replacement for how
we use rsync. Assuming we switch over to ssync completely
someday and drop rsync, we will then be able to improve the
algorithms even further (removing local objects as we
successfully transfer each one rather than waiting for whole
partitions, using an index.db with hash-trees, etc., etc.)
For easier review, this commit can be thought of in distinct
parts:
1) New global_conf_callback functionality for allowing
services to perform setup code before workers, etc. are
launched. (This is then used by ssync in the object
server to create a cross-worker semaphore to restrict
concurrent incoming replication.)
2) A bit of shifting of items up from object server and
replicator to diskfile or DEFAULT conf sections for
better sharing of the same settings. conn_timeout,
node_timeout, client_timeout, network_chunk_size,
disk_chunk_size.
3) Modifications to the object server and replicator to
optionally use ssync in place of rsync. This is done in
a generic enough way that switching to FutureSync should
be easy someday.
4) The biggest part, and (at least for now) completely
optional part, are the new ssync_sender and
ssync_receiver files. Nice and isolated for easier
testing and visibility into test coverage, etc.
All the usual logging, statsd, recon, etc. instrumentation
is still there when using ssync, just as it is when using
rsync.
Beyond the essential error and exceptional condition
logging, I have not added any additional instrumentation at
this time. Unless there is something someone finds super
pressing to have added to the logging, I think such
additions would be better as separate change reviews.
FOR NOW, IT IS NOT RECOMMENDED TO USE SSYNC ON PRODUCTION
CLUSTERS. Some of us will be in a limited fashion to look
for any subtle issues, tuning, etc. but generally ssync is
an experimental feature. In its current implementation it is
probably going to be a bit slower than rsync, but if all
goes according to plan it will end up much faster.
There are no comparisions yet between ssync and rsync other
than some raw virtual machine testing I've done to show it
should compete well enough once we can put it in use in the
real world.
If you Tweet, Google+, or whatever, be sure to indicate it's
experimental. It'd be best to keep it out of deployment
guides, howtos, etc. until we all figure out if we like it,
find it to be stable, etc.
Change-Id: If003dcc6f4109e2d2a42f4873a0779110fff16d6
2013-08-28 16:10:43 +00:00
|
|
|
|
|
|
|
def test_global_conf_callback_replication_semaphore(self):
|
|
|
|
preloaded_app_conf = {'replication_concurrency': 123}
|
|
|
|
global_conf = {}
|
|
|
|
with mock.patch.object(
|
|
|
|
object_server.multiprocessing, 'BoundedSemaphore',
|
|
|
|
return_value='test1') as mocked_Semaphore:
|
|
|
|
object_server.global_conf_callback(preloaded_app_conf, global_conf)
|
|
|
|
self.assertEqual(preloaded_app_conf, {'replication_concurrency': 123})
|
|
|
|
self.assertEqual(global_conf, {'replication_semaphore': ['test1']})
|
|
|
|
mocked_Semaphore.assert_called_once_with(123)
|
|
|
|
|
2013-12-11 16:59:59 -05:00
|
|
|
def test_handling_of_replication_semaphore_config(self):
|
|
|
|
conf = {'devices': self.testdir, 'mount_check': 'false'}
|
|
|
|
objsrv = object_server.ObjectController(conf)
|
|
|
|
self.assertTrue(objsrv.replication_semaphore is None)
|
|
|
|
conf['replication_semaphore'] = ['sema']
|
|
|
|
objsrv = object_server.ObjectController(conf)
|
|
|
|
self.assertEqual(objsrv.replication_semaphore, 'sema')
|
|
|
|
|
2012-12-17 06:39:25 -05:00
|
|
|
def test_serv_reserv(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test replication_server flag was set from configuration file.
|
2012-12-17 06:39:25 -05:00
|
|
|
conf = {'devices': self.testdir, 'mount_check': 'false'}
|
|
|
|
self.assertEquals(
|
|
|
|
object_server.ObjectController(conf).replication_server, None)
|
|
|
|
for val in [True, '1', 'True', 'true']:
|
|
|
|
conf['replication_server'] = val
|
|
|
|
self.assertTrue(
|
|
|
|
object_server.ObjectController(conf).replication_server)
|
|
|
|
for val in [False, 0, '0', 'False', 'false', 'test_string']:
|
|
|
|
conf['replication_server'] = val
|
|
|
|
self.assertFalse(
|
|
|
|
object_server.ObjectController(conf).replication_server)
|
|
|
|
|
|
|
|
def test_list_allowed_methods(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test list of allowed_methods
|
2013-05-23 20:16:21 +04:00
|
|
|
obj_methods = ['DELETE', 'PUT', 'HEAD', 'GET', 'POST']
|
Object replication ssync (an rsync alternative)
For this commit, ssync is just a direct replacement for how
we use rsync. Assuming we switch over to ssync completely
someday and drop rsync, we will then be able to improve the
algorithms even further (removing local objects as we
successfully transfer each one rather than waiting for whole
partitions, using an index.db with hash-trees, etc., etc.)
For easier review, this commit can be thought of in distinct
parts:
1) New global_conf_callback functionality for allowing
services to perform setup code before workers, etc. are
launched. (This is then used by ssync in the object
server to create a cross-worker semaphore to restrict
concurrent incoming replication.)
2) A bit of shifting of items up from object server and
replicator to diskfile or DEFAULT conf sections for
better sharing of the same settings. conn_timeout,
node_timeout, client_timeout, network_chunk_size,
disk_chunk_size.
3) Modifications to the object server and replicator to
optionally use ssync in place of rsync. This is done in
a generic enough way that switching to FutureSync should
be easy someday.
4) The biggest part, and (at least for now) completely
optional part, are the new ssync_sender and
ssync_receiver files. Nice and isolated for easier
testing and visibility into test coverage, etc.
All the usual logging, statsd, recon, etc. instrumentation
is still there when using ssync, just as it is when using
rsync.
Beyond the essential error and exceptional condition
logging, I have not added any additional instrumentation at
this time. Unless there is something someone finds super
pressing to have added to the logging, I think such
additions would be better as separate change reviews.
FOR NOW, IT IS NOT RECOMMENDED TO USE SSYNC ON PRODUCTION
CLUSTERS. Some of us will be in a limited fashion to look
for any subtle issues, tuning, etc. but generally ssync is
an experimental feature. In its current implementation it is
probably going to be a bit slower than rsync, but if all
goes according to plan it will end up much faster.
There are no comparisions yet between ssync and rsync other
than some raw virtual machine testing I've done to show it
should compete well enough once we can put it in use in the
real world.
If you Tweet, Google+, or whatever, be sure to indicate it's
experimental. It'd be best to keep it out of deployment
guides, howtos, etc. until we all figure out if we like it,
find it to be stable, etc.
Change-Id: If003dcc6f4109e2d2a42f4873a0779110fff16d6
2013-08-28 16:10:43 +00:00
|
|
|
repl_methods = ['REPLICATE', 'REPLICATION']
|
2013-05-23 20:16:21 +04:00
|
|
|
for method_name in obj_methods:
|
|
|
|
method = getattr(self.object_controller, method_name)
|
|
|
|
self.assertFalse(hasattr(method, 'replication'))
|
|
|
|
for method_name in repl_methods:
|
|
|
|
method = getattr(self.object_controller, method_name)
|
|
|
|
self.assertEquals(method.replication, True)
|
2012-12-17 06:39:25 -05:00
|
|
|
|
|
|
|
def test_correct_allowed_method(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test correct work for allowed method using
|
|
|
|
# swift.obj.server.ObjectController.__call__
|
2012-12-17 06:39:25 -05:00
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
2013-12-11 16:59:59 -05:00
|
|
|
self.object_controller = object_server.app_factory(
|
2013-05-23 20:16:21 +04:00
|
|
|
{'devices': self.testdir, 'mount_check': 'false',
|
|
|
|
'replication_server': 'false'})
|
2012-12-17 06:39:25 -05:00
|
|
|
|
|
|
|
def start_response(*args):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Sends args to outbuf
|
2012-12-17 06:39:25 -05:00
|
|
|
outbuf.writelines(args)
|
|
|
|
|
2013-05-23 20:16:21 +04:00
|
|
|
method = 'PUT'
|
2012-12-17 06:39:25 -05:00
|
|
|
env = {'REQUEST_METHOD': method,
|
|
|
|
'SCRIPT_NAME': '',
|
2013-12-11 16:59:59 -05:00
|
|
|
'PATH_INFO': '/sda1/p/a/c/o',
|
2012-12-17 06:39:25 -05:00
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False}
|
|
|
|
|
2013-05-23 20:16:21 +04:00
|
|
|
method_res = mock.MagicMock()
|
2014-05-27 01:17:13 -07:00
|
|
|
mock_method = public(lambda x:
|
|
|
|
mock.MagicMock(return_value=method_res))
|
2012-12-17 06:39:25 -05:00
|
|
|
with mock.patch.object(self.object_controller, method,
|
2013-05-23 20:16:21 +04:00
|
|
|
new=mock_method):
|
2012-12-17 06:39:25 -05:00
|
|
|
response = self.object_controller.__call__(env, start_response)
|
2013-05-23 20:16:21 +04:00
|
|
|
self.assertEqual(response, method_res)
|
2012-12-17 06:39:25 -05:00
|
|
|
|
|
|
|
def test_not_allowed_method(self):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Test correct work for NOT allowed method using
|
|
|
|
# swift.obj.server.ObjectController.__call__
|
2012-12-17 06:39:25 -05:00
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
2013-05-23 20:16:21 +04:00
|
|
|
self.object_controller = object_server.ObjectController(
|
|
|
|
{'devices': self.testdir, 'mount_check': 'false',
|
2013-12-11 16:59:59 -05:00
|
|
|
'replication_server': 'false'}, logger=FakeLogger())
|
2012-12-17 06:39:25 -05:00
|
|
|
|
|
|
|
def start_response(*args):
|
2013-08-31 22:36:58 -04:00
|
|
|
# Sends args to outbuf
|
2012-12-17 06:39:25 -05:00
|
|
|
outbuf.writelines(args)
|
|
|
|
|
2013-05-23 20:16:21 +04:00
|
|
|
method = 'PUT'
|
2012-12-17 06:39:25 -05:00
|
|
|
|
|
|
|
env = {'REQUEST_METHOD': method,
|
|
|
|
'SCRIPT_NAME': '',
|
2013-12-11 16:59:59 -05:00
|
|
|
'PATH_INFO': '/sda1/p/a/c/o',
|
2012-12-17 06:39:25 -05:00
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False}
|
|
|
|
|
|
|
|
answer = ['<html><h1>Method Not Allowed</h1><p>The method is not '
|
|
|
|
'allowed for this resource.</p></html>']
|
2013-05-23 20:16:21 +04:00
|
|
|
mock_method = replication(public(lambda x: mock.MagicMock()))
|
2012-12-17 06:39:25 -05:00
|
|
|
with mock.patch.object(self.object_controller, method,
|
2013-05-23 20:16:21 +04:00
|
|
|
new=mock_method):
|
|
|
|
mock_method.replication = True
|
2013-12-11 16:59:59 -05:00
|
|
|
with mock.patch('time.gmtime',
|
|
|
|
mock.MagicMock(side_effect=[gmtime(10001.0)])):
|
|
|
|
with mock.patch('time.time',
|
|
|
|
mock.MagicMock(side_effect=[10000.0,
|
|
|
|
10001.0])):
|
2014-07-07 17:11:38 -07:00
|
|
|
with mock.patch('os.getpid',
|
|
|
|
mock.MagicMock(return_value=1234)):
|
|
|
|
response = self.object_controller.__call__(
|
|
|
|
env, start_response)
|
|
|
|
self.assertEqual(response, answer)
|
|
|
|
self.assertEqual(
|
|
|
|
self.object_controller.logger.log_dict['info'],
|
|
|
|
[(('None - - [01/Jan/1970:02:46:41 +0000] "PUT'
|
|
|
|
' /sda1/p/a/c/o" 405 - "-" "-" "-" 1.0000 "-"'
|
|
|
|
' 1234',),
|
|
|
|
{})])
|
2013-12-11 16:59:59 -05:00
|
|
|
|
|
|
|
def test_not_utf8_and_not_logging_requests(self):
|
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
|
|
|
self.object_controller = object_server.ObjectController(
|
|
|
|
{'devices': self.testdir, 'mount_check': 'false',
|
|
|
|
'replication_server': 'false', 'log_requests': 'false'},
|
|
|
|
logger=FakeLogger())
|
|
|
|
|
|
|
|
def start_response(*args):
|
|
|
|
# Sends args to outbuf
|
|
|
|
outbuf.writelines(args)
|
|
|
|
|
|
|
|
method = 'PUT'
|
|
|
|
|
|
|
|
env = {'REQUEST_METHOD': method,
|
|
|
|
'SCRIPT_NAME': '',
|
|
|
|
'PATH_INFO': '/sda1/p/a/c/\x00%20/%',
|
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False}
|
|
|
|
|
|
|
|
answer = ['Invalid UTF8 or contains NULL']
|
|
|
|
mock_method = public(lambda x: mock.MagicMock())
|
|
|
|
with mock.patch.object(self.object_controller, method,
|
|
|
|
new=mock_method):
|
2012-12-17 06:39:25 -05:00
|
|
|
response = self.object_controller.__call__(env, start_response)
|
|
|
|
self.assertEqual(response, answer)
|
2013-12-11 16:59:59 -05:00
|
|
|
self.assertEqual(self.object_controller.logger.log_dict['info'],
|
|
|
|
[])
|
|
|
|
|
|
|
|
def test__call__returns_500(self):
|
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
|
|
|
self.object_controller = object_server.ObjectController(
|
|
|
|
{'devices': self.testdir, 'mount_check': 'false',
|
|
|
|
'replication_server': 'false', 'log_requests': 'false'},
|
|
|
|
logger=FakeLogger())
|
|
|
|
|
|
|
|
def start_response(*args):
|
|
|
|
# Sends args to outbuf
|
|
|
|
outbuf.writelines(args)
|
|
|
|
|
|
|
|
method = 'PUT'
|
|
|
|
|
|
|
|
env = {'REQUEST_METHOD': method,
|
|
|
|
'SCRIPT_NAME': '',
|
|
|
|
'PATH_INFO': '/sda1/p/a/c/o',
|
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False}
|
|
|
|
|
|
|
|
@public
|
|
|
|
def mock_put_method(*args, **kwargs):
|
|
|
|
raise Exception()
|
|
|
|
|
|
|
|
with mock.patch.object(self.object_controller, method,
|
|
|
|
new=mock_put_method):
|
|
|
|
response = self.object_controller.__call__(env, start_response)
|
|
|
|
self.assertTrue(response[0].startswith(
|
|
|
|
'Traceback (most recent call last):'))
|
|
|
|
self.assertEqual(
|
|
|
|
self.object_controller.logger.log_dict['exception'],
|
|
|
|
[(('ERROR __call__ error with %(method)s %(path)s ',
|
|
|
|
{'method': 'PUT', 'path': '/sda1/p/a/c/o'}),
|
|
|
|
{},
|
|
|
|
'')])
|
|
|
|
self.assertEqual(self.object_controller.logger.log_dict['INFO'],
|
|
|
|
[])
|
|
|
|
|
|
|
|
def test_PUT_slow(self):
|
|
|
|
inbuf = StringIO()
|
|
|
|
errbuf = StringIO()
|
|
|
|
outbuf = StringIO()
|
|
|
|
self.object_controller = object_server.ObjectController(
|
|
|
|
{'devices': self.testdir, 'mount_check': 'false',
|
|
|
|
'replication_server': 'false', 'log_requests': 'false',
|
|
|
|
'slow': '10'},
|
|
|
|
logger=FakeLogger())
|
|
|
|
|
|
|
|
def start_response(*args):
|
|
|
|
# Sends args to outbuf
|
|
|
|
outbuf.writelines(args)
|
|
|
|
|
|
|
|
method = 'PUT'
|
|
|
|
|
|
|
|
env = {'REQUEST_METHOD': method,
|
|
|
|
'SCRIPT_NAME': '',
|
|
|
|
'PATH_INFO': '/sda1/p/a/c/o',
|
|
|
|
'SERVER_NAME': '127.0.0.1',
|
|
|
|
'SERVER_PORT': '8080',
|
|
|
|
'SERVER_PROTOCOL': 'HTTP/1.0',
|
|
|
|
'CONTENT_LENGTH': '0',
|
|
|
|
'wsgi.version': (1, 0),
|
|
|
|
'wsgi.url_scheme': 'http',
|
|
|
|
'wsgi.input': inbuf,
|
|
|
|
'wsgi.errors': errbuf,
|
|
|
|
'wsgi.multithread': False,
|
|
|
|
'wsgi.multiprocess': False,
|
|
|
|
'wsgi.run_once': False}
|
|
|
|
|
|
|
|
mock_method = public(lambda x: mock.MagicMock())
|
|
|
|
with mock.patch.object(self.object_controller, method,
|
|
|
|
new=mock_method):
|
|
|
|
with mock.patch('time.time',
|
|
|
|
mock.MagicMock(side_effect=[10000.0,
|
|
|
|
10001.0])):
|
|
|
|
with mock.patch('swift.obj.server.sleep',
|
|
|
|
mock.MagicMock()) as ms:
|
|
|
|
self.object_controller.__call__(env, start_response)
|
|
|
|
ms.assert_called_with(9)
|
|
|
|
self.assertEqual(
|
|
|
|
self.object_controller.logger.log_dict['info'], [])
|
2012-12-17 06:39:25 -05:00
|
|
|
|
2014-03-26 22:55:55 +00:00
|
|
|
def test_log_line_format(self):
|
|
|
|
req = Request.blank(
|
|
|
|
'/sda1/p/a/c/o',
|
|
|
|
environ={'REQUEST_METHOD': 'HEAD', 'REMOTE_ADDR': '1.2.3.4'})
|
|
|
|
self.object_controller.logger = FakeLogger()
|
|
|
|
with mock.patch(
|
|
|
|
'time.gmtime', mock.MagicMock(side_effect=[gmtime(10001.0)])):
|
|
|
|
with mock.patch(
|
|
|
|
'time.time',
|
|
|
|
mock.MagicMock(side_effect=[10000.0, 10001.0, 10002.0])):
|
2014-07-07 17:11:38 -07:00
|
|
|
with mock.patch(
|
|
|
|
'os.getpid', mock.MagicMock(return_value=1234)):
|
|
|
|
req.get_response(self.object_controller)
|
2014-03-26 22:55:55 +00:00
|
|
|
self.assertEqual(
|
|
|
|
self.object_controller.logger.log_dict['info'],
|
|
|
|
[(('1.2.3.4 - - [01/Jan/1970:02:46:41 +0000] "HEAD /sda1/p/a/c/o" '
|
2014-07-07 17:11:38 -07:00
|
|
|
'404 - "-" "-" "-" 2.0000 "-" 1234',), {})])
|
2014-03-26 22:55:55 +00:00
|
|
|
|
2014-05-27 01:17:13 -07:00
|
|
|
@patch_policies([storage_policy.StoragePolicy(0, 'zero', True),
|
|
|
|
storage_policy.StoragePolicy(1, 'one', False)])
|
|
|
|
def test_dynamic_datadir(self):
|
|
|
|
timestamp = normalize_timestamp(time())
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test',
|
|
|
|
'Foo': 'fooheader',
|
|
|
|
'Baz': 'bazheader',
|
2014-06-23 12:52:50 -07:00
|
|
|
'X-Backend-Storage-Policy-Index': 1,
|
2014-05-27 01:17:13 -07:00
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-Two': 'Two'})
|
|
|
|
req.body = 'VERIFY'
|
|
|
|
object_dir = self.testdir + "/sda1/objects-1"
|
|
|
|
self.assertFalse(os.path.isdir(object_dir))
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
self.assertTrue(os.path.isdir(object_dir))
|
|
|
|
|
|
|
|
# make sure no idx in header uses policy 0 data_dir
|
|
|
|
req = Request.blank('/sda1/p/a/c/o', environ={'REQUEST_METHOD': 'PUT'},
|
|
|
|
headers={'X-Timestamp': timestamp,
|
|
|
|
'Content-Type': 'application/x-test',
|
|
|
|
'Foo': 'fooheader',
|
|
|
|
'Baz': 'bazheader',
|
|
|
|
'X-Object-Meta-1': 'One',
|
|
|
|
'X-Object-Meta-Two': 'Two'})
|
|
|
|
req.body = 'VERIFY'
|
|
|
|
object_dir = self.testdir + "/sda1/objects"
|
|
|
|
self.assertFalse(os.path.isdir(object_dir))
|
|
|
|
with mock.patch.object(POLICIES, 'get_by_index',
|
|
|
|
lambda _: True):
|
|
|
|
resp = req.get_response(self.object_controller)
|
|
|
|
self.assertEquals(resp.status_int, 201)
|
|
|
|
self.assertTrue(os.path.isdir(object_dir))
|
2012-12-17 06:39:25 -05:00
|
|
|
|
2010-07-12 17:03:45 -05:00
|
|
|
if __name__ == '__main__':
|
|
|
|
unittest.main()
|