9e65dd9ceb
fallocate() allows for reserving disk space for a particular inode and fd. Hence, a client can be sure that he won't see a ENOSPC (eventually a 507 HTTP response) later during writes. Swift's object server has had fallocate() support from a long time. P.S: Older versions of glusterfs (<3.6) did not support fallocate because FUSE did not support it earlier. http://review.gluster.org/4969 http://fuse.996288.n3.nabble.com/fallocate-support-in-FUSE-td10668.html Change-Id: Ida4b16357901707d624f92bf1b2dc8f07da4f1ad Signed-off-by: Prashanth Pai <ppai@redhat.com>
69 lines
2.8 KiB
Plaintext
69 lines
2.8 KiB
Plaintext
[DEFAULT]
|
|
#
|
|
# Default swiftonfile mount point to be used for object server
|
|
# This mount point is the parent directory to the device specified
|
|
# in this swiftonfile object ring.
|
|
devices = /mnt/swiftonfile
|
|
#
|
|
# You must always have the volume properly mounted *before* starting the
|
|
# SwiftOnFile object-server.
|
|
mount_check = false
|
|
bind_port = 6050
|
|
#
|
|
# Maximum number of clients one worker can process simultaneously (it will
|
|
# actually accept N + 1). Setting this to one (1) will only handle one request
|
|
# at a time, without accepting another request concurrently. By increasing the
|
|
# number of workers to a much higher value, one can prevent slow file system
|
|
# operations for one request from starving other requests.
|
|
max_clients = 1024
|
|
#
|
|
# If not doing the above, setting this value initially to match the number of
|
|
# CPUs is a good starting point for determining the right value.
|
|
workers = 1
|
|
# You can set disable_fallocate to true to turn off usage of fallocate()
|
|
# sys call to reserve space during a PUT operation.
|
|
# disable_fallocate = false
|
|
#
|
|
# You can set fallocate_reserve to the number of bytes you'd like fallocate to
|
|
# reserve, whether there is space for the given file size or not.
|
|
# fallocate_reserve = 0
|
|
|
|
[pipeline:main]
|
|
pipeline = object-server
|
|
|
|
[app:object-server]
|
|
use = egg:swiftonfile#object
|
|
user = <your-user-name>
|
|
log_facility = LOG_LOCAL2
|
|
log_level = WARN
|
|
# For performance, after ensuring things are running in a stable manner, you
|
|
# can turn off normal request logging for the object server to reduce the
|
|
# per-request overhead and unclutter the log files. Warnings and errors will
|
|
# still be logged.
|
|
log_requests = off
|
|
#
|
|
# Adjust this value to match the stripe width of the underlying storage array
|
|
# (not the stripe element size). This will provide a reasonable starting point
|
|
# for tuning this value.
|
|
disk_chunk_size = 65536
|
|
#
|
|
# Adjust this value match whatever is set for the disk_chunk_size initially.
|
|
# This will provide a reasonable starting point for tuning this value.
|
|
network_chunk_size = 65536
|
|
|
|
# In older versions of swiftonfile, metadata stored as xattrs of files
|
|
# were serialized using PICKLE format. The PICKLE format is vulnerable to
|
|
# exploits in deployments where a user has access to backend filesystem.
|
|
# Deserializing pickled metadata can result in malicious code being
|
|
# executed if an attacker has stored malicious code as xattr from filesystem
|
|
# interface. Although, new metadata is always serialized using JSON format,
|
|
# existing metadata already stored in PICKLE format can be loaded by setting
|
|
# the following option to 'on'.
|
|
# You can turn this option to 'off' once you have migrated all your metadata
|
|
# from PICKLE format to JSON format using swiftonfile-migrate-metadata tool.
|
|
# This conf option will be deprecated and eventualy removed in future releases
|
|
# read_pickled_metadata = off
|
|
|
|
[object-updater]
|
|
user = <your-user-name>
|