Thanks Strahil,
I'm attaching a text file with the result of volume info and one with
the result of set help.
I might not have made myself clear. My gluster cluster is running
glusterfs 5.13 (I know it's old, but I want to wait to retire the legacy
systems before upgrading). My other servers use different versions of
glusterfs from 3.7.9 to 6.0. The Ovirt server was using v. 7.9 without a
problem, when it upgraded to 8.4 it stopped working properly (mounting
and creating directories and copying files works, creating file does
not). I only want to downgrade the Ovirt server from 8.4 to 7.9 without
touching the gluster cluster or the other servers.
On 3/28/21 10:30 AM, Strahil Nikolov wrote:
> Can you provide the volume info ?
>
> I have previously managed to downgrade v7.0 to 6.6 , but it's risky -
> so nobody will recommend you. Previously, the shard size was 64MB by
> default. It will be interesting to see the current and default value
> (gluster volume set help).
>
>
> Best Regards,
> Strahil Nikolov
>
> On Sun, Mar 28, 2021 at 9:20, Valerio Luccio
> <valerio.luccio at nyu.edu> wrote:
>
> Hello all,
>
> I have a gluster storage that is still running gluster 5.
> Unfortunately I have some legacy systems that cannot be updated
> that use this storage (I'm hoping to replace these systems in the
> next 12 months, I do all of this by my lonesome so it takes some
> time).
>
> On a CentOS 8 server I run Ovirt to manage some VM's. It was using
> gluster 7 to connect to the storage and that worked. A couple of
> days ago I did an update and Ovirt updated my gluster packages to
> 8, after which I now get 'Failed to get
> trusted.glusterfs.shard.file-size' when trying to create files
> from the Ovirt server (the other servers don't have issues). Two
> related questions:
>
> 1. Is there a way to fix this, maybe changing some parameter in a
> conf file ?
> 2. If 1. is not possible, can you see pitfalls in downgrading the
> packages to gluster 7 after downloading them from
> http://mirror.centos.org <http://mirror.centos.org> ?
>
> Thanks in advance.
> --
> As a result of Coronavirus-related precautions, NYU and the Center
> for Brain Imaging operations will be managed remotely until
> further notice.
> All telephone calls and e-mail correspondence are being monitored
> remotely during our normal business hours of 9am-5pm, Monday
> through Friday.
> For MRI scanner-related emergency, please contact: Keith
> Sanzenbach at keith.sanzenbach at nyu.edu
> <mailto:keith.sanzenbach at nyu.edu> and/or Pablo Velasco at
> pablo.velasco at nyu.edu <mailto:pablo.velasco at nyu.edu>
> For computer/hardware/software emergency, please contact: Valerio
> Luccio at valerio.luccio at nyu.edu <mailto:valerio.luccio at
nyu.edu>
> For TMS/EEG-related emergency, please contact: Chrysa Papadaniil
> at chrysa at nyu.edu <mailto:chrysa at nyu.edu>
> For CBI-related administrative emergency, please contact: Jennifer
> Mangan at jennifer.mangan at nyu.edu <mailto:jennifer.mangan at
nyu.edu>
>
> Valerio Luccio (212) 998-8736
> Center for Brain Imaging 4 Washington Place, Room 158
> New York University New York, NY 10003
>
> "In an open world, who needs windows or gates ?"
>
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://meet.google.com/cpu-eiue-hvk
> <https://meet.google.com/cpu-eiue-hvk>
> Gluster-users mailing list
> Gluster-users at gluster.org <mailto:Gluster-users at
gluster.org>
> https://lists.gluster.org/mailman/listinfo/gluster-users
> <https://lists.gluster.org/mailman/listinfo/gluster-users>
>
--
As a result of Coronavirus-related precautions, NYU and the Center for
Brain Imaging operations will be managed remotely until further notice.
All telephone calls and e-mail correspondence are being monitored
remotely during our normal business hours of 9am-5pm, Monday through
Friday.
For MRI scanner-related emergency, please contact: Keith Sanzenbach at
keith.sanzenbach at nyu.edu and/or Pablo Velasco at pablo.velasco at nyu.edu
For computer/hardware/software emergency, please contact: Valerio Luccio
at valerio.luccio at nyu.edu
For TMS/EEG-related emergency, please contact: Chrysa Papadaniil at
chrysa at nyu.edu
For CBI-related administrative emergency, please contact: Jennifer
Mangan at jennifer.mangan at nyu.edu
Valerio Luccio (212) 998-8736
Center for Brain Imaging 4 Washington Place, Room 158
New York University New York, NY 10003
"In an open world, who needs windows or gates ?"
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://lists.gluster.org/pipermail/gluster-users/attachments/20210328/a31a9a05/attachment.html>
-------------- next part --------------
Volume Name: MRIData
Type: Distributed-Replicate
Volume ID: e051ac20-ead1-4648-9ac6-a29b531515ca
Status: Started
Snapshot Count: 0
Number of Bricks: 6 x (2 + 1) = 18
Transport-type: tcp
Bricks:
Brick1: hydra1:/gluster1/data
Brick2: hydra1:/gluster2/data
Brick3: hydra1:/arbiter/1 (arbiter)
Brick4: hydra1:/gluster3/data
Brick5: hydra2:/gluster1/data
Brick6: hydra1:/arbiter/2 (arbiter)
Brick7: hydra2:/gluster2/data
Brick8: hydra2:/gluster3/data
Brick9: hydra2:/arbiter/1 (arbiter)
Brick10: hydra3:/gluster1/data
Brick11: hydra3:/gluster2/data
Brick12: hydra3:/arbiter/1 (arbiter)
Brick13: hydra3:/gluster3/data
Brick14: hydra4:/gluster1/data
Brick15: hydra3:/arbiter/2 (arbiter)
Brick16: hydra4:/gluster2/data
Brick17: hydra4:/gluster3/data
Brick18: hydra4:/arbiter/1 (arbiter)
Options Reconfigured:
features.cache-invalidation: off
nfs.exports-auth-enable: on
nfs.disable: on
transport.address-family: inet
cluster.data-self-heal: on
cluster.metadata-self-heal: on
cluster.entry-self-heal: on
cluster.self-heal-daemon: on
cluster.quorum-type: auto
server.allow-insecure: on
network.ping-timeout: 10
auth.allow: *
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.low-prio-threads: 32
network.remote-dio: enable
cluster.eager-lock: enable
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
cluster.choose-local: off
storage.owner-uid: 36
storage.owner-gid: 36
-------------- next part --------------
Option: cluster.lookup-unhashed
Default Value: on
Description: This option if set to ON, does a lookup through all the
sub-volumes, in case a lookup didn't return any result from the hash
subvolume. If set to OFF, it does not do a lookup on the remaining subvolumes.
Option: cluster.lookup-optimize
Default Value: on
Description: This option if set to ON enables the optimization of -ve lookups,
by not doing a lookup on non-hashed subvolumes for files, in case the hashed
subvolume does not return any result. This option disregards the lookup-unhashed
setting, when enabled.
Option: cluster.min-free-disk
Default Value: 10%
Description: Percentage/Size of disk space, after which the process starts
balancing out the cluster, and logs will appear in log files
Option: cluster.min-free-inodes
Default Value: 5%
Description: after system has only N% of inodes, warnings starts to appear in
log files
Option: cluster.rebalance-stats
Default Value: off
Description: This option if set to ON displays and logs the time taken for
migration of each file, during the rebalance process. If set to OFF, the
rebalance logs will only display the time spent in each directory.
Option: cluster.subvols-per-directory
Default Value: (null)
Description: Specifies the directory layout spread. Takes number of subvolumes
as default value.
Option: cluster.readdir-optimize
Default Value: off
Description: This option if set to ON enables the optimization that allows DHT
to requests non-first subvolumes to filter out directory entries.
Option: cluster.rebal-throttle
Default Value: normal
Description: Sets the maximum number of parallel file migrations allowed on a
node during the rebalance operation. The default value is normal and allows a
max of [($(processing units) - 4) / 2), 2] files to be migrated at a time. Lazy
will allow only one file to be migrated at a time and aggressive will allow max
of [($(processing units) - 4) / 2), 4]
Option: cluster.lock-migration
Default Value: off
Description: If enabled this feature will migrate the posix locks associated
with a file during rebalance
Option: cluster.force-migration
Default Value: off
Description: If disabled, rebalance will not migrate files that are being
written to by an application
Option: cluster.weighted-rebalance
Default Value: on
Description: When enabled, files will be allocated to bricks with a probability
proportional to their size. Otherwise, all bricks will have the same
probability (legacy behavior).
Option: cluster.entry-change-log
Default Value: on
Description: This option exists only for backward compatibility and configuring
it doesn't have any effect
Option: cluster.read-subvolume
Default Value: (null)
Description: inode-read fops happen only on one of the bricks in replicate. Afr
will prefer the one specified using this option if it is not stale. Option value
must be one of the xlator names of the children. Ex: <volname>-client-0
till <volname>-client-<number-of-bricks - 1>
Option: cluster.read-subvolume-index
Default Value: -1
Description: inode-read fops happen only on one of the bricks in replicate. AFR
will prefer the one specified using this option if it is not stale. allowed
options include -1 till replica-count - 1
Option: cluster.read-hash-mode
Default Value: 1
Description: inode-read fops happen only on one of the bricks in replicate. AFR
will prefer the one computed using the method specified using this option.
0 = first readable child of AFR, starting from 1st child.
1 = hash by GFID of file (all clients use same subvolume).
2 = hash by GFID of file and client PID.
3 = brick having the least outstanding read requests.
Option: cluster.background-self-heal-count
Default Value: 8
Description: This specifies the number of per client self-heal jobs that can
perform parallel heals in the background.
Option: cluster.metadata-self-heal
Default Value: on
Description: Using this option we can enable/disable metadata i.e. Permissions,
ownerships, xattrs self-heal on the file/directory.
Option: cluster.data-self-heal
Default Value: on
Description: Using this option we can enable/disable data self-heal on the file.
"open" means data self-heal action will only be triggered by file open
operations.
Option: cluster.entry-self-heal
Default Value: on
Description: Using this option we can enable/disable entry self-heal on the
directory.
Option: cluster.self-heal-daemon
Default Value: on
Description: This option applies to only self-heal-daemon. Index directory crawl
and automatic healing of files will not be performed if this option is turned
off.
Option: cluster.heal-timeout
Default Value: 600
Description: time interval for checking the need to self-heal in
self-heal-daemon
Option: cluster.self-heal-window-size
Default Value: 1
Description: Maximum number blocks per file for which self-heal process would be
applied simultaneously.
Option: cluster.data-change-log
Default Value: on
Description: This option exists only for backward compatibility and configuring
it doesn't have any effect
Option: cluster.metadata-change-log
Default Value: on
Description: This option exists only for backward compatibility and configuring
it doesn't have any effect
Option: cluster.data-self-heal-algorithm
Default Value: (null)
Description: Select between "full", "diff". The
"full" algorithm copies the entire file from source to sink. The
"diff" algorithm copies to sink only those blocks whose checksums
don't match with those of source. If no option is configured the option is
chosen dynamically as follows: If the file does not exist on one of the sinks or
empty file exists or if the source file size is about the same as page size the
entire file will be read and written i.e "full" algo, otherwise
"diff" algo is chosen.
Option: cluster.eager-lock
Default Value: on
Description: Enable/Disable eager lock for replica volume. Lock phase of a
transaction has two sub-phases. First is an attempt to acquire locks in parallel
by broadcasting non-blocking lock requests. If lock acquisition fails on any
server, then the held locks are unlocked and we revert to a blocking locks mode
sequentially on one server after another. If this option is enabled the initial
broadcasting lock request attempts to acquire a full lock on the entire file. If
this fails, we revert back to the sequential "regional" blocking locks
as before. In the case where such an "eager" lock is granted in the
non-blocking phase, it gives rise to an opportunity for optimization. i.e, if
the next write transaction on the same FD arrives before the unlock phase of the
first transaction, it "takes over" the full file lock. Similarly if
yet another data transaction arrives before the unlock phase of the
"optimized" transaction, that in turn "takes over" the lock
as well. The actual unlock now happens at the end of the last
"optimized" transaction.
Option: disperse.eager-lock
Default Value: on
Description: Enable/Disable eager lock for regular files on a disperse volume.
If a fop takes a lock and completes its operation, it waits for next 1 second
before releasing the lock, to see if the lock can be reused for next fop from
the same client. If ec finds any lock contention within 1 second it releases the
lock immediately before time expires. This improves the performance of file
operations. However, as it takes lock on first brick, for few operations like
read, discovery of lock contention might take long time and can actually degrade
the performance. If eager lock is disabled, lock will be released as soon as fop
completes.
Option: disperse.other-eager-lock
Default Value: on
Description: It's equivalent to the eager-lock option but for non regular
files.
Option: disperse.eager-lock-timeout
Default Value: 1
Description: Maximum time (in seconds) that a lock on an inode is kept held if
no new operations on the inode are received.
Option: disperse.other-eager-lock-timeout
Default Value: 1
Description: It's equivalent to eager-lock-timeout option but for non
regular files.
Option: cluster.quorum-type
Default Value: none
Description: If value is "fixed" only allow writes if quorum-count
bricks are present. If value is "auto" only allow writes if more than
half of bricks, or exactly half including the first, are present.
Option: cluster.quorum-count
Default Value: (null)
Description: If quorum-type is "fixed" only allow writes if this many
bricks are present. Other quorum types will OVERWRITE this value.
Option: cluster.choose-local
Default Value: true
Description: Choose a local subvolume (i.e. Brick) to read from if
read-subvolume is not explicitly set.
Option: cluster.self-heal-readdir-size
Default Value: 1KB
Description: readdirp size for performing entry self-heal
Option: cluster.ensure-durability
Default Value: on
Description: Afr performs fsyncs for transactions if this option is on to make
sure the changelogs/data is written to the disk
Option: cluster.consistent-metadata
Default Value: no
Description: If this option is enabled, readdirp will force lookups on those
entries read whose read child is not the same as that of the parent. This will
guarantee that all read operations on a file serve attributes from the same
subvol as long as it holds a good copy of the file/dir.
Option: cluster.heal-wait-queue-length
Default Value: 128
Description: This specifies the number of heals that can be queued for the
parallel background self heal jobs.
Option: cluster.favorite-child-policy
Default Value: none
Description: This option can be used to automatically resolve split-brains using
various policies without user intervention. "size" picks the file with
the biggest size as the source. "ctime" and "mtime" pick the
file with the latest ctime and mtime respectively as the source.
"majority" picks a file with identical mtime and size in more than
half the number of bricks in the replica.
Option: cluster.stripe-block-size
Default Value: 128KB
Description: Size of the stripe unit that would be read from or written to the
striped servers.
Option: cluster.stripe-coalesce
Default Value: true
Description: Enable/Disable coalesce mode to flatten striped files as stored on
the server (i.e., eliminate holes caused by the traditional format).
Option: diagnostics.latency-measurement
Default Value: off
Description: If on stats related to the latency of each operation would be
tracked inside GlusterFS data-structures.
Option: diagnostics.dump-fd-stats
Default Value: off
Description: If on stats related to file-operations would be tracked inside
GlusterFS data-structures.
Option: diagnostics.brick-log-level
Default Value: INFO
Description: Changes the log-level of the bricks
Option: diagnostics.client-log-level
Default Value: INFO
Description: Changes the log-level of the clients
Option: diagnostics.brick-sys-log-level
Default Value: CRITICAL
Description: Gluster's syslog log-level
Option: diagnostics.client-sys-log-level
Default Value: CRITICAL
Description: Gluster's syslog log-level
Option: diagnostics.brick-logger
Default Value: (null)
Description: (null)
Option: diagnostics.client-logger
Default Value: (null)
Description: (null)
Option: diagnostics.brick-log-format
Default Value: (null)
Description: (null)
Option: diagnostics.client-log-format
Default Value: (null)
Description: (null)
Option: diagnostics.brick-log-buf-size
Default Value: 5
Description: (null)
Option: diagnostics.client-log-buf-size
Default Value: 5
Description: (null)
Option: diagnostics.brick-log-flush-timeout
Default Value: 120
Description: (null)
Option: diagnostics.client-log-flush-timeout
Default Value: 120
Description: (null)
Option: diagnostics.stats-dump-interval
Default Value: 0
Description: Interval (in seconds) at which to auto-dump statistics. Zero
disables automatic dumping.
Option: diagnostics.fop-sample-interval
Default Value: 0
Description: Interval in which we want to collect FOP latency samples. 2 means
collect a sample every 2nd FOP.
Option: diagnostics.stats-dump-format
Default Value: json
Description: The dump-format option specifies the format in which to dump the
statistics. Select between "text", "json", "dict"
and "samples". Default is "json".
Option: diagnostics.fop-sample-buf-size
Default Value: 65535
Description: The maximum size of our FOP sampling ring buffer.
Option: diagnostics.stats-dnscache-ttl-sec
Default Value: 86400
Description: The interval after wish a cached DNS entry will be re-validated.
Default: 24 hrs
Option: performance.cache-max-file-size
Default Value: 0
Description: Maximum file size which would be cached by the io-cache translator.
Option: performance.cache-min-file-size
Default Value: 0
Description: Minimum file size which would be cached by the io-cache translator.
Option: performance.cache-refresh-timeout
Default Value: 1
Description: The cached data for a file will be retained for
'cache-refresh-timeout' seconds, after which data re-validation is
performed.
Option: performance.cache-priority
Default Value:
Description: Assigns priority to filenames with specific patterns so that when a
page needs to be ejected out of the cache, the page of a file whose priority is
the lowest will be ejected earlier
Option: performance.cache-size
Default Value: 32MB
Description: Size of the read cache.
Option: performance.io-thread-count
Default Value: 16
Description: Number of threads in IO threads translator which perform concurrent
IO operations
Option: performance.high-prio-threads
Default Value: 16
Description: Max number of threads in IO threads translator which perform high
priority IO operations at a given time
Option: performance.normal-prio-threads
Default Value: 16
Description: Max number of threads in IO threads translator which perform normal
priority IO operations at a given time
Option: performance.low-prio-threads
Default Value: 16
Description: Max number of threads in IO threads translator which perform low
priority IO operations at a given time
Option: performance.least-prio-threads
Default Value: 1
Description: Max number of threads in IO threads translator which perform least
priority IO operations at a given time
Option: performance.enable-least-priority
Default Value: on
Description: Enable/Disable least priority
Option: performance.iot-watchdog-secs
Default Value: (null)
Description: Number of seconds a queue must be stalled before starting an
'emergency' thread.
Option: performance.iot-cleanup-disconnected-reqs
Default Value: off
Description: 'Poison' queued requests when a client disconnects
Option: performance.iot-pass-through
Default Value: false
Description: Enable/Disable io threads translator
Option: performance.io-cache-pass-through
Default Value: false
Description: Enable/Disable io cache translator
Option: performance.qr-cache-timeout
Default Value: 1
Description: (null)
Option: performance.cache-invalidation
Default Value: false
Description: When "on", invalidates/updates the metadata cache, on
receiving the cache-invalidation notifications
Option: performance.ctime-invalidation
Default Value: false
Description: Quick-read by default uses mtime to identify changes to file data.
However there are applications like rsync which explicitly set mtime making it
unreliable for the purpose of identifying change in file content . Since ctime
also changes when content of a file changes and it cannot be set explicitly, it
becomes suitable for identifying staleness of cached data. This option makes
quick-read to prefer ctime over mtime to validate its cache. However, using
ctime can result in false positives as ctime changes with just attribute changes
like permission without changes to file data. So, use this only when mtime is
not reliable
Option: performance.flush-behind
Default Value: on
Description: If this option is set ON, instructs write-behind translator to
perform flush in background, by returning success (or any errors, if any of
previous writes were failed) to application even before flush FOP is sent to
backend filesystem.
Option: performance.nfs.flush-behind
Default Value: on
Description: If this option is set ON, instructs write-behind translator to
perform flush in background, by returning success (or any errors, if any of
previous writes were failed) to application even before flush FOP is sent to
backend filesystem.
Option: performance.write-behind-window-size
Default Value: 1MB
Description: Size of the write-behind buffer for a single file (inode).
Option: performance.resync-failed-syncs-after-fsync
Default Value: (null)
Description: If sync of "cached-writes issued before fsync" (to
backend) fails, this option configures whether to retry syncing them after fsync
or forget them. If set to on, cached-writes are retried till a "flush"
fop (or a successful sync) on sync failures. fsync itself is failed irrespective
of the value of this option.
Option: performance.nfs.write-behind-window-size
Default Value: 1MB
Description: Size of the write-behind buffer for a single file (inode).
Option: performance.strict-o-direct
Default Value: off
Description: This option when set to off, ignores the O_DIRECT flag.
Option: performance.nfs.strict-o-direct
Default Value: off
Description: This option when set to off, ignores the O_DIRECT flag.
Option: performance.strict-write-ordering
Default Value: off
Description: Do not let later writes overtake earlier writes even if they do not
overlap
Option: performance.nfs.strict-write-ordering
Default Value: off
Description: Do not let later writes overtake earlier writes even if they do not
overlap
Option: performance.write-behind-trickling-writes
Default Value: on
Description: (null)
Option: performance.aggregate-size
Default Value: 128KB
Description: Will aggregate writes until data of specified size is fully filled
for a single file provided there are no dependent fops on cached writes. This
option just sets the aggregate size. Note that aggregation won't happen if
performance.write-behind-trickling-writes is turned on. Hence turn off
performance.write-behind.trickling-writes so that writes are aggregated till a
max of "aggregate-size" bytes
Option: performance.nfs.write-behind-trickling-writes
Default Value: on
Description: (null)
Option: performance.lazy-open
Default Value: yes
Description: Perform open in the backend only when a necessary FOP arrives (e.g
writev on the FD, unlink of the file). When option is disabled, perform backend
open right after unwinding open().
Option: performance.read-after-open
Default Value: yes
Description: read is sent only after actual open happens and real fd is
obtained, instead of doing on anonymous fd (similar to write)
Option: performance.open-behind-pass-through
Default Value: false
Description: Enable/Disable open behind translator
Option: performance.read-ahead-page-count
Default Value: 4
Description: Number of pages that will be pre-fetched
Option: performance.read-ahead-pass-through
Default Value: false
Description: Enable/Disable read ahead translator
Option: performance.readdir-ahead-pass-through
Default Value: false
Description: Enable/Disable readdir ahead translator
Option: performance.md-cache-pass-through
Default Value: false
Description: Enable/Disable md cache translator
Option: performance.md-cache-timeout
Default Value: 1
Description: Time period after which cache has to be refreshed
Option: performance.cache-swift-metadata
Default Value: (null)
Description: Cache swift metadata (user.swift.metadata xattr)
Option: performance.cache-samba-metadata
Default Value: (null)
Description: Cache samba metadata (user.DOSATTRIB, security.NTACL xattr)
Option: performance.cache-capability-xattrs
Default Value: (null)
Description: Cache xattrs required for capability based security
Option: performance.cache-ima-xattrs
Default Value: (null)
Description: Cache xattrs required for IMA (Integrity Measurement Architecture)
Option: performance.md-cache-statfs
Default Value: off
Description: Cache statfs information of filesystem on the client
Option: performance.xattr-cache-list
Default Value: (null)
Description: A comma separated list of xattrs that shall be cached by md-cache.
The only wildcard allowed is '*'
Option: performance.nl-cache-pass-through
Default Value: false
Description: Enable/Disable nl cache translator
Option: features.encryption
Default Value: off
Description: enable/disable client-side encryption for the volume.
Option: encryption.master-key
Default Value: (null)
Description: Pathname of regular file which contains master volume key
Option: encryption.data-key-size
Default Value: 256
Description: Data key size (bits)
Option: encryption.block-size
Default Value: 4096
Description: Atom size (bits)
Option: network.frame-timeout
Default Value: 1800
Description: Time frame after which the (file) operation would be declared as
dead, if the server does not respond for a particular (file) operation.
Option: network.ping-timeout
Default Value: 42
Description: Time duration for which the client waits to check if the server is
responsive.
Option: network.tcp-window-size
Default Value: (null)
Description: Specifies the window size for tcp socket.
Option: client.ssl
Default Value: (null)
Description: enable/disable client.ssl flag in the volume.
Option: network.remote-dio
Default Value: disable
Description: If enabled, in open/creat/readv/writev fops, O_DIRECT flag will be
filtered at the client protocol level so server will still continue to cache the
file. This works similar to NFS's behavior of O_DIRECT. Anon-fds can choose
to readv/writev using O_DIRECT
Option: client.event-threads
Default Value: 2
Description: Specifies the number of event threads to execute in parallel.
Larger values would help process responses faster, depending on available
processing power. Range 1-32 threads.
Option: network.inode-lru-limit
Default Value: 16384
Description: Specifies the limit on the number of inodes in the lru list of the
inode cache.
Option: auth.allow
Default Value: *
Description: Allow a comma separated list of addresses and/or hostnames to
connect to the server. Option auth.reject overrides this option. By default, all
connections are allowed.
Option: auth.reject
Default Value: (null)
Description: Reject a comma separated list of addresses and/or hostnames to
connect to the server. This option overrides the auth.allow option. By default,
all connections are allowed.
Option: server.allow-insecure
Default Value: on
Description: (null)
Option: server.root-squash
Default Value: off
Description: Map requests from uid/gid 0 to the anonymous uid/gid. Note that
this does not apply to any other uids or gids that might be equally sensitive,
such as user bin or group staff.
Option: server.anonuid
Default Value: 65534
Description: value of the uid used for the anonymous user/nfsnobody when
root-squash is enabled.
Option: server.anongid
Default Value: 65534
Description: value of the gid used for the anonymous user/nfsnobody when
root-squash is enabled.
Option: server.statedump-path
Default Value: /var/run/gluster
Description: Specifies directory in which gluster should save its statedumps.
Option: server.outstanding-rpc-limit
Default Value: 64
Description: Parameter to throttle the number of incoming RPC requests from a
client. 0 means no limit (can potentially run out of memory)
Option: server.ssl
Default Value: (null)
Description: enable/disable server.ssl flag in the volume.
Option: server.manage-gids
Default Value: off
Description: Resolve groups on the server-side.
Option: server.dynamic-auth
Default Value: on
Description: When 'on' perform dynamic authentication of volume options
in order to allow/terminate client transport connection immediately in response
to *.allow | *.reject volume set options.
Option: server.gid-timeout
Default Value: 300
Description: Timeout in seconds for the cached groups to expire.
Option: server.event-threads
Default Value: 1
Description: Specifies the number of event threads to execute in parallel.
Larger values would help process responses faster, depending on available
processing power.
Option: server.tcp-user-timeout
Default Value: 42
Description: (null)
Option: server.keepalive-time
Default Value: (null)
Description: (null)
Option: server.keepalive-interval
Default Value: (null)
Description: (null)
Option: server.keepalive-count
Default Value: (null)
Description: (null)
Option: transport.listen-backlog
Default Value: 1024
Description: This option uses the value of backlog argument that defines the
maximum length to which the queue of pending connections for socket fd may grow.
Option: ssl.own-cert
Default Value: (null)
Description: SSL certificate. Ignored if SSL is not enabled.
Option: ssl.private-key
Default Value: (null)
Description: SSL private key. Ignored if SSL is not enabled.
Option: ssl.ca-list
Default Value: (null)
Description: SSL CA list. Ignored if SSL is not enabled.
Option: ssl.crl-path
Default Value: (null)
Description: Path to directory containing CRL. Ignored if SSL is not enabled.
Option: ssl.certificate-depth
Default Value: (null)
Description: Maximum certificate-chain depth. If zero, the peer's
certificate itself must be in the local certificate list. Otherwise, there may
be up to N signing certificates between the peer's and the local list.
Ignored if SSL is not enabled.
Option: ssl.cipher-list
Default Value: (null)
Description: Allowed SSL ciphers. Ignored if SSL is not enabled.
Option: ssl.dh-param
Default Value: (null)
Description: DH parameters file. Ignored if SSL is not enabled.
Option: ssl.ec-curve
Default Value: (null)
Description: ECDH curve name. Ignored if SSL is not enabled.
Option: performance.write-behind
Default Value: on
Description: enable/disable write-behind translator in the volume.
Option: performance.read-ahead
Default Value: on
Description: enable/disable read-ahead translator in the volume.
Option: performance.readdir-ahead
Default Value: on
Description: enable/disable readdir-ahead translator in the volume.
Option: performance.io-cache
Default Value: on
Description: enable/disable io-cache translator in the volume.
Option: performance.quick-read
Default Value: on
Description: enable/disable quick-read translator in the volume.
Option: performance.open-behind
Default Value: on
Description: enable/disable open-behind translator in the volume.
Option: performance.nl-cache
Default Value: off
Description: enable/disable negative entry caching translator in the volume.
Enabling this option improves performance of 'create file/directory'
workload
Option: performance.stat-prefetch
Default Value: on
Description: enable/disable meta-data caching translator in the volume.
Option: performance.client-io-threads
Default Value: on
Description: enable/disable io-threads translator in the client graph of volume.
Option: performance.nfs.write-behind
Default Value: on
Description: enable/disable write-behind translator in the volume
Option: performance.force-readdirp
Default Value: true
Description: Convert all readdir requests to readdirplus to collect stat info on
each entry.
Option: performance.cache-invalidation
Default Value: false
Description: When "on", invalidates/updates the metadata cache, on
receiving the cache-invalidation notifications
Option: features.uss
Default Value: off
Description: enable/disable User Serviceable Snapshots on the volume.
Option: features.snapshot-directory
Default Value: .snaps
Description: Entry point directory for entering snapshot world. Value can have
only [0-9a-z-_] and starts with dot (.) and cannot exceed 255 character
Option: features.show-snapshot-directory
Default Value: off
Description: show entry point in readdir output of snapdir-entry-path which is
set by samba
Option: features.tag-namespaces
Default Value: off
Description: This option enables this translator's functionality that tags
every fop with a namespace hash for later throttling, stats collection, logging,
etc.
Option: network.compression
Default Value: off
Description: enable/disable network compression translator
Option: network.compression.window-size
Default Value: -15
Description: Size of the zlib history buffer.
Option: network.compression.mem-level
Default Value: 8
Description: Memory allocated for internal compression state. 1 uses minimum
memory but is slow and reduces compression ratio; memLevel=9 uses maximum memory
for optimal speed. The default value is 8.
Option: network.compression.min-size
Default Value: 0
Description: Data is compressed only when its size exceeds this.
Option: network.compression.compression-level
Default Value: -1
Description: Compression levels
0 : no compression, 1 : best speed,
9 : best compression, -1 : default compression
Option: features.quota-deem-statfs
Default Value: on
Description: If set to on, it takes quota limits into consideration while
estimating fs size. (df command) (Default is on).
Option: nfs.transport-type
Default Value: (null)
Description: Specifies the nfs transport type. Valid transport types are
'tcp' and 'rdma'.
Option: nfs.rdirplus
Default Value: (null)
Description: When this option is set to off NFS falls back to standard readdir
instead of readdirp
Option: features.read-only
Default Value: off
Description: When "on", makes a volume read-only. It is turned
"off" by default.
Option: features.worm
Default Value: off
Description: When "on", makes a volume get write once read many
feature. It is turned "off" by default.
Option: features.worm-file-level
Default Value: off
Description: When "on", activates the file level worm. It is turned
"off" by default.
Option: features.worm-files-deletable
Default Value: on
Description: When "off", doesn't allow the Worm filesto be
deleted. It is turned "on" by default.
Option: features.default-retention-period
Default Value: 120
Description: The default retention period for the files.
Option: features.retention-mode
Default Value: relax
Description: The mode of retention (relax/enterprise). It is relax by default.
Option: features.auto-commit-period
Default Value: 180
Description: Auto commit period for the files.
Option: storage.linux-aio
Default Value: off
Description: Support for native Linux AIO
Option: storage.batch-fsync-mode
Default Value: reverse-fsync
Description: Possible values:
- syncfs: Perform one syncfs() on behalf oa batchof fsyncs.
- syncfs-single-fsync: Perform one syncfs() on behalf of a batch of fsyncs and
one fsync() per batch.
- syncfs-reverse-fsync: Perform one syncfs() on behalf of a batch of fsyncs and
fsync() each file in the batch in reverse order.
in reverse order.
- reverse-fsync: Perform fsync() of each file in the batch in reverse order.
Option: storage.batch-fsync-delay-usec
Default Value: 0
Description: Num of usecs to wait for aggregating fsync requests
Option: storage.owner-uid
Default Value: -1
Description: Support for setting uid of brick's owner
Option: storage.owner-gid
Default Value: -1
Description: Support for setting gid of brick's owner
Option: storage.node-uuid-pathinfo
Default Value: off
Description: return glusterd's node-uuid in pathinfo xattr string instead of
hostname
Option: storage.health-check-interval
Default Value: 30
Description: Interval in seconds for a filesystem health check, set to 0 to
disable
Option: storage.build-pgfid
Default Value: off
Description: Enable placeholders for gfid to path conversion
Option: storage.gfid2path-separator
Default Value: :
Description: Path separator for glusterfs.gfidtopath virt xattr
Option: storage.reserve
Default Value: 1
Description: Percentage of disk space to be reserved. Set to 0 to disable
Option: storage.force-create-mode
Default Value: 0000
Description: Mode bit permission that will always be set on a file.
Option: storage.force-directory-mode
Default Value: 0000
Description: Mode bit permission that will be always set on directory
Option: storage.create-mask
Default Value: 0777
Description: Any bit not set here will be removed from themodes set on a file
when it is created
Option: storage.create-directory-mask
Default Value: 0777
Description: Any bit not set here will be removed from themodes set on a
directory when it is created
Option: storage.max-hardlinks
Default Value: 100
Description: max number of hardlinks allowed on any one inode.
0 is unlimited, 1 prevents any hardlinking at all.
Option: storage.ctime
Default Value: off
Description: When this option is enabled, time attributes (ctime,mtime,atime)
are stored in xattr to keep it consistent across replica and distribute set. The
time attributes stored at the backend are not considered
Option: storage.bd-aio
Default Value: off
Description: Support for native Linux AIO
Option: config.gfproxyd
Default Value: off
Description: If this option is enabled, the proxy client daemon called gfproxyd
will be started on all the trusted storage pool nodes
Option: cluster.server-quorum-type
Default Value: (null)
Description: This feature is on the server-side i.e. in glusterd. Whenever the
glusterd on a machine observes that the quorum is not met, it brings down the
bricks to prevent data split-brains. When the network connections are brought
back up and the quorum is restored the bricks in the volume are brought back up.
Option: cluster.server-quorum-ratio
Default Value: (null)
Description: Sets the quorum percentage for the trusted storage pool.
Option: changelog.changelog-barrier-timeout
Default Value: 120
Description: After 'timeout' seconds since the time 'barrier'
option was set to "on", unlink/rmdir/rename operations are no longer
blocked and previously blocked fops are allowed to go through
Option: features.barrier-timeout
Default Value: 120
Description: After 'timeout' seconds since the time 'barrier'
option was set to "on", acknowledgements to file operations are no
longer blocked and previously blocked acknowledgements are sent to the
application
Option: features.trash
Default Value: off
Description: Enable/disable trash translator
Option: features.trash-dir
Default Value: .trashcan
Description: Directory for trash files
Option: features.trash-eliminate-path
Default Value: (null)
Description: Eliminate paths to be excluded from trashing
Option: features.trash-max-filesize
Default Value: 5MB
Description: Maximum size of file that can be moved to trash
Option: features.trash-internal-op
Default Value: off
Description: Enable/disable trash translator for internal operations
Option: cluster.enable-shared-storage
Default Value: disable
Description: Create and mount the shared storage volume(gluster_shared_storage)
at /var/run/gluster/shared_storage on enabling this option. Unmount and delete
the shared storage volume on disabling this option.
Option: cluster.write-freq-threshold
Default Value: 0
Description: Defines the number of writes, in a promotion/demotion cycle, that
would mark a file HOT for promotion. Any file that has write hits less than this
value will be considered as COLD and will be demoted.
Option: cluster.read-freq-threshold
Default Value: 0
Description: Defines the number of reads, in a promotion/demotion cycle, that
would mark a file HOT for promotion. Any file that has read hits less than this
value will be considered as COLD and will be demoted.
Option: cluster.tier-pause
Default Value: off
Description: (null)
Option: cluster.tier-promote-frequency
Default Value: 120
Description: (null)
Option: cluster.tier-demote-frequency
Default Value: 3600
Description: (null)
Option: cluster.watermark-hi
Default Value: 90
Description: Upper % watermark for promotion. If hot tier fills above this
percentage, no promotion will happen and demotion will happen with high
probability.
Option: cluster.watermark-low
Default Value: 75
Description: Lower % watermark. If hot tier is less full than this, promotion
will happen and demotion will not happen. If greater than this,
promotion/demotion will happen at a probability relative to how full the hot
tier is.
Option: cluster.tier-mode
Default Value: cache
Description: Either 'test' or 'cache'. Test mode periodically
demotes or promotes files automatically based on access. Cache mode does so
based on whether the cache is full or not, as specified with watermarks.
Option: cluster.tier-max-promote-file-size
Default Value: 0
Description: The maximum file size in bytes that is promoted. If 0, there is no
maximum size (default).
Option: cluster.tier-max-mb
Default Value: 4000
Description: The maximum number of MB that may be migrated in any direction in a
given cycle by a single node.
Option: cluster.tier-max-files
Default Value: 10000
Description: The maximum number of files that may be migrated in any direction
in a given cycle by a single node.
Option: cluster.tier-compact
Default Value: on
Description: Activate or deactivate the compaction of the DB for the
volume's metadata.
Option: cluster.tier-hot-compact-frequency
Default Value: 604800
Description: Frequency to compact DBs on hot tier in system
Option: cluster.tier-cold-compact-frequency
Default Value: 604800
Description: Frequency to compact DBs on cold tier in system
Option: features.ctr-enabled
Default Value: off
Description: Enable CTR xlator
Option: features.record-counters
Default Value: off
Description: Its a Change Time Recorder Xlator option to enable recording write
and read heat counters. The default is disabled. If enabled,
"cluster.write-freq-threshold" and
"cluster.read-freq-threshold" defined the number of writes (or reads)
to a given file are needed before triggering migration.
Option: features.ctr-sql-db-cachesize
Default Value: 12500
Description: Defines the cache size of the sqlite database of changetimerecorder
xlator.The input to this option is in pages.Each page is 4096 bytes. Default
value is 12500 pages.The max value is 262144 pages i.e 1 GB and the min value is
1000 pages i.e ~ 4 MB.
Option: features.ctr-sql-db-wal-autocheckpoint
Default Value: 25000
Description: Defines the autocheckpoint of the sqlite database of
changetimerecorder. The input to this option is in pages. Each page is 4096
bytes. Default value is 25000 pages.The max value is 262144 pages i.e 1 GB and
the min value is 1000 pages i.e ~4 MB.
Option: locks.trace
Default Value: off
Description: Trace the different lock requests to logs.
Option: locks.mandatory-locking
Default Value: off
Description: Specifies the mandatory-locking mode. Valid options are
'file' to use linux style mandatory locks, 'forced' to use
volume strictly under mandatory lock semantics only and 'optimal' to
treat advisory and mandatory locks separately on their own.
Option: cluster.quorum-reads
Default Value: no
Description: This option has been removed. Reads are not allowed if quorum is
not met.
Option: features.timeout
Default Value: (null)
Description: Specifies the number of seconds the quiesce translator will wait
for a CHILD_UP event before force-unwinding the frames it has currently stored
for retry.
Option: features.failover-hosts
Default Value: (null)
Description: It is a comma separated list of hostname/IP addresses. It Specifies
the list of hosts where the gfproxy daemons are running, to which the the thin
clients can failover to.
Option: features.shard
Default Value: off
Description: enable/disable sharding translator on the volume.
Option: features.shard-block-size
Default Value: 64MB
Description: The size unit used to break a file into multiple chunks
Option: features.shard-deletion-rate
Default Value: 100
Description: The number of shards to send deletes on at a time
Option: features.cache-invalidation
Default Value: off
Description: When "on", sends cache-invalidation notifications.
Option: features.cache-invalidation-timeout
Default Value: 60
Description: After 'timeout' seconds since the time client accessed any
file, cache-invalidation notifications are no longer sent to that client.
Option: features.leases
Default Value: off
Description: When "on", enables leases support
Option: features.lease-lock-recall-timeout
Default Value: 60
Description: After 'timeout' seconds since the recall_lease request has
been sent to the client, the lease lock will be forcefully purged by the server.
Option: disperse.background-heals
Default Value: 8
Description: This option can be used to control number of parallel heals
Option: disperse.heal-wait-qlength
Default Value: 128
Description: This option can be used to control number of heals that can wait
Option: dht.force-readdirp
Default Value: on
Description: This option if set to ON, forces the use of readdirp, and hence
also displays the stats of the files.
Option: disperse.read-policy
Default Value: gfid-hash
Description: inode-read fops happen only on 'k' number of bricks in
n=k+m disperse subvolume. 'round-robin' selects the read subvolume using
round-robin algo. 'gfid-hash' selects read subvolume based on hash of
the gfid of that file/directory.
Option: cluster.shd-max-threads
Default Value: 1
Description: Maximum number of parallel heals SHD can do per local brick. This
can substantially lower heal times, but can also crush your bricks if you
don't have the storage hardware to support this.
Option: cluster.shd-wait-qlength
Default Value: 1024
Description: This option can be used to control number of heals that can wait in
SHD per subvolume
Option: cluster.locking-scheme
Default Value: full
Description: If this option is set to granular, self-heal will stop being
compatible with afr-v1, which helps afr be more granular while self-healing
Option: cluster.granular-entry-heal
Default Value: no
Description: If this option is enabled, self-heal will resort to granular way of
recording changelogs and doing entry self-heal.
Option: features.locks-revocation-secs
Default Value: 0
Description: Maximum time a lock can be taken out, beforebeing revoked.
Option: features.locks-revocation-clear-all
Default Value: false
Description: If set to true, will revoke BOTH granted and blocked (pending) lock
requests if a revocation threshold is hit.
Option: features.locks-revocation-max-blocked
Default Value: 0
Description: A number of blocked lock requests after which a lock will be
revoked to allow the others to proceed. Can be used in conjunction w/
revocation-clear-all.
Option: features.locks-notify-contention
Default Value: no
Description: When this option is enabled and a lock request conflicts with a
currently granted lock, an upcall notification will be sent to the current owner
of the lock to request it to be released as soon as possible.
Option: features.locks-notify-contention-delay
Default Value: 5
Description: This value determines the minimum amount of time (in seconds)
between upcall contention notifications on the same inode. If multiple lock
requests are received during this period, only one upcall will be sent.
Option: disperse.shd-max-threads
Default Value: 1
Description: Maximum number of parallel heals SHD can do per local brick. This
can substantially lower heal times, but can also crush your bricks if you
don't have the storage hardware to support this.
Option: disperse.shd-wait-qlength
Default Value: 1024
Description: This option can be used to control number of heals that can wait in
SHD per subvolume
Option: disperse.cpu-extensions
Default Value: auto
Description: force the cpu extensions to be used to accelerate the galois field
computations.
Option: disperse.self-heal-window-size
Default Value: 1
Description: Maximum number blocks(128KB) per file for which self-heal process
would be applied simultaneously.
Option: cluster.use-compound-fops
Default Value: no
Description: This option exists only for backward compatibility and configuring
it doesn't have any effect
Option: performance.parallel-readdir
Default Value: off
Description: If this option is enabled, the readdir operation is performed in
parallel on all the bricks, thus improving the performance of readdir. Note that
the performance improvement is higher in large clusters
Option: performance.rda-request-size
Default Value: 131072
Description: size of buffer in readdirp calls initiated by readdir-ahead
Option: performance.rda-cache-limit
Default Value: 10MB
Description: maximum size of cache consumed by readdir-ahead xlator. This value
is global and total memory consumption by readdir-ahead is capped by this value,
irrespective of the number/size of directories cached
Option: performance.nl-cache-positive-entry
Default Value: (null)
Description: enable/disable storing of entries that were lookedup and found to
be present in the volume, thus lookup on non existent file is served from the
cache
Option: performance.nl-cache-limit
Default Value: 131072
Description: the value over which caching will be disabled fora while and the
cache is cleared based on LRU
Option: performance.nl-cache-timeout
Default Value: 60
Description: Time period after which cache has to be refreshed
Option: cluster.brick-multiplex
Default Value: off
Description: This global option can be used to enable/disable brick
multiplexing. Brick multiplexing ensures that compatible brick instances can
share one single brick process.
Option: cluster.max-bricks-per-process
Default Value: 0
Description: This option can be used to limit the number of brick instances per
brick process when brick-multiplexing is enabled. If not explicitly set, this
tunable is set to 0 which denotes that brick-multiplexing can happen without any
limit on the number of bricks per process. Also this option can't be set
when the brick-multiplexing feature is disabled.
Option: cluster.halo-enabled
Default Value: False
Description: Enable Halo (geo) replication mode.
Option: cluster.halo-shd-max-latency
Default Value: 99999
Description: Maximum latency for shd halo replication in msec.
Option: cluster.halo-nfsd-max-latency
Default Value: 5
Description: Maximum latency for nfsd halo replication in msec.
Option: cluster.halo-max-latency
Default Value: 5
Description: Maximum latency for halo replication in msec.
Option: cluster.halo-max-replicas
Default Value: 99999
Description: The maximum number of halo replicas; replicas beyond this value
will be written asynchronouslyvia the SHD.
Option: cluster.halo-min-replicas
Default Value: 2
Description: The minimmum number of halo replicas, before adding out of region
replicas.
Option: features.utime
Default Value: off
Description: enable/disable utime translator on the volume.
Option: ctime.noatime
Default Value: on
Description: enable/disable noatime option with ctime enabled.
Option: feature.cloudsync-storetype
Default Value: (null)
Description: Defines which remote store is enabled