Displaying 20 results from an estimated 100 matches similar to: "GlusterFS don't expose iSCSI target for Window server"
2018 Apr 04
0
Invisible files and directories
Hi,
I'm currently facing the same behaviour.?
Today, one of my users tried to delete a folder. It failed, saying the directory wasn't empty. ls -lah showed an empty folder but on the bricks I
found some files. Renaming the directory caused it to reappear.
We're running gluster 3.12.7-1 on Debian 9 from the repositories provided by gluster.org, upgraded from 3.8 a while ago. The
2018 Mar 06
0
Multiple Volumes over diffrent Interfaces
Hi,
I'm trying to create two gluster volumes over two nodes with two
seperate networks:
The names are in the hosts file of each node:
root at gluster01 :~# cat /etc/hosts
127.0.0.1 localhost
127.0.1.1 gluster01.peiker-cee.de gluster01
10.0.2.54 gluster02g1.peiker-cee.de gluster02g1
10.0.7.54 gluster02g2.peiker-cee.de gluster02g2
10.0.2.53
2018 Jan 11
0
Creating cluster replica on 2 nodes 2 bricks each.
Hi Jose,
Gluster is working as expected. The Distribute-replicated type just means
that there are now 2 replica sets and files will be distributed across
them.
A volume of type Replicate (1xn where n is the number of bricks in the
replica set) indicates there is no distribution (all files on the
volume will be present on all the bricks in the volume).
A volume of type Distributed-Replicate
2013 Mar 20
1
About adding bricks ...
Hi @all,
I've created a Distributed-Replicated Volume consisting of 4 bricks on
2 servers.
# gluster volume create glusterfs replica 2 transport tcp \
gluster0{0..1}:/srv/gluster/exp0 gluster0{0..1}:/srv/gluster/exp1
Now I have the following very nice replication schema:
+-------------+ +-------------+
| gluster00 | | gluster01 |
+-------------+ +-------------+
| exp0 | exp1 |
2018 Jan 10
2
Creating cluster replica on 2 nodes 2 bricks each.
Hi Nithya
This is what i have so far, I have peer both cluster nodes together as replica, from node 1A and 1B , now when i tried to add it , i get the error that it is already part of a volume. when i run the cluster volume info , i see that has switch to distributed-replica.
Thanks
Jose
[root at gluster01 ~]# gluster volume status
Status of volume: scratch
Gluster process
2018 Jan 15
1
Creating cluster replica on 2 nodes 2 bricks each.
On Fri, 12 Jan 2018 at 21:16, Nithya Balachandran <nbalacha at redhat.com>
wrote:
> ---------- Forwarded message ----------
> From: Jose Sanchez <josesanc at carc.unm.edu>
> Date: 11 January 2018 at 22:05
> Subject: Re: [Gluster-users] Creating cluster replica on 2 nodes 2 bricks
> each.
> To: Nithya Balachandran <nbalacha at redhat.com>
> Cc: gluster-users
2018 Jan 12
0
Creating cluster replica on 2 nodes 2 bricks each.
---------- Forwarded message ----------
From: Jose Sanchez <josesanc at carc.unm.edu>
Date: 11 January 2018 at 22:05
Subject: Re: [Gluster-users] Creating cluster replica on 2 nodes 2 bricks
each.
To: Nithya Balachandran <nbalacha at redhat.com>
Cc: gluster-users <gluster-users at gluster.org>
Hi Nithya
Thanks for helping me with this, I understand now , but I have few
2018 Jan 11
3
Creating cluster replica on 2 nodes 2 bricks each.
Hi Nithya
Thanks for helping me with this, I understand now , but I have few questions.
When i had it setup in replica (just 2 nodes with 2 bricks) and tried to added , it failed.
> [root at gluster01 ~]# gluster volume add-brick scratch replica 2 gluster01ib:/gdata/brick2/scratch gluster02ib:/gdata/brick2/scratch
> volume add-brick: failed: /gdata/brick2/scratch is already part of a
2011 Oct 25
1
problems with gluster 3.2.4
Hi, we have 4 test machines (gluster01 to gluster04).
I've created a replicated volume with the 4 machines.
Then on the client machine i've executed:
mount -t glusterfs gluster01:/volume01 /mnt/gluster
And everything works ok.
The main problem occurs in every client machine that I do:
umount /mnt/gluster
and the
mount -t glusterfs gluster01:/volume01 /mnt/gluster
The client
2018 Apr 25
0
Turn off replication
Hello Karthik
Im having trouble adding the two bricks back online. Any help is appreciated
thanks
when i try to add-brick command this is what i get
[root at gluster01 ~]# gluster volume add-brick scratch gluster02ib:/gdata/brick2/scratch/
volume add-brick: failed: Pre Validation failed on gluster02ib. Brick: gluster02ib:/gdata/brick2/scratch not available. Brick may be containing or be
2018 Apr 07
0
Turn off replication
Hi Jose,
Thanks for providing the volume info. You have 2 subvolumes. Data is
replicated within the bricks of that subvolumes.
First one consisting of Node A's brick1 & Node B's brick1 and the second
one consisting of Node A's brick2 and Node B's brick2.
You don't have the same data on all the 4 bricks. Data are distributed
between these two subvolumes.
To remove the
2018 Apr 27
0
Turn off replication
Hi Jose,
Why are all the bricks visible in volume info if the pre-validation
for add-brick failed? I suspect that the remove brick wasn't done
properly.
You can provide the cmd_history.log to verify this. Better to get the
other log messages.
Also I need to know what are the bricks that were actually removed,
the command used and its output.
On Thu, Apr 26, 2018 at 3:47 AM, Jose Sanchez
2018 May 02
0
Turn off replication
Hi,
Removing data to speed up from rebalance is not something that is recommended.
Rebalance can be stopped but if started again it will start from the beginning
(will have to check and skip the files already moved).
Rebalance will take a while, better to let it run. It doesn't have any
down side.
Unless you touch the backend the data on gluster volume will be
available for usage
in spite of
2018 Apr 25
2
Turn off replication
Looking at the logs , it seems that it is trying to add using the same port was assigned for gluster01ib:
Any Ideas??
Jose
[2018-04-25 22:08:55.169302] I [MSGID: 106482] [glusterd-brick-ops.c:447:__glusterd_handle_add_brick] 0-management: Received add brick req
[2018-04-25 22:08:55.186037] I [run.c:191:runner_log] (-->/usr/lib64/glusterfs/3.8.15/xlator/mgmt/glusterd.so(+0x33045)
2018 Apr 30
2
Turn off replication
Hi All
We were able to get all 4 bricks are distributed , we can see the right amount of space. but we have been rebalancing since 4 days ago for 16Tb. and still only 8tb. is there a way to speed up. there is also data we can remove from it to speed it up, but what is the best procedures removing data , is it from the Gluster main export point or going on each brick and remove it . We would like
2013 May 10
1
samba4 with glusterfs
Hi Sambalist,
maybe someone with some glusterfs experience can help me.
I have a running samba4 installation here.
Now i tried to get glusterfs configured so I can use glusterfs as an
additional storagesystem.
The setup of glusterfs worked pretty easy. I tried multiple
configurations with up to 4 virtual machines (Virtual Box).
The bricks are xfs-filesystems.
Mounting glusterfs (mount -t
2013 Sep 24
1
files on bricks but not on client
Hi all -
We're running a 3-node distributed volume, using gluster 3.3.1.
We're seeing a rare but repeated issue, where files are written to the
volume, appears to be written OK, but are not accessible via NFS or the
gluster client. These files appear when we inspect the bricks directly.
For example, the path in question is
2017 Jun 02
0
libgfapi with encryption?
Hi,
I created an encrypted volume which appears to be working fine with
FUSE but the volume is supposed to store VM images (master key in
place). I noticed some references in libgfapi source code to
encryption so I decided to try it out. While attempting to create an
image:
# qemu-img create -f qcow2 gluster://gluster01/virt0/testing.img 30G
Formatting
2018 Apr 12
2
Turn off replication
On Wed, Apr 11, 2018 at 7:38 PM, Jose Sanchez <josesanc at carc.unm.edu> wrote:
> Hi Karthik
>
> Looking at the information you have provided me, I would like to make sure
> that I?m running the right commands.
>
> 1. gluster volume heal scratch info
>
If the count is non zero, trigger the heal and wait for heal info count to
become zero.
> 2. gluster volume
2018 Apr 04
2
Invisible files and directories
Right now the volume is running with
readdir-optimize off
parallel-readdir off
On Wed, Apr 4, 2018 at 1:29 AM, Nithya Balachandran <nbalacha at redhat.com>
wrote:
> Hi Serg,
>
> Do you mean that turning off readdir-optimize did not work? Or did you
> mean turning off parallel-readdir did not work?
>
>
>
> On 4 April 2018 at 10:48, Serg Gulko <s.gulko at