So glusterp3 is in a reject state,
[root at glusterp1 /]# gluster peer status
Number of Peers: 2
Hostname: glusterp2.graywitch.co.nz
Uuid: 93eebe2c-9564-4bb0-975f-2db49f12058b
State: Peer in Cluster (Connected)
Other names:
glusterp2
Hostname: glusterp3.graywitch.co.nz
Uuid: 5d59b704-e42f-46c6-8c14-cf052c489292
State: Peer Rejected (Connected)
Other names:
glusterp3
[root at glusterp1 /]#
=======
[root at glusterp2 /]# gluster peer status
Number of Peers: 2
Hostname: glusterp1.graywitch.co.nz
Uuid: 4ece8509-033e-48d1-809f-2079345caea2
State: Peer in Cluster (Connected)
Other names:
glusterp1
Hostname: glusterp3.graywitch.co.nz
Uuid: 5d59b704-e42f-46c6-8c14-cf052c489292
State: Peer Rejected (Connected)
Other names:
glusterp3
[root at glusterp2 /]#
=======
[root at glusterp3 /]# gluster peer status
Number of Peers: 2
Hostname: glusterp1.graywitch.co.nz
Uuid: 4ece8509-033e-48d1-809f-2079345caea2
State: Peer Rejected (Connected)
Other names:
glusterp1
Hostname: glusterp2.graywitch.co.nz
Uuid: 93eebe2c-9564-4bb0-975f-2db49f12058b
State: Peer Rejected (Connected)
Other names:
glusterp2
=========on glusterp3 gluster is dead and will not start,
[root at glusterp3 /]# systemctl status gluster
? gluster.service
Loaded: not-found (Reason: No such file or directory)
Active: inactive (dead)
[root at glusterp3 /]# systemctl restart gluster
Failed to restart gluster.service: Unit gluster.service failed to load: No
such file or directory.
[root at glusterp3 /]# systemctl enable gluster
Failed to execute operation: Access denied
[root at glusterp3 /]# systemctl enable gluster.service
Failed to execute operation: Access denied
[root at glusterp3 /]# systemctl start gluster.service
Failed to start gluster.service: Unit gluster.service failed to load: No
such file or directory.
=========
[root at glusterp3 /]# rpm -qa |grep gluster
glusterfs-client-xlators-3.8.4-1.el7.x86_64
glusterfs-server-3.8.4-1.el7.x86_64
nfs-ganesha-gluster-2.3.3-1.el7.x86_64
glusterfs-cli-3.8.4-1.el7.x86_64
glusterfs-api-3.8.4-1.el7.x86_64
glusterfs-fuse-3.8.4-1.el7.x86_64
glusterfs-ganesha-3.8.4-1.el7.x86_64
glusterfs-3.8.4-1.el7.x86_64
centos-release-gluster38-1.0-1.el7.centos.noarch
glusterfs-libs-3.8.4-1.el7.x86_64
[root at glusterp3 /]#
?
On 14 October 2016 at 12:31, Thing <thing.thing at gmail.com> wrote:
> Hmm seem I have something rather inconsistent,
>
> [root at glusterp1 /]# gluster volume create gv1 replica 3
> glusterp1:/brick1/gv1 glusterp2:/brick1/gv1 glusterp3:/brick1/gv1
> volume create: gv1: failed: Host glusterp3 is not in 'Peer in
Cluster'
> state
> [root at glusterp1 /]# gluster peer probe glusterp3
> peer probe: success. Host glusterp3 port 24007 already in peer list
> [root at glusterp1 /]# gluster peer probe glusterp2
> peer probe: success. Host glusterp2 port 24007 already in peer list
> [root at glusterp1 /]# gluster volume create gv1 replica 3
> glusterp1:/brick1/gv1 glusterp2:/brick1/gv1 glusterp3:/brick1/gv1
> volume create: gv1: failed: /brick1/gv1 is already part of a volume
> [root at glusterp1 /]# gluster volume show
> unrecognized word: show (position 1)
> [root at glusterp1 /]# gluster volume
> add-brick delete info quota
> reset status
> barrier geo-replication list rebalance
> set stop
> clear-locks heal log remove-brick
> start sync
> create help profile replace-brick
> statedump top
> [root at glusterp1 /]# gluster volume list
> volume1
> [root at glusterp1 /]# gluster volume start gv0
> volume start: gv0: failed: Volume gv0 does not exist
> [root at glusterp1 /]# gluster volume start gv1
> volume start: gv1: failed: Volume gv1 does not exist
> [root at glusterp1 /]# gluster volume status
> Status of volume: volume1
> Gluster process TCP Port RDMA Port Online
> Pid
> ------------------------------------------------------------
> ------------------
> Brick glusterp1.graywitch.co.nz:/data1 49152 0 Y
> 2958
> Brick glusterp2.graywitch.co.nz:/data1 49152 0 Y
> 2668
> NFS Server on localhost N/A N/A N
> N/A
> Self-heal Daemon on localhost N/A N/A Y
> 1038
> NFS Server on glusterp2.graywitch.co.nz N/A N/A N
> N/A
> Self-heal Daemon on glusterp2.graywitch.co.
> nz N/A N/A Y
> 676
>
> Task Status of Volume volume1
> ------------------------------------------------------------
> ------------------
> There are no active volume tasks
>
> [root at glusterp1 /]#
>
> On 14 October 2016 at 12:20, Thing <thing.thing at gmail.com> wrote:
>
>> I deleted a gluster volume gv0 as I wanted to make it thin provisioned.
>>
>> I have rebuilt "gv0" but I am getting a failure,
>>
>> =========>> [root at glusterp1 /]# df -h
>> Filesystem Size Used Avail Use% Mounted on
>> /dev/mapper/centos-root 20G 3.9G 17G 20% /
>> devtmpfs 1.8G 0 1.8G 0% /dev
>> tmpfs 1.8G 12K 1.8G 1% /dev/shm
>> tmpfs 1.8G 8.9M 1.8G 1% /run
>> tmpfs 1.8G 0 1.8G 0% /sys/fs/cgroup
>> /dev/mapper/centos-tmp 3.9G 33M 3.9G 1% /tmp
>> /dev/mapper/centos-home 50G 41M 50G 1% /home
>> /dev/mapper/centos-data1 120G 33M 120G 1% /data1
>> /dev/sda1 997M 312M 685M 32% /boot
>> /dev/mapper/centos-var 20G 401M 20G 2% /var
>> tmpfs 368M 0 368M 0% /run/user/1000
>> /dev/mapper/vol_brick1-brick1 100G 33M 100G 1% /brick1
>> [root at glusterp1 /]# mkdir /brick1/gv0
>> [root at glusterp1 /]# gluster volume create gv0 replica 3
>> glusterp1:/brick1/gv0 glusterp2:/brick1/gv0 glusterp3:/brick1/gv0
>> volume create: gv0: failed: Host glusterp3 is not in 'Peer in
Cluster'
>> state
>> [root at glusterp1 /]# gluster peer probe glusterp3
>> peer probe: success. Host glusterp3 port 24007 already in peer list
>> [root at glusterp1 /]# gluster volume create gv0 replica 3
>> glusterp1:/brick1/gv0 glusterp2:/brick1/gv0 glusterp3:/brick1/gv0
>> volume create: gv0: failed: /brick1/gv0 is already part of a volume
>> [root at glusterp1 /]# gluster volume start gv0
>> volume start: gv0: failed: Volume gv0 does not exist
>> [root at glusterp1 /]# gluster volume create gv0 replica 3
>> glusterp1:/brick1/gv0 glusterp2:/brick1/gv0 glusterp3:/brick1/gv0
--force
>> unrecognized option --force
>> [root at glusterp1 /]# gluster volume create gv0 replica 3
>> glusterp1:/brick1/gv0 glusterp2:/brick1/gv0 glusterp3:/brick1/gv0
>> volume create: gv0: failed: /brick1/gv0 is already part of a volume
>> [root at glusterp1 /]#
>> =========>>
>> Obviously something isnt happy here but I have no idea what.......
>>
>> how to fix this please?
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://www.gluster.org/pipermail/gluster-users/attachments/20161014/ae8b9c0d/attachment.html>