I have a problem. I have run two servers and two clients. On both clients
operate in the background loop:
for ((j=0; j< $RANDOM; j++)) {
PLIK=$RANDOM.$RANDOM
dd if=/dev/urandom of=/mnt/glusterfs/$KAT/$PLIK bs=$RANDOM count=1
dd if=/mnt/glusterfs/$KAT/$PLIK of=/dev/null
rm -f /mnt/glusterfs/$KAT/$PLIK
}
If both servers are connected to everything is fine. If one stops working
after several minutes will go back to the server, a client I have:
dd: opening `/ mnt/glusterfs/24427/30087.20476 ': No such file or directory
dd: opening `/ mnt/glusterfs/24427/30087.20476 ': No such file or directory
dd: opening `/ mnt/glusterfs/24427/18649.25895 ': No such file or directory
after a few seconds, everything is working again.
I think that the client is trying to read the file from the new server. I
think this is not work?
client:
volume client1
type protocol/client
option transport-type tcp/client
option remote-host 10.0.1.130
option remote-port 6996
option remote-subvolume posix1
end-volume
volume client2
type protocol/client
option transport-type tcp/client
option remote-host 10.0.1.131
option remote-port 6996
option remote-subvolume posix2
end-volume
volume afr
type cluster/afr
subvolumes client1 client2
end-volume
volume rh
type performance/read-ahead
option page-size 100KB
option page-count 3
subvolumes afr
end-volume
volume wh
type performance/write-behind
option aggregate-size 1MB
option flush-behind on
subvolumes rh
end-volume
server:
volume posix1
type storage/posix
option directory /var/storage/glusterfs
option debug on
end-volume
volume posix-locks
type features/posix-locks
option mandatory on
subvolumes posix1
end-volume
volume io-thr
type performance/io-threads
option thread-count 2
option cache-size 64MB
subvolumes posix-locks
end-volume
volume server
type protocol/server
option transport-type tcp/server
option listen-port 6996
subvolumes io-thr
option auth.ip.posix1.allow 10.*.*.*
end-volume
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://supercolony.gluster.org/pipermail/gluster-users/attachments/20081210/271c7bdd/attachment.html>
On Wed, Dec 10, 2008 at 9:21 PM, <a_pirania at poczta.onet.pl> wrote:> I have a problem. I have run two servers and two clients. On both clients > operate in the background loop: > > for ((j=0; j< $RANDOM; j++)) { > PLIK=$RANDOM.$RANDOM > dd if=/dev/urandom of=/mnt/glusterfs/$KAT/$PLIK bs=$RANDOM count=1 > dd if=/mnt/glusterfs/$KAT/$PLIK of=/dev/null > rm -f /mnt/glusterfs/$KAT/$PLIK > } > > > > If both servers are connected to everything is fine. If one stops working > after several minutes will go back to the server, a client I have: > > dd: opening `/ mnt/glusterfs/24427/30087.20476 ': No such file or directory > dd: opening `/ mnt/glusterfs/24427/30087.20476 ': No such file or directory > dd: opening `/ mnt/glusterfs/24427/18649.25895 ': No such file or directory > > > after a few seconds, everything is working again. > > I think that the client is trying to read the file from the new server. I > think this is not work? > > > client: > > volume client1 > type protocol/client > option transport-type tcp/client > option remote-host 10.0.1.130 > option remote-port 6996 > option remote-subvolume posix1 > end-volume > > volume client2 > type protocol/client > option transport-type tcp/client > option remote-host 10.0.1.131 > option remote-port 6996 > option remote-subvolume posix2 > end-volume >On 10.0.1.131 are you sure that it is posix2 and not posix1? Do log files give any clue? Can you paste it? Krishna
Hi,
I've been trying to reproduce your problem. Some observations,
* I've run into 'No such file or directory' errors, but it was due
to dd not
creating the file due to block size being 0 (bs=0).
* As per dd error msgs, there is a space between '/' and 'mnt'
(/ mnt).
dd: opening `/ mnt/glusterfs/24427/30087.20476 ': No such file or directory
please make sure that the file is getting created in first place (file may
not be created due to invalid parameters to dd, like bs=0 in above case).
* since I am not able to reproduce it on my setup, is it possible for you to
try out the test with
- afr-self-heal turned on
- afr-self-heal turned off
- following options control afr self heal
option data-self-heal off
option metadata-self-heal off
option entry-self-heal off
regards,
On Wed, Dec 10, 2008 at 7:51 PM, <a_pirania at poczta.onet.pl> wrote:
> I have a problem. I have run two servers and two clients. On both clients
> operate in the background loop:
>
> for ((j=0; j< $RANDOM; j++)) {
> PLIK=$RANDOM.$RANDOM
> dd if=/dev/urandom of=/mnt/glusterfs/$KAT/$PLIK bs=$RANDOM count=1
> dd if=/mnt/glusterfs/$KAT/$PLIK of=/dev/null
> rm -f /mnt/glusterfs/$KAT/$PLIK
> }
>
>
>
> If both servers are connected to everything is fine. If one stops working
> after several minutes will go back to the server, a client I have:
>
> dd: opening `/ mnt/glusterfs/24427/30087.20476 ': No such file or
directory
> dd: opening `/ mnt/glusterfs/24427/30087.20476 ': No such file or
directory
> dd: opening `/ mnt/glusterfs/24427/18649.25895 ': No such file or
directory
>
>
> after a few seconds, everything is working again.
>
> I think that the client is trying to read the file from the new server. I
> think this is not work?
>
>
> client:
>
> volume client1
> type protocol/client
> option transport-type tcp/client
> option remote-host 10.0.1.130
> option remote-port 6996
> option remote-subvolume posix1
> end-volume
>
> volume client2
> type protocol/client
> option transport-type tcp/client
> option remote-host 10.0.1.131
> option remote-port 6996
> option remote-subvolume posix2
> end-volume
>
> volume afr
> type cluster/afr
> subvolumes client1 client2
> end-volume
>
> volume rh
> type performance/read-ahead
> option page-size 100KB
> option page-count 3
> subvolumes afr
> end-volume
>
> volume wh
> type performance/write-behind
> option aggregate-size 1MB
> option flush-behind on
> subvolumes rh
> end-volume
>
>
> server:
>
> volume posix1
> type storage/posix
> option directory /var/storage/glusterfs
> option debug on
> end-volume
>
> volume posix-locks
> type features/posix-locks
> option mandatory on
> subvolumes posix1
> end-volume
>
> volume io-thr
> type performance/io-threads
> option thread-count 2
> option cache-size 64MB
> subvolumes posix-locks
> end-volume
>
> volume server
> type protocol/server
> option transport-type tcp/server
> option listen-port 6996
> subvolumes io-thr
> option auth.ip.posix1.allow 10.*.*.*
> end-volume
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
>
>
--
Raghavendra G
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://supercolony.gluster.org/pipermail/gluster-users/attachments/20081218/d41776b8/attachment.html>