Displaying 20 results from an estimated 1731 matches for "handshakes".
Did you mean:
handshake
2006 Jan 29
1
I gonna understand one day
Hi,
I'm still struggling with dovecot and Unixware:
On my main server, I have two problems:
1- imap-login has great difficulties to connect to 'default'.
I have traced it to accept in auth_master_listener.c witch fails with
errno=71 (EPROTO) weird...
2- It some times works and I can log-in but imap gets killed by signal 6
(abort?)
See the logs:
Jan 24 12:40:25 dovecot: Dovecot
2019 Sep 25
3
[nbdkit PATCH 0/2] more protocol.h tweaks
More nbd-protocol.h improvements
Eric Blake (2):
common/protocol: Switch nbdmagic to uint64_t
common/protocol: Declare additional constants
common/protocol/nbd-protocol.h | 16 ++++++++++------
server/protocol-handshake-newstyle.c | 2 +-
server/protocol-handshake-oldstyle.c | 2 +-
plugins/nbd/nbd-standalone.c | 2 +-
tests/test-layers.c | 2 +-
5 files
2019 Sep 24
0
[PATCH nbdkit 3/4] common/protocol: Update nbd-protocol.h so it matches libnbd’s copy.
Diff against libnbd’s copy of this file, and change this one until it
matches.
---
common/protocol/nbd-protocol.h | 76 +++++++++++++++++-----------
server/protocol-handshake-newstyle.c | 26 +++++-----
server/protocol-handshake-oldstyle.c | 4 +-
server/protocol.c | 25 ++++-----
tests/test-layers.c | 14 ++---
5 files changed, 81 insertions(+), 64
2012 Jun 22
1
Fedora 17 GlusterFS 3.3.0 problmes
When I do a NFS mount and do a ls I get:
[root at ovirt share]# ls
ls: reading directory .: Too many levels of symbolic links
[root at ovirt share]# ls -fl
ls: reading directory .: Too many levels of symbolic links
total 3636
drwxr-xr-x 3 root root 16384 Jun 21 19:34 .
dr-xr-xr-x. 21 root root 4096 Jun 21 19:29 ..
drwxr-xr-x 3 root root 16384 Jun 21 19:34 .
dr-xr-xr-x. 21 root root 4096
2016 Sep 26
1
[PATCH] nbdkit: flags are 32 bits for oldstyle connections
# HG changeset patch
# User Carl-Daniel Hailfinger <Carl-Daniel.Hailfinger@inf.h-brs.de>
# Date 1474903734 -7200
# Mon Sep 26 17:28:54 2016 +0200
# Node ID dbd1ea0a401cedcfa426097a289d852971b14f1e
# Parent d7d5078d08c711032837dcac79a4450226ec2ce5
nbdkit: Fix flags in old-style connection
diff -r d7d5078d08c7 -r dbd1ea0a401c src/connections.c
--- a/src/connections.c Sun Sep 25 05:04:02
2019 Sep 12
0
[nbdkit PATCH 2/2] server: Add --mask-handshake option for integration testing
Similar to --no-sr, it can be handy for testing a client
implementation to have a server that can easily be forced into older
behaviors, without having to recompile a one-off hack into a server or
dig up an older server binary that lacked a newer feature.
To see the patch in action, try things like:
$ ./nbdkit -U - -fv --mask-handshake=0 null \
--run 'qemu-nbd --list -k $unixsocket'
2010 Oct 21
4
2.0.5 Error: Timeout waiting for handshake from auth server.
Hello,
Upgraded from 1.2.x to 2.0.5, converted configuration as described in
http://wiki2.dovecot.org/Upgrading/2.0 and now I'm unable to log in to
imap server. Dovecot error.log shows:
Oct 21 09:52:48 imap-login: Error: Timeout waiting for handshake from
auth server. my pid=13338, input bytes=0
Oct 21 09:52:48 imap-login: Error: Timeout waiting for handshake from
auth server. my
2019 Sep 24
11
[PATCH nbdkit 0/4] common/protocol: Unify public <nbd-protocol.h>
We should have only one NBD protocol file. Let's make nbdkit's
version the canonical one, and use it in libnbd.
Rich.
2017 Dec 05
4
SAMBA VFS module for GlusterFS crashes
Hello,
I'm trying to set up a SAMBA server serving a GlusterFS volume.
Everything works fine if I locally mount the GlusterFS volume (`mount
-t glusterfs ...`) and then serve the mounted FS through SAMBA, but
the performance is slower by a 2x/3x compared to a SAMBA server with a
local ext4 filesystem.
I gather that SAMBA vfs_glusterfs module can give better
performance. However, as soon as I
2018 Apr 09
2
Gluster cluster on two networks
Hi all!
I have setup a replicated/distributed gluster cluster 2 x (2 + 1).
Centos 7 and gluster version 3.12.6 on server.
All machines have two network interfaces and connected to two different networks,
10.10.0.0/16 (with hostnames in /etc/hosts, gluster version 3.12.6)
192.168.67.0/24 (with ldap, gluster version 3.13.1)
Gluster cluster was created on the 10.10.0.0/16 net, gluster peer
2018 May 10
0
Tinc 1.1pre15 double-crash
Hello,
this morning I apparently had tinc crash on me.
In 2 independent tinc clusters of 3 nodes each (but located in the same datacenter), one tinc process crashed in each of the clusters.
One process apparently with `status=6/ABRT`, the other with `status=11/SEGV`.
Interestingly, they crashed with only 5 minutes difference.
The only thing I can come up with that might explain this correlation
2018 Apr 10
0
Gluster cluster on two networks
Marcus,
Can you share server-side gluster peer probe and client-side mount
command-lines.
On Tue, Apr 10, 2018 at 12:36 AM, Marcus Peders?n <marcus.pedersen at slu.se>
wrote:
> Hi all!
>
> I have setup a replicated/distributed gluster cluster 2 x (2 + 1).
>
> Centos 7 and gluster version 3.12.6 on server.
>
> All machines have two network interfaces and connected to
2012 Jan 12
2
Request for obfuscating the handshake
Dear OpenSSH team,
First of all thanks a lot for your good work on developing such a usable
peace of software. Nice job.
As you may know, we have some issues using OpenSSH in Iran. Recently the
government did some packet filtering on some protocols, including SSH. We
don't know what exactly they've done but it appears to be some routing rule
that prevents SSH protocol to initiate a
2019 Sep 12
3
[nbdkit PATCH 0/2] Make client fallback testing easier
This is similar to the recent --no-sr option - it's a change that is
unlikely to ever be used except by someone testing whether a client is
compliant to the protocol, but in that niche case, it can be quite
handy (it's a lot nicer to be able to purposefully cripple a server
from the command line than from a one-off compile, when testing if a
client's fallback for a spec-compliant but
2019 Sep 28
11
[nbdkit PATCH v2 0/7] Spec compliance patches
Since the v1 series (0/4, at [1]), I've applied patches 1 and 2,
rewritten patch 3 [Forbid NUL in export and context names] into patch
4 here, patch 4 there turned into patch 6 here, and everything else
here is new.
[1]https://www.redhat.com/archives/libguestfs/2019-September/msg00180.html
I don't know if there is a handy reusable function for checking
whether a string contains valid
2018 Apr 10
1
Gluster cluster on two networks
Yes,
In first server (urd-gds-001):
gluster peer probe urd-gds-000
gluster peer probe urd-gds-002
gluster peer probe urd-gds-003
gluster peer probe urd-gds-004
gluster pool list (from urd-gds-001):
UUID Hostname State
bdbe4622-25f9-4ef1-aad1-639ca52fc7e0 urd-gds-002 Connected
2a48a3b9-efa0-4fb7-837f-c800f04bf99f urd-gds-003 Connected
ad893466-ad09-47f4-8bb4-4cea84085e5b urd-gds-004
2017 Dec 05
0
SAMBA VFS module for GlusterFS crashes
Keep in mind a local disk is 3,6,12 Gbps but a network connection is typically 1Gbps. A local disk quad in raid 10 will outperform a 10G ethernet (especially using SAS drives).
On December 5, 2017 6:11:38 AM EST, Riccardo Murri <riccardo.murri at uzh.ch> wrote:
>Hello,
>
>I'm trying to set up a SAMBA server serving a GlusterFS volume.
>Everything works fine if I locally
2018 Apr 10
0
Gluster cluster on two networks
Hi all!
I have setup a replicated/distributed gluster cluster 2 x (2 + 1).
Centos 7 and gluster version 3.12.6 on server.
All machines have two network interfaces and connected to two different networks,
10.10.0.0/16 (with hostnames in /etc/hosts, gluster version 3.12.6)
192.168.67.0/24 (with ldap, gluster version 3.13.1)
Gluster cluster was created on the 10.10.0.0/16 net, gluster peer probe
2019 Sep 24
0
[PATCH nbdkit 1/4] common/protocol: Rename protocol.h to nbd-protocol.h.
In preparation for installing this header as a common public header
for use by other projects, rename it.
---
common/protocol/Makefile.am | 10 +++++-----
common/protocol/{protocol.h => nbd-protocol.h} | 6 +++---
common/protocol/protostrings.sed | 6 +++---
plugins/nbd/nbd-standalone.c | 2 +-
server/protocol-handshake-newstyle.c
2017 Dec 06
0
SAMBA VFS module for GlusterFS crashes
On Tue, 2017-12-05 at 11:11 +0000, Riccardo Murri wrote:
> Hello,
>
> I'm trying to set up a SAMBA server serving a GlusterFS volume.
> Everything works fine if I locally mount the GlusterFS volume (`mount
> -t glusterfs ...`) and then serve the mounted FS through SAMBA, but
> the performance is slower by a 2x/3x compared to a SAMBA server with a
> local ext4 filesystem.