search for: failbacks

Displaying 20 results from an estimated 57 matches for "failbacks".

Did you mean: failback
2010 Jul 28
1
remus - failback?
does remus provide failback mechanism? _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
2017 Aug 14
6
Failback mailboxes?
Hi! Have been using Fedora as my dovecot server for some time and am struggling with systemd at every update. Fedora insists on setting ProtectSystem=full in both dovecot.service and postfix.service at every update of the packages. This makes my mailstore which is in /usr/local/var/mail Read-only. And this makes the incoming emails delivered through dovecot-lda disappear into /dev/null until I
2017 Aug 14
0
Failback mailboxes?
14.08.2017 09:24 Dag Nygren kirjutas: > > Hi! > > Have been using Fedora as my dovecot server for > some time and am struggling with systemd > at every update. > Fedora insists on setting > ProtectSystem=full in both dovecot.service and postfix.service > at every update of the packages. > > This makes my mailstore which is in /usr/local/var/mail > Read-only.
2017 Aug 16
0
Failback mailboxes?
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 On Wed, 16 Aug 2017, Matt Bryant wrote: > hmm if message cannot be written to disk surely it remains on mda queue > as not delviered and does not just disappear ? or am i reading this > wrong ?! as Matt writes your MDA (aka dovecot-lda) returns with an exit code != 0 and your MTA should queue the message for later re-delivery. IMHO, you
2017 Aug 15
3
Failback mailboxes?
hmm if message cannot be written to disk surely it remains on mda queue as not delviered and does not just disappear ? or am i reading this wrong ?! > Dag Nygren <mailto:dag at newtech.fi> > 16 August 2017 at 7:14 am > Thanks for all the advice on how to configure systemd > not to loose my emails after every update. Much appreciated. > > But there could be other reasons
2017 Aug 14
1
Failback mailboxes?
On Monday 14 August 2017 10:22:54 Sander Lepik wrote: > 14.08.2017 09:24 Dag Nygren kirjutas: > > PS! I really hate systemd - Destroys the UNIX way of > > doing things with a heavy axe.... > > Don't hate it, better learn to use it: > https://wiki.archlinux.org/index.php/systemd#Drop-in_files Cannot find a way to "remove" the ProtectSystem setting as there
2017 Aug 15
0
Failback mailboxes?
Thanks for all the advice on how to configure systemd not to loose my emails after every update. Much appreciated. But there could be other reasons for the mailboxes not being writable and what I am really asking for is for dovecot-lda not to loose the incoming emails into thin air in these cases. Could we have some kind of collective place/places where they would be saved in this case and then
2019 Feb 25
2
glusterfs + ctdb + nfs-ganesha , unplug the network cable of serving node, takes around ~20 mins for IO to resume
Hi all We did some failover/failback tests on 2 nodes��A and B�� with architecture 'glusterfs + ctdb(public address) + nfs-ganesha'�� 1st: During write, unplug the network cable of serving node A ->NFS Client took a few seconds to recover to conitinue writing. After some minutes, plug the network cable of serving node A ->NFS Client also took a few seconds to recover
2005 Jun 20
2
Fallback
Thanks for the answer. How can I do a "failback" for a relayed stream ? The idea is to set up a relay that can use for the same mountpoint two connection. Karl Heyes a ?crit : >On Mon, 2005-06-20 at 13:34, EISELE Pascal wrote: > > >>Hi, >> >>I'm trying the following settings but it seams that it's not working :( >>While I try to switch down
2010 Jun 21
0
Seriously degraded SAS multipathing performance
I''m seeing seriously degraded performance with round-robin SAS multipathing. I''m hoping you guys can help me achieve full throughput across both paths. My System Config: OpenSolaris snv_134 2 x E5520 2.4 GHz Xeon Quad-Core Processors 48 GB RAM 2 x LSI SAS 9200-8e (eight-port external 6Gb/s SATA and SAS PCIe 2.0 HBA) 1 X Mellanox 40 Gb/s dual port card PCIe 2.0 1 x JBOD:
2018 Nov 21
2
relay backup file
Hi, I have install 2 Icecast servers. We not using master relay. But we use specific mountrelay for fallback: server1 - pc1 encoder over dsl server2 - pc2 encoder over cable internet We use the same mountpoints on both icecast servers and use the failback to a relay mountpoint from the other server. Works fine when we shutdown pc1 or pc2 there is on both icecast server music. We like to have
2007 Feb 17
8
ZFS with SAN Disks and mutipathing
Hi, I just deploy the ZFS on an SAN attach disk array and it''s working fine. How do i get dual pathing advantage of the disk ( like DMP in Veritas). Can someone point to correct doc and setup. Thanks in Advance. Rgds Vikash Gupta This message posted from opensolaris.org
2016 Nov 10
1
CTDB IP takeover/failover tunables - do you use them?
...will always be hosted by node Y. The cost of using deterministic IP address assignment is that it disables part of the logic where ctdb tries to reduce the number of public IP assignment changes in the cluster. This tunable may increase the number of IP failover/failbacks that are performed on the cluster by a small margin. LCP2PublicIPs Default: 1 When set to 1, ctdb uses the LCP2 ip allocation algorithm. I plan to replace these with a single tunable to select the algorithm (0 = deterministic, 1 = non-deterministic, 2 = LCP2 (default))....
2009 Sep 17
1
multipath using defaults rather than multipath.conf contents for some devices (?) - why ?
hi all We have a rh linux server connected to two HP SAN controllers, one an HSV200 (on the way out), the other an HSV400 (on the way in). (Via a Qlogic HBA). /etc/multipath.conf contains this : device { vendor "(COMPAQ|HP)" product "HSV1[01]1|HSV2[01]0|HSV300|HSV4[05]0" getuid_callout "/sbin/scsi_id -g -u -s /block/%n"
2019 Mar 04
0
glusterfs + ctdb + nfs-ganesha , unplug the network cable of serving node, takes around ~20 mins for IO to resume
Hi Dan, On Mon, 25 Feb 2019 02:43:31 +0000, "Liu, Dan via samba" <samba at lists.samba.org> wrote: > We did some failover/failback tests on 2 nodes(A and B) with > architecture 'glusterfs + ctdb(public address) + nfs-ganesha'。 > > 1st: > During write, unplug the network cable of serving node A > ->NFS Client took a few seconds to recover to conitinue
2017 Jan 31
1
multipath show config different in CentOS 7?
Hello, suppose I want to use a special configuration for my IBM/1814 storage array luns, then I put something like this in multipath.conf devices { device { vendor "IBM" product "^1814" product_blacklist "Universal Xport" path_grouping_policy "group_by_prio" path_checker
2019 Mar 31
1
mountpoint configuration beyond the basics?
Hello, I'm new to icecast, but in few time I setup a streaming system using a raspberry pi as stream source through darkice towards my icecast server. All seemed fine with the basic mountpoint configuration, but when I tried to try some of the more "advanced" features I failed badly. More specifically: 1. I tried to force a failback mountpoint like this:
2006 Apr 13
1
device-mapper multipath
I am attempting to get multipath working with device-mapper (CentOS 4.2 and 4.3). It works on EVERY install of mine from RH (also v4.2, 4.3), but the same multipath.conf imported to all my installs of CentOS do not work. Note that I have tested a working 4.2 configuration file from RH on CentOS 4.2 and a working 4.3 configuration (it changed slightly) on CentOS 4.3. Neither worked. Our production
2008 Feb 08
4
List of supported multipath drivers
Where can I find a list of supported multipath drivers for ZFS? Keith McAndrew Senior Systems Engineer Northern California SUN Microsystems - Data Management Group <mailto:Keith.McAndrew at SUN.com> Keith.McAndrew at SUN.com 916 715 8352 Cell CONFIDENTIALITY NOTICE The information contained in this transmission may contain privileged and confidential information of SUN
2008 Mar 04
0
Device-mapper-multipath not working correctly with GNBD devices
Hi all, I am trying to configure a failover multipath between 2 GNBD devices. I have a 4 nodes Redhat Cluster Suite (RCS) cluster. 3 of them are used for running services, 1 of them for central storage. In the future I am going to introduce another machine for central storage. The 2 storage machine are going to share/export the same disk. The idea is not to have a single point of failure