Displaying 7 results from an estimated 7 matches for "pod2".
Did you mean:
pod
2015 Jul 20
3
dovecot proxy/director and high availability design
...ts to use and how to order them in the network path from client to mail store.
If I have say 1,000 users, all stored in MySQL (or LDAP) and have 4 mail stores, configured into 2, 2 node pods.
MS1 and MS2 are pod1 and are configured with replication (dsync) and host users 0-500. MS3 and MS4 are pod2 and are configured with replication between them and host users 501-1000. Ideally the active connections in pod1 would be split 50/50 between MS1 and MS2. When maintenance is performed obviously all active connections/users would be moved to the other node in the pod and then rebalanced once mai...
2015 Jul 21
3
dovecot proxy/director and high availability design
...) --> Dovecot Proxy(DP) --> Dovecot Director(DD) --> MS1 / MS2.
When DP checks say user100 it'll find a host=DD-POD1 that returns two IPs, those of the two DD that sit in front of POD1. This DD pair is the only pair in the ring and only responsible for POD1. Another pair will handle POD2. When DD looks up the host value for a user it'll find the same name, but the IPs returned will be different. Instead have both IPs of the mail stores returned.
I believe this will achieve what I'm after. HAProxy will do the load balancing of the DP instances. DP will balance the DDs...
2015 Jul 21
0
dovecot proxy/director and high availability design
...t Proxy(DP) --> Dovecot Director(DD) --> MS1 / MS2.
>
>
> When DP checks say user100 it'll find a host=DD-POD1 that returns two IPs, those of the two DD that sit in front of POD1. This DD pair is the only pair in the ring and only responsible for POD1. Another pair will handle POD2. When DD looks up the host value for a user it'll find the same name, but the IPs returned will be different. Instead have both IPs of the mail stores returned.
>
> I believe this will achieve what I'm after. HAProxy will do the load balancing of the DP instances. DP will balan...
2015 Jul 21
2
dovecot proxy/director and high availability design
...) --> Dovecot Director(DD) --> MS1 / MS2.
>>
>>
>> When DP checks say user100 it'll find a host=DD-POD1 that returns two IPs, those of the two DD that sit in front of POD1. This DD pair is the only pair in the ring and only responsible for POD1. Another pair will handle POD2. When DD looks up the host value for a user it'll find the same name, but the IPs returned will be different. Instead have both IPs of the mail stores returned.
>>
>> I believe this will achieve what I'm after. HAProxy will do the load balancing of the DP instances. DP will...
2015 Jul 20
0
dovecot proxy/director and high availability design
...m in the network path from client to mail store.
>
>
> If I have say 1,000 users, all stored in MySQL (or LDAP) and have 4 mail stores, configured into 2, 2 node pods.
>
>
> MS1 and MS2 are pod1 and are configured with replication (dsync) and host users 0-500. MS3 and MS4 are pod2 and are configured with replication between them and host users 501-1000. Ideally the active connections in pod1 would be split 50/50 between MS1 and MS2. When maintenance is performed obviously all active connections/users would be moved to the other node in the pod and then rebalanced once mai...
2012 Dec 13
0
ANNOUNCE: libguestfs 1.20 - tools for accessing and modifying virtual machine disk images
...c/launch-libvirt.c.
* Generated action code is now split over several src/action*.c
files, for faster compilation.
* The huge src/guestfs.c file is now split into smaller logical
units.
POD (documentation) is now generated using a rewritten Perl program
instead of pod2* + shell scripts.
Man pages now contain stable dates (Hilko Bengen).
Skipped tests now exit with code 77, so they appear as SKIP: in make
check output.
The parallel mount-local test has been rewritten in C (RHBZ#838081).
Ruby 1.8.5 is now supported (though Ruby ? 1.9 is pref...
2013 Jul 25
19
[PATCH hivex 00/19] Fix read/write handling of li-records.
This is, hopefully, a full fix for handling of li-records.
See:
https://bugzilla.redhat.com/show_bug.cgi?id=717583
https://bugzilla.redhat.com/show_bug.cgi?id=987463
Rich.