Hello, I have a couple of Dell 2950 III, both of them with CentOS 5.3, Xen, drbd 8.2 and cluster suite. Hardware: 32DB RAM, RAID 5 with 6 SAS disks (one hot spare) on a PERC/6 controller. I configured DRBD to use the main network interfaces (bnx2 driver), with bonding and crossover cables to have a direct link. The normal network traffic uses two different network cards. There are two DRBD resources for a total of a little less than 1TB. When the two hosts are in sync, if I activate more than a few (six or seven) xen guests, the master server crashes spectacularly and reboots. I've seen a kernel dump over the serial console, but the machine restarts immediately so I didn't write it down. Unfortunately I cannot experiment because I have production services on those machines (and they are working fine until I start drbd on the slave). drbd configuration is attached. Anybody has an idea of the problem? The crash is perfectly reproducible, and drbd seems to be the problem (maybe the Xen kernel helps?). Thanks in advance, Andrea -- Question: ?Is there a God?? Answer: ?No.? From the Official God FAQ, http://www.400monkeys.com/God/ -------------- next part -------------- # # At most ONE global section is allowed. # It must precede any resource section. # global { # minor-count 64; # dialog-refresh 5; # 5 seconds # disable-ip-verification; usage-count no; } common { syncer { rate 50M; } } # # this need not be r#, you may use phony resource names, # like "resource web" or "resource mail", too # resource virtual1 { protocol C; handlers { pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; #outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater -t 5"; #pri-lost "echo pri-lost. Have a look at the log files. | mail -s 'DRBD Alert' root"; #split-brain "echo split-brain. drbdadm -- --discard-my-data connect $DRBD_RESOURCE ? | mail -s 'DRBD Alert' root"; #out-of-sync "echo out-of-sync. drbdadm down $DRBD_RESOURCE. drbdadm ::::0 set-gi $DRBD_RESOURCE. drbdadm up $DRBD_RESOURCE. | mail -s 'DRBD Alert' root"; #before-resync-target "/usr/lib/drbd/snapshot-resync-target-lvm.sh -p 15 -- -c 16k"; #after-resync-target /usr/lib/drbd/unsnapshot-resync-target-lvm.sh; } startup { wfc-timeout 120; degr-wfc-timeout 120; # 2 minutes. # wait-after-sb; } disk { on-io-error pass_on; fencing resource-and-stonith; # ONLY USE THIS OPTION IF YOU KNOW WHAT YOU ARE DOING. # no-disk-flushes; # no-md-flushes; max-bio-bvecs 1; } net { # max-buffers 2048; # unplug-watermark 128; # max-epoch-size 2048; # ko-count 4; cram-hmac-alg "sha1"; shared-secret "bah"; after-sb-0pri discard-younger-primary; after-sb-1pri call-pri-lost-after-sb; after-sb-2pri call-pri-lost-after-sb; rr-conflict disconnect; # data-integrity-alg "md5"; } syncer { rate 50M; #after "r2"; al-extents 257; # cpu-mask 15; } on server-1 { device /dev/drbd1; disk /dev/sda2; address 192.168.2.1:7788; meta-disk internal; } on server-2 { device /dev/drbd1; disk /dev/sda4; address 192.168.2.2:7788; meta-disk internal; } } resource servizi0 { #********** # protocol C; handlers { pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; #outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater -t 5"; #pri-lost "echo pri-lost. Have a look at the log files. | mail -s 'DRBD Alert' root"; # Notify someone in case DRBD split brained. #split-brain "echo split-brain. drbdadm -- --discard-my-data connect $DRBD_RESOURCE ? | mail -s 'DRBD Alert' root"; # Notify someone in case an online verify run found the backing devices out of sync. #out-of-sync "echo out-of-sync. drbdadm down $DRBD_RESOURCE. drbdadm ::::0 set-gi $DRBD_RESOURCE. drbdadm up $DRBD_RESOURCE. | mail -s 'DRBD Alert' root"; # #before-resync-target "/usr/lib/drbd/snapshot-resync-target-lvm.sh -p 15 -- -c 16k"; #after-resync-target /usr/lib/drbd/unsnapshot-resync-target-lvm.sh; } startup { wfc-timeout 120; degr-wfc-timeout 120; # 2 minutes. # wait-after-sb; # become-primary-on both; } disk { on-io-error pass_on; fencing resource-and-stonith; # size 10G; # no-disk-flushes; # no-md-flushes; max-bio-bvecs 1; } net { # sndbuf-size 512k; # timeout 60; # 6 seconds (unit = 0.1 seconds) # connect-int 10; # 10 seconds (unit = 1 second) # ping-int 10; # 10 seconds (unit = 1 second) # ping-timeout 5; # 500 ms (unit = 0.1 seconds) # max-buffers 2048; # unplug-watermark 128; # max-epoch-size 2048; # ko-count 4; # allow-two-primaries; cram-hmac-alg "sha1"; shared-secret "serviziDRBDcosasegretissima"; after-sb-0pri discard-younger-primary; after-sb-1pri call-pri-lost-after-sb; after-sb-2pri call-pri-lost-after-sb; rr-conflict disconnect; # data-integrity-alg "md5"; } syncer { rate 50M; #after "r2"; al-extents 257; # cpu-mask 15; } on server-1 { device /dev/drbd0; disk /dev/sda5; address 192.168.2.1:7789; meta-disk internal; } on server-2 { device /dev/drbd0; disk /dev/sda3; address 192.168.2.2:7789; meta-disk internal; } } -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 307 bytes Desc: This is a digitally signed message part URL: <http://lists.centos.org/pipermail/centos/attachments/20090728/0f4dc1cb/attachment-0003.sig>
On Tue, 2009-07-28 at 20:11 +0200, Andrea Dell'Amico wrote:> Hello, > I have a couple of Dell 2950 III, both of them with CentOS 5.3, Xen, > drbd 8.2 and cluster suite. > Hardware: 32DB RAM, RAID 5 with 6 SAS disks (one hot spare) on a PERC/6 > controller. > > I configured DRBD to use the main network interfaces (bnx2 driver), with > bonding and crossover cables to have a direct link. > The normal network traffic uses two different network cards. > There are two DRBD resources for a total of a little less than 1TB. > > When the two hosts are in sync, if I activate more than a few (six or > seven) xen guests, the master server crashes spectacularly and reboots. > > I've seen a kernel dump over the serial console, but the machine > restarts immediately so I didn't write it down.If you have an available pc, hook it up in place of the serial console and start a terminal emulator, e.g. minicom or whatever you prefer, and turn on full logging. This should save everyting in a file that you can then review. If it's a Windows based, just remember to get rid of the ^M with dos2unix, or equivalent, after you send it to a *IX box. I don't know anything about the rest of your problem, sorry.> > Unfortunately I cannot experiment because I have production services on > those machines (and they are working fine until I start drbd on the > slave). > > drbd configuration is attached. > > Anybody has an idea of the problem? The crash is perfectly reproducible, > and drbd seems to be the problem (maybe the Xen kernel helps?). > > Thanks in advance, > Andrea > <snip sig stuff>HTH -- Bill
On Jul 29, 2009, at 7:52 AM, "Andrea Dell'Amico" <adellam at sevenseas.org> wrote:> On Tue, 2009-07-28 at 14:31 -0400, William L. Maltby wrote: > >>> When the two hosts are in sync, if I activate more than a few (six >>> or >>> seven) xen guests, the master server crashes spectacularly and >>> reboots. >>> >>> I've seen a kernel dump over the serial console, but the machine >>> restarts immediately so I didn't write it down. >> >> If you have an available pc, hook it up in place of the serial >> console >> and start a terminal emulator, e.g. minicom or whatever you prefer, >> and >> turn on full logging. This should save everyting in a file that you >> can >> then review. > > Uhm. The console is on the DRAC5 card. I think I would need to > activate > some network kernel crash dump feature. > >> If it's a Windows based, just remember to get rid of the ^M with >> dos2unix, or equivalent, after you send it to a *IX box. >> >> I don't know anything about the rest of your problem, sorry. > > As I wrote, it's a production server. I cannot stop it when I want, I > need to reserve a weekend session. > In the meantime, I was asking if there's a known problem with a setup > like mine.I read on another forum how a user using iSCSI for domUs was experiencing network hangs due to the fact that dom0 didn't have enough scheduler credits to handle the network throughput. That might be related. http://lists.centos.org/pipermail/centos-virt/2009-June/001021.html -Ross
Reasonably Related Threads
- Recover botched drdb gfs2 setup .
- I/O errors in domU with LVM on DRBD
- drbd
- Problem with mdadm + lvm + drbd + ocfs ( sounds obvious, eh ? :) )
- Using XEN 4.1 to run Windows 7 DOMU on SERVER1, replicated the HVM to SERVER2 using DRBD. Win7 DOMU doesn't boot in SERVER2 after replication