This is using Xen 3.0 testing (8259). SMP was disabled in the domU kernel to allow GFS to compile. I''m not sure this is a Xen issue but I thought there were some people running GFS under Xen so it might be worth posting. update...I tried OCFS2 and FUSE (also in unstable/8438) with the same result. I''ve heard that others have been able to run GFS2 in domU so maybe it just requires the right versions and configs? I''d be happy to pay for help. --kyler ==============================================================Lock_Harness 1.01.00 (built Dec 22 2005 16:27:41) installed GFS 1.01.00 (built Dec 22 2005 16:28:33) installed Gulm 1.01.00 (built Dec 22 2005 16:27:59) installed GFS: Trying to join cluster "lock_gulm", "eng:00" GFS: fsid=eng:00.0: Joined cluster. Now mounting FS... GFS: fsid=eng:00.0: jid=0: Trying to acquire journal lock... GFS: fsid=eng:00.0: jid=0: Looking at journal... GFS: fsid=eng:00.0: jid=0: Done GFS: fsid=eng:00.0: jid=1: Trying to acquire journal lock... GFS: fsid=eng:00.0: jid=1: Looking at journal... GFS: fsid=eng:00.0: jid=1: Done GFS: fsid=eng:00.0: jid=2: Trying to acquire journal lock... GFS: fsid=eng:00.0: jid=2: Looking at journal... GFS: fsid=eng:00.0: jid=2: Done GFS: fsid=eng:00.0: jid=3: Trying to acquire journal lock... GFS: fsid=eng:00.0: jid=3: Looking at journal... GFS: fsid=eng:00.0: jid=3: Done GFS: fsid=eng:00.0: jid=4: Trying to acquire journal lock... GFS: fsid=eng:00.0: jid=4: Looking at journal... GFS: fsid=eng:00.0: jid=4: Done Unable to handle kernel paging request at ffff8f000db49000 RIP: <ffffffff88072ee5>{:gfs:gfs_meta_header_out+4} PGD 0 Oops: 0002 [1] CPU 0 Modules linked in: lock_gulm gfs lock_harness md5 ipv6 dm_mod Pid: 490, comm: mount Not tainted 2.6.12.6-xenU RIP: e030:[<ffffffff88072ee5>] <ffffffff88072ee5>{:gfs:gfs_meta_header_out+4} RSP: e02b:ffff88000e54d840 EFLAGS: 00010296 RAX: 0000000070191601 RBX: ffff88000e54d888 RCX: 0000000000001000 RDX: 0000000000000008 RSI: ffff8f000db49000 RDI: ffff88000e54d888 RBP: ffff88000e54d9e8 R08: 0000000000000010 R09: 0000000000000000 R10: ffff88000e54d888 R11: 0000000000000068 R12: ffff8f000db49000 R13: 0000000000000000 R14: ffffc20000084718 R15: 0000000000000001 FS: 00002aaaab00e6d0(0000) GS:ffffffff8036a300(0000) knlGS:0000000000000000 CS: e033 DS: 0000 ES: 0000 Process mount (pid: 490, threadinfo ffff88000e54c000, task ffff88000ed0a170) Stack: ffffffff88072f3e ffff88000e54d9e8 0000000000000190 ffff88000db26b00 ffffffff88070aa4 0000000801161970 ffff88000e54da20 ffffc2000005c000 0000000000000001 0000000901161970 Call Trace:<ffffffff88072f3e>{:gfs:gfs_desc_out+18} <ffffffff88070aa4>{:gfs:unlinked_build_bhlist+230} <ffffffff8014f3b2>{__alloc_pages+190} <ffffffff8806f007>{:gfs:disk_commit+161} <ffffffff80151a4d>{cache_alloc_refill+769} <ffffffff880701cd>{:gfs:gfs_log_dump+945} <ffffffff88081782>{:gfs:gfs_make_fs_rw+235} <ffffffff88077900>{:gfs:gfs_get_sb+2759} <ffffffff80151727>{kmem_cache_alloc+57} <ffffffff8016ea27>{do_kern_mount+86} <ffffffff80181d3d>{do_mount+1488} <ffffffff8015a9d0>{do_no_page+1316} <ffffffff8014f273>{buffered_rmqueue+558} <ffffffff8014f3b2>{__alloc_pages+190} <ffffffff8014f8f4>{__get_free_pages+30} <ffffffff80181e43>{sys_mount+133} <ffffffff801111d1>{system_call+117} <ffffffff8011115c>{system_call+0} Code: 89 06 8b 47 04 0f c8 89 46 04 8b 47 10 0f c8 89 46 10 8b 47 RIP <ffffffff88072ee5>{:gfs:gfs_meta_header_out+4} RSP <ffff88000e54d840> CR2: ffff8f000db49000 _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users