Jason J. W. Williams
2006-Dec-07 22:18 UTC
[zfs-discuss] Re: zpool import takes to long with large numbers of file systems
Hi Luke, That''s terrific! You know you might be able to tell ZFS which disks to look at. I''m not sure. It would be interesting, if anyone with a Thumper could comment on whether or not they see the import time issue. What are your load times now with MPXIO? Best Regards, Jason On 12/7/06, Luke Schwab <lukeschwab at yahoo.com> wrote:> Jason, > > Sorry, I don''t have IM. I did make some progress on > testing. > > The solaris 10 OS allows you to kind of lun mask > within the fp.conf file by creating a list of luns you > don''t want to see. This has improved my time greatly. > I can now create/export within a second and import > only takes about 10 seconds. What a differnce compared > to the 5-8 minutes I''ve been seeing! but this not good > if my machine needs to see LOTs of luns. > > It would be nice if there was a feature in zfs where > you could specify which disk the the pool resides > instead of zfs looking through every disk attached to > the machine. > > Luke Schwab > > > --- "Jason J. W. Williams" <jasonjwwilliams at gmail.com> > wrote: > > > Hey Luke, > > > > Do you have IM? > > > > My Yahoo IM ID is jasonjwwilliams at yahoo.com > > -J > > > > On 12/6/06, Luke Schwab <lukeschwab at yahoo.com> > > wrote: > > > Rats, I think I know where your going? We use LSIs > > > exclusively. > > > > > > LSI performs lun masking as the driver level. You > > can > > > specifically tell the LSI HBA to only bind to > > specific > > > luns on the array. The array doesn''t appear to > > support > > > lun masking by itself. > > > > > > I believe you can also mask at the SAN switch but > > most > > > of our connections are direct connect to the > > array. > > > > > > I just thought you know a quick and easy way to > > mask > > > in Solaris 10. I tried the fp.conf file with a > > > "blackout list" to prevent certain luns from being > > > viewed but I couldn''t get the OS to have a list of > > > luns that it can only allow. > > > > > > Thanks, > > > Luke > > > > > > > > > --- "Jason J. W. Williams" > > <jasonjwwilliams at gmail.com> > > > wrote: > > > > > > > Hi Luke, > > > > > > > > Who makes your array? IBM, SGI or StorageTek? > > > > > > > > Best Regards, > > > > Jason > > > > > > > > On 12/6/06, Luke Schwab <lukeschwab at yahoo.com> > > > > wrote: > > > > > Jason, > > > > > > > > > > could you give me a tip on how to do lun > > masking. > > > > I > > > > > used to do it via the /kernel/drv/ssd.conf > > file > > > > with > > > > > an LSI HBA. Now I have Emulex HBAs with the > > > > Leadville. > > > > > > > > > > > > > > > I saw on sunsolve a way to mask using a "black > > > > list" > > > > > in the /kernel/drv/fp.conf file but that isn''t > > > > what I > > > > > was looking for. > > > > > > > > > > Do you know any other ways? > > > > > > > > > > Thanks, > > > > > > > > > > Luke Scwhab > > > > > --- "Jason J. W. Williams" > > > > <jasonjwwilliams at gmail.com> > > > > > wrote: > > > > > > > > > > > Hi Luke, > > > > > > > > > > > > I think you''ll really like it. We moved from > > > > UFS/SVM > > > > > > and its a night > > > > > > and day management difference. Though I > > > > understand > > > > > > SVM itself is > > > > > > easier to deal with than VxVM, so it may be > > an > > > > order > > > > > > of magnitude > > > > > > easier. > > > > > > > > > > > > Best Regards, > > > > > > Jason > > > > > > > > > > > > On 12/6/06, Luke Schwab > > <lukeschwab at yahoo.com> > > > > > > wrote: > > > > > > > The 4884 as well as the V280 server is > > using 2 > > > > > > ports > > > > > > > each. I don''t have any FS''s at this point. > > I''m > > > > > > trying > > > > > > > to keep it simple for now. > > > > > > > > > > > > > > We are beta testing to go away from VxVM > > and > > > > such. > > > > > > > > > > > > > > > > > > > > > > > > > > > > --- "Jason J. W. Williams" > > > > > > <jasonjwwilliams at gmail.com> > > > > > > > wrote: > > > > > > > > > > > > > > > Hi Luke, > > > > > > > > > > > > > > > > Is the 4884 using two or four ports? > > Also, > > > > how > > > > > > many > > > > > > > > FSs are involved? > > > > > > > > > > > > > > > > Best Regards, > > > > > > > > Jason > > > > > > > > > > > > > > > > On 12/6/06, Luke Schwab > > > > <lukeschwab at yahoo.com> > > > > > > > > wrote: > > > > > > > > > I, too, experienced a long delay while > > > > > > importing a > > > > > > > > zpool on a second machine. I do not have > > any > > > > > > > > filesystems in the pool. Just the > > Solaris 10 > > > > > > > > Operating system, Emulex 10002DC HBA, > > and a > > > > 4884 > > > > > > LSI > > > > > > > > array (dual attached). > > > > > > > > > > > > > > > > > > I don''t have any file systems created > > but > > > > when > > > > > > > > STMS(mpxio) is enabled I see > > > > > > > > > > > > > > > > > > # time zpool import testpool > > > > > > > > > real 6m41.01s > > > > > > > > > user 0m.30s > > > > > > > > > sys 0m0.14s > > > > > > > > > > > > > > > > > > When I disable STMS(mpxio), the times > > are > > > > much > > > > > > > > better but still not that great? > > > > > > > > > > > > > > > > > > # time zpool import testpool > > > > > > > > > real 1m15.01s > > > > > > > > > user 0m.15s > > > > > > > > > sys 0m0.35s > > > > > > > > > > > > > > > > > > Are these normal symproms?? > > > > > > > > > > > > > > > > > > Can anyone explain why I too see > > delays > > > > even > > > > > > > > though I don''t have any file systems in > > the > > > > > > zpool? > > > > > > > > > > > > > > > > > > > > > > > > > > > This message posted from > > opensolaris.org > > > > > > > > > > > > > > > > > _______________________________________________ > > > > > > > > > zfs-discuss mailing list > > > > > > > > > zfs-discuss at opensolaris.org > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > __________________________________________________ > > > > > > > Do You Yahoo!? > > > > > > > Tired of spam? Yahoo! Mail has the best > > spam > > > > > > protection around > > > > > > > http://mail.yahoo.com > > > > > > > > > > > > > > > > > > > > > > > > > > === message truncated ==> > > > > ____________________________________________________________________________________ > Have a burning question? > Go to www.Answers.yahoo.com and get answers from real people who know. >
Roch - PAE
2006-Dec-08 09:14 UTC
[zfs-discuss] Re: zpool import takes to long with large numbers of file systems
I got around that some time ago with a little hack; Maintain a directory with soft links to disks of interest: ls -l .../mydsklist total 50 lrwxrwxrwx 1 cx158393 staff 17 Apr 29 2006 c1t0d0s1 -> /dev/dsk/c1t0d0s1 lrwxrwxrwx 1 cx158393 staff 18 Apr 29 2006 c1t16d0s1 -> /dev/dsk/c1t16d0s1 lrwxrwxrwx 1 cx158393 staff 18 Apr 29 2006 c1t17d0s1 -> /dev/dsk/c1t17d0s1 Then: zpool import -d ..../mydkslist mypool Hope that helps... -r Jason J. W. Williams writes: > Hi Luke, > > That''s terrific! > > You know you might be able to tell ZFS which disks to look at. I''m not > sure. It would be interesting, if anyone with a Thumper could comment > on whether or not they see the import time issue. What are your load > times now with MPXIO? > > Best Regards, > Jason > > > > On 12/7/06, Luke Schwab <lukeschwab at yahoo.com> wrote: > > Jason, > > > > Sorry, I don''t have IM. I did make some progress on > > testing. > > > > The solaris 10 OS allows you to kind of lun mask > > within the fp.conf file by creating a list of luns you > > don''t want to see. This has improved my time greatly. > > I can now create/export within a second and import > > only takes about 10 seconds. What a differnce compared > > to the 5-8 minutes I''ve been seeing! but this not good > > if my machine needs to see LOTs of luns. > > > > It would be nice if there was a feature in zfs where > > you could specify which disk the the pool resides > > instead of zfs looking through every disk attached to > > the machine. > > > > Luke Schwab > > > > > > --- "Jason J. W. Williams" <jasonjwwilliams at gmail.com> > > wrote: > > > > > Hey Luke, > > > > > > Do you have IM? > > > > > > My Yahoo IM ID is jasonjwwilliams at yahoo.com > > > -J > > > > > > On 12/6/06, Luke Schwab <lukeschwab at yahoo.com> > > > wrote: > > > > Rats, I think I know where your going? We use LSIs > > > > exclusively. > > > > > > > > LSI performs lun masking as the driver level. You > > > can > > > > specifically tell the LSI HBA to only bind to > > > specific > > > > luns on the array. The array doesn''t appear to > > > support > > > > lun masking by itself. > > > > > > > > I believe you can also mask at the SAN switch but > > > most > > > > of our connections are direct connect to the > > > array. > > > > > > > > I just thought you know a quick and easy way to > > > mask > > > > in Solaris 10. I tried the fp.conf file with a > > > > "blackout list" to prevent certain luns from being > > > > viewed but I couldn''t get the OS to have a list of > > > > luns that it can only allow. > > > > > > > > Thanks, > > > > Luke > > > > > > > > > > > > --- "Jason J. W. Williams" > > > <jasonjwwilliams at gmail.com> > > > > wrote: > > > > > > > > > Hi Luke, > > > > > > > > > > Who makes your array? IBM, SGI or StorageTek? > > > > > > > > > > Best Regards, > > > > > Jason > > > > > > > > > > On 12/6/06, Luke Schwab <lukeschwab at yahoo.com> > > > > > wrote: > > > > > > Jason, > > > > > > > > > > > > could you give me a tip on how to do lun > > > masking. > > > > > I > > > > > > used to do it via the /kernel/drv/ssd.conf > > > file > > > > > with > > > > > > an LSI HBA. Now I have Emulex HBAs with the > > > > > Leadville. > > > > > > > > > > > > > > > > > > I saw on sunsolve a way to mask using a "black > > > > > list" > > > > > > in the /kernel/drv/fp.conf file but that isn''t > > > > > what I > > > > > > was looking for. > > > > > > > > > > > > Do you know any other ways? > > > > > > > > > > > > Thanks, > > > > > > > > > > > > Luke Scwhab > > > > > > --- "Jason J. W. Williams" > > > > > <jasonjwwilliams at gmail.com> > > > > > > wrote: > > > > > > > > > > > > > Hi Luke, > > > > > > > > > > > > > > I think you''ll really like it. We moved from > > > > > UFS/SVM > > > > > > > and its a night > > > > > > > and day management difference. Though I > > > > > understand > > > > > > > SVM itself is > > > > > > > easier to deal with than VxVM, so it may be > > > an > > > > > order > > > > > > > of magnitude > > > > > > > easier. > > > > > > > > > > > > > > Best Regards, > > > > > > > Jason > > > > > > > > > > > > > > On 12/6/06, Luke Schwab > > > <lukeschwab at yahoo.com> > > > > > > > wrote: > > > > > > > > The 4884 as well as the V280 server is > > > using 2 > > > > > > > ports > > > > > > > > each. I don''t have any FS''s at this point. > > > I''m > > > > > > > trying > > > > > > > > to keep it simple for now. > > > > > > > > > > > > > > > > We are beta testing to go away from VxVM > > > and > > > > > such. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --- "Jason J. W. Williams" > > > > > > > <jasonjwwilliams at gmail.com> > > > > > > > > wrote: > > > > > > > > > > > > > > > > > Hi Luke, > > > > > > > > > > > > > > > > > > Is the 4884 using two or four ports? > > > Also, > > > > > how > > > > > > > many > > > > > > > > > FSs are involved? > > > > > > > > > > > > > > > > > > Best Regards, > > > > > > > > > Jason > > > > > > > > > > > > > > > > > > On 12/6/06, Luke Schwab > > > > > <lukeschwab at yahoo.com> > > > > > > > > > wrote: > > > > > > > > > > I, too, experienced a long delay while > > > > > > > importing a > > > > > > > > > zpool on a second machine. I do not have > > > any > > > > > > > > > filesystems in the pool. Just the > > > Solaris 10 > > > > > > > > > Operating system, Emulex 10002DC HBA, > > > and a > > > > > 4884 > > > > > > > LSI > > > > > > > > > array (dual attached). > > > > > > > > > > > > > > > > > > > > I don''t have any file systems created > > > but > > > > > when > > > > > > > > > STMS(mpxio) is enabled I see > > > > > > > > > > > > > > > > > > > > # time zpool import testpool > > > > > > > > > > real 6m41.01s > > > > > > > > > > user 0m.30s > > > > > > > > > > sys 0m0.14s > > > > > > > > > > > > > > > > > > > > When I disable STMS(mpxio), the times > > > are > > > > > much > > > > > > > > > better but still not that great? > > > > > > > > > > > > > > > > > > > > # time zpool import testpool > > > > > > > > > > real 1m15.01s > > > > > > > > > > user 0m.15s > > > > > > > > > > sys 0m0.35s > > > > > > > > > > > > > > > > > > > > Are these normal symproms?? > > > > > > > > > > > > > > > > > > > > Can anyone explain why I too see > > > delays > > > > > even > > > > > > > > > though I don''t have any file systems in > > > the > > > > > > > zpool? > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > This message posted from > > > opensolaris.org > > > > > > > > > > > > > > > > > > > > _______________________________________________ > > > > > > > > > > zfs-discuss mailing list > > > > > > > > > > zfs-discuss at opensolaris.org > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > __________________________________________________ > > > > > > > > Do You Yahoo!? > > > > > > > > Tired of spam? Yahoo! Mail has the best > > > spam > > > > > > > protection around > > > > > > > > http://mail.yahoo.com > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > === message truncated == > > > > > > > > > > ____________________________________________________________________________________ > > Have a burning question? > > Go to www.Answers.yahoo.com and get answers from real people who know. > > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss at opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Jason J. W. Williams
2006-Dec-08 09:16 UTC
[zfs-discuss] Re: zpool import takes to long with large numbers of file systems
Hi Roch, That''s a pretty cool idea! -J On 12/8/06, Roch - PAE <Roch.Bourbonnais at sun.com> wrote:> > I got around that some time ago with a little hack; > Maintain a directory with soft links to disks of interest: > > > ls -l .../mydsklist > total 50 > lrwxrwxrwx 1 cx158393 staff 17 Apr 29 2006 c1t0d0s1 -> /dev/dsk/c1t0d0s1 > lrwxrwxrwx 1 cx158393 staff 18 Apr 29 2006 c1t16d0s1 -> /dev/dsk/c1t16d0s1 > lrwxrwxrwx 1 cx158393 staff 18 Apr 29 2006 c1t17d0s1 -> /dev/dsk/c1t17d0s1 > > > Then: > > > zpool import -d ..../mydkslist mypool > > Hope that helps... > > -r > > Jason J. W. Williams writes: > > Hi Luke, > > > > That''s terrific! > > > > You know you might be able to tell ZFS which disks to look at. I''m not > > sure. It would be interesting, if anyone with a Thumper could comment > > on whether or not they see the import time issue. What are your load > > times now with MPXIO? > > > > Best Regards, > > Jason > > > > > > > > On 12/7/06, Luke Schwab <lukeschwab at yahoo.com> wrote: > > > Jason, > > > > > > Sorry, I don''t have IM. I did make some progress on > > > testing. > > > > > > The solaris 10 OS allows you to kind of lun mask > > > within the fp.conf file by creating a list of luns you > > > don''t want to see. This has improved my time greatly. > > > I can now create/export within a second and import > > > only takes about 10 seconds. What a differnce compared > > > to the 5-8 minutes I''ve been seeing! but this not good > > > if my machine needs to see LOTs of luns. > > > > > > It would be nice if there was a feature in zfs where > > > you could specify which disk the the pool resides > > > instead of zfs looking through every disk attached to > > > the machine. > > > > > > Luke Schwab > > > > > > > > > --- "Jason J. W. Williams" <jasonjwwilliams at gmail.com> > > > wrote: > > > > > > > Hey Luke, > > > > > > > > Do you have IM? > > > > > > > > My Yahoo IM ID is jasonjwwilliams at yahoo.com > > > > -J > > > > > > > > On 12/6/06, Luke Schwab <lukeschwab at yahoo.com> > > > > wrote: > > > > > Rats, I think I know where your going? We use LSIs > > > > > exclusively. > > > > > > > > > > LSI performs lun masking as the driver level. You > > > > can > > > > > specifically tell the LSI HBA to only bind to > > > > specific > > > > > luns on the array. The array doesn''t appear to > > > > support > > > > > lun masking by itself. > > > > > > > > > > I believe you can also mask at the SAN switch but > > > > most > > > > > of our connections are direct connect to the > > > > array. > > > > > > > > > > I just thought you know a quick and easy way to > > > > mask > > > > > in Solaris 10. I tried the fp.conf file with a > > > > > "blackout list" to prevent certain luns from being > > > > > viewed but I couldn''t get the OS to have a list of > > > > > luns that it can only allow. > > > > > > > > > > Thanks, > > > > > Luke > > > > > > > > > > > > > > > --- "Jason J. W. Williams" > > > > <jasonjwwilliams at gmail.com> > > > > > wrote: > > > > > > > > > > > Hi Luke, > > > > > > > > > > > > Who makes your array? IBM, SGI or StorageTek? > > > > > > > > > > > > Best Regards, > > > > > > Jason > > > > > > > > > > > > On 12/6/06, Luke Schwab <lukeschwab at yahoo.com> > > > > > > wrote: > > > > > > > Jason, > > > > > > > > > > > > > > could you give me a tip on how to do lun > > > > masking. > > > > > > I > > > > > > > used to do it via the /kernel/drv/ssd.conf > > > > file > > > > > > with > > > > > > > an LSI HBA. Now I have Emulex HBAs with the > > > > > > Leadville. > > > > > > > > > > > > > > > > > > > > > I saw on sunsolve a way to mask using a "black > > > > > > list" > > > > > > > in the /kernel/drv/fp.conf file but that isn''t > > > > > > what I > > > > > > > was looking for. > > > > > > > > > > > > > > Do you know any other ways? > > > > > > > > > > > > > > Thanks, > > > > > > > > > > > > > > Luke Scwhab > > > > > > > --- "Jason J. W. Williams" > > > > > > <jasonjwwilliams at gmail.com> > > > > > > > wrote: > > > > > > > > > > > > > > > Hi Luke, > > > > > > > > > > > > > > > > I think you''ll really like it. We moved from > > > > > > UFS/SVM > > > > > > > > and its a night > > > > > > > > and day management difference. Though I > > > > > > understand > > > > > > > > SVM itself is > > > > > > > > easier to deal with than VxVM, so it may be > > > > an > > > > > > order > > > > > > > > of magnitude > > > > > > > > easier. > > > > > > > > > > > > > > > > Best Regards, > > > > > > > > Jason > > > > > > > > > > > > > > > > On 12/6/06, Luke Schwab > > > > <lukeschwab at yahoo.com> > > > > > > > > wrote: > > > > > > > > > The 4884 as well as the V280 server is > > > > using 2 > > > > > > > > ports > > > > > > > > > each. I don''t have any FS''s at this point. > > > > I''m > > > > > > > > trying > > > > > > > > > to keep it simple for now. > > > > > > > > > > > > > > > > > > We are beta testing to go away from VxVM > > > > and > > > > > > such. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --- "Jason J. W. Williams" > > > > > > > > <jasonjwwilliams at gmail.com> > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > Hi Luke, > > > > > > > > > > > > > > > > > > > > Is the 4884 using two or four ports? > > > > Also, > > > > > > how > > > > > > > > many > > > > > > > > > > FSs are involved? > > > > > > > > > > > > > > > > > > > > Best Regards, > > > > > > > > > > Jason > > > > > > > > > > > > > > > > > > > > On 12/6/06, Luke Schwab > > > > > > <lukeschwab at yahoo.com> > > > > > > > > > > wrote: > > > > > > > > > > > I, too, experienced a long delay while > > > > > > > > importing a > > > > > > > > > > zpool on a second machine. I do not have > > > > any > > > > > > > > > > filesystems in the pool. Just the > > > > Solaris 10 > > > > > > > > > > Operating system, Emulex 10002DC HBA, > > > > and a > > > > > > 4884 > > > > > > > > LSI > > > > > > > > > > array (dual attached). > > > > > > > > > > > > > > > > > > > > > > I don''t have any file systems created > > > > but > > > > > > when > > > > > > > > > > STMS(mpxio) is enabled I see > > > > > > > > > > > > > > > > > > > > > > # time zpool import testpool > > > > > > > > > > > real 6m41.01s > > > > > > > > > > > user 0m.30s > > > > > > > > > > > sys 0m0.14s > > > > > > > > > > > > > > > > > > > > > > When I disable STMS(mpxio), the times > > > > are > > > > > > much > > > > > > > > > > better but still not that great? > > > > > > > > > > > > > > > > > > > > > > # time zpool import testpool > > > > > > > > > > > real 1m15.01s > > > > > > > > > > > user 0m.15s > > > > > > > > > > > sys 0m0.35s > > > > > > > > > > > > > > > > > > > > > > Are these normal symproms?? > > > > > > > > > > > > > > > > > > > > > > Can anyone explain why I too see > > > > delays > > > > > > even > > > > > > > > > > though I don''t have any file systems in > > > > the > > > > > > > > zpool? > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > This message posted from > > > > opensolaris.org > > > > > > > > > > > > > > > > > > > > > > > _______________________________________________ > > > > > > > > > > > zfs-discuss mailing list > > > > > > > > > > > zfs-discuss at opensolaris.org > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > __________________________________________________ > > > > > > > > > Do You Yahoo!? > > > > > > > > > Tired of spam? Yahoo! Mail has the best > > > > spam > > > > > > > > protection around > > > > > > > > > http://mail.yahoo.com > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > === message truncated ==> > > > > > > > > > > > > > > ____________________________________________________________________________________ > > > Have a burning question? > > > Go to www.Answers.yahoo.com and get answers from real people who know. > > > > > _______________________________________________ > > zfs-discuss mailing list > > zfs-discuss at opensolaris.org > > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > >
Robert Milkowski
2006-Dec-12 18:14 UTC
[zfs-discuss] Re: zpool import takes to long with large numbers of file systems
Hello Jason, Thursday, December 7, 2006, 11:18:17 PM, you wrote: JJWW> Hi Luke, JJWW> That''s terrific! JJWW> You know you might be able to tell ZFS which disks to look at. I''m not JJWW> sure. It would be interesting, if anyone with a Thumper could comment JJWW> on whether or not they see the import time issue. What are your load JJWW> times now with MPXIO? On x4500 importing a pool made of 44 disks takes about 13 seconds. -- Best regards, Robert mailto:rmilkowski at task.gda.pl http://milek.blogspot.com