Hey folks, I have a bit of an anomaly... Lustre quotas are only working on a portion of my OSTs despite having the correct parameters set... [root at balki ~]# lfs quota -u mjl19 -v /scratch Disk quotas for user mjl19 (uid 1552540): Filesystem kbytes quota limit grace files quota limit grace /scratch 3095592* 10 20 - 3 1000000 1001000 - scratch-MDT0000_UUID 264 - 1024 - 3 - 5120 - scratch-OST0000_UUID 0 - 1024 - - - - - scratch-OST0001_UUID 6936* - 1024 - - - - - scratch-OST0002_UUID 0 - 1024 - - - - - .... scratch-OST001d_UUID 3095560 - 0 - - - - - scratch-OST001e_UUID 0 - 0 - - - - - .... [root at oss2 ~]# tunefs.lustre --dryrun --param /dev/mapper/ost_scratch_29 checking for existing Lustre data: found CONFIGS/mountdata Reading CONFIGS/mountdata Read previous values: Target: scratch-OST001d Index: 29 Lustre FS: scratch Mount type: ldiskfs Flags: 0x1002 (OST no_primnode ) Persistent mount opts: errors=remount-ro,extents,mballoc,nodelalloc,nobarrier Parameters: mgsnode=10.0.1.240 at o2ib mgsnode=10.0.1.239 at o2ib failover.node=10.0.1.236 at o2ib failover.node=10.0.1.235 at o2ib ost.quota_type=ug I even tried to reset the user quota (thinking that perhaps this OSS was offline during the initial setting)... no change and this isn''t an anomaly for the user but for all users on those OSTs Does this mean I have to run a lfs quotacheck across the entire filesystem? If yes, how dangerous is it to do on a live system? Thanks, Mike
Hi, What version of lustre are you running? -cf Michael Losapio <mike.losapio at nyu.edu> wrote:>Hey folks, > >I have a bit of an anomaly... > >Lustre quotas are only working on a portion of my OSTs despite having >the correct parameters set... > >[root at balki ~]# lfs quota -u mjl19 -v /scratch >Disk quotas for user mjl19 (uid 1552540): > Filesystem kbytes quota limit grace files quota limit grace > /scratch 3095592* 10 20 - 3 1000000 1001000 - >scratch-MDT0000_UUID > 264 - 1024 - 3 - 5120 - >scratch-OST0000_UUID > 0 - 1024 - - - - - >scratch-OST0001_UUID > 6936* - 1024 - - - - - >scratch-OST0002_UUID > 0 - 1024 - - - - - >.... >scratch-OST001d_UUID > 3095560 - 0 - - - - - >scratch-OST001e_UUID > 0 - 0 - - - - - >.... > >[root at oss2 ~]# tunefs.lustre --dryrun --param /dev/mapper/ost_scratch_29 >checking for existing Lustre data: found CONFIGS/mountdata >Reading CONFIGS/mountdata > > Read previous values: >Target: scratch-OST001d >Index: 29 >Lustre FS: scratch >Mount type: ldiskfs >Flags: 0x1002 > (OST no_primnode ) >Persistent mount opts: errors=remount-ro,extents,mballoc,nodelalloc,nobarrier >Parameters: mgsnode=10.0.1.240 at o2ib mgsnode=10.0.1.239 at o2ib >failover.node=10.0.1.236 at o2ib failover.node=10.0.1.235 at o2ib >ost.quota_type=ug > >I even tried to reset the user quota (thinking that perhaps this OSS >was offline during the initial setting)... no change and this isn''t an >anomaly for the user but for all users on those OSTs > >Does this mean I have to run a lfs quotacheck across the entire >filesystem? If yes, how dangerous is it to do on a live system? > >Thanks, > >Mike >_______________________________________________ >Lustre-discuss mailing list >Lustre-discuss at lists.lustre.org >http://lists.lustre.org/mailman/listinfo/lustre-discuss-------------- next part -------------- An HTML attachment was scrubbed... URL: http://lists.lustre.org/pipermail/lustre-discuss/attachments/20130227/8fa05da9/attachment.html
1.8.7 Wham Cloud release... lustre-1.8.7-wc1_2.6.18_274.3.1.el5 On Wed, Feb 27, 2013 at 8:53 PM, Colin Faber <colin_faber at xyratex.com> wrote:> Hi, > > What version of lustre are you running? > > -cf > > > Michael Losapio <mike.losapio at nyu.edu> wrote: > > Hey folks, > > I have a bit of an anomaly... > > Lustre quotas are only working on a portion of my OSTs despite having > the correct parameters set... > > [root at balki ~]# lfs quota -u mjl19 -v /scratch > Disk quotas for user mjl19 (uid 1552540): > Filesystem kbytes quota limit grace files quota limit > grace > /scratch 3095592* 10 20 - 3 1000000 1001000 > - > scratch-MDT0000_UUID > 264 - 1024 - 3 - 5120 > - > scratch-OST0000_UUID > 0 - 1024 - - - - > - > scratch-OST0001_UUID > 6936* - 1024 - - - - > - > scratch-OST0002_UUID > 0 - 1024 - - - - > - > .... > scratch-OST001d_UUID > 3095560 - 0 - - - - > - > scratch-OST001e_UUID > 0 - 0 - - - - > - > .... > > [root at oss2 ~]# tunefs.lustre --dryrun --param /dev/mapper/ost_scratch_29 > checking for existing Lustre data: found CONFIGS/mountdata > Reading CONFIGS/mountdata > > Read previous values: > Target: scratch-OST001d > Index: 29 > Lustre FS: scratch > Mount type: ldiskfs > Flags: 0x1002 > (OST no_primnode ) > Persistent mount opts: > errors=remount-ro,extents,mballoc,nodelalloc,nobarrier > Parameters: mgsnode=10.0.1.240 at o2ib mgsnode=10.0.1.239 at o2ib > failover.node=10.0.1.236 at o2ib failover.node=10.0.1.235 at o2ib > ost.quota_type=ug > > I even tried to reset the user quota (thinking that perhaps this OSS > was offline during the initial setting)... no change and this isn''t an > anomaly for the user but for all users on those OSTs > > Does this mean I have to run a lfs quotacheck across the entire > filesystem? If yes, how dangerous is it to do on a live system? > > Thanks, > > Mike > _______________________________________________ > Lustre-discuss mailing list > Lustre-discuss at lists.lustre.org > http://lists.lustre.org/mailman/listinfo/lustre-discuss
Hi Mike, You should be fine running quotacheck. The warning in the manual still holds true though, if you''re actively writing data to the file system while QC is running your accounting may not be 100% correct, but if `close enough'' works for you, then you should have no issues =) -cf On 02/28/2013 06:32 AM, Michael Losapio wrote:> 1.8.7 Wham Cloud release... > > lustre-1.8.7-wc1_2.6.18_274.3.1.el5 > > > > On Wed, Feb 27, 2013 at 8:53 PM, Colin Faber <colin_faber at xyratex.com> wrote: >> Hi, >> >> What version of lustre are you running? >> >> -cf >> >> >> Michael Losapio <mike.losapio at nyu.edu> wrote: >> >> Hey folks, >> >> I have a bit of an anomaly... >> >> Lustre quotas are only working on a portion of my OSTs despite having >> the correct parameters set... >> >> [root at balki ~]# lfs quota -u mjl19 -v /scratch >> Disk quotas for user mjl19 (uid 1552540): >> Filesystem kbytes quota limit grace files quota limit >> grace >> /scratch 3095592* 10 20 - 3 1000000 1001000 >> - >> scratch-MDT0000_UUID >> 264 - 1024 - 3 - 5120 >> - >> scratch-OST0000_UUID >> 0 - 1024 - - - - >> - >> scratch-OST0001_UUID >> 6936* - 1024 - - - - >> - >> scratch-OST0002_UUID >> 0 - 1024 - - - - >> - >> .... >> scratch-OST001d_UUID >> 3095560 - 0 - - - - >> - >> scratch-OST001e_UUID >> 0 - 0 - - - - >> - >> .... >> >> [root at oss2 ~]# tunefs.lustre --dryrun --param /dev/mapper/ost_scratch_29 >> checking for existing Lustre data: found CONFIGS/mountdata >> Reading CONFIGS/mountdata >> >> Read previous values: >> Target: scratch-OST001d >> Index: 29 >> Lustre FS: scratch >> Mount type: ldiskfs >> Flags: 0x1002 >> (OST no_primnode ) >> Persistent mount opts: >> errors=remount-ro,extents,mballoc,nodelalloc,nobarrier >> Parameters: mgsnode=10.0.1.240 at o2ib mgsnode=10.0.1.239 at o2ib >> failover.node=10.0.1.236 at o2ib failover.node=10.0.1.235 at o2ib >> ost.quota_type=ug >> >> I even tried to reset the user quota (thinking that perhaps this OSS >> was offline during the initial setting)... no change and this isn''t an >> anomaly for the user but for all users on those OSTs >> >> Does this mean I have to run a lfs quotacheck across the entire >> filesystem? If yes, how dangerous is it to do on a live system? >> >> Thanks, >> >> Mike >> _______________________________________________ >> Lustre-discuss mailing list >> Lustre-discuss at lists.lustre.org >> http://lists.lustre.org/mailman/listinfo/lustre-discuss