search for: ost0001

Displaying 9 results from an estimated 9 matches for "ost0001".

2007 Dec 25
0
lustre performance question
Hi, We have one Lustre volume that is getting full and some other volumes that are totally empty. The one that is full is a little sluggish at times with the following messages appearing in syslog on the OSS - Lustre: 5809:0:(filter_io_26.c:698:filter_commitrw_write()) data1-OST0001: slow i_mutex 82s Lustre: 5809:0:(filter_io_26.c:711:filter_commitrw_write()) data1-OST0001: slow brw_start 82s Lustre: 5809:0:(filter_io_26.c:763:filter_commitrw_write()) data1-OST0001: slow direct_io 82s Lustre: 5809:0:(filter_io_26.c:776:filter_commitrw_write()) data1-OST0001: slow commitrw...
2007 Mar 20
15
How to bypass failed OST without blocking?
Hi I want my lustre do such things during OST failed: if some file has stripe data on th failed OST, any operation on the file will return IO error without blocking, and also at this moment I can create and read/write new file or read/write files which have no stripe data on the failed OST without blocking. What should I do ? How to configure? thanks! swin -------------- next part
2007 Nov 23
2
How to remove OST permanently?
All, I''ve added a new 2.2 TB OST to my cluster easily enough, but this new disk array is meant to replace several smaller OSTs that I used to have of which were only 120 GB, 500 GB, and 700 GB. Adding an OST is easy, but how do I REMOVE the small OSTs that I no longer want to be part of my cluster? Is there a command to tell luster to move all the file stripes off one of the nodes?
2007 Nov 07
9
How To change server recovery timeout
...server to wait before failing recovery. We performed that experiment on our test lustre installation with one OST. storage02 is our OSS [root at storage02 ~]# lctl dl 0 UP mgc MGC10.143.245.3 at tcp 31259d9b-e655-cdc4-c760-45d3df426d86 5 1 UP ost OSS OSS_uuid 3 2 UP obdfilter home-md-OST0001 home-md-OST0001_UUID 7 [root at storage02 ~]# lctl --device 2 set_timeout 600 set_timeout has been deprecated. Use conf_param instead. e.g. conf_param lustre-MDT0000 obd_timeout=50 usage: conf_param obd_timeout=<secs> run <command> after connecting to device <devno> --device <...
2010 Aug 06
1
Depreciated client still shown on OST exports
Some clients have been removed several weeks ago but are still listed in: ls -l /proc/fs/lustre/obdfilter/*/exports/ This was found after tracing back mystery tcp packets to the OSS. Although this is causing no damage, it raises the question of when former clients will be cleared from the OSS. Is there a way to manually remove these exports from the OSS? -- Regards, David
2008 Feb 14
9
how do you mount mountconf (i.e. 1.6) lustre on your servers?
As any of you using version 1.6 of Lustre knows, Lustre servers can now be started simply my mounting the devices it is using. Even an /etc/fstab entry can be used if you can have the mount delayed until the network is started. Given this change, you have also notices that we have eliminated the initscript for Lustre that used to exist for releases prior to 1.6. I''d like to take a
2013 Oct 17
3
Speeding up configuration log regeneration?
Hi, We run four-node Lustre 2.3, and I needed to both change hardware under MGS/MDS and reassign an OSS ip. Just the same, I added a brand new 10GE network to the system, which was the reason for MDS hardware change. I ran tunefs.lustre --writeconf as per chapter 14.4 in Lustre Manual, and everything mounts fine. Log regeneration apparently works, since it seems to do something, but
2013 Feb 12
2
Lost folders after changing MDS
...B> I deactivated those two OSTs below. [root at mgs data]# lfs df -h UUID bytes Used Available Use% Mounted on AC3-MDT0000_UUID 37.5G 499.5M 34.5G 1% /data[MDT:0] AC3-OST0000_UUID 16.4T 2.2T 13.3T 14% /data[OST:0] AC3-OST0001_UUID 16.4T 1.8T 13.7T 12% /data[OST:1] AC3-OST0002_UUID 6.4T 6.0T 49.2G 99% /data[OST:2] AC3-OST0003_UUID 6.4T 6.1T 912.9M 100% /data[OST:3] AC3-OST0004_UUID 4.3T 4.1T 17.2G 100% /data[OST:4] AC3-OST0005...
2007 Nov 06
4
Checksum Algorithm
Hi, We have seen a huge performance drop in 1.6.3, due to the checksum being enabled by default. I looked at the algorithm being used, and it is actually a CRC32, which is a very strong algorithm for detecting all sorts of problems, such as single bit errors, swapped bytes, and missing bytes. I''ve been experimenting with using a simple XOR algorithm. I''ve been able to recover