search for: mdt0000

Displaying 17 results from an estimated 17 matches for "mdt0000".

2013 Feb 12
2
Lost folders after changing MDS
...OGS OBJECTS/* on the new MDT partition. I also upgraded from 1.8.8 to 2. I managed to mount the Lustre filesystem and if I do lfs df -h, I get: NB> I deactivated those two OSTs below. [root at mgs data]# lfs df -h UUID bytes Used Available Use% Mounted on AC3-MDT0000_UUID 37.5G 499.5M 34.5G 1% /data[MDT:0] AC3-OST0000_UUID 16.4T 2.2T 13.3T 14% /data[OST:0] AC3-OST0001_UUID 16.4T 1.8T 13.7T 12% /data[OST:1] AC3-OST0002_UUID 6.4T 6.0T 49.2G 99% /data[OST:2] AC3-OST0003...
2008 Jan 10
4
1.6.4.1 - active client evicted
...nt: ----------------------------8<------------------------ Jan 10 12:40:38 LustreError: 11-0: an error occurred while communicating with 130.239.78.233 at tcp. The ldlm_enqueue operation failed with -107 Jan 10 12:40:38 LustreError: Skipped 1 previous similar message Jan 10 12:40:38 Lustre: hpfs-MDT0000-mdc-ffff8100016d2c00: Connection to service hpfs-MDT0000 via nid 130.239.78.233 at tcp was lost; in progress operations using this service will wait for recovery to complete. Jan 10 12:40:38 LustreError: 167-0: This client was evicted by hpfs-MDT0000; in progress operations using this service will...
2007 Nov 07
9
How To change server recovery timeout
...ctl dl 0 UP mgc MGC10.143.245.3 at tcp 31259d9b-e655-cdc4-c760-45d3df426d86 5 1 UP ost OSS OSS_uuid 3 2 UP obdfilter home-md-OST0001 home-md-OST0001_UUID 7 [root at storage02 ~]# lctl --device 2 set_timeout 600 set_timeout has been deprecated. Use conf_param instead. e.g. conf_param lustre-MDT0000 obd_timeout=50 usage: conf_param obd_timeout=<secs> run <command> after connecting to device <devno> --device <devno> <command [args ...]> [root at storage02 ~]# lctl --device 1 conf_param obd_timeout=600 No device found for name MGS: Invalid argument error: conf_par...
2013 Dec 17
2
Setting up a lustre zfs dual mgs/mdt over tcp - help requested
...s1@tcp0 --servicenode=mds2@tcp0 --param sys.timeout=5000 --backfstype=zfs --mgsnode=mds1@tcp0 --mgsnode=mds2@tcp0  lustre-mdt0/mdt0 warning: lustre-mdt0/mdt0: for Lustre 2.4 and later, the target index must be specified with --index    Permanent disk data: Target:     fs0:MDT0000 Index:      0 Lustre FS:  fs0 Mount type: zfs Flags:      0x1061               (MDT first_time update no_primnode ) Persistent mount opts: Parameters: failover.node=10.0.0.22@tcp failover.node=10.0.0.23@tcp sys.timeout=5000 mgsnode=10.0.0.22@tcp mgsnode=10.0....
2007 Dec 11
2
lustre + nfs + alphas
...g it. On the nfs export server i see these messages-- Lustre: 4224:0:(o2iblnd_cb.c:412:kiblnd_handle_rx()) PUT_NACK from 192.168.64.70 at o2ib LustreError: 4400:0:(client.c:969:ptlrpc_expire_one_request()) @@@ timeout (sent at 1197415542, 100s ago) req at ffff810827bfbc00 x38827/t0 o36->data-MDT0000_UUID at 192.168.64.70@o2ib:12 lens 14256/672 ref 1 fl Rpc:/0/0 rc 0/-22 Lustre: data-MDT0000-mdc-ffff81082d702000: Connection to service data-MDT0000 via nid 192.168.64.70 at o2ib was lost; in progress operations using this service will wait for recovery to complete. A trace of the hung nfs deamon...
2013 Oct 17
3
Speeding up configuration log regeneration?
Hi, We run four-node Lustre 2.3, and I needed to both change hardware under MGS/MDS and reassign an OSS ip. Just the same, I added a brand new 10GE network to the system, which was the reason for MDS hardware change. I ran tunefs.lustre --writeconf as per chapter 14.4 in Lustre Manual, and everything mounts fine. Log regeneration apparently works, since it seems to do something, but
2008 Mar 07
2
Multihomed question: want Lustre over IB andEthernet
...ID found for 36.121.255.201 at tcp LustreError: 11043:0:(client.c:58:ptlrpc_uuid_to_connection()) cannot find peer 36.121.255.201 at tcp! LustreError: 11043:0:(ldlm_lib.c:312:client_obd_setup()) can''t add initial connection LustreError: 11043:0:(obd_config.c:325:class_setup()) setup ddnlfs-MDT0000-mdc-0000010430934400 failed (-2) LustreError: 11043:0:(obd_config.c:1062:class_config_llog_handler()) Err -2 on cfg command: LustreError: 11141:0:(connection.c:142:ptlrpc_put_connection()) NULL connection Lustre: cmd=cf003 0:ddnlfs-MDT0000-mdc 1:ddnlfs-MDT0000_UUID 2:36.121.255.201 at tcp Lustr...
2008 Feb 04
32
Luster clients getting evicted
on our cluster that has been running lustre for about 1 month. I have 1 MDT/MGS and 1 OSS with 2 OST''s. Our cluster uses all Gige and has about 608 nodes 1854 cores. We have allot of jobs that die, and/or go into high IO wait, strace shows processes stuck in fstat(). The big problem is (i think) I would like some feedback on it that of these 608 nodes 209 of them have in dmesg
2013 Sep 15
0
Lustre 2.4 MDT: LustreError: Communicating with 0@lo: operation mds_connect failed with -11
I''m a Lustre newbie who just joined this list. I''d appreciate any help on the following Lustre 2.4 issue I''m running into: Every time I mount the MDT, the mount appears to succeed but /var/log/messages contains the message: "LustreError: 11-0: lustre-MDT0000-lwp-MDT0000: Communicating with 0@lo, operation mds_connect failed with -11". The MDT uses 4 local drives in a RAID10 configuration. Each OSS has their own RAID10 of 36 drives each. The OSS''s mount correctly without any errors. I''ve seen this error mentioned in countless...
2014 Nov 13
0
OST acting up
...eem to be written to. When I check the MDS with 'lctl dl' I do not see that OST in the list. However when I check the OSS that OST belongs to I can see it is mounted and up; 0 UP osd-zfs l2-OST0003-osd l2-OST0003-osd_UUID 5 3 UP obdfilter l2-OST0003 l2-OST0003_UUID 5 4 UP lwp l2-MDT0000-lwp-OST0003 l2-MDT0000-lwp-OST0003_UUID 5 Since it isn't written to (the MDS doesn't seem to know about it, I created a directory. The index of that OST is 3 so I did a "lfs setstripe -i 3 -c 1 /mnt/l2-lustre/test-37" to force stuff that is written in that directory to be wr...
2007 Nov 23
2
How to remove OST permanently?
All, I''ve added a new 2.2 TB OST to my cluster easily enough, but this new disk array is meant to replace several smaller OSTs that I used to have of which were only 120 GB, 500 GB, and 700 GB. Adding an OST is easy, but how do I REMOVE the small OSTs that I no longer want to be part of my cluster? Is there a command to tell luster to move all the file stripes off one of the nodes?
2010 Sep 18
0
no failover with failover MDS
...tre: Added LNI 10.12.68.195 at tcp [8/256/0/180] Lustre: Accept secure, port 988 Lustre: Lustre Client File System; http://www.lustre.org/ Lustre: MGC10.12.112.28 at tcp: Reactivating import Lustre: 14530:0:(client.c:1476:ptlrpc_expire_one_request()) @@@ Request x1347247522447397 sent from gsilust-MDT0000-mdc-ffff81033d489400 to NID 10.12.115.120 at tcp 5s ago has timed out (5s prior to deadline). req at ffff8103312da400 x1347247522447397/t0 o38->gsilust-MDT0000_UUID at 10.12.115.120@tcp:12/10 lens 368/584 e 0 to 1 dl 1284835365 ref 1 fl Rpc:N/0/0 rc 0/0 Obviously the clients stubbornly t...
2008 Feb 05
2
obdfilter/datafs-OST0000/recovery_status
.../sda1 datafs-MDTffff + e2label /dev/sda2 datafs-OSTffff + mount.lustre /dev/sda1 /mnt/data/mdt + mount.lustre /dev/sda2 /mnt/data/ost0 + dmesg -c + mount.lustre pool4 at tcp:/datafs /mnt/datafs + dmesg -c + umount /mnt/datafs + umount /mnt/data/ost0 + umount /mnt/data/mdt + e2label /dev/sda1 datafs-MDT0000 + e2label /dev/sda2 datafs-OST0000 + dmesg -c + mount.lustre /dev/sda1 /mnt/data/mdt + mount.lustre /dev/sda2 /mnt/data/ost0 + dmesg -c + cat /proc/fs/lustre/obdfilter/datafs-OST0000/recovery_status + egrep ''RECOVERING|time remaining'' status: RECOVERING time remaining: 250 + sleep...
2007 Nov 12
8
More failover issues
In 1.6.0, when creating a MDT, you could specify multiple --mgsnode options and it would failover between them. 1.6.3 only seems to take the last one and --mgsnode=192.168.1.252 at o2ib:192.168.1.253 at o2ib doesn''t seem to failover to the other node. Any ideas how to get around this? Robert Robert LeBlanc College of Life Sciences Computer Support Brigham Young University leblanc at
2008 Feb 14
9
how do you mount mountconf (i.e. 1.6) lustre on your servers?
As any of you using version 1.6 of Lustre knows, Lustre servers can now be started simply my mounting the devices it is using. Even an /etc/fstab entry can be used if you can have the mount delayed until the network is started. Given this change, you have also notices that we have eliminated the initscript for Lustre that used to exist for releases prior to 1.6. I''d like to take a
2008 Feb 12
0
Lustre-discuss Digest, Vol 25, Issue 17
...>>> lock callback timer expired: evicting cl >>>>>>>> ient >>>>>>>> 2faf3c9e-26fb-64b7-ca6c-7c5b09374e67 at NET_0x200000aa4008d_UUID >>>>>>>> nid 10.164.0.141 at tcp ns: mds-nobackup >>>>>>>> -MDT0000_UUID lock: 00000100476df240/0xbc269e05c512de3a lrc: >>>>>>>> 1/0,0 mode: CR/CR res: 11240142/324715850 bi >>>>>>>> ts 0x5 rrc: 2 type: IBT flags: 20 remote: 0x4e54bc800174cd08 >>>>>>>> expref: 372 pid 26925 >>>>&g...
2008 Jan 02
9
lustre quota problems
Hello, I''ve several problems with quota on our testcluster: When I set the quota for a person to a given value (e.g. the values which are provided in the operations manual), I''m able to write exact the amount which is set with setquota. But when I delete the files(file) I''m not able to use this space again. Here is what I''ve done in detail: lfs checkquota