similar to: Lost folders after changing MDS

Displaying 20 results from an estimated 400 matches similar to: "Lost folders after changing MDS"

2007 Nov 07
9
How To change server recovery timeout
Hi, Our lustre environment is: 2.6.9-55.0.9.EL_lustre.1.6.3smp I would like to change recovery timeout from default value 250s to something longer I tried example from manual: set_timeout <secs> Sets the timeout (obd_timeout) for a server to wait before failing recovery. We performed that experiment on our test lustre installation with one OST. storage02 is our OSS [root at
2008 Jan 10
4
1.6.4.1 - active client evicted
Hi! We''ve started to poke and prod at Lustre 1.6.4.1, and it seems to mostly work (we haven''t had it OOPS on us yet like the earlier 1.6-versions did). However, we had this weird incident where an active client (it was copying 4GB files and running ls at the time) got evicted by the MDS and all OST''s. After a while logs indicate that it did recover the connection
2013 Oct 17
3
Speeding up configuration log regeneration?
Hi, We run four-node Lustre 2.3, and I needed to both change hardware under MGS/MDS and reassign an OSS ip. Just the same, I added a brand new 10GE network to the system, which was the reason for MDS hardware change. I ran tunefs.lustre --writeconf as per chapter 14.4 in Lustre Manual, and everything mounts fine. Log regeneration apparently works, since it seems to do something, but
2007 Nov 23
2
How to remove OST permanently?
All, I''ve added a new 2.2 TB OST to my cluster easily enough, but this new disk array is meant to replace several smaller OSTs that I used to have of which were only 120 GB, 500 GB, and 700 GB. Adding an OST is easy, but how do I REMOVE the small OSTs that I no longer want to be part of my cluster? Is there a command to tell luster to move all the file stripes off one of the nodes?
2008 Feb 12
0
Lustre-discuss Digest, Vol 25, Issue 17
Hi, i just want to know whether there are any alternative file systems for HP SFS. I heard that there is Cluster Gateway from Polyserve. Can anybody plz help me in finding more abt this Cluster Gateway. Thanks and Regards, Ashok Bharat -----Original Message----- From: lustre-discuss-bounces at lists.lustre.org on behalf of lustre-discuss-request at lists.lustre.org Sent: Tue 2/12/2008 3:18 AM
2006 Dec 09
2
Cannot create writeable share - tried many different things
Hi, I appreciate any help that can be given. For the life of me, I cannot create a writeable share on a specific box - a fairly new install of Fedora Core 5. I have posted on several forums over the past 2 weeks, but haven't gotten anywhere with the advice given, so I'm turning to this list as my last hope. This box runs as an NIS client, but I have had no problems with NIS. The problem
2010 Aug 11
3
lfs --obd discrepancy to lctl dl (1.8.3)
Hello, lfs prints different obd(idx) compared to lctl dl. We use single striping. cluster1 tmp # lfs find --obd scia-OST0017_UUID /data/scia/L0/V0.00/20100327/SCI_NL__0PNPDE20100327_193441_000040582088_00071_42209_1158.N1 /data/scia/L0/V0.00/20100327/SCI_NL__0PNPDE20100327_193441_000040582088_00071_42209_1158.N1 cluster1 tmp # lfs getstripe
2010 May 19
0
Hi - Regarding xend - xm create Error
Hi, I am able to get Xend running on Fedora 8 but when I try to create any VM it fails. The xm create osc.cfg throws following error - [root@asds174 vm_images]# xm create osc.cfg Modified arg=osc.cfg stop3 Using config file "./osc.cfg". group_id=None [''create'', ''osc.cfg''] Stop 1 None [''vm'', [''name'',
2013 Dec 17
2
Setting up a lustre zfs dual mgs/mdt over tcp - help requested
Hi all, Here is the situation: I have 2 nodes MDS1 , MDS2 (10.0.0.22 , 10.0.0.23) I wish to use as failover MGS, active/active MDT with zfs. I have a jbod shelf with 12 disks, seen by both nodes as das (the shelf has 2 sas ports, connected to a sas hba on each node), and I am using lustre 2.4 on centos 6.4 x64 I have created 3 zfs pools: 1. mgs: # zpool
2007 Dec 11
2
lustre + nfs + alphas
This is the strangest problem I have seen. I have a lustre filesystem mounted on a linux server and its being exported to various alpha systems. The alphas mount it just fine however under heavy load the NFS server stops responding, as does the lustre mount on the export server. The weird thing is that if i mount the nfs export on another nfs server and run the same benchmark (bonnie) everything
2013 Sep 15
0
Lustre 2.4 MDT: LustreError: Communicating with 0@lo: operation mds_connect failed with -11
I''m a Lustre newbie who just joined this list. I''d appreciate any help on the following Lustre 2.4 issue I''m running into: Every time I mount the MDT, the mount appears to succeed but /var/log/messages contains the message: "LustreError: 11-0: lustre-MDT0000-lwp-MDT0000: Communicating with 0@lo, operation mds_connect failed with -11". The MDT uses 4 local
2014 Nov 13
0
OST acting up
whoops, sent from wrong email address, form right address now: Hello, I am using Lustre 2.4.2 and have an OST that doesn't seem to be written to. When I check the MDS with 'lctl dl' I do not see that OST in the list. However when I check the OSS that OST belongs to I can see it is mounted and up; 0 UP osd-zfs l2-OST0003-osd l2-OST0003-osd_UUID 5 3 UP obdfilter l2-OST0003
2008 Sep 05
1
Rmpi on Linux (SuSE) with OpenMPI
Guys, I'm having troubles with Rmpi on Linux. I got it to work on windows, but on Linux (SuSE 64 bit), using OpenMPI, I can't seem to get it to work :( When I do library(Rmpi) I get the warning: mca: base: component_find: unable to open osc pt2pt: file not found (ignored) Interestingly enough all the examples from the snow tutorial ( http://www.sfu.ca/~sblay/R/snow.html) work despite
2008 Mar 07
2
Multihomed question: want Lustre over IB andEthernet
Chris, Perhaps you need to perform some write_conf like command. I''m not sure if this is needed in 1.6 or not. Shane ----- Original Message ----- From: lustre-discuss-bounces at lists.lustre.org <lustre-discuss-bounces at lists.lustre.org> To: lustre-discuss <lustre-discuss at lists.lustre.org> Sent: Fri Mar 07 12:03:17 2008 Subject: Re: [Lustre-discuss] Multihomed
2006 Jul 14
1
(Solved) Re: How to configure omega.cgi to search multiple flint dbs?
Folks, OK found it, omega.cc looks for multiple DB definitions and adds it to the search list. If the dbs are only one directory lower (e.g. not in default as in beta/default, but only "beta"), then modifying database_dir setting as follows will work: database_dir /svr/hda1/xapian then, ./omega.cgi 'P=pda' FMT='xml' DB="beta" DB="gamma"
2012 Dec 28
6
problem with installing lustre and ofed
Hello, I am having trouble installing the server modules for lustre 2.1.4 and use mellanox''s OFED distribution so we may use infiniband. Would you folks look at my procedure and results below and let me know what you think? Thanks very much! The mellanox ofed installation builds and installs some kernel modules too, so I used this method to ensure OFED compiled against the correct
2008 Jan 10
2
How to PXE Boot
Folks, Are there updated instructions on how to set up PXE boot under CentOS 5.x? I found that the system-config-netboot RPM has been absent, and the instructions written reference it, with a bug report not yet closed on Red Hat. Thanks, OSC http://www.centos.org/docs/5/html/Installation_Guide-en-US/s1-netboot-pxe-config.html http://rhn.redhat.com/errata/RHBA-2006-0380.html -- Are we headed
2011 Sep 19
1
new client : DeeFuzzer
DeeFuzzer is an icecast2 source client fully written in python and based on shout-python. Features : * MP3 and OGG (audio & video) file streaming over internet (Icecast) * Full metadata encapsulation and management * RSS podcast generator (current tracks and playlists) * M3U playlist generator * Recursive, random (shuffled) or pre-defined playlists * Multi-threaded architecture : multiple
2010 Sep 18
0
no failover with failover MDS
Hi all, we have two servers A, B as a failover MGS/MDT pair, with IPs A=10.12.112.28 and B=10.12.115.120 over tcp. When server B crashes, MGS and MDT are mounted on A. Recovery times out with only one out of 445 clients recovered. Afterwards, the MDT lists all its OSTs as UP and in the logs of the OSTs I see: Lustre: MGC10.12.112.28 at tcp: Connection restored to service MGS using nid
2006 Jul 19
1
Q: T2000: raidctl vs. zpool status
Hi all, IHACWHAC (I have a colleague who has a customer - hello, if you''re listening :-) who''s trying to build and test a scenario where he can salvage the data off the (internal ?) disks of a T2000 in case the sysboard and with it the on-board raid controller dies. If I understood correctly, he replaces the motherboard, does some magic to get the raid config back, but even