Displaying 20 results from an estimated 200 matches similar to: "The mds_connect operation failed with -11"
2008 Mar 07
2
Multihomed question: want Lustre over IB andEthernet
Chris,
Perhaps you need to perform some write_conf like command. I''m not sure if this is needed in 1.6 or not.
Shane
----- Original Message -----
From: lustre-discuss-bounces at lists.lustre.org <lustre-discuss-bounces at lists.lustre.org>
To: lustre-discuss <lustre-discuss at lists.lustre.org>
Sent: Fri Mar 07 12:03:17 2008
Subject: Re: [Lustre-discuss] Multihomed
2013 Sep 15
0
Lustre 2.4 MDT: LustreError: Communicating with 0@lo: operation mds_connect failed with -11
I''m a Lustre newbie who just joined this list. I''d appreciate any help on
the following Lustre 2.4 issue I''m running into:
Every time I mount the MDT, the mount appears to succeed but
/var/log/messages contains the message: "LustreError: 11-0:
lustre-MDT0000-lwp-MDT0000: Communicating with 0@lo, operation mds_connect
failed with -11". The MDT uses 4 local
2008 Feb 22
0
lustre error
Dear All,
Yesterday evening or cluster has stopped.
Two of our nodes tried to take the resource from each other, they
haven''t seen the other side, if I saw well.
I stopped heartbeat, resources, start it again, and back to online,
worked fine.
This morning I saw this in logs:
Feb 22 03:25:07 node4 kernel: Lustre:
7:0:(linux-debug.c:98:libcfs_run_upcall()) Invoked LNET upcall
2007 Sep 28
0
llog_origin_handle_cancel and other LustreErrors
Hi again!
Same setup as before (Lustre 1.6.2 + 2.6.18 kernel).
This time things suddenly started to be very slow (as in periodically
stalling), and we found a bunch of llog_ LustreErrors on the MDS. Some
time later stuff had automagically recovered and is back to normal
speed.
Any idea on the meaning/cause of these errors?
What are the seriousness of "LustreError" errors in
2010 Aug 11
0
OSS: IMP_CLOSED errors
Hello.
OS CentOS 5.4
uname -a
Linux oss0 2.6.18-128.7.1.el5_lustre.1.8.1.1 #1 SMP Tue Oct 6 05:48:57 MDT 2009 x86_64 x86_64 x86_64 GNU/Linux
Lustre 1.8.1.1
OSS server.
A lot of errors in /var/log/messages:
Aug 10 14:46:34 oss0 kernel: LustreError: 2802:0:(client.c:775:ptlrpc_import_delay_req()) Skipped 1 previous similar message
Aug 10 15:07:01 oss0 kernel: LustreError:
2008 Jan 10
4
1.6.4.1 - active client evicted
Hi!
We''ve started to poke and prod at Lustre 1.6.4.1, and it seems to
mostly work (we haven''t had it OOPS on us yet like the earlier
1.6-versions did).
However, we had this weird incident where an active client (it was
copying 4GB files and running ls at the time) got evicted by the MDS
and all OST''s. After a while logs indicate that it did recover the
connection
2007 Nov 07
1
ll_cfg_requeue process timeouts
Hi,
Our environment is: 2.6.9-55.0.9.EL_lustre.1.6.3smp
I am getting following errors from two OSS''s
...
Nov 7 10:39:51 storage09.beowulf.cluster kernel: LustreError:
23045:0:(client.c:519:ptlrpc_import_delay_req()) @@@ IMP_INVALID
req at 00000100b410be00 x4190687/t0 o101->MGS at MGC10.143.245.201@tcp_0:26
lens 232/240 ref 1 fl Rpc:/0/0 rc 0/0
Nov 7 10:39:51
2007 Oct 25
1
Error message
I''m seeing this error message on one of my OSS''s but not the other
three. Any idea what is causing it?
Oct 25 13:58:56 oss2 kernel: LustreError:
3228:0:(client.c:519:ptlrpc_import_delay_req()) @@@ IMP_INVALID
req at f6b13200 x18040/t0 o101->MGS at MGC192.168.0.200@tcp_0:26 lens 176/184
ref 1 fl Rpc:/0/0 rc 0/0
Oct 25 13:58:56 oss2 kernel: LustreError:
2013 Apr 29
1
OSTs inactive on one client (only)
Hi everyone,
I have seen this question here before, but without a very
satisfactory answer. One of our half a dozen clients has
lost access to a set of OSTs:
> lfs osts
OBDS::
0: lustre-OST0000_UUID ACTIVE
1: lustre-OST0001_UUID ACTIVE
2: lustre-OST0002_UUID INACTIVE
3: lustre-OST0003_UUID INACTIVE
4: lustre-OST0004_UUID INACTIVE
5: lustre-OST0005_UUID ACTIVE
6: lustre-OST0006_UUID ACTIVE
2008 Mar 06
2
strange lustre errors
Hi,
On a few of the hpc cluster nodes, i am seeing a new lustre
error that is pasted below. The volumes are working fine and there
is nothing on the oss and mds to report.
LustreError: 5080:0:(import.c:607:ptlrpc_connect_interpret())
data3-OST0000_UUID at 192.168.2.98@tcp changed handle from
0xfe51139158c64fae to 0xfe511392a35878b3; copying, but this may
foreshadow disaster
2008 Feb 12
0
Lustre-discuss Digest, Vol 25, Issue 17
Hi,
i just want to know whether there are any alternative file systems for HP SFS.
I heard that there is Cluster Gateway from Polyserve. Can anybody plz help me in finding more abt this Cluster Gateway.
Thanks and Regards,
Ashok Bharat
-----Original Message-----
From: lustre-discuss-bounces at lists.lustre.org on behalf of lustre-discuss-request at lists.lustre.org
Sent: Tue 2/12/2008 3:18 AM
2010 Sep 16
2
Lustre module not getting loaded in MDS
Hello All,
I have installed and configured Lustre 1.8.4 on SuSe 11.0 and everything
works fine if i run modprobe lustre and when the lustre module is getting
loaded. But when the server reboots it is not getting loaded. Kindly help.
Lnet is configured in /etc/modprobe.conf.local as below.
options lnet networks=tcp0(eth0) accept=all
For loading lustre module i tried including lustre module in
2008 Jan 31
2
lustre+samba
Dear All,
I try to use our cluster though samba share. Everything work fine, but
I think, we should have -o flock at lustre mount time.
Great, it''s work. But when I want to save a file on the share, I get
this on the logs:
Jan 31 10:45:24 opteron-ren-11 kernel: LustreError: 24836:0:(file.c:2309:ll_file_flock()) unknown fcntl lock type: 32
Jan 31 10:45:24 opteron-ren-11 kernel:
2005 Oct 05
1
how can i let the user in 1th Asterisk can call the user in 2nd Asterisk?
Hi list,
I set up two asterisk servers , 1001 is the first asterisk server's sip
user, and 2001 is the second asterisk server's sip user. Each of them work
well, but I don't konw how to connect them. I want to let the user in 1th
Asterisk can call the user in 2nd Asterisk.
First asterisk server ip : 192.168.3.101
Second asterisk server ip : 192.168.3.102
can someone
2013 Dec 17
2
Setting up a lustre zfs dual mgs/mdt over tcp - help requested
Hi all,
Here is the situation:
I have 2 nodes MDS1 , MDS2 (10.0.0.22 , 10.0.0.23) I wish to use as
failover MGS, active/active MDT with zfs.
I have a jbod shelf with 12 disks, seen by both nodes as das (the
shelf has 2 sas ports, connected to a sas hba on each node), and I
am using lustre 2.4 on centos 6.4 x64
I have created 3 zfs pools:
1. mgs:
# zpool
2007 Nov 07
9
How To change server recovery timeout
Hi,
Our lustre environment is:
2.6.9-55.0.9.EL_lustre.1.6.3smp
I would like to change recovery timeout from default value 250s to
something longer
I tried example from manual:
set_timeout <secs> Sets the timeout (obd_timeout) for a server
to wait before failing recovery.
We performed that experiment on our test lustre installation with one
OST.
storage02 is our OSS
[root at
2010 Sep 04
0
Set quota on Lustre system file client, reboots MDS/MGS node
Hi
I used lustre-1.8.3 for Centos5.4. I patched the kernel according to Lustre
1.8 operations manual.pdf.
I have a problem when I want to implement quota.
My cluster configuration is:
1. one MGS/MDS host (with two devices: sda and sdb,respectively)
with the following commands:
1) mkfs.lustre --mgs /dev/sda
2) mount -t lustre /dev/sda /mnt/mgt
3) mkfs.lustre --fsname=lustre
2008 Mar 03
1
Quota setup fails because of OST ordering
Hi all,
after installing a Lustre test file system consisting of 34 OSTs, I
encountered a strange error when trying to set up quotas:
lfs quotacheck gave me an "Input/Output error", while in
/var/log/kern.log I found a Lustre error
LustreError: 20807:0:(quota_check.c:227:lov_quota_check()) lov idx 32
inactive
Indeed, in /proc/fs/lustre/lov/.../target_obd all 34 OSTs were listed
2007 Dec 11
2
lustre + nfs + alphas
This is the strangest problem I have seen. I have a lustre filesystem mounted on a linux server and its being exported to various alpha systems. The alphas mount it just fine however under heavy load the NFS server stops responding, as does the lustre mount on the export server. The weird thing is that if i mount the nfs export on another nfs server and run the same benchmark (bonnie) everything
2008 Jan 15
19
How do you make an MGS/OSS listen on 2 NICs?
I am running on CentOS 5 distribution without adding any updates from CentOS. I am using the lustre 1.6.4.1 kernel and software.
I have two NICs that run though different switches.
I have the lustre options in my modprobe.conf to look like this:
options lnet networks=tcp0(eth1,eth0)
My MGS seems to be only listening on the first interface however.
When I try and ping the 1st interface (eth1)