ren yufei
2010-Nov-03 18:30 UTC
[Lustre-discuss] Lustre FS in Infiniband - client mount problem
Dear all, I have setup some nodes (include MDT/MGS/OSS/Client) with Mellanox 40G RNIC connected via Mellanox MTS3600 switch, and deployed lustre FS in this cluster. The MDT/MGS node and OSS nodes works but client could not mount on this FS. The error information is as follows. Client: 192.168.1.23 MDS: 192.168.1.11 -- error information Client # mount -t lustre 192.168.1.11 at o2ib0:/lustre /mnt/lustre mount.lustre: mount 192.168.1.11 at o2ib0:/lustre at /mnt/lustre failed: No such device Are the lustre modules loaded? Check /etc/modprobe.conf and /proc/filesystems Note ''alias lustre llite'' should be removed from modprobe.conf Client # ls -l /mnt/lustre total 0 Client # dmesg | tail LustreError: 165-2: Nothing registered for client mount! Is the ''lustre'' module loaded? LustreError: 5116:0:(obd_mount.c:2045:lustre_fill_super()) Unable to mount (-19) -- environment information. Client: # lctl list_nids 192.168.1.23 at o2ib # lctl ping 192.168.1.11 at o2ib0 12345-0 at lo 12345-192.168.1.11 at o2ib MDS: # lctl list_nids 192.168.1.11 at o2ib lctl > device_list 0 UP mgs MGS MGS 13 1 UP mgc MGC192.168.1.11 at o2ib bb9cf87d-fd14-b679-85ce-f0fa1a866aff 5 2 UP mdt MDS MDS_uuid 3 3 UP lov lustre-mdtlov lustre-mdtlov_UUID 4 4 UP mds lustre-MDT0000 lustre-MDT0000_UUID 3 5 UP osc lustre-OST0000-osc lustre-mdtlov_UUID 5 6 UP osc lustre-OST0001-osc lustre-mdtlov_UUID 5 ... By the way, All these nodes could connect or access to each other via ping/iperf(TCP)/ibv_rc_pingpong/ibv_ud_pingpong/ib_write_lat.However, the ''rping'' client, based on librdmacm, could not connect to the server side. error is: # rping -c 192.168.1.11 cq completion failed status 5 wait for CONNECTED state 10 connect error -1 cma event RDMA_CM_EVENT_REJECTED, error 8 Thank you very much. Yufei -------------- next part -------------- An HTML attachment was scrubbed... URL: http://lists.lustre.org/pipermail/lustre-discuss/attachments/20101104/a7ba1787/attachment.html
Wang Yibin
2010-Nov-04 02:03 UTC
[Lustre-discuss] Lustre FS in Infiniband - client mount problem
As the error message said, the lustre modules probably were not loaded when you were trying to mount lustre client. Please provide information - specifically, the loaded modules and more dmesg log. ? 2010-11-4???2:30? ren yufei ???> Dear all, > > I have setup some nodes (include MDT/MGS/OSS/Client) with Mellanox 40G RNIC connected via Mellanox MTS3600 switch, and deployed lustre FS in this cluster. The MDT/MGS node and OSS nodes works but client could not mount on this FS. The error information is as follows. > > Client: 192.168.1.23 > MDS: 192.168.1.11 > > -- error information > Client # mount -t lustre 192.168.1.11 at o2ib0:/lustre /mnt/lustre > mount.lustre: mount 192.168.1.11 at o2ib0:/lustre at /mnt/lustre failed: No such device > Are the lustre modules loaded? > Check /etc/modprobe.conf and /proc/filesystems > Note ''alias lustre llite'' should be removed from modprobe.conf > > Client # ls -l /mnt/lustre > total 0 > > Client # dmesg | tail > LustreError: 165-2: Nothing registered for client mount! Is the ''lustre'' module loaded? > LustreError: 5116:0:(obd_mount.c:2045:lustre_fill_super()) Unable to mount (-19) > > -- environment information. > > Client: > # lctl list_nids > 192.168.1.23 at o2ib > # lctl ping 192.168.1.11 at o2ib0 > 12345-0 at lo > 12345-192.168.1.11 at o2ib > > MDS: > # lctl list_nids > 192.168.1.11 at o2ib > > lctl > device_list > 0 UP mgs MGS MGS 13 > 1 UP mgc MGC192.168.1.11 at o2ib bb9cf87d-fd14-b679-85ce-f0fa1a866aff 5 > 2 UP mdt MDS MDS_uuid 3 > 3 UP lov lustre-mdtlov lustre-mdtlov_UUID 4 > 4 UP mds lustre-MDT0000 lustre-MDT0000_UUID 3 > 5 UP osc lustre-OST0000-osc lustre-mdtlov_UUID 5 > 6 UP osc lustre-OST0001-osc lustre-mdtlov_UUID 5 > ... > > By the way, All these nodes could connect or access to each other via ping/iperf(TCP)/ibv_rc_pingpong/ibv_ud_pingpong/ib_write_lat. However, the ''rping'' client, based on librdmacm, could not connect to the server side. error is: > > # rping -c 192.168.1.11 > cq completion failed status 5 > wait for CONNECTED state 10 > connect error -1 > cma event RDMA_CM_EVENT_REJECTED, error 8 > > > Thank you very much. > > Yufei > > > _______________________________________________ > Lustre-discuss mailing list > Lustre-discuss at lists.lustre.org > http://lists.lustre.org/mailman/listinfo/lustre-discuss-------------- next part -------------- An HTML attachment was scrubbed... URL: http://lists.lustre.org/pipermail/lustre-discuss/attachments/20101104/5f29b181/attachment-0001.html
rishi pathak
2010-Nov-04 06:12 UTC
[Lustre-discuss] Lustre FS in Infiniband - client mount problem
Try using mount -t lustre 192.168.1.11 at o2ib:/lustre /mnt/lustre On Thu, Nov 4, 2010 at 12:00 AM, ren yufei <renyufei83 at yahoo.com.cn> wrote:> Dear all, > > I have setup some nodes (include MDT/MGS/OSS/Client) with Mellanox 40G RNIC > connected via Mellanox MTS3600 switch, and deployed lustre FS in this > cluster. The MDT/MGS node and OSS nodes works but client could not mount > on this FS. The error information is as follows. > > Client: 192.168.1.23 > MDS: 192.168.1.11 > > -- error information > Client # mount -t lustre 192.168.1.11 at o2ib0:/lustre /mnt/lustre > mount.lustre: mount 192.168.1.11 at o2ib0:/lustre at /mnt/lustre failed: No > such device > Are the lustre modules loaded? > Check /etc/modprobe.conf and /proc/filesystems > Note ''alias lustre llite'' should be removed from modprobe.conf > > Client # ls -l /mnt/lustre > total 0 > > Client # dmesg | tail > LustreError: 165-2: Nothing registered for client mount! Is the ''lustre'' > module loaded? > LustreError: 5116:0:(obd_mount.c:2045:lustre_fill_super()) Unable to mount > (-19) > > -- environment information. > > Client: > # lctl list_nids > 192.168.1.23 at o2ib > # lctl ping 192.168.1.11 at o2ib0 > 12345-0 at lo > 12345-192.168.1.11 at o2ib > > MDS: > # lctl list_nids > 192.168.1.11 at o2ib > > lctl > device_list > 0 UP mgs MGS MGS 13 > 1 UP mgc MGC192.168.1.11 at o2ib bb9cf87d-fd14-b679-85ce-f0fa1a866aff 5 > 2 UP mdt MDS MDS_uuid 3 > 3 UP lov lustre-mdtlov lustre-mdtlov_UUID 4 > 4 UP mds lustre-MDT0000 lustre-MDT0000_UUID 3 > 5 UP osc lustre-OST0000-osc lustre-mdtlov_UUID 5 > 6 UP osc lustre-OST0001-osc lustre-mdtlov_UUID 5 > ... > > By the way, All these nodes could connect or access to each other via > ping/iperf(TCP)/ibv_rc_pingpong/ibv_ud_pingpong/ib_write_lat. However, *the > ''rping'' client, based on librdmacm, could not connect to the server side. > *error is: > > # rping -c 192.168.1.11 > cq completion failed status 5 > wait for CONNECTED state 10 > connect error -1 > cma event RDMA_CM_EVENT_REJECTED, error 8 > > > Thank you very much. > > Yufei > > > > _______________________________________________ > Lustre-discuss mailing list > Lustre-discuss at lists.lustre.org > http://lists.lustre.org/mailman/listinfo/lustre-discuss > >-- Regards-- Rishi Pathak National PARAM Supercomputing Facility Center for Development of Advanced Computing(C-DAC) Pune University Campus,Ganesh Khind Road Pune-Maharastra -------------- next part -------------- An HTML attachment was scrubbed... URL: http://lists.lustre.org/pipermail/lustre-discuss/attachments/20101104/ded812a4/attachment.html
ren yufei
2010-Nov-04 16:22 UTC
[Lustre-discuss] Lustre FS in Infiniband - client mount problem
Thank you. After the self compiled ''lustre-modules-1.8.3-2.6.18_164.11.1.el5_lustre.1.8.3_201011012110'' installed on the client side, the problem resolved. Yufei ________________________________ From: Wang Yibin <wang.yibin at oracle.com> To: ren yufei <renyufei83 at yahoo.com.cn> Cc: lustre-discuss at lists.lustre.org Sent: Wed, November 3, 2010 10:03:31 PM Subject: Re: [Lustre-discuss] Lustre FS in Infiniband - client mount problem As the error message said, the lustre modules probably were not loaded when you were trying to mount lustre client. Please provide information - specifically, the loaded modules and more dmesg log. ? 2010-11-4???2:30? ren yufei ??? Dear all,> >I have setup some nodes (include MDT/MGS/OSS/Client) with Mellanox 40G RNIC >connected via Mellanox MTS3600 switch, and deployed lustre FS in this cluster. >The MDT/MGS node and OSS nodes works but client could not mount on this FS. The >error information is as follows. > >Client: 192.168.1.23 >MDS: 192.168.1.11 > >-- error information >Client # mount -t lustre 192.168.1.11 at o2ib0:/lustre /mnt/lustre >mount.lustre: mount 192.168.1.11 at o2ib0:/lustre at /mnt/lustre failed: No such >device >Are the lustre modules loaded? >Check /etc/modprobe.conf and /proc/filesystems >Note ''alias lustre llite'' should be removed from modprobe.conf > >Client # ls -l /mnt/lustre >total 0 > >Client # dmesg | tail >LustreError: 165-2: Nothing registered for client mount! Is the ''lustre'' module >loaded? >LustreError: 5116:0:(obd_mount.c:2045:lustre_fill_super()) Unable to mount >(-19) > >-- environment information. > >Client: ># lctl list_nids >192.168.1.23 at o2ib ># lctl ping 192.168.1.11 at o2ib0 >12345-0 at lo >12345-192.168.1.11 at o2ib > >MDS: ># lctl list_nids >192.168.1.11 at o2ib > >lctl > device_list > 0 UP mgs MGS MGS 13 > 1 UP mgc MGC192.168.1.11 at o2ib bb9cf87d-fd14-b679-85ce-f0fa1a866aff 5 > 2 UP mdt MDS MDS_uuid 3 > 3 UP lov lustre-mdtlov lustre-mdtlov_UUID 4 > 4 UP mds lustre-MDT0000 lustre-MDT0000_UUID 3 > 5 UP osc lustre-OST0000-osc lustre-mdtlov_UUID 5 > 6 UP osc lustre-OST0001-osc lustre-mdtlov_UUID 5 >... > >By the way, All these nodes could connect or access to each other via >ping/iperf(TCP)/ibv_rc_ pingpong/ibv_ud_pingpong/ib_ write_lat. However, the >''rping'' client, based on librdmacm, could not connect to the server side. error >is: > ># rping -c 192.168.1.11 >cq completion failed status 5 >wait for CONNECTED state 10 >connect error -1 >cma event RDMA_CM_EVENT_REJECTED, error 8 > > >Thank you very much. > >Yufei > > > >_______________________________________________ >Lustre-discuss mailing list >Lustre-discuss at lists.lustre.org >http://lists.lustre.org/mailman/listinfo/lustre-discuss >-------------- next part -------------- An HTML attachment was scrubbed... URL: http://lists.lustre.org/pipermail/lustre-discuss/attachments/20101105/37a4661c/attachment-0001.html