similar to: Jeremy''s git tags

Displaying 20 results from an estimated 2000 matches similar to: "Jeremy''s git tags"

2010 Sep 03
1
Compiling lustre-client 2.0.0.1 on RHEL 4
Hi, I tried to compile lustre-client 2.0.0.1 on RHEL4 with kernel 2.6.9-89.0.28.EL-x86_64 and I got 3 errors and 1 warning during the compile. The compile is executed with -Werror option, and it fails in all 4 cases * Error: lustre_compat25.h CC [M] /usr/src/redhat/BUILD/lustre-2.0.0.1/lustre/fid/fid_handler.o In file included from
2007 Nov 26
15
bad 1.6.3 striped write performance
Hi, I''m seeing what can only be described as dismal striped write performance from lustre 1.6.3 clients :-/ 1.6.2 and 1.6.1 clients are fine. 1.6.4rc3 clients (from cvs a couple of days ago) are also terrible. the below shows that the OS (centos4.5/5) or fabric (gigE/IB) or lustre version on the servers doesn''t matter - the problem is with the 1.6.3 and 1.6.4rc3 client kernels
2013 Nov 26
2
Lustre 1.8 client on EL 6.5?
Hello, in preparation for CentOS 6.5 I was trying to build Lustre 1.8 client on CentOS 6.4 updated with 2.6.32-431 kernel. Seems like recent changes in fs.h (/usr/src/kernels/2.6.32-431.el6.x86_64/include/linux/fs.h) is causing problems. make[5]: Entering directory `/usr/src/kernels/2.6.32-431.el6.x86_64'' cc1: warnings being treated as errors
2007 Nov 16
5
Lustre Debug level
Hi, Lustre manual 1.6 v18 says that that in production lustre debug level should be set to fairly low. Manual also says that I can verify that level by running following commands: # sysctl portals.debug This gives ne following error error: ''portals.debug'' is an unknown key cat /proc/sys/lnet/debug gives output: ioctl neterror warning error emerg ha config console cat
2010 Aug 11
3
Failure when mounting Lustre
Hi, I get the following error when I try to mount lustre on the clients. Permanent disk data: Target: lustre-OSTffff Index: unassigned Lustre FS: lustre Mount type: ldiskfs Flags: 0x72 (OST needs_index first_time update ) Persistent mount opts: errors=remount-ro,extents,mballoc Parameters: mgsnode=164.107.119.231 at tcp sh: losetup: command not found mkfs.lustre: error 32512 on losetup:
2008 Feb 04
32
Luster clients getting evicted
on our cluster that has been running lustre for about 1 month. I have 1 MDT/MGS and 1 OSS with 2 OST''s. Our cluster uses all Gige and has about 608 nodes 1854 cores. We have allot of jobs that die, and/or go into high IO wait, strace shows processes stuck in fstat(). The big problem is (i think) I would like some feedback on it that of these 608 nodes 209 of them have in dmesg
2008 Mar 25
2
patchless kernel
Dear All, make[5]: Entering directory `/usr/src/kernels/2.6.23.15-80.fc7-x86_64'' /usr/src/redhat/BUILD/lustre-1.6.4.3/lustre/llite/lloop.c:142: warning: ''request_queue_t'' is deprecated /usr/src/redhat/BUILD/lustre-1.6.4.3/lustre/llite/lloop.c:273: warning: ''request_queue_t'' is deprecated /usr/src/redhat/BUILD/lustre-1.6.4.3/lustre/llite/lloop.c:312:
2016 Jun 30
17
[PATCH v2 00/12] gendisk: Generate uevent after attribute available
The race condition is noticed between disk_add() and disk attributes, on virtio-blk hotplug. Userspace listens to the KOBJ_ADD uevent generated in add_disk(). At that point we haven't created the serial attribute file, therefore depending on how fast udev reacts, the /dev/disk/by-id/ entry doesn't always get created. As pointed out by Christoph Hellwig in the specific fix [1], virtio-blk
2016 Jun 30
17
[PATCH v2 00/12] gendisk: Generate uevent after attribute available
The race condition is noticed between disk_add() and disk attributes, on virtio-blk hotplug. Userspace listens to the KOBJ_ADD uevent generated in add_disk(). At that point we haven't created the serial attribute file, therefore depending on how fast udev reacts, the /dev/disk/by-id/ entry doesn't always get created. As pointed out by Christoph Hellwig in the specific fix [1], virtio-blk
2010 Aug 11
3
Version mismatch of Lustre client and server
Hello, I am planning on deploying a few more clients in my lustre environment and was wondering which client version to install. I know it is okay to run a newer client version than your lustre server for upgrade purposes. However, would it be okay to be in this state for a longer period of time (for the life of this filesystem)? My lustre server is currently running 1.8.1.1 on RHEL 5.3 and I
2010 Apr 30
3
Announce: Lustre 1.8.3 is available!
Hi all, Lustre 1.8.3 is available on the Sun Download Center Site. http://www.sun.com/software/products/lustre/get.jsp Our forwarding link has not yet been updated but should be shortly, so you can either find your way in through: http://www.sun.com/downloads/index.jsp?tab=2#L and look for Lustre 1.8.3 near the bottom of the "L" section. or
2010 Apr 30
3
Announce: Lustre 1.8.3 is available!
Hi all, Lustre 1.8.3 is available on the Sun Download Center Site. http://www.sun.com/software/products/lustre/get.jsp Our forwarding link has not yet been updated but should be shortly, so you can either find your way in through: http://www.sun.com/downloads/index.jsp?tab=2#L and look for Lustre 1.8.3 near the bottom of the "L" section. or
2010 Jul 13
4
Enable async journals
Hi all, we use SLES 11 and Lustre 1.8.1.1 + patches and like convert a lustre FS using external journals to one with async journals enabled. Question is whether the procedure: umount <filesystem> on all clients umount <osts> on all OSSes e2fsck <ost-device> on all OSSes for all all OSTs tune2fs -O ^has_journal <ost-device> on all
2010 Apr 19
20
Lustre Client - Memory Issue
Hi Guys, My users are reporting some issues with memory on our lustre 1.8.1 clients. It looks like when they submit a single job at a time the run time was about 4.5 minutes. However, when they ran multiple jobs (10 or less) on a client with 192GB of memory on a single node the run time for each job was exceeding 3-4X the run time for the single process. They also noticed that the swap space
2010 Sep 10
11
Large directory performance
We have been struggling with our Lustre performance for some time now especially with large directories. I recently did some informal benchmarking (on a live system so I know results are not scientifically valid) and noticed a huge drop in performance of reads(stat operations) past 20k files in a single directory. I''m using bonnie++, disabling IO testing (-s 0) and just creating, reading,
2013 Mar 18
1
OST0006 : inactive device
I installed 1 MDS , 2 OSS/OST and 2 Lustre Client. My MDS shows: [code] [root at MDS ~]# lctl list_nids 10.94.214.185 at tcp [root at MDS ~]# [/code] On Lustre Client1: [code] [root at lustreclient1 lustre]# lfs df -h UUID bytes Used Available Use% Mounted on lustre-MDT0000_UUID 4.5G 274.3M 3.9G 6% /mnt/lustre[MDT:0] lustre-OST0000_UUID
2008 Jun 12
13
Announce: Lustre 1.6.5 is available!
Hi all, At long last, Lustre 1.6.5 is available on the Sun Download Center Site. http://www.sun.com/software/products/lustre/get.jsp The change log and release notes can be read here: http://wiki.lustre.org/index.php?title=Change_Log_1.6 Thank you for your assistance; as always, you can report issues via Bugzilla (https://bugzilla.lustre.org/) Happy downloading! -- The Lustre Team --
2008 Jun 12
13
Announce: Lustre 1.6.5 is available!
Hi all, At long last, Lustre 1.6.5 is available on the Sun Download Center Site. http://www.sun.com/software/products/lustre/get.jsp The change log and release notes can be read here: http://wiki.lustre.org/index.php?title=Change_Log_1.6 Thank you for your assistance; as always, you can report issues via Bugzilla (https://bugzilla.lustre.org/) Happy downloading! -- The Lustre Team --
2013 Mar 18
1
lustre showing inactive devices
I installed 1 MDS , 2 OSS/OST and 2 Lustre Client. My MDS shows: [code] [root at MDS ~]# lctl list_nids 10.94.214.185 at tcp [root at MDS ~]# [/code] On Lustre Client1: [code] [root at lustreclient1 lustre]# lfs df -h UUID bytes Used Available Use% Mounted on lustre-MDT0000_UUID 4.5G 274.3M 3.9G 6% /mnt/lustre[MDT:0]
2010 Jul 08
5
No space left on device on not full filesystem
Hello, We have running lustre 1.8.1 and have met "No space lest on device" error when uploading 500 Gb small files (less then 100 Kb each). The problem seems to depends on the number of files. If we remove one file, we can create one new file, even with Gb size; but if we haven''t remove something we can''t create even very little file, as an example using touch