similar to: Samba files_struct: "bool modified" and "bool update_write_time_on_close"

Displaying 20 results from an estimated 1000 matches similar to: "Samba files_struct: "bool modified" and "bool update_write_time_on_close""

2008 Dec 23
0
[jra@samba.org: Patch to improve Samba write speeds on Linux ext3 with 3.2.x]
----- Forwarded message from Jeremy Allison <jra@samba.org> ----- From: Jeremy Allison <jra@samba.org> To: samba-technical@samba.org Cc: jra@samba.org Subject: Patch to improve Samba write speeds on Linux ext3 with 3.2.x Reply-To: Jeremy Allison <jra@samba.org> Hi all, Intel recently did some research on Samba write speeds, tested in Linux using ext3 resulting in the following
2003 Apr 27
4
Bogus rsync "Success" message when out of disk space
Patches welcome, eh, Paul? Upon further (belated) investigation, there are 2 affected places in receiver.c with this error message. Both call write_file(). And write_file is called only in those two places. So that is the appropriate location to patch. Especially since the obvious fix is to use the rewrite code already there for the sparse file writes.
2002 Dec 09
2
Rsync performance increase through buffering
I've been studying the read and write buffering in rsync and it turns out most I/O is done just a couple of bytes at a time. This means there are lots of system calls, and also most network traffic comprises lots of small packets. The behavior is most extreme when sending/receiving file deltas of identical files. The main case where I/O is buffered is writes from the server (when io
2003 Jan 03
0
[Fwd: Re: rsync windows -> unix still hanging :(]
Author of the message didn't include rsync@lists.samba.org in the reply, and I think this message is in topic. -------- Original Message -------- Subject: Re: rsync windows -> unix still hanging :( Date: Mon, 30 Dec 2002 16:10:32 -0800 From: Jim Kleckner <jek-cygwin@kleckner.net> To: Mike Rubel <mrubel@galcit.caltech.edu>, cygwin@cygwin.com Mike - Greger Cronquist and I
2008 Mar 23
1
[PATCH] allow to change the block size used to handle sparse files
In some filesystems, typically optimized for large I/O throughputs (like IBM GPFS, IBM SAN FS, or distributed filesystems in general) a lot of lseek() operations can strongly impact on performances. In this cases it can be helpful to enlarge the block size used to handle sparse files directly from a command line parameter. For example, using a sparse write size of 32KB, I've been able to
2004 Aug 02
4
reducing memmoves
Attached is a patch that makes window strides constant when files are walked with a constant block size. In these cases, it completely avoids all memmoves. In my simple local test of rsyncing 57MB of 10 local files, memmoved bytes went from 18MB to zero. I haven't tested this for a big variety of file cases. I think that this will always reduce the memmoves involved with walking a large
2004 Feb 17
0
VFS module programmieren
############################# In German : Hallo, gibt es da Jemanden der mir den folgenden Text vielleicht ein bischen besser erkl?ren kann? Habe dies aus dem Samba Developers Guide. Warum mach ich das nun? Ich habe ein Problem: Ich will das jedes mal wenn ein User versucht einen Ordner in der Freigabe zu verschieben oder umzubenenner, der nicht leer ist, eine Fehlermeldung auf dem Client
2004 Feb 09
1
[patch] Add `--link-by-hash' option.
This patch adds the --link-by-hash=DIR option, which hard links received files in a link farm arranged by MD4 file hash. The result is that the system will only store one copy of the unique contents of each file, regardless of the file's name. Anyone have an example of an MD4 collision so I can test that case? :) Patch Summary: -1 +1 Makefile.in -0 +304 hashlink.c (new)
2017 Mar 02
1
[PATCH] generator: Move some deprecated functions to actions_core_deprecated.ml.
Fixes commit 97773d2bbee8e28830fd689deef9e9f63ce0c18e. --- generator/actions_core.ml | 83 ------------------------------------ generator/actions_core_deprecated.ml | 83 ++++++++++++++++++++++++++++++++++++ 2 files changed, 83 insertions(+), 83 deletions(-) diff --git a/generator/actions_core.ml b/generator/actions_core.ml index e662827..ed89f74 100644 --- a/generator/actions_core.ml
2012 Mar 16
2
plotting border over map
I am using following codes to plot map library(sp) library(rgdal) library(maps) library(gplots) library(clim.pact) library(fields) source("/R/PlotGridded2DMap.R") source("/R/image.plot.fix.R") source("/R/image.plot.plt.fix.r") seasonal_plot<-function(input,lonll=-180,latll=-90,lonres=5.,latres=3.75,write_file=TRUE,The_title=NULL){ if(is.null(The_title)){
2004 Feb 23
0
[patch] Add `--link-by-hash' option (rev 4).
This patch adds the --link-by-hash=DIR option, which hard links received files in a link farm arranged by MD4 file hash. The result is that the system will only store one copy of the unique contents of each file, regardless of the file's name. (rev 4) * Updated for committed robust_rename() patch, other changes in CVS. (rev 3) * Don't link empty files. * Roll over to new file when
2004 Feb 17
0
[patch] Add `--link-by-hash' option (rev 3).
This patch adds the --link-by-hash=DIR option, which hard links received files in a link farm arranged by MD4 file hash. The result is that the system will only store one copy of the unique contents of each file, regardless of the file's name. (rev 3) * Don't link empty files. * Roll over to new file when filesystem maximum link count is reached. * If link fails for another reason, leave
2004 Feb 23
0
[patch] Add `--link-by-hash' option (rev 5).
This patch adds the --link-by-hash=DIR option, which hard links received files in a link farm arranged by MD4 file hash. The result is that the system will only store one copy of the unique contents of each file, regardless of the file's name. (rev 5) * Fixed silly logic error. (rev 4) * Updated for committed robust_rename() patch, other changes in CVS. (rev 3) * Don't link empty
2004 Feb 16
1
[patch] Add `--link-by-hash' option (rev 2).
This patch adds the --link-by-hash=DIR option, which hard links received files in a link farm arranged by MD4 file hash. The result is that the system will only store one copy of the unique contents of each file, regardless of the file's name. (rev 2) * This revision is actually against CVS HEAD (I didn't realize I was working from a stale rsync'd CVS). * Apply permissions after
2020 Sep 15
0
[PATCH 01/18] media/v4l2: remove V4L2-FLAG-MEMORY-NON-CONSISTENT flag
From: Sergey Senozhatsky <sergey.senozhatsky at gmail.com> The patch partially reverts some of the UAPI bits of the buffer cache management hints. Namely, the queue consistency (memory coherency) user-space hint because, as it turned out, the kernel implementation of this feature was misusing DMA_ATTR_NON_CONSISTENT. The patch revers both kernel and user space parts: removes the DMA
2003 May 20
0
patch for better handling of write failures (disk full)
I've been having problems trying to sync two small partitions (128MB) that may be near to full. If rsync gets a write error (such as is caused when you fill up a partition) during a sync without the use of "-T", it will stop with this error: rsync: writefd_unbuffered failed to write 4 bytes: phase "unknown": Broken pipe rsync error: error in rsync protocol data stream
2003 May 23
1
PATCH: better handling for write failures (disk full)
[I sent this the other day, but it never got approved for the list] I've been having problems trying to sync two small partitions (128MB) that are usually near being full. The rsync would fail with this cryptic error: rsync: writefd_unbuffered failed to write 4 bytes: phase "unknown": Broken pipe rsync error: error in rsync protocol data stream (code 12) at io.c(515) It ends up
2008 Nov 13
3
Creating a file and saving in public directory
Hello, I''m fairly new to Ruby on Rails. I have a model called Report and I''m trying to create a text file that is saved in a /public/report directory. I''ve had a look at the ruby api but I can''t seem to get this working. The current code I have is: #report.rb class Report < ActiveRecord::Base after_save :write_file def write_file if @file_data
2002 Aug 05
5
[patch] read-devices
Greetings, I'd like to propose a new option to rsync, which causes it to read device files as if they were regular files. This includes pipes, character devices and block devices (I'm not sure about sockets). The main motivation is cases where you need to synchronize a large amount of data that is not available as regular files, as in the following scenarios: * Keep a copy of a block
2003 Aug 24
1
readdir() and read() errors ignored
We've just been hit rather badly by a very nasty bug that can cause rsync to silently discard files or fill them with zeroes. It happens when e.g. opendir() succeeds but readdir() returns an error; rsync does not check for an error from readdir() and so simply ignores the error (along with any remaining files in the directory). No error is reported to the user, who will then happily