similar to: [PATCH nbdkit] Add new filter for rate-limiting connections.

Displaying 20 results from an estimated 2000 matches similar to: "[PATCH nbdkit] Add new filter for rate-limiting connections."

2019 Jul 31
13
[nbdkit PATCH 0/8] fd leak safety
There's enough here to need a review; some of it probably needs backporting to stable-1.12. This probably breaks tests on Haiku or other platforms that have not been as on-the-ball about atomic CLOEXEC; feel free to report issues that arise, and I'll help come up with workarounds (even if we end up leaving a rare fd leak on less-capable systems). Meanwhile, I'm still working on my
2019 Mar 05
0
[PATCH nbdkit] Add new filter for rate-limiting connections.
--- filters/delay/nbdkit-delay-filter.pod | 4 +- filters/rate/nbdkit-rate-filter.pod | 84 +++++++++ configure.ac | 2 + filters/rate/bucket.h | 62 +++++++ filters/rate/bucket.c | 173 +++++++++++++++++++ filters/rate/rate.c | 235 ++++++++++++++++++++++++++ TODO | 9 +
2020 Aug 07
7
[nbdkit PATCH 0/3] Content differentiation during --tls=on
Patch 3 still needs tests added, but it is at least working from my simple command line tests. Eric Blake (3): server: Implement nbdkit_is_tls for use during .open server: Expose final thread_model to filter's .get_ready tlsdummy: New filter docs/nbdkit-filter.pod | 21 +- docs/nbdkit-plugin.pod | 34 ++- docs/nbdkit-tls.pod
2008 Oct 28
14
[PATCH 0/13] ocfs2: xattr bucket API
When the extended attribute namespace grows to a b-tree, the leaf clusters are organized by means of 'buckets'. Each bucket is 4K in size, regardless of blocksize. Thus, a bucket may be made of more than one block. fs/ocfs2/xattr.c has a nice little abstraction to wrap this, struct ocfs2_xattr_bucket. It contains a list of buffer_heads representing these blocks, and there is even an
2006 Oct 13
3
HTB has 2 bucket?
in HTB use 2 bucket for manage 2 rate??? first bucket -> keep token for sending with rate second bucket -> keep ctoken for sending with ceil rate Is it true?? may be i''m misunderstand about token/bucket thoery _______________________________________________ LARTC mailing list LARTC@mailman.ds9a.nl http://mailman.ds9a.nl/cgi-bin/mailman/listinfo/lartc
2005 Jun 20
6
Bucketting data
Hi, Am sure this is a trivial question but for some reason, haven't been able to figure it out. I want to bucket data in a vector, and then iterate over the buckets. Say the data set is: > cleandata[,4] [1] 26 26 26 26 26 26 26 26 26 26 26 26 61 61 61 61 61 61 61 61 61 61 61 89 89 89 89 89 89 89 180 180 180 180 362 544 544 544 [39] 544 544 544 544 544 544
2008 Oct 26
1
[PATCH 1/1] ocfs2/xattr: Proper hash collision handle in bucket division.v3
Modification from V2 to V3: Use a more pefect code suggested by Joel. Thank Joel for it. In ocfs2/xattr, we must make sure the xattrs which have the same hash value exist in the same bucket so that the search schema can work. But in the old implementation, when we want to extend a bucket, we just move half number of xattrs to the new bucket. This works in most cases, but if we are lucky enough we
2009 Mar 09
4
[PATCH] ocfs2: Use xs->bucket to set xattr value outside.
Tristan, could you please run your xattr test against it? xs->base used to be allocated a 4K size and all the contents in the bucket are copied to the it. So in ocfs2_xattr_bucket_set_value_outside, we are safe to use xs->base + offset. Now we use ocfs2_xattr_bucket to abstract xattr bucket and xs->base is initialized to the start of the bu_bhs[0]. So xs->base + offset will overflow
2007 May 08
3
Token Bucket Filter and Dropping
I am trying to create my own Token Bucket Filter. However, I have a problem with packet dropping. Scenario : I got two streams 20KB/s each. I got one bucket with rate 20KB/s I put both streams into this bucket. When buffer is full packets need to be dropped. The problem is that only every other packet needs to be dropped in this scenario. Streams are the same so queue looks like that : S1 |
2009 Apr 22
1
[PATCH 1/1] OCFS2: fasten dlm_lock_resource hash_table lookups
#backporting the 3 patches at http://kernel.us.oracle.com/~smushran/srini/ to 1.2. enlarge hash_table capacity to fasten hash_table lookups. Signed-off-by: Wengang wang <wen.gang.wang at oracle.com> -- diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c --- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c.orig 2009-04-22 11:00:37.000000000 +0800 +++
2012 Nov 17
4
survfit & number of variables != number of variable names
This works ok: > cox = coxph(surv ~ bucket*(today + accor + both) + activity, data = data) > fit = survfit(cox, newdata=data[1:100,]) but using strata leads to problems: > cox.s = coxph(surv ~ bucket*(today + accor + both) + strata(activity), > data = data) > fit.s = survfit(cox.s, newdata=data[1:100,]) Error in model.frame.default(data = data[1:100, ], formula = ~bucket + :
2015 Apr 01
2
[PATCH 8/9] qspinlock: Generic paravirt support
On Wed, Apr 01, 2015 at 03:58:58PM -0400, Waiman Long wrote: > On 04/01/2015 02:48 PM, Peter Zijlstra wrote: > I am sorry that I don't quite get what you mean here. My point is that in > the hashing step, a cpu will need to scan an empty bucket to put the lock > in. In the interim, an previously used bucket before the empty one may get > freed. In the lookup step for that lock,
2015 Apr 01
2
[PATCH 8/9] qspinlock: Generic paravirt support
On Wed, Apr 01, 2015 at 03:58:58PM -0400, Waiman Long wrote: > On 04/01/2015 02:48 PM, Peter Zijlstra wrote: > I am sorry that I don't quite get what you mean here. My point is that in > the hashing step, a cpu will need to scan an empty bucket to put the lock > in. In the interim, an previously used bucket before the empty one may get > freed. In the lookup step for that lock,
2012 Oct 01
5
s3 as mysql directory
Hello list, I am soliciting opinion here as opposed technical help with an idea I have. I've setup a bacula backup system on an AWS volume. Bacula stores a LOT of information in it's mysql database (in my setup, you can also use postgres or sqlite if you chose). Since I've started doing this I notice that the mysql data directory has swelled to over 700GB! That's quite a lot and
2010 Aug 08
2
paperclip save to disk and s3
I have a standard Paperclip setup that saves a file to my disk. In addition I would also like the file saved to my amazon s3 bucket. [code] after_save :copy_to_s3 def copy_to_s3 has_attached_file :photo, :storage => :s3, :s3_credentials => "#{RAILS_ROOT}/config/s3.yml", :styles => { :thumb => "100x100#", :small => "750x750>"
2016 Dec 02
3
Asterisk Can't start with the default configs
Hi, I compiled the asterisk 14.0.2 to my ARMv5 NAS, however I just have enough time to test it now. But with the default config (I only edited the http.conf), it won't start, but gives the following: Sorcery registered wizard 'bucket' Sorcery registered wizard 'bucket_file' Parsing /ffp/etc/asterisk/sorcery.conf Parsing '/ffp/etc/asterisk/sorcery.conf': Found Cannot
2012 Jun 25
4
[LLVMdev] AddressSanitizer+CMake unittest question
Context: I'm trying to implement support for ASan's unittest suite in CMake. This is ... quite challenging. I think I can get it to work with one significant caveat: it will require manual dependency management. None of the automatic header tracking. I think this is fine in some cases, and not so fine in other cases. Let me explain. It feels like these tests are really comprised of two
2006 Feb 09
1
[Bug 445] New: ipt_account reports: sleeping function called from invalid context at mm/slab.c:2063
https://bugzilla.netfilter.org/bugzilla/show_bug.cgi?id=445 Summary: ipt_account reports: sleeping function called from invalid context at mm/slab.c:2063 Product: netfilter/iptables Version: patch-o-matic-ng Platform: All OS/Version: other Status: NEW Severity: normal Priority: P2
2011 Aug 19
4
Using dtrace to follow a kernel linked list
I''m looking for an example of how one could write a dtrace probe that could follow something like a NULL terminated linked list. For example: struct list { struct list *next; void *data; }; struct list *list_root; Assuming I had "list_root" available, but I wanted to know how long the linked list is, how can I do this in a dtrace probe? If I had looping
2003 Nov 14
1
rate vs peakrate - TBF.
Hi, Can you please explain to me the difference between rate and peakrate, in conjucture to TBF. IMHO, rate : controls the no. of pkts flowing through the pipe, and peakrate : controls the no. of tokens flowing through the bucket. A funtion(?) of peakrate and mtu is rate. If this is true, why do I need to specify both peakrate and rate? Regards, suraj. __________________________________ Do