Displaying 9 results from an estimated 9 matches similar to: "[PATCH] blktap: ensure vma->vm_mm''s mmap_sem is being held whenever it is being modified"
2013 Sep 17
0
Bug#717157: blktap-dkms: Fails to build against Linux 3.10
Package: blktap-dkms
Followup-For: Bug #717157
I picked the commits from upstream. You should consider pushing it to
the archive.
root at ibmx3650m2-210-147:/usr/src/blktap-2.0.91# diff -Naru ring.c
/tmp/ring.c
--- ring.c 2013-09-17 11:47:27.468200000 -0400
+++ /tmp/ring.c 2013-09-17 11:47:56.836200000 -0400
@@ -210,7 +210,7 @@
pgoff = 1 + request->usr_idx *
2005 Nov 06
2
Bug in use of grant tables in blkback.c error path?
In dispatch_rw_block_io after a call to HYPERVISOR_grant_table_op, there
is the following code which calls fast_flush_area and breaks out of the
loop early if one of the handles returned from HYPERVISOR_grant_table_op
is negative:
for (i = 0; i < nseg; i++) {
if (unlikely(map[i].handle < 0)) {
DPRINTK("invalid buffer -- could not remap it\n");
fast_flush_area(pending_idx,
2010 Sep 15
15
xenpaging fixes for kernel and hypervisor
Patrick,
there following patches fix xenpaging for me.
Granttable handling is incomplete. If a page is gone, a GNTST_eagain
should be returned to the caller to inidcate the hypercall has to be
retried after a while, until the page is available again.
Please review.
Olaf
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xensource.com
2005 Nov 07
1
[PATCH] fast_flush_area in blkback.c still broken after 55194bd55b86
I think you need this patch otherwise you will be passing uninitialised
data from the stack to GNTTABOP_unmap_grant_ref in fast_flush_area.
Also, I think the memset of the pending_grant_handles array should
happen before you start using them so I moved it to before
blkif_xenbus_init. I''m not sure if this is necessary or it might
possibly have to go even earlier.
Signed off by Harry
2011 Apr 04
0
[PATCH] linux-2.6.18/backends: use xenbus_be.ko interfaces instead of open-coding them
Also remove unused xenbus_{,un}map_ring(), adjust types, and clean up
header inclusion.
Signed-off-by: Jan Beulich <jbeulich@novell.com>
--- a/drivers/xen/blkback/blkback.c
+++ b/drivers/xen/blkback/blkback.c
@@ -41,6 +41,8 @@
#include <linux/list.h>
#include <linux/delay.h>
#include <xen/balloon.h>
+#include <xen/evtchn.h>
+#include <xen/gnttab.h>
2011 Jun 21
13
VM disk I/O limit patch
Hi all,
I add a blkback QoS patch.
You can config(dynamic/static) different I/O speed for different VM disk
by this patch.
----------------------------------------------------------------------------
diff -urNp blkback/blkback.c blkback-qos/blkback.c
--- blkback/blkback.c 2011-06-22 07:54:19.000000000 +0800
+++ blkback-qos/blkback.c 2011-06-22 07:53:18.000000000 +0800
@@ -44,6 +44,11 @@
2011 Jun 21
13
VM disk I/O limit patch
Hi all,
I add a blkback QoS patch.
You can config(dynamic/static) different I/O speed for different VM disk
by this patch.
----------------------------------------------------------------------------
diff -urNp blkback/blkback.c blkback-qos/blkback.c
--- blkback/blkback.c 2011-06-22 07:54:19.000000000 +0800
+++ blkback-qos/blkback.c 2011-06-22 07:53:18.000000000 +0800
@@ -44,6 +44,11 @@
2011 May 25
2
[PATCH linux-2.6.18-xen] blkback: don''t call vbd_size() if bd_disk is NULL
...because vbd_size() dereferences bd_disk if bd_part is NULL.
Signed-off-by: Laszlo Ersek<lersek@redhat.com>
---
drivers/xen/blkback/vbd.c | 3 ++-
1 file changed, 2 insertions(+), 1 deletion(-)
diff -r 415a9b435fef drivers/xen/blkback/vbd.c
--- a/drivers/xen/blkback/vbd.c Mon May 23 18:36:33 2011 +0100
+++ b/drivers/xen/blkback/vbd.c Wed May 25 12:15:26 2011 +0200
@@ -73,7 +73,6 @@
2011 May 02
32
[PATCH] blkback: Fix block I/O latency issue
In blkback driver, after I/O requests are submitted to Dom-0 block I/O subsystem, blkback goes to ''sleep'' effectively without letting blkfront know about it (req_event isn''t set appropriately). Hence blkfront doesn''t notify blkback when it submits a new I/O thus delaying the ''dispatch'' of the new I/O to Dom-0 block I/O subsystem. The new I/O is