Displaying 20 results from an estimated 500 matches similar to: "[PATCH 0/6] Various Coverity fixes #2"
2016 May 25
4
[PATCH 0/4] qemu: Use sqlite to store qemu detection data.
Patches 1 & 2 were posted previously here:
https://www.redhat.com/archives/libguestfs/2016-May/msg00134.html
Patch 3 is a hack so I can test this using my own version of qemu (the
`-L ?' stuff is not upstream).
Patch 4 is where the real action takes place: Replace the caching of
qemu features in blob-like files with a sqlite database. Probably the
best way to approach this patch is to
2017 Sep 11
4
[PATCH 0/4] lib: qemu: Add test for mandatory locking.
The patch I posted last week to disable mandatory locking for readonly
drives
(https://www.redhat.com/archives/libguestfs/2017-September/msg00013.html)
was wrong in a couple of respects. Firstly it didn't work, which I
didn't detect because my tests were testing the wrong thing. Oops.
Secondly it used a simple version number check to detect qemu binaries
implementing mandatory locking.
2017 Sep 12
8
[PATCH v3 0/6] launch: direct: Disable qemu locking when opening drives readonly.
v2 -> v3:
- I addressed everything that Pino mentioned last time.
- It's tricky to get a stable run when multiple copies of qemu are
involved, because the same cache files get overwritten by parallel
libguestfs. So I changed the names of the cache files to include
the qemu binary key (size, mtime), which removes this conflict.
This is in new patch 4/6.
Rich.
2017 Sep 12
9
[PATCH v2 0/5] launch: direct: Disable qemu locking when opening drives readonly (RHBZ#1417306)
Patches 1-4 are almost the same as they are when previously posted
here:
https://www.redhat.com/archives/libguestfs/2017-September/msg00039.html
Patch 5 actually uses the mandatory locking test to turn off locking
in the narrow case where a drive is opened readonly, and then only for
the drive being inspected.
Passes ordinary tests (‘check-direct’ and ‘check-valgrind-direct’).
Rich.
2016 May 18
2
[PATCH v2 0/2] lib: qemu: Memoize qemu feature detection.
v1 -> v2:
- Rebase on top of Pino's version work.
Two patches went upstream, these are the two remaining patches.
Note the generation number is still inside the qemu.stat file. We
could put it in the filename, I have no particular preference.
Rich.
2016 May 12
7
[PATCH 0/4] lib: qemu: Memoize qemu feature detection.
Doing qemu feature detection in the direct backend takes ~100ms
because we need to run `qemu -help' and `qemu -devices ?', and each of
those interacts with glibc's very slow link loader.
Fixing the link loader is really hard. Instead memoize the
output of those two commands.
This patch series first separates all the code dealing with qemu into
a separate module (src/qemu.c) and
2016 May 17
2
[PATCH 0/2] Use -bios bios-fast.bin where supported.
NOTE: Not for upstream, yet.
This depends on 3 non-upstream patches.
- The qemu rework in libguestfs, which in turn depends on what
Pino is up to.
- Support for '-L ?' in qemu:
https://lists.gnu.org/archive/html/qemu-devel/2016-05/threads.html#02596
- Support for bios-fast.bin in qemu:
https://lists.gnu.org/archive/html/qemu-devel/2016-05/threads.html#02408
This commit
2016 May 18
2
[PATCH v2 0/2] Use -bios bios-fast.bin where supported.
This commit uses -bios bios-fast.bin if available, which basically
stops SeaBIOS from trying to do PCI probing during boot, which is a
waste of time when using the -kernel option.
v1 -> v2:
- Rebase on top of Pino's work.
This still has 3 dependencies:
- The qemu memoization work (v2).
- Support for '-L ?' in qemu:
2016 May 12
0
[PATCH 4/4] lib: qemu: Memoize qemu feature detection.
qemu feature detection takes about 95ms on my laptop. The overhead is
almost all due to the time taken by the glibc link loader opening the
170+ libraries that qemu is linked to (×2 because we need to run qemu
twice).
Fixing that is seriously hard work.
Therefore memoize the results of guestfs_int_test_qemu.
This is keyed on the size and mtime of the qemu binary, so if the user
changes the
2017 Sep 12
0
[PATCH v2 2/5] lib: qemu: Factor out common code for reading and writing cache files.
The previous code duplicated a lot of common code for reading and
writing the cache file per data field. This change simply factors out
that common code. This makes it simpler to add new tests in future.
This is just refactoring, it should have no effect.
---
lib/qemu.c | 375 +++++++++++++++++++++++++++++++++++++++----------------------
1 file changed, 238 insertions(+), 137 deletions(-)
2017 Jan 12
3
proposed change to ssh_connect_direct()
On Sat, Jan 7, 2017 at 2:30 PM, Peter Moody <mindrot at hda3.com> wrote:
> so I spent a bit of time looking at this and it seems like the only
> way to go, at least if I want to keep it in ssh_connect_direct(), is
> to use pthreads. further, it seems like getting that accepted is
> something of a long shot:
Sorry, pthreads is a non-starter.
I would have thought that using
2017 Feb 03
5
[PATCH 0/5] Support socket activation in virt-p2v.
As the subject says, support socket activation in virt-p2v.
I have added upstream support for socket activation to nbdkit already:
https://github.com/libguestfs/nbdkit/commit/7ff39d028c6359f5c0925ed2cf4a2c4c751af2e4
I posted a patch for qemu-nbd, still waiting on more reviews for that
one:
https://www.mail-archive.com/qemu-devel@nongnu.org/msg427246.html
I tested this against old and new qemu
2014 Dec 15
2
[PATCH v2 0/1] p2v: avoid connecting to ourself while probing qemu-nbd
Changes since v1:
- Set probing source port to be nbd_local_port+1 instead of always
using 50124 to deal with multi-disk scenario.
- Set SO_REUSEADDR on client socket to avoid issues with old
connections in TIME_WAIT.
I've been running this for a few hours now using the updated
multi-disk test and haven't seen any problems.
2002 Apr 03
1
[PATCH] connect() timeout
Here is a version of this widely used patch specific for OpenSSH 3.1p1,
as it is still not in the main tree (perhap one day...)
The patch avoids spending too much time when doing an ssh()/scp() on a
down host, as it does not depend off the default TCP timeout used
by connect().
Patch was tested on Linux, Solaris and HP-UX.
The patch can also be found on:
2002 Jan 26
5
[PATCH] Connect timeout
The attached patch adds a new 'ConnectTimeout' option (man page updated
in patch) to avoid wasting time when the target host is down. I needed that
because I was using rsync/rdist over ssh for massive files update and the
default connect() took too long for my purpose.
The patch was tested on Linux only, but I used a similar one for ssh 1.2.XX
on Linux, Solaris and HP-UX without
2014 Feb 10
3
What magic did openssh do to handle programs like VIM?
Hi,
I'm trying to write my own remote shell program,
It simply connects to a remote server then copy the sock fd to stdin,
stdout and stderr. And it works.
But I found several problems:
1. It doesn't handle characters like ^A ^E or tab (any bash/readline stuff)
-> openssh itself doesn't seem to incorporate readline, so the
magic is not on readline
2. If I run
2012 Apr 12
8
Background jobs with #fork
Hi
I''ve migrated from Passenger to Unicorn about a week ago. It''s great.
Great transparency and management, thanks for this great software!
A few of my Rails applications start background jobs using
Kernel#fork. Of course, the ActiveRecord connections are closed and
reopened again in the parent and child processes. The child process
also does its job.
Unfortunately, it seems
2004 May 12
8
How do fix a good solution against spam..
Heya folks
First of all, sorry if this isn't the correct list, but yet, I think spam is a kind of
network attack and should be treated as a security issue.. I run a working mail server
using Postfix, MySQL, Courier-IMAP, SpamAssassin and ClamAV (amavisd-new) ..
I've checked the configuration file for SpamAssassin, but yet I havn't find any good
solution for spam.. Sure, spam will
2017 Jan 26
3
[PATCH 0/3] p2v: Allow nbdkit as an alternative NBD server.
qemu-nbd is a fine NBD server. However it's not easy to compile a
reliably working version on RHEL 5.
This patch series allows nbdkit to be used as an alternative (qemu-nbd
is still the default).
Rich.
2018 Nov 30
2
[RFC] Discuss about an new idea "Vsock over Virtio-net"
On Fri, Nov 30, 2018 at 09:10:03PM +0800, Jason Wang wrote:
>
> On 2018/11/30 ??8:55, Jason Wang wrote:
> >
> > On 2018/11/30 ??8:52, Michael S. Tsirkin wrote:
> > > > > ?? If you want to compare it with
> > > > > something that would be TCP or QUIC.? The fundamental
> > > > > difference between
> > > > > virtio-vsock