Displaying 20 results from an estimated 100 matches similar to: "Re: Appliance image. Why raw?"
2017 Jun 20
0
Re: Appliance image. Why raw?
On Tue, Jun 20, 2017 at 06:34:28PM +0300, Pavel Butsykin wrote:
> On 20.06.2017 17:47, Richard W.M. Jones wrote:
> >On Tue, Jun 20, 2017 at 05:04:53PM +0300, Pavel Butsykin wrote:
> >>I noticed that raw is intentionally used as image format for appliance
> >>image. So I would like to ask, is there any reason to use raw as image
> >>format for appliance?
>
2017 Jun 20
0
Re: Appliance image. Why raw?
On Tue, Jun 20, 2017 at 06:48:16PM +0300, Pavel Butsykin wrote:
> On 20.06.2017 18:41, Richard W.M. Jones wrote:
> >On Tue, Jun 20, 2017 at 06:34:28PM +0300, Pavel Butsykin wrote:
> >>On 20.06.2017 17:47, Richard W.M. Jones wrote:
> >>>On Tue, Jun 20, 2017 at 05:04:53PM +0300, Pavel Butsykin wrote:
> >>>>I noticed that raw is intentionally used as image
2000 Mar 31
1
R: one bananna aov() question
Hello world,
I'm trying to do an
anova on data in data.set, dependent variable is a column
named "dep.var", grouping variable is in a column called "indep.var", and
is.factor(indep.var) is TRUE...
why can't I just do aov(dep.var ~ indep.var, data = data.set)?
What have I done to deserve this?! What gives? Am I missing something
totlly obvious?
R-base-1.0.0-1,
2016 May 16
0
[PATCH] launch: direct: Add DAX root filesystem support.
Allow the appliance / root filesystem to be placed on a virtual NVDIMM
and accessed directly by the guest kernel (DAX).
This requires corresponding changes in supermin.
---
src/guestfs-internal.h | 1 +
src/launch-direct.c | 68 ++++++++++++++++++++++++++++++++++++++++----------
src/launch.c | 8 +++++-
3 files changed, 63 insertions(+), 14 deletions(-)
diff --git
2016 Apr 18
0
Re: [PATCH] launch: direct: specify format for appliance drive
On Mon, Apr 18, 2016 at 02:49:07PM +0200, Pino Toscano wrote:
> The drive used for the appliance is a raw (sparse) disk: specify that
> explicitly in its -drive qemu command line options, so qemu can skip the
> autodetection of its format and save a tiny bit of time.
> ---
> src/launch-direct.c | 2 +-
> 1 file changed, 1 insertion(+), 1 deletion(-)
>
> diff --git
2016 Apr 18
2
[PATCH] launch: direct: specify format for appliance drive
The drive used for the appliance is a raw (sparse) disk: specify that
explicitly in its -drive qemu command line options, so qemu can skip the
autodetection of its format and save a tiny bit of time.
---
src/launch-direct.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/src/launch-direct.c b/src/launch-direct.c
index ee0a855..8521e5a 100644
--- a/src/launch-direct.c
+++
2015 Jan 21
0
[PATCH] aarch64: appliance: Use AAVMF (UEFI) if available for running the appliance.
From: "Richard W.M. Jones" <rjones@redhat.com>
AAVMF is an open source UEFI implementation for aarch64 based on OVMF.
As aarch64 is heading for requiring UEFI even inside guests, if the
AAVMF firmware is installed on the host, use it as a hint that we
should boot the guest using AAVMF instead of the default "empty
machine".
Note this requires very recent AAVMF, libvirt,
2016 Jul 12
0
Re: Help: Can I access network inside libguestfs?
On Tue, Jul 12, 2016 at 04:48:06PM +0800, Baochuan Wu wrote:
> Thanks Pino for quick response.
> Actually I use the libguestfs APIs in C++ program. Following your idea, I
> think I should try guestfs_set_network for this.
Yes this is correct. Calling:
guestfs_set_network (g, 1);
should enable the network inside the appliance.
For further tips on using yum, have a look at how
2016 Dec 07
5
[PATCH 0/3] Miscellaneous improvements to supermin.
Document what each module does, using *.mli files.
Remove the --dtb option, it's obsolete.
Rename modules according to their purpose.
Rich.
2017 Apr 19
2
[PATCH] lib: direct: Remove support for virtio-blk as the default.
virtio-scsi has been supported in qemu since 2012, and it is superior
in every respect to virtio-blk. There's no reason to still be using
virtio-blk.
virtio-scsi support was initially added in 2012
(commit 0c0a7d0d868d153adf0600189f771459e1068b0a).
You can still use virtio-blk using the (deprecated) iface parameter,
but don't do that in new code.
---
lib/guestfs-internal.h | 1 -
2018 Aug 09
0
Using SPDK as QEMU's rootfs disk && A patch for libguestfs to support SPDK
Hi guys,
During the last few weeks, I've been looking at how to use SPDK as QEMU's
bootable rootfs disk.
I managed to boot up a SPDK rootfs disk by using OVMF UEFI boot-loader. And
in order to deploy the guest OS before start-up(which has a unrecognizable
filesystem to the host), I have written a small patch for the libguestfs.
It was based on Redhat's CentOS libguestfs-1.36.10 RPM,
2011 Jan 15
3
respond_with javascript
My ajax stopped working when I switched to using respond_with. For my
Votes, I have the create action and the corresponding create.js.erb,
and respond_to :html, :js, :xml in the controller. Heres the log when
I try and create a vote:
Started POST "/stories/3-asdfasdf1111/votes" for 127.0.0.1 at Fri Jan
14 20:46:36 -0800 2011
Processing by VotesController#create as */*
Parameters:
2016 Jul 12
4
Re: Help: Can I access network inside libguestfs?
Thanks Pino for quick response.
Actually I use the libguestfs APIs in C++ program. Following your idea, I
think I should try guestfs_set_network for this.
Please correct me if I have any misunderstanding.
Thanks,
Allen
2016-07-12 16:38 GMT+08:00 Pino Toscano <ptoscano@redhat.com>:
> Hi,
>
> On Tuesday, 12 July 2016 16:30:17 CEST Baochuan Wu wrote:
> > I mounted virtual
2009 Apr 08
3
Rails 2.3 memcache performance drop
After much effort I upgraded our fairly large Rails app from 2.1 to
2.3. After deployment yesterday I noticed an across the board
increase in response times. It seems that every call to memcache now
takes 10x longer than before. Here are some example numbers from my
development log (below), with memcache running locally. We see
similar scale of performance drop in production as well - average
2015 Jun 25
2
[PATCH] launch: rework handling of --enable-valgrind-daemon
Instead of forcing valgrind to be run when --enable-valgrind-daemon is
passed to configure, enable it only when the backend setting
"valgrind_daemon" is set. This allows developers to keep building with
--enable-valgrind-daemon, which unconditionally adds valgrind in the
appliance but using only when requested.
When --enable-valgrind-daemon is not passed (typical for release
builds),
2011 Feb 01
4
Delay Loading Rails Views
Hi All,
I''m having trouble with a Rails app and I can''t figure out what /
where the problem is.
From the logging and testing I''ve done, it looks like there is a
consistent 3 second delay between the time that Rails finishes
rendering the view and the time that the browser actually receives the
data. That delay doesn''t show up on my laptop, running in
2015 Jun 25
0
[PATCH v2 1/9] build: Remove ./configure --enable-valgrind-daemon.
If you've ever tried to use this option, you'll know that it didn't
work well. It broke random things (probably RHBZ#1020216, definitely
RHBZ#1023630), and caused random failures generally, while often not
actually failing when valgrind itself found problems.
This does not remove the guestfs_internal_exit API. That will
instead be modified in a future commit.
---
2018 Jan 14
0
[PATCH v2 1/3] appliance: init: Avoid running degraded md devices
The issue:
- raid1 will be in degraded state if one of its components is logical volume (LV)
- raid0 will be inoperable at all (inacessible from within appliance) if one of its component is LV
- raidN: you can expect the same issue for any raid level depends on how many components are inaccessible at the time mdadm is running and raid redundency.
It happens because mdadm is launched prior to lvm
2009 Apr 13
6
Memcached 1.6.5 (Rails 2.3) 10x slower
The move to memcached_client 1.6.5 in Rails 2.3 seems to have made the
Rails cache about 10x slower. Since that''s the opposite effect I
would expect, I was hoping somebody would explain where I''m misreading
these numbers?
I noticed my fragment caching was slow -- it shouldn''t take 2ms just
to read a 2k string from a localhost memcached server:
Cached fragment hit:
2009 Aug 13
5
First hit on app takes a long time
So, I''ve deployed my first rails app, however I''m the only one hitting
it for now.
I''ve noticed that the first time I hit the app after some long period,
it takes a really long time to respond (2-3 seconds). After that,
it''s very fast.
At first, I thought it might be my VPS "waking up" from some sleep
state, but Rails does log the long load time