Displaying 20 results from an estimated 1000 matches similar to: "CentOS as Guest OS on Red Hat Virtualisation 4.x"
2017 Sep 28
2
Bandwidth and latency requirements
Interesting table Karan!,
Could you please tell us how you did the benchmark? fio or iozone
orsimilar?
thanks
Arman.
On Wed, Sep 27, 2017 at 1:20 PM, Karan Sandha <ksandha at redhat.com> wrote:
> Hi Collin,
>
> During our arbiter latency testing for completion of ops we found the
> below results:- an arbiter node in another data centre and both the data
> bricks in the
2017 Sep 27
0
Bandwidth and latency requirements
Hi Collin,
During our arbiter latency testing for completion of ops we found the below
results:- an arbiter node in another data centre and both the data bricks
in the same data centre,
1) File-size 1 KB (10000 files )
2) mkdir
Latency
5ms
10ms
20ms
50ms
100ms
200ms
Ops
Create
755 secs
1410 secs
2717 secs
5874 secs
12908 sec
26113 sec
Mkdir
922 secs
1725 secs
3325 secs
8127
2017 Oct 30
0
Poor gluster performance on large files.
Hi Brandon,
Can you please turn OFF client-io-threads as we have seen degradation of
performance with io-threads ON on sequential read/writes, random
read/writes. Server event threads is 1 and client event threads are 2 by
default.
Thanks & Regards
On Fri, Oct 27, 2017 at 12:17 PM, Brandon Bates <brandon at brandonbates.com>
wrote:
> Hi gluster users,
> I've spent several
2004 May 30
1
What's wrong with this simple code???
Hi, all
I can not figure this out, please have a look and help me out.
thank you!
Note: this is in SPLUS, not R.
I have following code
***********************************
modfit<-function(yir,yew, ft) {
n<-length(yew)
yew<-yew[1:(n-1)]
yy<-yir-ft
xx<-yew-ft
n<-length(xx)
xx0<-xx[2:n]
yy0 <-yy [2:n]
xx1<-xx[1:(n-1)]
fit <- garch(yy0~xx0 + xx1+var.in.mean,
2017 Sep 29
0
Bandwidth and latency requirements
It was simple emulation of network packets on the port of the server node
using tc tool tc qdisc add dev <port> root netem delay <time>ms. The files
were created using dd tool (in-built in linux) and mkdir. Post the IO's we
verified with no pending heals.
Thanks & Regards
On Thu, Sep 28, 2017 at 2:06 PM, Arman Khalatyan <arm2arm at gmail.com> wrote:
> Interesting
2017 Oct 12
0
Bandwidth and latency requirements
Apologies for the late reply.
Further to this, if my Linux clients are connecting uing glusterfs-fuse and
I have my volumes defined like this:
dc1srv1:/gv_fileshare dc2srv1:/gv_fileshare dc1srv2:/gv_fileshare
dc2srv2:/gv_fileshare (replica 2)
How do I ensure that clients in dc1 prefer dc1srv1 and dc1srv2 while
clients in dc2 prefer the dc2 servers?
Is it simply a matter of ordering in
2017 Sep 25
2
Bandwidth and latency requirements
Hi all
I've googled but can't find an answer to my question.
I have two data centers. Currently, I have a replica (count of 2 plus
arbiter) in one data center but is used by both.
I want to change this to be a distributed replica across the two data
centers.
There is a 20Mbps pipe and approx 22 ms latency. Is this sufficient?
I really don't want to do the geo-replication in its
2011 Apr 26
1
Public Apology to Minister Mentor Lee Kuan Yew and Prime Minister Lee Hsien Loong, Singapore
I was misconstrued as having insulted Minister Mentor Lee Kuan Yew. My
words were twisted and misinterpreted and I feel that there is a need to
explain myself and set the record straight.
[b]What really happened[/b]
It was sometime in Aug/Sep in the year 2009. The setting was in the
Tampines Central office of Asiasoft Online Pte Ltd. I was having a
***verbal*** conversation with Melvin Lee,
2017 Oct 27
5
Poor gluster performance on large files.
Hi gluster users,
I've spent several months trying to get any kind of high performance out of
gluster. The current XFS/samba array is used for video editing and
300-400MB/s for at least 4 clients is minimum (currently a single windows
client gets at least 700/700 for a single client over samba, peaking to 950
at times using blackmagic speed test). Gluster has been getting me as low
as
2019 Nov 18
15
[PATCH v2 00/11] rvh-upload: Various fixes and cleanups
This series extract oVirt SDK and imageio code to make it eaiser to follow the
code and improve error handing in open() and close().
Tested with virt-v2v master.
Changes since v1:
- Rebase on merged patches from v1
- Fix regression introduced by "rhv-upload: Fix cleanup after errors"
- Remove "rhv-upload: Try to remove disk on timeout" since it cannot
succeed
- Add more
2019 Sep 19
2
[PATCH] v2v: -o rhv-upload: add -oo rhv-disk-uuid option
This way it is possible to override the UUIDs of the uploaded disks,
instead of letting RHV generate them.
This can be useful to force certain UUIDs, and to specify the disks in
--no-copy mode (which now can be used).
---
v2v/output_rhv_upload.ml | 43 ++++++++++++++++++++++++++++++++-----
v2v/rhv-upload-plugin.py | 2 ++
v2v/virt-v2v-output-rhv.pod | 23 ++++++++++++++++++++
3 files
2019 Nov 17
23
[PATCH 00/18] rvh-upload: Various fixes and cleanups
This series extract oVirt SDK and imageio code to make it eaiser to follow the
code and improve error handing in open() and close().
The first small patches can be consider as fixes for downstream.
Tested based on libguestfs v1.41.5, since I had trouble building virt-v2v and
libguestfs from master.
Nir Soffer (18):
rhv-upload: Remove unused exception class
rhv-upload: Check status more
2018 Apr 05
2
[PATCH v8] v2v: Add -o rhv-upload output mode (RHBZ#1557273).
v7 was here:
https://www.redhat.com/archives/libguestfs/2018-March/msg00143.html
Since then:
- Earlier patches are now upstream.
- The to-do list is moved from the commit message to the TODO file.
- This version forces -of raw + -oa sparse and gives an error in
any other mode. We intend to lift these restrictions later.
- Tested against latest imageio which supports longer timeouts,
2018 Apr 10
2
[PATCH v9] v2v: Add -o rhv-upload output mode (RHBZ#1557273).
v8 was here:
https://www.redhat.com/archives/libguestfs/2018-April/msg00022.html
v8 -> v9:
- Addresses the only feedback from Tomáš.
Rich.
2016 Dec 07
2
[PATCH v2] v2v: Rename RHEV to RHV throughout.
v2:
- Fix virt-p2v messages too.
Rich.
2018 Mar 21
2
[PATCH v6] v2v: Add -o rhv-upload output mode.
v5 was here:
https://www.redhat.com/archives/libguestfs/2018-March/msg00032.html
There is only a single patch in this version because the other
patches went upstream.
This patch adds the virt-v2v -o rhv-upload mode
(https://bugzilla.redhat.com/show_bug.cgi?id=1557273).
Compared to v5, this adds the ability to make zero, trim and flush
requests to the oVirt imageio server
2018 Mar 08
6
[PATCH v5 0/4] v2v: Add -o rhv-upload output mode.
Mainly minor fixes and code cleanups over the v4 patch.
There are still several problems with this patch, but it is in a
reviewable state, especially the Python code.
Rich.
2019 Sep 19
1
Re: [PATCH] v2v: -o rhv-upload: add -oo rhv-disk-uuid option
On Thursday, 19 September 2019 13:16:01 CEST Martin Kletzander wrote:
> On Thu, Sep 19, 2019 at 12:38:41PM +0200, Pino Toscano wrote:
> >This way it is possible to override the UUIDs of the uploaded disks,
> >instead of letting RHV generate them.
> >
> >This can be useful to force certain UUIDs, and to specify the disks in
> >--no-copy mode (which now can be used).
2018 Sep 19
4
[PATCH 0/3] v2v: -o rhv-upload: Add a test.
This adds a test of -o rhv-upload.
Obviously for an upstream test we cannot require a working oVirt
server. This test works by faking the ovirtsdk4 Python module,
setting PYTHONPATH so that the fake module is picked up instead of the
real module (if installed).
However it's more complex than that because the nbdkit plugin also
expects to talk to a working imageio HTTPS server. Therefore
2020 Jan 10
7
[v2v PATCH 0/6] Various Python pycodestyle fixes
Fixes the majority of the pycodestyle issues in the Python scripts, and
fix the existing test-v2v-python-syntax.sh to use pycodestyle to
actually perform style checks.
Pino Toscano (6):
PEP 8: adapt whitespaces in lines
PEP 8: move imports at the top
PEP 8: adapt empty lines
tests: find all the Python scripts for syntax checks
-o rhv-upload: remove unused Python imports
Revamp check