Displaying 15 results from an estimated 15 matches for "kvm_stat".
Did you mean:
kvm_state
2023 Mar 10
0
[PATCH v2 0/6] use canonical ftrace path whenever possible
...t; and I've only sent people the patches that I think they care about to
> avoid spamming everyone.
>
> Ross Zwisler (6):
> tracing: always use canonical ftrace path
> selftests: use canonical ftrace path
> leaking_addresses: also skip canonical ftrace path
> tools/kvm_stat: use canonical ftrace path
> tools/virtio: use canonical ftrace path
> tools/virtio: fix typo in README instructions
>
> include/linux/kernel.h | 2 +-
> include/linux/tracepoint.h | 4 ++--
> kernel/trace/Kconfig...
2016 Mar 11
0
Re: 100% CPU when using nested virtualization
...things to identify where the problem might
be:
- `mpstat` shows processor details and CPU utilization, probably you
might want to run that to get a general view. There are several
values it presents: %guest (guest code); %usr (QEMU device
emulation), etc. Refer its man page.
- `kvm_stat` (provided by 'qemu-kvm-tools' package on Fedora-based
systems), a `top`-like tool to show runtime statics of KVM events.
An example of what the results look like[1].
- Or use the 'perf' to record nested virtualization related KVM
events:
$ perf record -a -e...
2016 Mar 11
2
100% CPU when using nested virtualization
Hi all,
I got a new laptop recently and what worked before no longer works
(Fedora 23 on the laptops in both cases)...
I'm trying to get nested virtualization to work because I use the VMs
on the laptop to simulate an HA cluster that itself hosts VMs. I don't
care much at all about the performance of the nested VM, it's just there
so that I can work on the cluster's code.
2014 Apr 10
1
[PATCH RFC V2 4/4] tools: virtio: add a top-like utility for displaying vhost satistics
On Fri, 03/21 17:41, Jason Wang wrote:
> This patch adds simple python to display vhost satistics of vhost, the codes
> were based on kvm_stat script from qemu. As work function has been recored,
> filters could be used to distinguish which kinds of work are being executed or
> queued:
>
> vhost statistics
>
> vhost_virtio_update_used_idx 1215215 0
> vhost_virtio_get_vq_desc 1...
2014 Apr 10
1
[PATCH RFC V2 4/4] tools: virtio: add a top-like utility for displaying vhost satistics
On Fri, 03/21 17:41, Jason Wang wrote:
> This patch adds simple python to display vhost satistics of vhost, the codes
> were based on kvm_stat script from qemu. As work function has been recored,
> filters could be used to distinguish which kinds of work are being executed or
> queued:
>
> vhost statistics
>
> vhost_virtio_update_used_idx 1215215 0
> vhost_virtio_get_vq_desc 1...
2014 Mar 21
0
[PATCH RFC V2 4/4] tools: virtio: add a top-like utility for displaying vhost satistics
This patch adds simple python to display vhost satistics of vhost, the codes
were based on kvm_stat script from qemu. As work function has been recored,
filters could be used to distinguish which kinds of work are being executed or
queued:
vhost statistics
vhost_virtio_update_used_idx 1215215 0
vhost_virtio_get_vq_desc 1215215 0
vhost_work_queue_wa...
2011 Aug 18
0
profiling KVM i/o on CentOS 5
Hi,
I am wondering what tools are available to troubleshoot KVM I/O problems on
CentOS 5, given that the kernel is 2.6.18 with some backports to support
virtualisation?
kvm_stat / kvmtrace - These seem to be available in the kvm-tools package.
ftrace related things seem not to be available.
OProfile is available, can it be used?
Also, SystemTap is available, can it be used?
Anything else?
I don't currently have any experience using any of those, except for a br...
2014 Mar 21
5
[PATCH RFC V2 0/4] Adding tracepoints to vhost/net
Recent debugging on vhost net zerocopy shows the need of
tracepoints. So to help in vhost{net} debugging and performance
analyzing, the following series adding basic tracepoints to
vhost. Operations of both vhost and vhost_net were traced in current
implementation.
A top-like satistics displaying script were introduced to help the
troubleshooting:
vhost statistics
vhost_virtio_update_used_idx
2014 Mar 21
5
[PATCH RFC V2 0/4] Adding tracepoints to vhost/net
Recent debugging on vhost net zerocopy shows the need of
tracepoints. So to help in vhost{net} debugging and performance
analyzing, the following series adding basic tracepoints to
vhost. Operations of both vhost and vhost_net were traced in current
implementation.
A top-like satistics displaying script were introduced to help the
troubleshooting:
vhost statistics
vhost_virtio_update_used_idx
2012 Apr 10
3
[PATCH 0/2] adding tracepoints to vhost
To help in vhost analyzing, the following series adding basic tracepoints to
vhost. Operations of both virtqueues and vhost works were traced in current
implementation, net code were untouched. A top-like satistics displaying script
were introduced to help the troubleshooting.
TODO:
- net specific tracepoints?
---
Jason Wang (2):
vhost: basic tracepoints
tools: virtio: add a
2012 Apr 10
3
[PATCH 0/2] adding tracepoints to vhost
To help in vhost analyzing, the following series adding basic tracepoints to
vhost. Operations of both virtqueues and vhost works were traced in current
implementation, net code were untouched. A top-like satistics displaying script
were introduced to help the troubleshooting.
TODO:
- net specific tracepoints?
---
Jason Wang (2):
vhost: basic tracepoints
tools: virtio: add a
2008 Jun 27
2
PCI device assignment to guests (userspace)
Userspace patches for the pci-passthrough functionality.
The major updates since the last post are:
- Loop to add passthrough devices in pc_init1
- Handle errors in read/write calls
- Allow invocation without irq number for in-kernel irqchip
Other than this, several small things were fixed according to review comments received last time.
2008 Jun 27
2
PCI device assignment to guests (userspace)
Userspace patches for the pci-passthrough functionality.
The major updates since the last post are:
- Loop to add passthrough devices in pc_init1
- Handle errors in read/write calls
- Allow invocation without irq number for in-kernel irqchip
Other than this, several small things were fixed according to review comments received last time.
2012 Aug 10
1
virtio-scsi <-> vhost multi lun/adapter performance results with 3.6-rc0
Hi folks,
The following are initial virtio-scsi + target vhost benchmark results
using multiple target LUNs per vhost and multiple virtio PCI adapters to
scale the total number of virtio-scsi LUNs into a single KVM guest.
The test setup is currently using 4x SCSI LUNs per vhost WWPN, with 8x
virtio PCI adapters for a total of 32x 500MB ramdisk LUNs into a single
guest, along with each backend
2012 Aug 10
1
virtio-scsi <-> vhost multi lun/adapter performance results with 3.6-rc0
Hi folks,
The following are initial virtio-scsi + target vhost benchmark results
using multiple target LUNs per vhost and multiple virtio PCI adapters to
scale the total number of virtio-scsi LUNs into a single KVM guest.
The test setup is currently using 4x SCSI LUNs per vhost WWPN, with 8x
virtio PCI adapters for a total of 32x 500MB ramdisk LUNs into a single
guest, along with each backend