Ryan Harper
2006-Oct-23 18:09 UTC
[Xen-devel] [PATCH 6/6] xen,tools: add numa stat tool and hcalls
This patch adds a tool used to display how memory for each domain was allocated (which nodes does a domain''s memory reside within). It also dumps information about the Xen heap, including how many pages are in each zone and node. HEAPINFO: TOTAL_FREE_PAGES: 5824956 HEAPINFO: NR_ZONES: 3 HEAPINFO: NR_NODES: 4 HEAPINFO: ZONE_XEN: NODE0: FREE_PAGES: 3111 HEAPINFO: ZONE_XEN: NODE1: FREE_PAGES: 0 HEAPINFO: ZONE_XEN: NODE2: FREE_PAGES: 0 HEAPINFO: ZONE_XEN: NODE3: FREE_PAGES: 0 HEAPINFO: ZONE_DOM: NODE0: FREE_PAGES: 860698 HEAPINFO: ZONE_DOM: NODE1: FREE_PAGES: 1507327 HEAPINFO: ZONE_DOM: NODE2: FREE_PAGES: 1507327 HEAPINFO: ZONE_DOM: NODE3: FREE_PAGES: 1507327 HEAPINFO: ZONE_DMA: NODE0: FREE_PAGES: 439166 HEAPINFO: ZONE_DMA: NODE1: FREE_PAGES: 0 HEAPINFO: ZONE_DMA: NODE2: FREE_PAGES: 0 HEAPINFO: ZONE_DMA: NODE3: FREE_PAGES: 0 DOM0: NODE0: PAGES: 130757 DOM0: NODE1: PAGES: 0 DOM0: NODE2: PAGES: 0 DOM0: NODE3: PAGES: 0 -- Ryan Harper Software Engineer; Linux Technology Center IBM Corp., Austin, Tx (512) 838-9253 T/L: 678-9253 ryanh@us.ibm.com diffstat output: b/tools/xen_numastat/Makefile | 35 +++++ b/tools/xen_numastat/xen_numastat.1 | 22 +++ b/tools/xen_numastat/xen_numastat.c | 215 ++++++++++++++++++++++++++++++++++++ tools/Makefile | 1 tools/libxc/xc_domain.c | 45 +++++++ tools/libxc/xenctrl.h | 15 ++ xen/common/domctl.c | 72 ++++++++++++ xen/common/page_alloc.c | 5 xen/include/public/domctl.h | 25 ++++ xen/include/xen/mm.h | 5 10 files changed, 435 insertions(+), 5 deletions(-) Signed-off-by: Ryan Harper <ryanh@us.ibm.com> --- Add Xen NUMA statistics/heap probe tool/hcalls diff -r b11096ca847e tools/Makefile --- a/tools/Makefile Mon Oct 23 12:15:44 2006 -0500 +++ b/tools/Makefile Mon Oct 23 12:15:45 2006 -0500 @@ -13,6 +13,7 @@ SUBDIRS-y += console SUBDIRS-y += console SUBDIRS-y += xenmon SUBDIRS-y += guest-headers +SUBDIRS-y += xen_numastat SUBDIRS-$(VTPM_TOOLS) += vtpm_manager SUBDIRS-$(VTPM_TOOLS) += vtpm SUBDIRS-y += xenstat diff -r b11096ca847e tools/libxc/xc_domain.c --- a/tools/libxc/xc_domain.c Mon Oct 23 12:15:44 2006 -0500 +++ b/tools/libxc/xc_domain.c Mon Oct 23 12:15:45 2006 -0500 @@ -552,6 +552,51 @@ int xc_domain_iomem_permission(int xc_ha domctl.u.iomem_permission.allow_access = allow_access; return do_domctl(xc_handle, &domctl); +} + +int xc_availheap(int xc_handle, + int zone, + int node, + uint32_t *nr_zones, + uint32_t *nr_nodes, + uint64_t *pages) +{ + DECLARE_DOMCTL; + int rc = 0; + + domctl.cmd = XEN_DOMCTL_availheap; + domctl.u.availheap.zone = zone; + domctl.u.availheap.node = node; + + rc = do_domctl(xc_handle, &domctl); + if ( rc >= 0 ) { + if (nr_zones) + *nr_zones = domctl.u.availheap.nr_zones; + if (nr_nodes) + *nr_nodes = domctl.u.availheap.nr_nodes; + *pages = domctl.u.availheap.pages; + } + + return rc; +} + +int xc_domain_getdomnodestat(int xc_handle, + int domid, + uint32_t node, + uint64_t *pages) +{ + DECLARE_DOMCTL; + int rc = 0; + + domctl.cmd = XEN_DOMCTL_getdomnodestat; + domctl.u.getdomnodestat.domain = domid; + domctl.u.getdomnodestat.node = node; + + rc = do_domctl(xc_handle, &domctl); + if ( rc >= 0 ) + *pages = domctl.u.getdomnodestat.pages; + + return rc; } /* diff -r b11096ca847e tools/libxc/xenctrl.h --- a/tools/libxc/xenctrl.h Mon Oct 23 12:15:44 2006 -0500 +++ b/tools/libxc/xenctrl.h Mon Oct 23 12:15:45 2006 -0500 @@ -552,6 +552,21 @@ int xc_get_pfn_type_batch(int xc_handle, /* Get current total pages allocated to a domain. */ long xc_get_tot_pages(int xc_handle, uint32_t domid); +/** + * This function retrieves the the number of pages in the + * specified domain that are on the specified node. + * + * @parm xc_handle a handle to an open hypervisor interface + * @parm domid the domain to query, -1 for per-node free list + * @parm node the node to query + * @parm *pages caller variable to put output + * @return 0 on success, <0 on failure. + */ +int xc_domain_getdomnodestat(int xc_handle, int domid, uint32_t node, + uint64_t *pages); + +int xc_availheap(int xc_handle, int zone, int node, + uint32_t *nr_zones, uint32_t *nr_nodes, uint64_t *pages); /* * Trace Buffer Operations diff -r b11096ca847e xen/common/domctl.c --- a/xen/common/domctl.c Mon Oct 23 12:15:44 2006 -0500 +++ b/xen/common/domctl.c Mon Oct 23 12:15:45 2006 -0500 @@ -21,6 +21,7 @@ #include <asm/current.h> #include <public/domctl.h> #include <acm/acm_hooks.h> +#include <asm/numa.h> extern long arch_do_domctl( struct xen_domctl *op, XEN_GUEST_HANDLE(xen_domctl_t) u_domctl); @@ -640,6 +641,77 @@ long do_domctl(XEN_GUEST_HANDLE(xen_domc } break; + case XEN_DOMCTL_availheap: + { + ret = -EINVAL; + if ( op->u.availheap.node >= num_online_nodes() ) + break; + if ( op->u.availheap.zone >= NR_ZONES ) + break; + + /* indicate the number of zones/nodes queried. + * NB: -1 is wild card for all zones/nodes */ + ( op->u.availheap.zone < 0 ) ? + (op->u.availheap.nr_zones=NR_ZONES) : + (op->u.availheap.nr_zones=1); + + ( op->u.availheap.node < 0 ) ? + (op->u.availheap.nr_nodes=num_online_nodes()) : + (op->u.availheap.nr_nodes=1); + + op->u.availheap.pages + avail_heap_pages(op->u.availheap.zone, op->u.availheap.node); + + if ( copy_to_guest(u_domctl, op, 1) ) + ret = -EFAULT; + else + ret = 0; + } + break; + + case XEN_DOMCTL_getdomnodestat: + { + struct domain *d; + struct page_info *pg; + + ret = -EINVAL; + if ( op->u.getdomnodestat.node >= num_online_nodes() ) + break; + + ret = -ESRCH; + d = find_domain_by_id(op->u.getdomnodestat.domain); + + if ( d != NULL ) + { + /* clear out pages count */ + op->u.getdomnodestat.pages = 0; + + /* walk domain''s page list and count pages on node */ + spin_lock(&d->page_alloc_lock); + list_for_each_entry(pg, &d->page_list, list) + { + if ( phys_to_nid(page_to_maddr(pg)) =+ (int) op->u.getdomnodestat.node ) + op->u.getdomnodestat.pages++; + } + list_for_each_entry(pg, &d->xenpage_list, list) + { + if ( phys_to_nid(page_to_maddr(pg)) =+ (int) op->u.getdomnodestat.node ) + op->u.getdomnodestat.pages++; + } + spin_unlock(&d->page_alloc_lock); + + put_domain(d); + + if ( copy_to_guest(u_domctl, op, 1) ) + ret = -EFAULT; + else + ret = 0; + } + } + break; + default: ret = arch_do_domctl(op, u_domctl); break; diff -r b11096ca847e xen/common/page_alloc.c --- a/xen/common/page_alloc.c Mon Oct 23 12:15:44 2006 -0500 +++ b/xen/common/page_alloc.c Mon Oct 23 12:15:45 2006 -0500 @@ -242,11 +242,6 @@ unsigned long alloc_boot_pages(unsigned * BINARY BUDDY ALLOCATOR */ -#define MEMZONE_XEN 0 -#define MEMZONE_DOM 1 -#define MEMZONE_DMADOM 2 -#define NR_ZONES 3 - #define pfn_dom_zone_type(_pfn) \ (((_pfn) <= MAX_DMADOM_PFN) ? MEMZONE_DMADOM : MEMZONE_DOM) diff -r b11096ca847e xen/include/public/domctl.h --- a/xen/include/public/domctl.h Mon Oct 23 12:15:44 2006 -0500 +++ b/xen/include/public/domctl.h Mon Oct 23 12:15:45 2006 -0500 @@ -347,6 +347,29 @@ struct xen_domctl_settimeoffset { }; typedef struct xen_domctl_settimeoffset xen_domctl_settimeoffset_t; DEFINE_XEN_GUEST_HANDLE(xen_domctl_settimeoffset_t); + +#define XEN_DOMCTL_availheap 26 +struct xen_domctl_availheap { + /* in */ + int zone; /* query available pages in zone, -1 for all */ + int node; /* query available pages in node, -1 for all */ + /* out */ + uint32_t nr_zones; /* number of zones queried */ + uint32_t nr_nodes; /* number of nodes queried */ + uint64_t pages; +}; +typedef struct xen_domctl_availheap xen_domctl_availheap_t; +DEFINE_XEN_GUEST_HANDLE(xen_domctl_availheap_t); + +#define XEN_DOMCTL_getdomnodestat 27 +struct xen_domctl_getdomnodestat { + int domain; /* domain to query, -1 to query per-node free list */ + int32_t node; /* node to query */ + uint64_t pages; /* number of pages on node */ +}; +typedef struct xen_domctl_getdomnodestat xen_domctl_getdomnodestat_t; +DEFINE_XEN_GUEST_HANDLE(xen_domctl_getdomnodestat_t); + struct xen_domctl { uint32_t cmd; @@ -373,6 +396,8 @@ struct xen_domctl { struct xen_domctl_hypercall_init hypercall_init; struct xen_domctl_arch_setup arch_setup; struct xen_domctl_settimeoffset settimeoffset; + struct xen_domctl_availheap availheap; + struct xen_domctl_getdomnodestat getdomnodestat; uint8_t pad[128]; } u; }; diff -r b11096ca847e xen/include/xen/mm.h --- a/xen/include/xen/mm.h Mon Oct 23 12:15:44 2006 -0500 +++ b/xen/include/xen/mm.h Mon Oct 23 12:15:45 2006 -0500 @@ -32,6 +32,11 @@ #include <xen/types.h> #include <xen/list.h> #include <xen/spinlock.h> + +#define MEMZONE_XEN 0 +#define MEMZONE_DOM 1 +#define MEMZONE_DMADOM 2 +#define NR_ZONES 3 struct domain; struct page_info; diff -r b11096ca847e tools/xen_numastat/Makefile --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/tools/xen_numastat/Makefile Mon Oct 23 12:15:45 2006 -0500 @@ -0,0 +1,35 @@ +INSTALL = install +INSTALL_PROG = $(INSTALL) -m0755 +INSTALL_DIR = $(INSTALL) -d -m0755 +INSTALL_DATA = $(INSTALL) -m0644 + +XEN_ROOT=../.. +include $(XEN_ROOT)/tools/Rules.mk + +CFLAGS += -Wall -Werror -g + +CFLAGS += -I $(XEN_XC) +CFLAGS += -I $(XEN_LIBXC) + +HDRS = $(wildcard *.h) +OBJS = $(patsubst %.c,%.o,$(wildcard *.c)) + +BIN = xen_numastat +MAN1 = $(wildcard *.1) +LIBBIN = + +all: build +build: $(BIN) + +install: build + [ -d $(DESTDIR)/usr/bin ] || $(INSTALL_DIR) $(DESTDIR)/usr/bin + $(INSTALL_PROG) $(BIN) $(DESTDIR)/usr/bin + [ -d $(DESTDIR)/usr/share/man/man1 ] || \ + $(INSTALL_DIR) $(DESTDIR)/usr/share/man/man1 + $(INSTALL_DATA) $(MAN1) $(DESTDIR)/usr/share/man/man1 + +clean: + $(RM) *.a *.so *.o $(BIN) + +%: %.c $(HDRS) Makefile + $(CC) $(CFLAGS) -o $@ $< -L$(XEN_LIBXC) -lxenctrl diff -r b11096ca847e tools/xen_numastat/xen_numastat.1 --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/tools/xen_numastat/xen_numastat.1 Mon Oct 23 12:15:45 2006 -0500 @@ -0,0 +1,22 @@ +.\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.35. +.TH XEN_NUMASTAT "1" "May 2006" "Usage: xen_numastat [OPTION]" "User Commands" +.SH NAME +xen_numastat - Displays NUMA statistics about machine and domains +.SH SYNOPSIS +.B xen_numastat +[\fIOPTIONS\fR] +.SH DESCRIPTION +Displays NUMA statistics about machine and domains +.TP +\fB\-h\fR, \fB\-\-help\fR +display this help and exit +.TP +\fB\-H\fR, \fB\-\-heapinfo\fR +display info about Xen HEAP +.TP +\fB\-d\fR, \fB\-\-domid\fR <D> +display NUMA info about domain D +.TP +\fB\-p\fR, \fB\-\-physinfo\fR +display NUMA info about this machine +.PP diff -r b11096ca847e tools/xen_numastat/xen_numastat.c --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/tools/xen_numastat/xen_numastat.c Mon Oct 23 12:18:50 2006 -0500 @@ -0,0 +1,215 @@ +/*\ + * Copyright (C) International Business Machines Corp., 2006 + * Author(s): Ryan Harper <ryanh@us.ibm.com> + * + * Xen Domain NUMA statistics tool + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License as published by + * the Free Software Foundation; under version 2 of the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. * + * You should have received a copy of the GNU General Public License + * along with this program; if not, write to the Free Software + * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA +\*/ + +#include <stdlib.h> +#include <stdio.h> +#include <string.h> +#include <err.h> +#include <errno.h> +#include <xenctrl.h> +#include <getopt.h> +#include <inttypes.h> + +#define DOMAIN 0 +#define PHYSICAL 1 +/* this is a bit arbitrary, might be nice to query from xen */ +#define MAX_DOMAINS 256 + +/* taken from xen/include/xen/mm.h */ +#define MEMZONE_XEN 0 +#define MEMZONE_DOM 1 +#define MEMZONE_DMADOM 2 + +static void usage(const char *program) { + printf("Usage: %s [OPTION]\n" + "Displays NUMA statistics about machine and domains \n" + "\n" + " -h, --help display this help and exit\n" + " -H, --heapinfo display info about Xen HEAP\n" + " -d, --domid <D> display NUMA info about domain D\n" + " -p, --physinfo display NUMA info about this machine\n" + , program); +} + +int str_to_domid(const char* domidstr) { + char *end = NULL; + int domid; + + domid = strtol(domidstr, &end, 10); + if (end && *end) { + fprintf(stderr, "Invalid DOMID `%s''\n", domidstr); + fprintf(stderr, "*end -> %c\n", *end); + exit(EINVAL); + } + + return domid; +} + +void display_dominfo(int xc_handle, int domid, int nr_nodes) +{ + xc_domaininfo_t *dominfo =NULL; + uint64_t pages; + unsigned int i, j, num_domains = 1; + + /* malloc a large buffer for info */ + if ( !(dominfo = malloc(MAX_DOMAINS * sizeof(xc_domaininfo_t))) ) + err(errno, "failed to malloc memory"); + + /* get actual number of domains */ + num_domains = xc_domain_getinfolist(xc_handle, 0, MAX_DOMAINS, dominfo); + + for (i=0; i<num_domains; i++) { + /* show all domains, or only requested domain */ + if (domid == -1 || domid == dominfo[i].domain) { + /* print domain per-node info */ + for (j=0; j < nr_nodes; j++) { + if (xc_domain_getdomnodestat(xc_handle, + dominfo[i].domain, j, &pages) < 0) + err(errno, "xc_domain_getdomnodestat()"); + printf("DOM%d: NODE%d: PAGES: %"PRIu64"\n", + dominfo[i].domain, j, pages); + } + } + } + + free(dominfo); +} + +void display_heapinfo(int xc_handle) +{ + int i,j; + uint32_t nr_zones, nr_nodes; + uint64_t pages; + char zone[8]; + + /* get zones/nodes and total free pages */ + if( xc_availheap(xc_handle, -1, -1, &nr_zones, &nr_nodes, &pages) ) + err(errno, "xc_availheap()"); + + printf("HEAPINFO: TOTAL_FREE_PAGES: %"PRIu64"\n", pages); + printf("HEAPINFO: NR_ZONES: %d\n", nr_zones); + printf("HEAPINFO: NR_NODES: %d\n", nr_nodes); + + for (i=0; i<nr_zones; i++) { + switch(i) { + case MEMZONE_XEN: + sprintf(zone, "XEN"); + break; + case MEMZONE_DOM: + sprintf(zone, "DOM"); + break; + case MEMZONE_DMADOM: + sprintf(zone, "DMA"); + break; + } + for (j=0; j<nr_nodes; j++) { + if( xc_availheap(xc_handle, i, j, NULL, NULL, &pages) ) + err(errno, "xc_availheap()"); + printf("HEAPINFO: ZONE_%s: NODE%d: FREE_PAGES: %"PRIu64"\n", + zone, j, pages); + } + } +} + +void display_physinfo(int xc_handle, xc_physinfo_t *info) +{ + + printf("PHYSINFO: NR_NODES: %d\n", info->nr_nodes); + +} + +int main(int argc, char * argv[]) +{ + + char *sopt = "d:Hhp"; + int ch; + int opt_ind=0, heapinfo=0, dominfo=0, physinfo=0, domid=-1; + int xc_handle, rc; + xc_physinfo_t info; + xc_memory_chunk_t *chunks; + xc_cpu_to_node_t *map; + struct option lopt[] = { + { "domid", 1, 0, ''d'' }, + { "heapinfo", 0, 0, ''H'' }, + { "help", 0, 0, ''h'' }, + { "physinfo", 0, 0, ''p'' }, + { 0, 0, 0, 0 } + }; + + while((ch = getopt_long(argc, argv, sopt, lopt, &opt_ind)) != -1) { + switch(ch) { + case ''d'': + /* convert domid */ + domid = str_to_domid(optarg); + dominfo = 1; + break; + case ''H'': + heapinfo = 1; + break; + case ''h'': + usage(argv[0]); + exit(0); + break; + case ''p'': + physinfo = 1; + break; + } + } + + /* display help if no options are passed */ + if (dominfo == 0 && physinfo == 0 && heapinfo == 0) { + usage(argv[0]); + exit(0); + } + + /* set up */ + xc_handle = xc_interface_open(); + if (xc_handle == -1) + err(errno, "xc_interface_open()"); + + /* make space for mem chunks */ + chunks = (xc_memory_chunk_t *) + malloc( sizeof(xc_memory_chunk_t) * PUBLIC_MAXCHUNKS ); + set_xen_guest_handle(info.memory_chunks, chunks); + + /* make space for cpu_to_node mapping */ + map = (xc_cpu_to_node_t *) + malloc( sizeof(xc_cpu_to_node_t) * PUBLIC_MAX_CPUS ); + set_xen_guest_handle(info.cpu_to_node, map); + + /* get the machine physinfo */ + rc = xc_physinfo(xc_handle, &info); + if (rc < 0) + err(errno, "xc_physinfo()"); + + if (physinfo > 0) + display_physinfo(xc_handle, &info); + if (heapinfo > 0) + display_heapinfo(xc_handle); + if (dominfo > 0) + display_dominfo(xc_handle, domid, info.nr_nodes); + + /* free malloc''d memory */ + free(chunks); + free(map); + + xc_interface_close(xc_handle); + + return 0; +} _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com http://lists.xensource.com/xen-devel
John Levon
2006-Oct-23 20:49 UTC
Re: [Xen-devel] [PATCH 6/6] xen,tools: add numa stat tool and hcalls
On Mon, Oct 23, 2006 at 01:09:28PM -0500, Ryan Harper wrote:> This patch adds a tool used to display how memory for each domain was > allocated (which nodes does a domain''s memory reside within). It also > dumps information about the Xen heap, including how many pages are in > each zone and node.+ struct xen_domctl_availheap availheap; + struct xen_domctl_getdomnodestat getdomnodestat; uint8_t pad[128]; Don''t you need to alter the pad to compensate? john _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com http://lists.xensource.com/xen-devel
Ryan Harper
2006-Oct-23 21:11 UTC
Re: [Xen-devel] [PATCH 6/6] xen,tools: add numa stat tool and hcalls
* John Levon <levon@movementarian.org> [2006-10-23 15:53]:> On Mon, Oct 23, 2006 at 01:09:28PM -0500, Ryan Harper wrote: > > > This patch adds a tool used to display how memory for each domain was > > allocated (which nodes does a domain''s memory reside within). It also > > dumps information about the Xen heap, including how many pages are in > > each zone and node. > > + struct xen_domctl_availheap availheap; > + struct xen_domctl_getdomnodestat getdomnodestat; > uint8_t pad[128]; > > Don''t you need to alter the pad to compensate?AFAIK, I don''t. My understanding is that the pad[128] is the maximum size for any member of the union. Neither of the two added hypercalls exceed 128. -- Ryan Harper Software Engineer; Linux Technology Center IBM Corp., Austin, Tx (512) 838-9253 T/L: 678-9253 ryanh@us.ibm.com _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com http://lists.xensource.com/xen-devel