search for: cpumasks

Displaying 20 results from an estimated 419 matches for "cpumasks".

Did you mean: cpumask
2005 Jan 12
3
syslinux build fails with 2.6 linux headers
As the subject says, building syslinux fails when built against 2.6.8.1 kernel headers. Anybody out there have a patch? The compile errors are below. set -e ; for i in memdisk dos win32 mtools unix extlinux sample com32 ; do make DATE=0x41e470ae HEXDATE=0x41e470ae -C $i all ; done make[1]: Entering directory `/usr/src/syslinux-3.07-pre1/memdisk' make[1]: Nothing to be done for `all'.
2019 Jul 19
0
[PATCH v3 4/9] x86/mm/tlb: Flush remote and local TLBs concurrently
To improve TLB shootdown performance, flush the remote and local TLBs concurrently. Introduce flush_tlb_multi() that does so. Introduce paravirtual versions of flush_tlb_multi() for KVM, Xen and hyper-v (Xen and hyper-v are only compile-tested). While the updated smp infrastructure is capable of running a function on a single local core, it is not optimized for this case. The multiple function
2019 Jun 13
4
[PATCH 4/9] x86/mm/tlb: Flush remote and local TLBs concurrently
To improve TLB shootdown performance, flush the remote and local TLBs concurrently. Introduce flush_tlb_multi() that does so. The current flush_tlb_others() interface is kept, since paravirtual interfaces need to be adapted first before it can be removed. This is left for future work. In such PV environments, TLB flushes are not performed, at this time, concurrently. Add a static key to tell
2019 Jun 13
4
[PATCH 4/9] x86/mm/tlb: Flush remote and local TLBs concurrently
To improve TLB shootdown performance, flush the remote and local TLBs concurrently. Introduce flush_tlb_multi() that does so. The current flush_tlb_others() interface is kept, since paravirtual interfaces need to be adapted first before it can be removed. This is left for future work. In such PV environments, TLB flushes are not performed, at this time, concurrently. Add a static key to tell
2019 Jul 02
0
[PATCH v2 4/9] x86/mm/tlb: Flush remote and local TLBs concurrently
To improve TLB shootdown performance, flush the remote and local TLBs concurrently. Introduce flush_tlb_multi() that does so. Introduce paravirtual versions of flush_tlb_multi() for KVM, Xen and hyper-v (Xen and hyper-v are only compile-tested). While the updated smp infrastructure is capable of running a function on a single local core, it is not optimized for this case. The multiple function
2009 Jan 07
4
[PATCH 3/5] cpumask: convert misc driver functions
An embedded and charset-unspecified text was scrubbed... Name: cpumask:convert-drivers.patch Url: http://lists.linux-foundation.org/pipermail/virtualization/attachments/20090107/f459eb35/attachment.txt
2009 Jan 07
4
[PATCH 3/5] cpumask: convert misc driver functions
An embedded and charset-unspecified text was scrubbed... Name: cpumask:convert-drivers.patch Url: http://lists.linux-foundation.org/pipermail/virtualization/attachments/20090107/f459eb35/attachment.txt
2019 May 25
3
[RFC PATCH 5/6] x86/mm/tlb: Flush remote and local TLBs concurrently
To improve TLB shootdown performance, flush the remote and local TLBs concurrently. Introduce flush_tlb_multi() that does so. The current flush_tlb_others() interface is kept, since paravirtual interfaces need to be adapted first before it can be removed. This is left for future work. In such PV environments, TLB flushes are not performed, at this time, concurrently. Add a static key to tell
2019 May 25
3
[RFC PATCH 5/6] x86/mm/tlb: Flush remote and local TLBs concurrently
To improve TLB shootdown performance, flush the remote and local TLBs concurrently. Introduce flush_tlb_multi() that does so. The current flush_tlb_others() interface is kept, since paravirtual interfaces need to be adapted first before it can be removed. This is left for future work. In such PV environments, TLB flushes are not performed, at this time, concurrently. Add a static key to tell
2019 Jul 02
2
[PATCH v2 0/9] x86: Concurrent TLB flushes
Currently, local and remote TLB flushes are not performed concurrently, which introduces unnecessary overhead - each INVLPG can take 100s of cycles. This patch-set allows TLB flushes to be run concurrently: first request the remote CPUs to initiate the flush, then run it locally, and finally wait for the remote CPUs to finish their work. In addition, there are various small optimizations to avoid
2019 Jun 25
0
[PATCH 4/9] x86/mm/tlb: Flush remote and local TLBs concurrently
...n that there are no nested TLB > + * flushes, an assumption that is already made in > + * flush_tlb_mm_range(). > + */ > + struct cpumask *cond_cpumask = this_cpu_ptr(&flush_tlb_mask); This is logically a stack-local variable, right? But, since we've got preempt off and cpumasks can be huge, we don't want to allocate it on the stack. That might be worth a comment somewhere. > + int cpu; > + > + cpumask_clear(cond_cpumask); > + > + for_each_cpu(cpu, cpumask) { > + if (tlb_is_not_lazy(cpu)) > + __cpumask_set_cpu(cpu, cond_cpumask); > +...
2019 May 31
2
[RFC PATCH v2 04/12] x86/mm/tlb: Flush remote and local TLBs concurrently
To improve TLB shootdown performance, flush the remote and local TLBs concurrently. Introduce flush_tlb_multi() that does so. The current flush_tlb_others() interface is kept, since paravirtual interfaces need to be adapted first before it can be removed. This is left for future work. In such PV environments, TLB flushes are not performed, at this time, concurrently. Add a static key to tell
2019 May 31
2
[RFC PATCH v2 04/12] x86/mm/tlb: Flush remote and local TLBs concurrently
To improve TLB shootdown performance, flush the remote and local TLBs concurrently. Introduce flush_tlb_multi() that does so. The current flush_tlb_others() interface is kept, since paravirtual interfaces need to be adapted first before it can be removed. This is left for future work. In such PV environments, TLB flushes are not performed, at this time, concurrently. Add a static key to tell
2019 Jul 19
5
[PATCH v3 0/9] x86: Concurrent TLB flushes
[ Cover-letter is identical to v2, including benchmark results, excluding the change log. ] Currently, local and remote TLB flushes are not performed concurrently, which introduces unnecessary overhead - each INVLPG can take 100s of cycles. This patch-set allows TLB flushes to be run concurrently: first request the remote CPUs to initiate the flush, then run it locally, and finally wait for
2019 Jun 26
2
[PATCH 4/9] x86/mm/tlb: Flush remote and local TLBs concurrently
...>> + * flushes, an assumption that is already made in >> + * flush_tlb_mm_range(). >> + */ >> + struct cpumask *cond_cpumask = this_cpu_ptr(&flush_tlb_mask); > > This is logically a stack-local variable, right? But, since we've got > preempt off and cpumasks can be huge, we don't want to allocate it on > the stack. That might be worth a comment somewhere. I will add a comment here. > >> + int cpu; >> + >> + cpumask_clear(cond_cpumask); >> + >> + for_each_cpu(cpu, cpumask) { >> + if (tlb_is_not_lazy...
2019 Jun 26
2
[PATCH 4/9] x86/mm/tlb: Flush remote and local TLBs concurrently
...>> + * flushes, an assumption that is already made in >> + * flush_tlb_mm_range(). >> + */ >> + struct cpumask *cond_cpumask = this_cpu_ptr(&flush_tlb_mask); > > This is logically a stack-local variable, right? But, since we've got > preempt off and cpumasks can be huge, we don't want to allocate it on > the stack. That might be worth a comment somewhere. I will add a comment here. > >> + int cpu; >> + >> + cpumask_clear(cond_cpumask); >> + >> + for_each_cpu(cpu, cpumask) { >> + if (tlb_is_not_lazy...
2009 Jan 09
1
[PULL]: tip/cpus4096 updates
Hi Ingo, Please pull a few more cpus4096 changes. Most are fairly lightweight with the biggest change being to use cpumask_var_t's for the domain and pending_mask cpumask's in irq_desc. It also deals with this cpu_evtchn_mask bss space bump: 65536 +2031616 2097152 +3100% cpu_evtchn_mask(.bss) Everything merges into tip/master cleanly. Thanks! Mike The following changes since
2013 Sep 27
4
[PATCH net-next] virtio-net: switch to use XPS to choose txq
We used to use a percpu structure vq_index to record the cpu to queue mapping, this is suboptimal since it duplicates the work of XPS and loses all other XPS functionality such as allowing use to configure their own transmission steering strategy. So this patch switches to use XPS and suggest a default mapping when the number of cpus is equal to the number of queues. With XPS support, there's
2013 Sep 27
4
[PATCH net-next] virtio-net: switch to use XPS to choose txq
We used to use a percpu structure vq_index to record the cpu to queue mapping, this is suboptimal since it duplicates the work of XPS and loses all other XPS functionality such as allowing use to configure their own transmission steering strategy. So this patch switches to use XPS and suggest a default mapping when the number of cpus is equal to the number of queues. With XPS support, there's
2009 Jan 09
1
[PULL]: tip/cpus4096 updates
Hi Ingo, Please pull a few more cpus4096 changes. Most are fairly lightweight with the biggest change being to use cpumask_var_t's for the domain and pending_mask cpumask's in irq_desc. It also deals with this cpu_evtchn_mask bss space bump: 65536 +2031616 2097152 +3100% cpu_evtchn_mask(.bss) Everything merges into tip/master cleanly. Thanks! Mike The following changes since