Displaying 10 results from an estimated 10 matches for "0xfful".
Did you mean:
0xffful
2004 Aug 06
0
Re: does installed lib support _int()s ?
...2A3
Then you can define compile-time MACROS for
#define SPEEX_GET_FULL_VERSION (speex_lib_version())
#define SPEEX_GET_MAJOR_MINOR_VERSION (speex_lib_version() >> 24)
#define SPEEX_GET_MAJOR_VERSION (speex_lib_version() >> 16)
#define SPEEX_GET_RELEASE_LEVEL (speex_lib_version() & 0xFFUL)
#define SPEEX_GET_FULL_MINOR_VERSION (speex_lib_version() & 0xffffUL)
#define SPEEX_GET_MINOR_VERSION ((speex_lib_version() >> 8) & 0xFFUL)
2) A *RUNTIME* call into the library which returns the text-version string
of the library in use.
peex_lib_version_string() perhaps?
=MB=
---...
2004 Aug 06
4
Re: does installed lib support _int()s ?
Hi,
Right now, I'm thinking of adding a speex_lib_ctl() call that would
support SPEEX_GET_VERSION (and return a string) or
SPEEX_GET_MAJOR_VERSION and SPEEX_GET_MINOR_VERSION (and return ints).
I'm open to other suggestions though. If there's anything you'd like to
see in the API for 1.2, say it now. ...and no, I won't add a
speex_do_all_the_work_for_me() call :)
2016 Dec 08
0
[PATCH 2/2] x86, paravirt: Fix bool return type for PVOP_CALL
...irt_types.h
+++ b/arch/x86/include/asm/paravirt_types.h
@@ -508,6 +508,18 @@ int paravirt_disable_iospace(void);
#define PVOP_TEST_NULL(op) ((void)op)
#endif
+#define PVOP_RETMASK(rettype) \
+ ({ unsigned long __mask = ~0UL; \
+ switch (sizeof(rettype)) { \
+ case 1: __mask = 0xffUL; break; \
+ case 2: __mask = 0xffffUL; break; \
+ case 4: __mask = 0xffffffffUL; break; \
+ default: break; \
+ } \
+ __mask; \
+ })
+
+
#define ____PVOP_CALL(rettype, op, clbr, call_clbr, extra_clbr, \
pre, post, ...) \
({ \
@@ -535,7 +547,7...
2020 Jul 01
0
[PATCH v3 2/5] mm/hmm: add hmm_mapping order
...(BITS_PER_LONG - 2),
HMM_PFN_ERROR = 1UL << (BITS_PER_LONG - 3),
+ HMM_PFN_ORDER_SHIFT = (BITS_PER_LONG - 8),
/* Input flags */
HMM_PFN_REQ_FAULT = HMM_PFN_VALID,
HMM_PFN_REQ_WRITE = HMM_PFN_WRITE,
- HMM_PFN_FLAGS = HMM_PFN_VALID | HMM_PFN_WRITE | HMM_PFN_ERROR,
+ HMM_PFN_FLAGS = 0xFFUL << HMM_PFN_ORDER_SHIFT,
};
/*
@@ -61,6 +62,25 @@ static inline struct page *hmm_pfn_to_page(unsigned long hmm_pfn)
return pfn_to_page(hmm_pfn & ~HMM_PFN_FLAGS);
}
+/*
+ * hmm_pfn_to_map_order() - return the CPU mapping size order
+ *
+ * The hmm_pfn entry returned by hmm_range_fa...
2016 Dec 08
3
[PATCH 0/2] Fix paravirt fail
Two patches that cure fallout from commit:
3cded4179481 ("x86/paravirt: Optimize native pv_lock_ops.vcpu_is_preempted()")
2016 Dec 08
3
[PATCH 0/2] Fix paravirt fail
Two patches that cure fallout from commit:
3cded4179481 ("x86/paravirt: Optimize native pv_lock_ops.vcpu_is_preempted()")
2024 Oct 15
5
[PATCH v1 0/4] GPU Direct RDMA (P2P DMA) for Device Private Pages
From: Yonatan Maman <Ymaman at Nvidia.com>
This patch series aims to enable Peer-to-Peer (P2P) DMA access in
GPU-centric applications that utilize RDMA and private device pages. This
enhancement is crucial for minimizing data transfer overhead by allowing
the GPU to directly expose device private page data to devices such as
NICs, eliminating the need to traverse system RAM, which is the
2013 Mar 21
27
[PATCH 0/4] xen/arm: guest SMP support
Hi all,
this small patch series implement guest SMP support for ARM, using the
ARM PSCI interface for secondary cpu bringup.
Stefano Stabellini (4):
xen/arm: basic PSCI support, implement cpu_on
xen/arm: support for guest SGI
xen/arm: support vcpu_op hypercalls
xen: move VCPUOP_register_vcpu_info to common code
xen/arch/arm/domain.c | 66 ++++++++++++++++++++++++
2020 Jul 01
8
[PATCH v3 0/5] mm/hmm/nouveau: add PMD system memory mapping
The goal for this series is to introduce the hmm_pfn_to_map_order()
function. This allows a device driver to know that a given 4K PFN is
actually mapped by the CPU using a larger sized CPU page table entry and
therefore the device driver can safely map system memory using larger
device MMU PTEs.
The series is based on 5.8.0-rc3 and is intended for Jason Gunthorpe's
hmm tree. These were
2020 Aug 19
39
a saner API for allocating DMA addressable pages
Hi all,
this series replaced the DMA_ATTR_NON_CONSISTENT flag to dma_alloc_attrs
with a separate new dma_alloc_pages API, which is available on all
platforms. In addition to cleaning up the convoluted code path, this
ensures that other drivers that have asked for better support for
non-coherent DMA to pages with incurring bounce buffering over can finally
be properly supported.
I'm still a