Displaying 8 results from an estimated 8 matches for "rx_desc".
2020 Sep 03
1
[PATCH 22/28] sgiseeq: convert from dma_cache_sync to dma_sync_single_for_device
On Tue, Sep 01, 2020 at 07:38:10PM +0200, Thomas Bogendoerfer wrote:
> this is the problem:
>
> /* Always check for received packets. */
> sgiseeq_rx(dev, sp, hregs, sregs);
>
> so the driver will look at the rx descriptor on every interrupt, so
> we cache the rx descriptor on the first interrupt and if there was
> $no rx packet, we will only see it, if
2020 Sep 15
0
[PATCH 12/18] sgiseeq: convert to dma_alloc_noncoherent
...a..37ff25a84030eb 100644
--- a/drivers/net/ethernet/seeq/sgiseeq.c
+++ b/drivers/net/ethernet/seeq/sgiseeq.c
@@ -112,14 +112,18 @@ struct sgiseeq_private {
static inline void dma_sync_desc_cpu(struct net_device *dev, void *addr)
{
- dma_cache_sync(dev->dev.parent, addr, sizeof(struct sgiseeq_rx_desc),
- DMA_FROM_DEVICE);
+ struct sgiseeq_private *sp = netdev_priv(dev);
+
+ dma_sync_single_for_cpu(dev->dev.parent, VIRT_TO_DMA(sp, addr),
+ sizeof(struct sgiseeq_rx_desc), DMA_BIDIRECTIONAL);
}
static inline void dma_sync_desc_dev(struct net_device *dev, void *addr)
{
- dma_cache...
2020 Sep 14
2
[PATCH 11/17] sgiseeq: convert to dma_alloc_noncoherent
...a..37ff25a84030eb 100644
--- a/drivers/net/ethernet/seeq/sgiseeq.c
+++ b/drivers/net/ethernet/seeq/sgiseeq.c
@@ -112,14 +112,18 @@ struct sgiseeq_private {
static inline void dma_sync_desc_cpu(struct net_device *dev, void *addr)
{
- dma_cache_sync(dev->dev.parent, addr, sizeof(struct sgiseeq_rx_desc),
- DMA_FROM_DEVICE);
+ struct sgiseeq_private *sp = netdev_priv(dev);
+
+ dma_sync_single_for_cpu(dev->dev.parent, VIRT_TO_DMA(sp, addr),
+ sizeof(struct sgiseeq_rx_desc), DMA_BIDIRECTIONAL);
}
static inline void dma_sync_desc_dev(struct net_device *dev, void *addr)
{
- dma_cache...
2009 Feb 10
1
[PATCH 1/2] Fix ixgbe RSS operation
...Signed-off-by: Mitch Williams <mitch.a.williams@intel.com>
diff -r fcc044a90d40 drivers/net/ixgbe/ixgbe_main.c
--- a/drivers/net/ixgbe/ixgbe_main.c Thu Jan 29 10:46:35 2009 +0000
+++ b/drivers/net/ixgbe/ixgbe_main.c Tue Feb 10 10:13:32 2009 -0800
@@ -432,15 +432,11 @@
u16 tag = le16_to_cpu(rx_desc->wb.upper.vlan);
#ifdef CONFIG_XEN_NETDEV2_BACKEND
- if(ring->queue_index) {
+ if ((adapter->flags & IXGBE_FLAG_VMDQ_ENABLED) && ring->queue_index) {
/* This is a VMDq packet destined for a VM. */
vmq_netif_rx(skb, ring->queue_index);
return;
}
- else {
- ne...
2009 Jan 27
5
[PATCH 2/2] Add VMDq support to ixgbe
...e_driver_version[] = DRV_VERSION;
static char ixgbe_copyright[] = "Copyright (c) 1999-2008 Intel Corporation.";
/* ixgbe_pci_tbl - PCI Device ID Table
@@ -431,6 +431,17 @@ static void ixgbe_receive_skb(struct ixg
bool is_vlan = (status & IXGBE_RXD_STAT_VP);
u16 tag = le16_to_cpu(rx_desc->wb.upper.vlan);
+#ifdef CONFIG_XEN_NETDEV2_BACKEND
+ if(ring->queue_index) {
+ /* This is a VMDq packet destined for a VM. */
+ vmq_netif_rx(skb, ring->queue_index);
+ return;
+ }
+ else {
+ netif_rx(skb);
+ return;
+ }
+#endif
#ifndef IXGBE_NO_INET_LRO
if (adapter->netdev->...
2009 Feb 10
3
[PATCH 2/2] Use correct config option for ixgbe VMDq
...xgbe/ixgbe_main.c
--- a/drivers/net/ixgbe/ixgbe_main.c 2009-02-10 14:15:53.000000000 -0800
+++ b/drivers/net/ixgbe/ixgbe_main.c 2009-02-10 14:32:47.000000000 -0800
@@ -431,7 +431,7 @@ static void ixgbe_receive_skb(struct ixg
bool is_vlan = (status & IXGBE_RXD_STAT_VP);
u16 tag = le16_to_cpu(rx_desc->wb.upper.vlan);
-#ifdef CONFIG_XEN_NETDEV2_BACKEND
+#ifdef CONFIG_XEN_NETDEV2_VMQ
if ((adapter->flags & IXGBE_FLAG_VMDQ_ENABLED) && ring->queue_index) {
/* This is a VMDq packet destined for a VM. */
vmq_netif_rx(skb, ring->queue_index);
@@ -519,7 +519,7 @@ static...
2020 Sep 14
20
a saner API for allocating DMA addressable pages v2
Hi all,
this series replaced the DMA_ATTR_NON_CONSISTENT flag to dma_alloc_attrs
with a separate new dma_alloc_pages API, which is available on all
platforms. In addition to cleaning up the convoluted code path, this
ensures that other drivers that have asked for better support for
non-coherent DMA to pages with incurring bounce buffering over can finally
be properly supported.
I'm still a
2020 Sep 15
32
a saner API for allocating DMA addressable pages v3
Hi all,
this series replaced the DMA_ATTR_NON_CONSISTENT flag to dma_alloc_attrs
with a separate new dma_alloc_pages API, which is available on all
platforms. In addition to cleaning up the convoluted code path, this
ensures that other drivers that have asked for better support for
non-coherent DMA to pages with incurring bounce buffering over can finally
be properly supported.
As a follow up I