search for: rx_desc

Displaying 8 results from an estimated 8 matches for "rx_desc".

2020 Sep 03
1
[PATCH 22/28] sgiseeq: convert from dma_cache_sync to dma_sync_single_for_device
On Tue, Sep 01, 2020 at 07:38:10PM +0200, Thomas Bogendoerfer wrote: > this is the problem: > > /* Always check for received packets. */ > sgiseeq_rx(dev, sp, hregs, sregs); > > so the driver will look at the rx descriptor on every interrupt, so > we cache the rx descriptor on the first interrupt and if there was > $no rx packet, we will only see it, if
2020 Sep 15
0
[PATCH 12/18] sgiseeq: convert to dma_alloc_noncoherent
...a..37ff25a84030eb 100644 --- a/drivers/net/ethernet/seeq/sgiseeq.c +++ b/drivers/net/ethernet/seeq/sgiseeq.c @@ -112,14 +112,18 @@ struct sgiseeq_private { static inline void dma_sync_desc_cpu(struct net_device *dev, void *addr) { - dma_cache_sync(dev->dev.parent, addr, sizeof(struct sgiseeq_rx_desc), - DMA_FROM_DEVICE); + struct sgiseeq_private *sp = netdev_priv(dev); + + dma_sync_single_for_cpu(dev->dev.parent, VIRT_TO_DMA(sp, addr), + sizeof(struct sgiseeq_rx_desc), DMA_BIDIRECTIONAL); } static inline void dma_sync_desc_dev(struct net_device *dev, void *addr) { - dma_cache...
2020 Sep 14
2
[PATCH 11/17] sgiseeq: convert to dma_alloc_noncoherent
...a..37ff25a84030eb 100644 --- a/drivers/net/ethernet/seeq/sgiseeq.c +++ b/drivers/net/ethernet/seeq/sgiseeq.c @@ -112,14 +112,18 @@ struct sgiseeq_private { static inline void dma_sync_desc_cpu(struct net_device *dev, void *addr) { - dma_cache_sync(dev->dev.parent, addr, sizeof(struct sgiseeq_rx_desc), - DMA_FROM_DEVICE); + struct sgiseeq_private *sp = netdev_priv(dev); + + dma_sync_single_for_cpu(dev->dev.parent, VIRT_TO_DMA(sp, addr), + sizeof(struct sgiseeq_rx_desc), DMA_BIDIRECTIONAL); } static inline void dma_sync_desc_dev(struct net_device *dev, void *addr) { - dma_cache...
2009 Feb 10
1
[PATCH 1/2] Fix ixgbe RSS operation
...Signed-off-by: Mitch Williams <mitch.a.williams@intel.com> diff -r fcc044a90d40 drivers/net/ixgbe/ixgbe_main.c --- a/drivers/net/ixgbe/ixgbe_main.c Thu Jan 29 10:46:35 2009 +0000 +++ b/drivers/net/ixgbe/ixgbe_main.c Tue Feb 10 10:13:32 2009 -0800 @@ -432,15 +432,11 @@ u16 tag = le16_to_cpu(rx_desc->wb.upper.vlan); #ifdef CONFIG_XEN_NETDEV2_BACKEND - if(ring->queue_index) { + if ((adapter->flags & IXGBE_FLAG_VMDQ_ENABLED) && ring->queue_index) { /* This is a VMDq packet destined for a VM. */ vmq_netif_rx(skb, ring->queue_index); return; } - else { - ne...
2009 Jan 27
5
[PATCH 2/2] Add VMDq support to ixgbe
...e_driver_version[] = DRV_VERSION; static char ixgbe_copyright[] = "Copyright (c) 1999-2008 Intel Corporation."; /* ixgbe_pci_tbl - PCI Device ID Table @@ -431,6 +431,17 @@ static void ixgbe_receive_skb(struct ixg bool is_vlan = (status & IXGBE_RXD_STAT_VP); u16 tag = le16_to_cpu(rx_desc->wb.upper.vlan); +#ifdef CONFIG_XEN_NETDEV2_BACKEND + if(ring->queue_index) { + /* This is a VMDq packet destined for a VM. */ + vmq_netif_rx(skb, ring->queue_index); + return; + } + else { + netif_rx(skb); + return; + } +#endif #ifndef IXGBE_NO_INET_LRO if (adapter->netdev-&gt...
2009 Feb 10
3
[PATCH 2/2] Use correct config option for ixgbe VMDq
...xgbe/ixgbe_main.c --- a/drivers/net/ixgbe/ixgbe_main.c 2009-02-10 14:15:53.000000000 -0800 +++ b/drivers/net/ixgbe/ixgbe_main.c 2009-02-10 14:32:47.000000000 -0800 @@ -431,7 +431,7 @@ static void ixgbe_receive_skb(struct ixg bool is_vlan = (status & IXGBE_RXD_STAT_VP); u16 tag = le16_to_cpu(rx_desc->wb.upper.vlan); -#ifdef CONFIG_XEN_NETDEV2_BACKEND +#ifdef CONFIG_XEN_NETDEV2_VMQ if ((adapter->flags & IXGBE_FLAG_VMDQ_ENABLED) && ring->queue_index) { /* This is a VMDq packet destined for a VM. */ vmq_netif_rx(skb, ring->queue_index); @@ -519,7 +519,7 @@ static...
2020 Sep 14
20
a saner API for allocating DMA addressable pages v2
Hi all, this series replaced the DMA_ATTR_NON_CONSISTENT flag to dma_alloc_attrs with a separate new dma_alloc_pages API, which is available on all platforms. In addition to cleaning up the convoluted code path, this ensures that other drivers that have asked for better support for non-coherent DMA to pages with incurring bounce buffering over can finally be properly supported. I'm still a
2020 Sep 15
32
a saner API for allocating DMA addressable pages v3
Hi all, this series replaced the DMA_ATTR_NON_CONSISTENT flag to dma_alloc_attrs with a separate new dma_alloc_pages API, which is available on all platforms. In addition to cleaning up the convoluted code path, this ensures that other drivers that have asked for better support for non-coherent DMA to pages with incurring bounce buffering over can finally be properly supported. As a follow up I