Jason Wang
2021-Jan-04 06:55 UTC
[PATCH V3 16/19] virtio-pci: introduce modern device module
Signed-off-by: Jason Wang <jasowang at redhat.com>
---
drivers/virtio/Kconfig | 10 +-
drivers/virtio/Makefile | 1 +
drivers/virtio/virtio_pci_common.h | 27 +-
drivers/virtio/virtio_pci_modern.c | 617 -------------------------
drivers/virtio/virtio_pci_modern_dev.c | 599 ++++++++++++++++++++++++
include/linux/virtio_pci_modern.h | 111 +++++
6 files changed, 721 insertions(+), 644 deletions(-)
create mode 100644 drivers/virtio/virtio_pci_modern_dev.c
create mode 100644 include/linux/virtio_pci_modern.h
diff --git a/drivers/virtio/Kconfig b/drivers/virtio/Kconfig
index 7b41130d3f35..6b9b81f4b8c2 100644
--- a/drivers/virtio/Kconfig
+++ b/drivers/virtio/Kconfig
@@ -12,6 +12,14 @@ config ARCH_HAS_RESTRICTED_VIRTIO_MEMORY_ACCESS
This option is selected if the architecture may need to enforce
VIRTIO_F_ACCESS_PLATFORM
+config VIRTIO_PCI_MODERN
+ tristate "Modern Virtio PCI Device"
+ depends on PCI
+ help
+ Modern PCI device implementation. This module implements the
+ basic probe and control for devices which are based on modern
+ PCI device with possible vendor specific extensions.
+
menuconfig VIRTIO_MENU
bool "Virtio drivers"
default y
@@ -20,7 +28,7 @@ if VIRTIO_MENU
config VIRTIO_PCI
tristate "PCI driver for virtio devices"
- depends on PCI
+ depends on VIRTIO_PCI_MODERN
select VIRTIO
help
This driver provides support for virtio based paravirtual device
diff --git a/drivers/virtio/Makefile b/drivers/virtio/Makefile
index 591e6f72aa54..f097578aaa8f 100644
--- a/drivers/virtio/Makefile
+++ b/drivers/virtio/Makefile
@@ -1,5 +1,6 @@
# SPDX-License-Identifier: GPL-2.0
obj-$(CONFIG_VIRTIO) += virtio.o virtio_ring.o
+obj-$(CONFIG_VIRTIO_PCI_MODERN) += virtio_pci_modern_dev.o
obj-$(CONFIG_VIRTIO_MMIO) += virtio_mmio.o
obj-$(CONFIG_VIRTIO_PCI) += virtio_pci.o
virtio_pci-y := virtio_pci_modern.o virtio_pci_common.o
diff --git a/drivers/virtio/virtio_pci_common.h
b/drivers/virtio/virtio_pci_common.h
index f35ff5b6b467..beec047a8f8d 100644
--- a/drivers/virtio/virtio_pci_common.h
+++ b/drivers/virtio/virtio_pci_common.h
@@ -25,6 +25,7 @@
#include <linux/virtio_config.h>
#include <linux/virtio_ring.h>
#include <linux/virtio_pci.h>
+#include <linux/virtio_pci_modern.h>
#include <linux/highmem.h>
#include <linux/spinlock.h>
@@ -39,32 +40,6 @@ struct virtio_pci_vq_info {
unsigned msix_vector;
};
-struct virtio_pci_modern_device {
- struct pci_dev *pci_dev;
-
- struct virtio_pci_common_cfg __iomem *common;
- /* Device-specific data (non-legacy mode) */
- void __iomem *device;
- /* Base of vq notifications (non-legacy mode). */
- void __iomem *notify_base;
- /* Where to read and clear interrupt */
- u8 __iomem *isr;
-
- /* So we can sanity-check accesses. */
- size_t notify_len;
- size_t device_len;
-
- /* Capability for when we need to map notifications per-vq. */
- int notify_map_cap;
-
- /* Multiply queue_notify_off by this value. (non-legacy mode). */
- u32 notify_offset_multiplier;
-
- int modern_bars;
-
- struct virtio_device_id id;
-};
-
/* Our device structure */
struct virtio_pci_device {
struct virtio_device vdev;
diff --git a/drivers/virtio/virtio_pci_modern.c
b/drivers/virtio/virtio_pci_modern.c
index a5e3a5e40323..fbd4ebc00eb6 100644
--- a/drivers/virtio/virtio_pci_modern.c
+++ b/drivers/virtio/virtio_pci_modern.c
@@ -19,158 +19,6 @@
#define VIRTIO_RING_NO_LEGACY
#include "virtio_pci_common.h"
-/*
- * Type-safe wrappers for io accesses.
- * Use these to enforce at compile time the following spec requirement:
- *
- * The driver MUST access each field using the ?natural? access
- * method, i.e. 32-bit accesses for 32-bit fields, 16-bit accesses
- * for 16-bit fields and 8-bit accesses for 8-bit fields.
- */
-static inline u8 vp_ioread8(const u8 __iomem *addr)
-{
- return ioread8(addr);
-}
-static inline u16 vp_ioread16 (const __le16 __iomem *addr)
-{
- return ioread16(addr);
-}
-
-static inline u32 vp_ioread32(const __le32 __iomem *addr)
-{
- return ioread32(addr);
-}
-
-static inline void vp_iowrite8(u8 value, u8 __iomem *addr)
-{
- iowrite8(value, addr);
-}
-
-static inline void vp_iowrite16(u16 value, __le16 __iomem *addr)
-{
- iowrite16(value, addr);
-}
-
-static inline void vp_iowrite32(u32 value, __le32 __iomem *addr)
-{
- iowrite32(value, addr);
-}
-
-static void vp_iowrite64_twopart(u64 val,
- __le32 __iomem *lo, __le32 __iomem *hi)
-{
- vp_iowrite32((u32)val, lo);
- vp_iowrite32(val >> 32, hi);
-}
-
-/*
- * vp_modern_map_capability - map a part of virtio pci capability
- * @mdev: the modern virtio-pci device
- * @off: offset of the capability
- * @minlen: minimal length of the capability
- * @align: align requirement
- * @start: start from the capability
- * @size: map size
- * @len: the length that is actually mapped
- *
- * Returns the io address of for the part of the capability
- */
-void __iomem *vp_modern_map_capability(struct virtio_pci_modern_device *mdev,
int off,
- size_t minlen,
- u32 align,
- u32 start, u32 size,
- size_t *len)
-{
- struct pci_dev *dev = mdev->pci_dev;
- u8 bar;
- u32 offset, length;
- void __iomem *p;
-
- pci_read_config_byte(dev, off + offsetof(struct virtio_pci_cap,
- bar),
- &bar);
- pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, offset),
- &offset);
- pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, length),
- &length);
-
- if (length <= start) {
- dev_err(&dev->dev,
- "virtio_pci: bad capability len %u (>%u expected)\n",
- length, start);
- return NULL;
- }
-
- if (length - start < minlen) {
- dev_err(&dev->dev,
- "virtio_pci: bad capability len %u (>=%zu expected)\n",
- length, minlen);
- return NULL;
- }
-
- length -= start;
-
- if (start + offset < offset) {
- dev_err(&dev->dev,
- "virtio_pci: map wrap-around %u+%u\n",
- start, offset);
- return NULL;
- }
-
- offset += start;
-
- if (offset & (align - 1)) {
- dev_err(&dev->dev,
- "virtio_pci: offset %u not aligned to %u\n",
- offset, align);
- return NULL;
- }
-
- if (length > size)
- length = size;
-
- if (len)
- *len = length;
-
- if (minlen + offset < minlen ||
- minlen + offset > pci_resource_len(dev, bar)) {
- dev_err(&dev->dev,
- "virtio_pci: map virtio %zu@%u "
- "out of range on bar %i length %lu\n",
- minlen, offset,
- bar, (unsigned long)pci_resource_len(dev, bar));
- return NULL;
- }
-
- p = pci_iomap_range(dev, bar, offset, length);
- if (!p)
- dev_err(&dev->dev,
- "virtio_pci: unable to map virtio %u@%u on bar %i\n",
- length, offset, bar);
- return p;
-}
-
-/*
- * vp_modern_get_features - get features from device
- * @mdev: the modern virtio-pci device
- *
- * Returns the features read from the device
- */
-static u64 vp_modern_get_features(struct virtio_pci_modern_device *mdev)
-{
- struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
-
- u64 features;
-
- vp_iowrite32(0, &cfg->device_feature_select);
- features = vp_ioread32(&cfg->device_feature);
- vp_iowrite32(1, &cfg->device_feature_select);
- features |= ((u64)vp_ioread32(&cfg->device_feature) << 32);
-
- return features;
-}
-
-/* virtio config->get_features() implementation */
static u64 vp_get_features(struct virtio_device *vdev)
{
struct virtio_pci_device *vp_dev = to_vp_device(vdev);
@@ -188,149 +36,6 @@ static void vp_transport_features(struct virtio_device
*vdev, u64 features)
__virtio_set_bit(vdev, VIRTIO_F_SR_IOV);
}
-/*
- * vp_modern_set_features - set features to device
- * @mdev: the modern virtio-pci device
- * @features: the features set to device
- */
-static void vp_modern_set_features(struct virtio_pci_modern_device *mdev,
- u64 features)
-{
- struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
-
- vp_iowrite32(0, &cfg->guest_feature_select);
- vp_iowrite32((u32)features, &cfg->guest_feature);
- vp_iowrite32(1, &cfg->guest_feature_select);
- vp_iowrite32(features >> 32, &cfg->guest_feature);
-}
-
-/*
- * vp_modern_queue_vector - set the MSIX vector for a specific virtqueue
- * @mdev: the modern virtio-pci device
- * @index: queue index
- * @vector: the config vector
- *
- * Returns the config vector read from the device
- */
-static u16 vp_modern_queue_vector(struct virtio_pci_modern_device *mdev,
- u16 index, u16 vector)
-{
- struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
-
- vp_iowrite16(index, &cfg->queue_select);
- vp_iowrite16(vector, &cfg->queue_msix_vector);
- /* Flush the write out to device */
- return vp_ioread16(&cfg->queue_msix_vector);
-}
-
-/*
- * vp_modern_queue_address - set the virtqueue address
- * @mdev: the modern virtio-pci device
- * @index: the queue index
- * @desc_addr: address of the descriptor area
- * @driver_addr: address of the driver area
- * @device_addr: address of the device area
- */
-static void vp_modern_queue_address(struct virtio_pci_modern_device *mdev,
- u16 index, u64 desc_addr, u64 driver_addr,
- u64 device_addr)
-{
- struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
-
- vp_iowrite16(index, &cfg->queue_select);
-
- vp_iowrite64_twopart(desc_addr, &cfg->queue_desc_lo,
- &cfg->queue_desc_hi);
- vp_iowrite64_twopart(driver_addr, &cfg->queue_avail_lo,
- &cfg->queue_avail_hi);
- vp_iowrite64_twopart(device_addr, &cfg->queue_used_lo,
- &cfg->queue_used_hi);
-}
-
-/*
- * vp_modern_set_queue_enable - enable a virtqueue
- * @mdev: the modern virtio-pci device
- * @index: the queue index
- * @enable: whether the virtqueue is enable or not
- */
-static void vp_modern_set_queue_enable(struct virtio_pci_modern_device *mdev,
- u16 index, bool enable)
-{
- vp_iowrite16(index, &mdev->common->queue_select);
- vp_iowrite16(enable, &mdev->common->queue_enable);
-}
-
-/*
- * vp_modern_get_queue_enable - enable a virtqueue
- * @mdev: the modern virtio-pci device
- * @index: the queue index
- *
- * Returns whether a virtqueue is enabled or not
- */
-static bool vp_modern_get_queue_enable(struct virtio_pci_modern_device *mdev,
- u16 index)
-{
- vp_iowrite16(index, &mdev->common->queue_select);
-
- return vp_ioread16(&mdev->common->queue_enable);
-}
-
-/*
- * vp_modern_set_queue_size - set size for a virtqueue
- * @mdev: the modern virtio-pci device
- * @index: the queue index
- * @size: the size of the virtqueue
- */
-static void vp_modern_set_queue_size(struct virtio_pci_modern_device *mdev,
- u16 index, u16 size)
-{
- vp_iowrite16(index, &mdev->common->queue_select);
- vp_iowrite16(size, &mdev->common->queue_size);
-
-}
-
-/*
- * vp_modern_get_queue_size - get size for a virtqueue
- * @mdev: the modern virtio-pci device
- * @index: the queue index
- *
- * Returns the size of the virtqueue
- */
-static u16 vp_modern_get_queue_size(struct virtio_pci_modern_device *mdev,
- u16 index)
-{
- vp_iowrite16(index, &mdev->common->queue_select);
-
- return vp_ioread16(&mdev->common->queue_size);
-
-}
-
-/*
- * vp_modern_get_num_queues - get the number of virtqueues
- * @mdev: the modern virtio-pci device
- *
- * Returns the number of virtqueues
- */
-static u16 vp_modern_get_num_queues(struct virtio_pci_modern_device *mdev)
-{
- return vp_ioread16(&mdev->common->num_queues);
-}
-
-/*
- * vp_modern_get_queue_notify_off - get notification offset for a virtqueue
- * @mdev: the modern virtio-pci device
- * @index: the queue index
- *
- * Returns the notification offset for a virtqueue
- */
-static u16 vp_modern_get_queue_notify_off(struct virtio_pci_modern_device
*mdev,
- u16 index)
-{
- vp_iowrite16(index, &mdev->common->queue_select);
-
- return vp_ioread16(&mdev->common->queue_notify_off);
-}
-
/* virtio config->finalize_features() implementation */
static int vp_finalize_features(struct virtio_device *vdev)
{
@@ -429,19 +134,6 @@ static void vp_set(struct virtio_device *vdev, unsigned
offset,
}
}
-/*
- * vp_modern_generation - get the device genreation
- * @mdev: the modern virtio-pci device
- *
- * Returns the genreation read from device
- */
-static u32 vp_modern_generation(struct virtio_pci_modern_device *mdev)
-{
- struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
-
- return vp_ioread8(&cfg->config_generation);
-}
-
static u32 vp_generation(struct virtio_device *vdev)
{
struct virtio_pci_device *vp_dev = to_vp_device(vdev);
@@ -449,19 +141,6 @@ static u32 vp_generation(struct virtio_device *vdev)
return vp_modern_generation(&vp_dev->mdev);
}
-/*
- * vp_modern_get_status - get the device status
- * @mdev: the modern virtio-pci device
- *
- * Returns the status read from device
- */
-static u8 vp_modern_get_status(struct virtio_pci_modern_device *mdev)
-{
- struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
-
- return vp_ioread8(&cfg->device_status);
-}
-
/* config->{get,set}_status() implementations */
static u8 vp_get_status(struct virtio_device *vdev)
{
@@ -470,19 +149,6 @@ static u8 vp_get_status(struct virtio_device *vdev)
return vp_modern_get_status(&vp_dev->mdev);
}
-/*
- * vp_modern_set_status - set status to device
- * @mdev: the modern virtio-pci device
- * @status: the status set to device
- */
-static void vp_modern_set_status(struct virtio_pci_modern_device *mdev,
- u8 status)
-{
- struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
-
- vp_iowrite8(status, &cfg->device_status);
-}
-
static void vp_set_status(struct virtio_device *vdev, u8 status)
{
struct virtio_pci_device *vp_dev = to_vp_device(vdev);
@@ -510,25 +176,6 @@ static void vp_reset(struct virtio_device *vdev)
vp_synchronize_vectors(vdev);
}
-/*
- * vp_modern_config_vector - set the vector for config interrupt
- * @mdev: the modern virtio-pci device
- * @vector: the config vector
- *
- * Returns the config vector read from the device
- */
-static u16 vp_modern_config_vector(struct virtio_pci_modern_device *mdev,
- u16 vector)
-{
- struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
-
- /* Setup the vector used for configuration events */
- vp_iowrite16(vector, &cfg->msix_config);
- /* Verify we had enough resources to assign the vector */
- /* Will also flush the write out to device */
- return vp_ioread16(&cfg->msix_config);
-}
-
static u16 vp_config_vector(struct virtio_pci_device *vp_dev, u16 vector)
{
return vp_modern_config_vector(&vp_dev->mdev, vector);
@@ -789,253 +436,6 @@ static const struct virtio_config_ops
virtio_pci_config_ops = {
.get_shm_region = vp_get_shm_region,
};
-/**
- * virtio_pci_find_capability - walk capabilities to find device info.
- * @dev: the pci device
- * @cfg_type: the VIRTIO_PCI_CAP_* value we seek
- * @ioresource_types: IORESOURCE_MEM and/or IORESOURCE_IO.
- * @bars: the bitmask of BARs
- *
- * Returns offset of the capability, or 0.
- */
-static inline int virtio_pci_find_capability(struct pci_dev *dev, u8 cfg_type,
- u32 ioresource_types, int *bars)
-{
- int pos;
-
- for (pos = pci_find_capability(dev, PCI_CAP_ID_VNDR);
- pos > 0;
- pos = pci_find_next_capability(dev, pos, PCI_CAP_ID_VNDR)) {
- u8 type, bar;
- pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap,
- cfg_type),
- &type);
- pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap,
- bar),
- &bar);
-
- /* Ignore structures with reserved BAR values */
- if (bar > 0x5)
- continue;
-
- if (type == cfg_type) {
- if (pci_resource_len(dev, bar) &&
- pci_resource_flags(dev, bar) & ioresource_types) {
- *bars |= (1 << bar);
- return pos;
- }
- }
- }
- return 0;
-}
-
-/* This is part of the ABI. Don't screw with it. */
-static inline void check_offsets(void)
-{
- /* Note: disk space was harmed in compilation of this function. */
- BUILD_BUG_ON(VIRTIO_PCI_CAP_VNDR !- offsetof(struct virtio_pci_cap,
cap_vndr));
- BUILD_BUG_ON(VIRTIO_PCI_CAP_NEXT !- offsetof(struct virtio_pci_cap,
cap_next));
- BUILD_BUG_ON(VIRTIO_PCI_CAP_LEN !- offsetof(struct virtio_pci_cap,
cap_len));
- BUILD_BUG_ON(VIRTIO_PCI_CAP_CFG_TYPE !- offsetof(struct virtio_pci_cap,
cfg_type));
- BUILD_BUG_ON(VIRTIO_PCI_CAP_BAR !- offsetof(struct virtio_pci_cap,
bar));
- BUILD_BUG_ON(VIRTIO_PCI_CAP_OFFSET !- offsetof(struct virtio_pci_cap,
offset));
- BUILD_BUG_ON(VIRTIO_PCI_CAP_LENGTH !- offsetof(struct virtio_pci_cap,
length));
- BUILD_BUG_ON(VIRTIO_PCI_NOTIFY_CAP_MULT !- offsetof(struct
virtio_pci_notify_cap,
- notify_off_multiplier));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_DFSELECT !- offsetof(struct
virtio_pci_common_cfg,
- device_feature_select));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_DF !- offsetof(struct
virtio_pci_common_cfg, device_feature));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_GFSELECT !- offsetof(struct
virtio_pci_common_cfg,
- guest_feature_select));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_GF !- offsetof(struct
virtio_pci_common_cfg, guest_feature));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_MSIX !- offsetof(struct
virtio_pci_common_cfg, msix_config));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_NUMQ !- offsetof(struct
virtio_pci_common_cfg, num_queues));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_STATUS !- offsetof(struct
virtio_pci_common_cfg, device_status));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_CFGGENERATION !- offsetof(struct
virtio_pci_common_cfg, config_generation));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SELECT !- offsetof(struct
virtio_pci_common_cfg, queue_select));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SIZE !- offsetof(struct
virtio_pci_common_cfg, queue_size));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_MSIX !- offsetof(struct
virtio_pci_common_cfg, queue_msix_vector));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_ENABLE !- offsetof(struct
virtio_pci_common_cfg, queue_enable));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_NOFF !- offsetof(struct
virtio_pci_common_cfg, queue_notify_off));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCLO !- offsetof(struct
virtio_pci_common_cfg, queue_desc_lo));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCHI !- offsetof(struct
virtio_pci_common_cfg, queue_desc_hi));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILLO !- offsetof(struct
virtio_pci_common_cfg, queue_avail_lo));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILHI !- offsetof(struct
virtio_pci_common_cfg, queue_avail_hi));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDLO !- offsetof(struct
virtio_pci_common_cfg, queue_used_lo));
- BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDHI !- offsetof(struct
virtio_pci_common_cfg, queue_used_hi));
-}
-
-/*
- * vp_modern_probe: probe the modern virtio pci device, note that the
- * caller is required to enable PCI device before calling this function.
- * @mdev: the modern virtio-pci device
- *
- * Return 0 on succeed otherwise fail
- */
-static int vp_modern_probe(struct virtio_pci_modern_device *mdev)
-{
- struct pci_dev *pci_dev = mdev->pci_dev;
- int err, common, isr, notify, device;
- u32 notify_length;
- u32 notify_offset;
-
- check_offsets();
-
- mdev->pci_dev = pci_dev;
-
- /* We only own devices >= 0x1000 and <= 0x107f: leave the rest. */
- if (pci_dev->device < 0x1000 || pci_dev->device > 0x107f)
- return -ENODEV;
-
- if (pci_dev->device < 0x1040) {
- /* Transitional devices: use the PCI subsystem device id as
- * virtio device id, same as legacy driver always did.
- */
- mdev->id.device = pci_dev->subsystem_device;
- } else {
- /* Modern devices: simply use PCI device id, but start from 0x1040. */
- mdev->id.device = pci_dev->device - 0x1040;
- }
- mdev->id.vendor = pci_dev->subsystem_vendor;
-
- /* check for a common config: if not, use legacy mode (bar 0). */
- common = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_COMMON_CFG,
- IORESOURCE_IO | IORESOURCE_MEM,
- &mdev->modern_bars);
- if (!common) {
- dev_info(&pci_dev->dev,
- "virtio_pci: leaving for legacy driver\n");
- return -ENODEV;
- }
-
- /* If common is there, these should be too... */
- isr = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_ISR_CFG,
- IORESOURCE_IO | IORESOURCE_MEM,
- &mdev->modern_bars);
- notify = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_NOTIFY_CFG,
- IORESOURCE_IO | IORESOURCE_MEM,
- &mdev->modern_bars);
- if (!isr || !notify) {
- dev_err(&pci_dev->dev,
- "virtio_pci: missing capabilities %i/%i/%i\n",
- common, isr, notify);
- return -EINVAL;
- }
-
- err = dma_set_mask_and_coherent(&pci_dev->dev, DMA_BIT_MASK(64));
- if (err)
- err = dma_set_mask_and_coherent(&pci_dev->dev,
- DMA_BIT_MASK(32));
- if (err)
- dev_warn(&pci_dev->dev, "Failed to enable 64-bit or 32-bit DMA.
Trying to continue, but this might not work.\n");
-
- /* Device capability is only mandatory for devices that have
- * device-specific configuration.
- */
- device = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_DEVICE_CFG,
- IORESOURCE_IO | IORESOURCE_MEM,
- &mdev->modern_bars);
-
- err = pci_request_selected_regions(pci_dev, mdev->modern_bars,
- "virtio-pci-modern");
- if (err)
- return err;
-
- err = -EINVAL;
- mdev->common = vp_modern_map_capability(mdev, common,
- sizeof(struct virtio_pci_common_cfg), 4,
- 0, sizeof(struct virtio_pci_common_cfg),
- NULL);
- if (!mdev->common)
- goto err_map_common;
- mdev->isr = vp_modern_map_capability(mdev, isr, sizeof(u8), 1,
- 0, 1,
- NULL);
- if (!mdev->isr)
- goto err_map_isr;
-
- /* Read notify_off_multiplier from config space. */
- pci_read_config_dword(pci_dev,
- notify + offsetof(struct virtio_pci_notify_cap,
- notify_off_multiplier),
- &mdev->notify_offset_multiplier);
- /* Read notify length and offset from config space. */
- pci_read_config_dword(pci_dev,
- notify + offsetof(struct virtio_pci_notify_cap,
- cap.length),
- ¬ify_length);
-
- pci_read_config_dword(pci_dev,
- notify + offsetof(struct virtio_pci_notify_cap,
- cap.offset),
- ¬ify_offset);
-
- /* We don't know how many VQs we'll map, ahead of the time.
- * If notify length is small, map it all now.
- * Otherwise, map each VQ individually later.
- */
- if ((u64)notify_length + (notify_offset % PAGE_SIZE) <= PAGE_SIZE) {
- mdev->notify_base = vp_modern_map_capability(mdev, notify,
- 2, 2,
- 0, notify_length,
- &mdev->notify_len);
- if (!mdev->notify_base)
- goto err_map_notify;
- } else {
- mdev->notify_map_cap = notify;
- }
-
- /* Again, we don't know how much we should map, but PAGE_SIZE
- * is more than enough for all existing devices.
- */
- if (device) {
- mdev->device = vp_modern_map_capability(mdev, device, 0, 4,
- 0, PAGE_SIZE,
- &mdev->device_len);
- if (!mdev->device)
- goto err_map_device;
- }
-
- return 0;
-
-err_map_device:
- if (mdev->notify_base)
- pci_iounmap(pci_dev, mdev->notify_base);
-err_map_notify:
- pci_iounmap(pci_dev, mdev->isr);
-err_map_isr:
- pci_iounmap(pci_dev, mdev->common);
-err_map_common:
- return err;
-}
-
/* the PCI probing function */
int virtio_pci_modern_probe(struct virtio_pci_device *vp_dev)
{
@@ -1063,23 +463,6 @@ int virtio_pci_modern_probe(struct virtio_pci_device
*vp_dev)
return 0;
}
-/*
- * vp_modern_probe: remove and cleanup the modern virtio pci device
- * @mdev: the modern virtio-pci device
- */
-static void vp_modern_remove(struct virtio_pci_modern_device *mdev)
-{
- struct pci_dev *pci_dev = mdev->pci_dev;
-
- if (mdev->device)
- pci_iounmap(pci_dev, mdev->device);
- if (mdev->notify_base)
- pci_iounmap(pci_dev, mdev->notify_base);
- pci_iounmap(pci_dev, mdev->isr);
- pci_iounmap(pci_dev, mdev->common);
- pci_release_selected_regions(pci_dev, mdev->modern_bars);
-}
-
void virtio_pci_modern_remove(struct virtio_pci_device *vp_dev)
{
struct virtio_pci_modern_device *mdev = &vp_dev->mdev;
diff --git a/drivers/virtio/virtio_pci_modern_dev.c
b/drivers/virtio/virtio_pci_modern_dev.c
new file mode 100644
index 000000000000..cbd667496bb1
--- /dev/null
+++ b/drivers/virtio/virtio_pci_modern_dev.c
@@ -0,0 +1,599 @@
+// SPDX-License-Identifier: GPL-2.0-or-later
+
+#include <linux/virtio_pci_modern.h>
+#include <linux/module.h>
+#include <linux/pci.h>
+
+/*
+ * vp_modern_map_capability - map a part of virtio pci capability
+ * @mdev: the modern virtio-pci device
+ * @off: offset of the capability
+ * @minlen: minimal length of the capability
+ * @align: align requirement
+ * @start: start from the capability
+ * @size: map size
+ * @len: the length that is actually mapped
+ *
+ * Returns the io address of for the part of the capability
+ */
+void __iomem *vp_modern_map_capability(struct virtio_pci_modern_device *mdev,
int off,
+ size_t minlen,
+ u32 align,
+ u32 start, u32 size,
+ size_t *len)
+{
+ struct pci_dev *dev = mdev->pci_dev;
+ u8 bar;
+ u32 offset, length;
+ void __iomem *p;
+
+ pci_read_config_byte(dev, off + offsetof(struct virtio_pci_cap,
+ bar),
+ &bar);
+ pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, offset),
+ &offset);
+ pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, length),
+ &length);
+
+ if (length <= start) {
+ dev_err(&dev->dev,
+ "virtio_pci: bad capability len %u (>%u expected)\n",
+ length, start);
+ return NULL;
+ }
+
+ if (length - start < minlen) {
+ dev_err(&dev->dev,
+ "virtio_pci: bad capability len %u (>=%zu expected)\n",
+ length, minlen);
+ return NULL;
+ }
+
+ length -= start;
+
+ if (start + offset < offset) {
+ dev_err(&dev->dev,
+ "virtio_pci: map wrap-around %u+%u\n",
+ start, offset);
+ return NULL;
+ }
+
+ offset += start;
+
+ if (offset & (align - 1)) {
+ dev_err(&dev->dev,
+ "virtio_pci: offset %u not aligned to %u\n",
+ offset, align);
+ return NULL;
+ }
+
+ if (length > size)
+ length = size;
+
+ if (len)
+ *len = length;
+
+ if (minlen + offset < minlen ||
+ minlen + offset > pci_resource_len(dev, bar)) {
+ dev_err(&dev->dev,
+ "virtio_pci: map virtio %zu@%u "
+ "out of range on bar %i length %lu\n",
+ minlen, offset,
+ bar, (unsigned long)pci_resource_len(dev, bar));
+ return NULL;
+ }
+
+ p = pci_iomap_range(dev, bar, offset, length);
+ if (!p)
+ dev_err(&dev->dev,
+ "virtio_pci: unable to map virtio %u@%u on bar %i\n",
+ length, offset, bar);
+ return p;
+}
+EXPORT_SYMBOL_GPL(vp_modern_map_capability);
+
+/**
+ * virtio_pci_find_capability - walk capabilities to find device info.
+ * @dev: the pci device
+ * @cfg_type: the VIRTIO_PCI_CAP_* value we seek
+ * @ioresource_types: IORESOURCE_MEM and/or IORESOURCE_IO.
+ * @bars: the bitmask of BARs
+ *
+ * Returns offset of the capability, or 0.
+ */
+static inline int virtio_pci_find_capability(struct pci_dev *dev, u8 cfg_type,
+ u32 ioresource_types, int *bars)
+{
+ int pos;
+
+ for (pos = pci_find_capability(dev, PCI_CAP_ID_VNDR);
+ pos > 0;
+ pos = pci_find_next_capability(dev, pos, PCI_CAP_ID_VNDR)) {
+ u8 type, bar;
+ pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap,
+ cfg_type),
+ &type);
+ pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap,
+ bar),
+ &bar);
+
+ /* Ignore structures with reserved BAR values */
+ if (bar > 0x5)
+ continue;
+
+ if (type == cfg_type) {
+ if (pci_resource_len(dev, bar) &&
+ pci_resource_flags(dev, bar) & ioresource_types) {
+ *bars |= (1 << bar);
+ return pos;
+ }
+ }
+ }
+ return 0;
+}
+
+/* This is part of the ABI. Don't screw with it. */
+static inline void check_offsets(void)
+{
+ /* Note: disk space was harmed in compilation of this function. */
+ BUILD_BUG_ON(VIRTIO_PCI_CAP_VNDR !+ offsetof(struct virtio_pci_cap,
cap_vndr));
+ BUILD_BUG_ON(VIRTIO_PCI_CAP_NEXT !+ offsetof(struct virtio_pci_cap,
cap_next));
+ BUILD_BUG_ON(VIRTIO_PCI_CAP_LEN !+ offsetof(struct virtio_pci_cap,
cap_len));
+ BUILD_BUG_ON(VIRTIO_PCI_CAP_CFG_TYPE !+ offsetof(struct virtio_pci_cap,
cfg_type));
+ BUILD_BUG_ON(VIRTIO_PCI_CAP_BAR !+ offsetof(struct virtio_pci_cap,
bar));
+ BUILD_BUG_ON(VIRTIO_PCI_CAP_OFFSET !+ offsetof(struct virtio_pci_cap,
offset));
+ BUILD_BUG_ON(VIRTIO_PCI_CAP_LENGTH !+ offsetof(struct virtio_pci_cap,
length));
+ BUILD_BUG_ON(VIRTIO_PCI_NOTIFY_CAP_MULT !+ offsetof(struct
virtio_pci_notify_cap,
+ notify_off_multiplier));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_DFSELECT !+ offsetof(struct
virtio_pci_common_cfg,
+ device_feature_select));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_DF !+ offsetof(struct
virtio_pci_common_cfg, device_feature));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_GFSELECT !+ offsetof(struct
virtio_pci_common_cfg,
+ guest_feature_select));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_GF !+ offsetof(struct
virtio_pci_common_cfg, guest_feature));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_MSIX !+ offsetof(struct
virtio_pci_common_cfg, msix_config));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_NUMQ !+ offsetof(struct
virtio_pci_common_cfg, num_queues));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_STATUS !+ offsetof(struct
virtio_pci_common_cfg, device_status));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_CFGGENERATION !+ offsetof(struct
virtio_pci_common_cfg, config_generation));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SELECT !+ offsetof(struct
virtio_pci_common_cfg, queue_select));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SIZE !+ offsetof(struct
virtio_pci_common_cfg, queue_size));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_MSIX !+ offsetof(struct
virtio_pci_common_cfg, queue_msix_vector));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_ENABLE !+ offsetof(struct
virtio_pci_common_cfg, queue_enable));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_NOFF !+ offsetof(struct
virtio_pci_common_cfg, queue_notify_off));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCLO !+ offsetof(struct
virtio_pci_common_cfg, queue_desc_lo));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCHI !+ offsetof(struct
virtio_pci_common_cfg, queue_desc_hi));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILLO !+ offsetof(struct
virtio_pci_common_cfg, queue_avail_lo));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILHI !+ offsetof(struct
virtio_pci_common_cfg, queue_avail_hi));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDLO !+ offsetof(struct
virtio_pci_common_cfg, queue_used_lo));
+ BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDHI !+ offsetof(struct
virtio_pci_common_cfg, queue_used_hi));
+}
+
+/*
+ * vp_modern_probe: probe the modern virtio pci device, note that the
+ * caller is required to enable PCI device before calling this function.
+ * @mdev: the modern virtio-pci device
+ *
+ * Return 0 on succeed otherwise fail
+ */
+int vp_modern_probe(struct virtio_pci_modern_device *mdev)
+{
+ struct pci_dev *pci_dev = mdev->pci_dev;
+ int err, common, isr, notify, device;
+ u32 notify_length;
+ u32 notify_offset;
+
+ check_offsets();
+
+ mdev->pci_dev = pci_dev;
+
+ /* We only own devices >= 0x1000 and <= 0x107f: leave the rest. */
+ if (pci_dev->device < 0x1000 || pci_dev->device > 0x107f)
+ return -ENODEV;
+
+ if (pci_dev->device < 0x1040) {
+ /* Transitional devices: use the PCI subsystem device id as
+ * virtio device id, same as legacy driver always did.
+ */
+ mdev->id.device = pci_dev->subsystem_device;
+ } else {
+ /* Modern devices: simply use PCI device id, but start from 0x1040. */
+ mdev->id.device = pci_dev->device - 0x1040;
+ }
+ mdev->id.vendor = pci_dev->subsystem_vendor;
+
+ /* check for a common config: if not, use legacy mode (bar 0). */
+ common = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_COMMON_CFG,
+ IORESOURCE_IO | IORESOURCE_MEM,
+ &mdev->modern_bars);
+ if (!common) {
+ dev_info(&pci_dev->dev,
+ "virtio_pci: leaving for legacy driver\n");
+ return -ENODEV;
+ }
+
+ /* If common is there, these should be too... */
+ isr = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_ISR_CFG,
+ IORESOURCE_IO | IORESOURCE_MEM,
+ &mdev->modern_bars);
+ notify = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_NOTIFY_CFG,
+ IORESOURCE_IO | IORESOURCE_MEM,
+ &mdev->modern_bars);
+ if (!isr || !notify) {
+ dev_err(&pci_dev->dev,
+ "virtio_pci: missing capabilities %i/%i/%i\n",
+ common, isr, notify);
+ return -EINVAL;
+ }
+
+ err = dma_set_mask_and_coherent(&pci_dev->dev, DMA_BIT_MASK(64));
+ if (err)
+ err = dma_set_mask_and_coherent(&pci_dev->dev,
+ DMA_BIT_MASK(32));
+ if (err)
+ dev_warn(&pci_dev->dev, "Failed to enable 64-bit or 32-bit DMA.
Trying to continue, but this might not work.\n");
+
+ /* Device capability is only mandatory for devices that have
+ * device-specific configuration.
+ */
+ device = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_DEVICE_CFG,
+ IORESOURCE_IO | IORESOURCE_MEM,
+ &mdev->modern_bars);
+
+ err = pci_request_selected_regions(pci_dev, mdev->modern_bars,
+ "virtio-pci-modern");
+ if (err)
+ return err;
+
+ err = -EINVAL;
+ mdev->common = vp_modern_map_capability(mdev, common,
+ sizeof(struct virtio_pci_common_cfg), 4,
+ 0, sizeof(struct virtio_pci_common_cfg),
+ NULL);
+ if (!mdev->common)
+ goto err_map_common;
+ mdev->isr = vp_modern_map_capability(mdev, isr, sizeof(u8), 1,
+ 0, 1,
+ NULL);
+ if (!mdev->isr)
+ goto err_map_isr;
+
+ /* Read notify_off_multiplier from config space. */
+ pci_read_config_dword(pci_dev,
+ notify + offsetof(struct virtio_pci_notify_cap,
+ notify_off_multiplier),
+ &mdev->notify_offset_multiplier);
+ /* Read notify length and offset from config space. */
+ pci_read_config_dword(pci_dev,
+ notify + offsetof(struct virtio_pci_notify_cap,
+ cap.length),
+ ¬ify_length);
+
+ pci_read_config_dword(pci_dev,
+ notify + offsetof(struct virtio_pci_notify_cap,
+ cap.offset),
+ ¬ify_offset);
+
+ /* We don't know how many VQs we'll map, ahead of the time.
+ * If notify length is small, map it all now.
+ * Otherwise, map each VQ individually later.
+ */
+ if ((u64)notify_length + (notify_offset % PAGE_SIZE) <= PAGE_SIZE) {
+ mdev->notify_base = vp_modern_map_capability(mdev, notify,
+ 2, 2,
+ 0, notify_length,
+ &mdev->notify_len);
+ if (!mdev->notify_base)
+ goto err_map_notify;
+ } else {
+ mdev->notify_map_cap = notify;
+ }
+
+ /* Again, we don't know how much we should map, but PAGE_SIZE
+ * is more than enough for all existing devices.
+ */
+ if (device) {
+ mdev->device = vp_modern_map_capability(mdev, device, 0, 4,
+ 0, PAGE_SIZE,
+ &mdev->device_len);
+ if (!mdev->device)
+ goto err_map_device;
+ }
+
+ return 0;
+
+err_map_device:
+ if (mdev->notify_base)
+ pci_iounmap(pci_dev, mdev->notify_base);
+err_map_notify:
+ pci_iounmap(pci_dev, mdev->isr);
+err_map_isr:
+ pci_iounmap(pci_dev, mdev->common);
+err_map_common:
+ return err;
+}
+EXPORT_SYMBOL_GPL(vp_modern_probe);
+
+/*
+ * vp_modern_probe: remove and cleanup the modern virtio pci device
+ * @mdev: the modern virtio-pci device
+ */
+void vp_modern_remove(struct virtio_pci_modern_device *mdev)
+{
+ struct pci_dev *pci_dev = mdev->pci_dev;
+
+ if (mdev->device)
+ pci_iounmap(pci_dev, mdev->device);
+ if (mdev->notify_base)
+ pci_iounmap(pci_dev, mdev->notify_base);
+ pci_iounmap(pci_dev, mdev->isr);
+ pci_iounmap(pci_dev, mdev->common);
+ pci_release_selected_regions(pci_dev, mdev->modern_bars);
+}
+EXPORT_SYMBOL_GPL(vp_modern_remove);
+
+/*
+ * vp_modern_get_features - get features from device
+ * @mdev: the modern virtio-pci device
+ *
+ * Returns the features read from the device
+ */
+u64 vp_modern_get_features(struct virtio_pci_modern_device *mdev)
+{
+ struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
+
+ u64 features;
+
+ vp_iowrite32(0, &cfg->device_feature_select);
+ features = vp_ioread32(&cfg->device_feature);
+ vp_iowrite32(1, &cfg->device_feature_select);
+ features |= ((u64)vp_ioread32(&cfg->device_feature) << 32);
+
+ return features;
+}
+EXPORT_SYMBOL_GPL(vp_modern_get_features);
+
+/*
+ * vp_modern_set_features - set features to device
+ * @mdev: the modern virtio-pci device
+ * @features: the features set to device
+ */
+void vp_modern_set_features(struct virtio_pci_modern_device *mdev,
+ u64 features)
+{
+ struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
+
+ vp_iowrite32(0, &cfg->guest_feature_select);
+ vp_iowrite32((u32)features, &cfg->guest_feature);
+ vp_iowrite32(1, &cfg->guest_feature_select);
+ vp_iowrite32(features >> 32, &cfg->guest_feature);
+}
+EXPORT_SYMBOL_GPL(vp_modern_set_features);
+
+/*
+ * vp_modern_generation - get the device genreation
+ * @mdev: the modern virtio-pci device
+ *
+ * Returns the genreation read from device
+ */
+u32 vp_modern_generation(struct virtio_pci_modern_device *mdev)
+{
+ struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
+
+ return vp_ioread8(&cfg->config_generation);
+}
+EXPORT_SYMBOL_GPL(vp_modern_generation);
+
+/*
+ * vp_modern_get_status - get the device status
+ * @mdev: the modern virtio-pci device
+ *
+ * Returns the status read from device
+ */
+u8 vp_modern_get_status(struct virtio_pci_modern_device *mdev)
+{
+ struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
+
+ return vp_ioread8(&cfg->device_status);
+}
+EXPORT_SYMBOL_GPL(vp_modern_get_status);
+
+/*
+ * vp_modern_set_status - set status to device
+ * @mdev: the modern virtio-pci device
+ * @status: the status set to device
+ */
+void vp_modern_set_status(struct virtio_pci_modern_device *mdev,
+ u8 status)
+{
+ struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
+
+ vp_iowrite8(status, &cfg->device_status);
+}
+EXPORT_SYMBOL_GPL(vp_modern_set_status);
+
+/*
+ * vp_modern_queue_vector - set the MSIX vector for a specific virtqueue
+ * @mdev: the modern virtio-pci device
+ * @index: queue index
+ * @vector: the config vector
+ *
+ * Returns the config vector read from the device
+ */
+u16 vp_modern_queue_vector(struct virtio_pci_modern_device *mdev,
+ u16 index, u16 vector)
+{
+ struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
+
+ vp_iowrite16(index, &cfg->queue_select);
+ vp_iowrite16(vector, &cfg->queue_msix_vector);
+ /* Flush the write out to device */
+ return vp_ioread16(&cfg->queue_msix_vector);
+}
+EXPORT_SYMBOL_GPL(vp_modern_queue_vector);
+
+/*
+ * vp_modern_config_vector - set the vector for config interrupt
+ * @mdev: the modern virtio-pci device
+ * @vector: the config vector
+ *
+ * Returns the config vector read from the device
+ */
+u16 vp_modern_config_vector(struct virtio_pci_modern_device *mdev,
+ u16 vector)
+{
+ struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
+
+ /* Setup the vector used for configuration events */
+ vp_iowrite16(vector, &cfg->msix_config);
+ /* Verify we had enough resources to assign the vector */
+ /* Will also flush the write out to device */
+ return vp_ioread16(&cfg->msix_config);
+}
+EXPORT_SYMBOL_GPL(vp_modern_config_vector);
+
+/*
+ * vp_modern_queue_address - set the virtqueue address
+ * @mdev: the modern virtio-pci device
+ * @index: the queue index
+ * @desc_addr: address of the descriptor area
+ * @driver_addr: address of the driver area
+ * @device_addr: address of the device area
+ */
+void vp_modern_queue_address(struct virtio_pci_modern_device *mdev,
+ u16 index, u64 desc_addr, u64 driver_addr,
+ u64 device_addr)
+{
+ struct virtio_pci_common_cfg __iomem *cfg = mdev->common;
+
+ vp_iowrite16(index, &cfg->queue_select);
+
+ vp_iowrite64_twopart(desc_addr, &cfg->queue_desc_lo,
+ &cfg->queue_desc_hi);
+ vp_iowrite64_twopart(driver_addr, &cfg->queue_avail_lo,
+ &cfg->queue_avail_hi);
+ vp_iowrite64_twopart(device_addr, &cfg->queue_used_lo,
+ &cfg->queue_used_hi);
+}
+EXPORT_SYMBOL_GPL(vp_modern_queue_address);
+
+/*
+ * vp_modern_set_queue_enable - enable a virtqueue
+ * @mdev: the modern virtio-pci device
+ * @index: the queue index
+ * @enable: whether the virtqueue is enable or not
+ */
+void vp_modern_set_queue_enable(struct virtio_pci_modern_device *mdev,
+ u16 index, bool enable)
+{
+ vp_iowrite16(index, &mdev->common->queue_select);
+ vp_iowrite16(enable, &mdev->common->queue_enable);
+}
+EXPORT_SYMBOL_GPL(vp_modern_set_queue_enable);
+
+/*
+ * vp_modern_get_queue_enable - enable a virtqueue
+ * @mdev: the modern virtio-pci device
+ * @index: the queue index
+ *
+ * Returns whether a virtqueue is enabled or not
+ */
+bool vp_modern_get_queue_enable(struct virtio_pci_modern_device *mdev,
+ u16 index)
+{
+ vp_iowrite16(index, &mdev->common->queue_select);
+
+ return vp_ioread16(&mdev->common->queue_enable);
+}
+EXPORT_SYMBOL_GPL(vp_modern_get_queue_enable);
+
+/*
+ * vp_modern_set_queue_size - set size for a virtqueue
+ * @mdev: the modern virtio-pci device
+ * @index: the queue index
+ * @size: the size of the virtqueue
+ */
+void vp_modern_set_queue_size(struct virtio_pci_modern_device *mdev,
+ u16 index, u16 size)
+{
+ vp_iowrite16(index, &mdev->common->queue_select);
+ vp_iowrite16(size, &mdev->common->queue_size);
+
+}
+EXPORT_SYMBOL_GPL(vp_modern_set_queue_size);
+
+/*
+ * vp_modern_get_queue_size - get size for a virtqueue
+ * @mdev: the modern virtio-pci device
+ * @index: the queue index
+ *
+ * Returns the size of the virtqueue
+ */
+u16 vp_modern_get_queue_size(struct virtio_pci_modern_device *mdev,
+ u16 index)
+{
+ vp_iowrite16(index, &mdev->common->queue_select);
+
+ return vp_ioread16(&mdev->common->queue_size);
+
+}
+EXPORT_SYMBOL_GPL(vp_modern_get_queue_size);
+
+/*
+ * vp_modern_get_num_queues - get the number of virtqueues
+ * @mdev: the modern virtio-pci device
+ *
+ * Returns the number of virtqueues
+ */
+u16 vp_modern_get_num_queues(struct virtio_pci_modern_device *mdev)
+{
+ return vp_ioread16(&mdev->common->num_queues);
+}
+EXPORT_SYMBOL_GPL(vp_modern_get_num_queues);
+
+/*
+ * vp_modern_get_queue_notify_off - get notification offset for a virtqueue
+ * @mdev: the modern virtio-pci device
+ * @index: the queue index
+ *
+ * Returns the notification offset for a virtqueue
+ */
+u16 vp_modern_get_queue_notify_off(struct virtio_pci_modern_device *mdev,
+ u16 index)
+{
+ vp_iowrite16(index, &mdev->common->queue_select);
+
+ return vp_ioread16(&mdev->common->queue_notify_off);
+}
+EXPORT_SYMBOL_GPL(vp_modern_get_queue_notify_off);
+
+MODULE_VERSION("0.1");
+MODULE_DESCRIPTION("Modern Virtio PCI Device");
+MODULE_AUTHOR("Jason Wang <jasowang at redhat.com>");
+MODULE_LICENSE("GPL");
diff --git a/include/linux/virtio_pci_modern.h
b/include/linux/virtio_pci_modern.h
new file mode 100644
index 000000000000..f26acbeec965
--- /dev/null
+++ b/include/linux/virtio_pci_modern.h
@@ -0,0 +1,111 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+#ifndef _LINUX_VIRTIO_PCI_MODERN_H
+#define _LINUX_VIRTIO_PCI_MODERN_H
+
+#include <linux/pci.h>
+#include <linux/virtio_pci.h>
+
+struct virtio_pci_modern_device {
+ struct pci_dev *pci_dev;
+
+ struct virtio_pci_common_cfg __iomem *common;
+ /* Device-specific data (non-legacy mode) */
+ void __iomem *device;
+ /* Base of vq notifications (non-legacy mode). */
+ void __iomem *notify_base;
+ /* Where to read and clear interrupt */
+ u8 __iomem *isr;
+
+ /* So we can sanity-check accesses. */
+ size_t notify_len;
+ size_t device_len;
+
+ /* Capability for when we need to map notifications per-vq. */
+ int notify_map_cap;
+
+ /* Multiply queue_notify_off by this value. (non-legacy mode). */
+ u32 notify_offset_multiplier;
+
+ int modern_bars;
+
+ struct virtio_device_id id;
+};
+
+/*
+ * Type-safe wrappers for io accesses.
+ * Use these to enforce at compile time the following spec requirement:
+ *
+ * The driver MUST access each field using the ?natural? access
+ * method, i.e. 32-bit accesses for 32-bit fields, 16-bit accesses
+ * for 16-bit fields and 8-bit accesses for 8-bit fields.
+ */
+static inline u8 vp_ioread8(const u8 __iomem *addr)
+{
+ return ioread8(addr);
+}
+static inline u16 vp_ioread16 (const __le16 __iomem *addr)
+{
+ return ioread16(addr);
+}
+
+static inline u32 vp_ioread32(const __le32 __iomem *addr)
+{
+ return ioread32(addr);
+}
+
+static inline void vp_iowrite8(u8 value, u8 __iomem *addr)
+{
+ iowrite8(value, addr);
+}
+
+static inline void vp_iowrite16(u16 value, __le16 __iomem *addr)
+{
+ iowrite16(value, addr);
+}
+
+static inline void vp_iowrite32(u32 value, __le32 __iomem *addr)
+{
+ iowrite32(value, addr);
+}
+
+static inline void vp_iowrite64_twopart(u64 val,
+ __le32 __iomem *lo,
+ __le32 __iomem *hi)
+{
+ vp_iowrite32((u32)val, lo);
+ vp_iowrite32(val >> 32, hi);
+}
+
+u64 vp_modern_get_features(struct virtio_pci_modern_device *mdev);
+void vp_modern_set_features(struct virtio_pci_modern_device *mdev,
+ u64 features);
+u32 vp_modern_generation(struct virtio_pci_modern_device *mdev);
+u8 vp_modern_get_status(struct virtio_pci_modern_device *mdev);
+void vp_modern_set_status(struct virtio_pci_modern_device *mdev,
+ u8 status);
+u16 vp_modern_queue_vector(struct virtio_pci_modern_device *mdev,
+ u16 idx, u16 vector);
+u16 vp_modern_config_vector(struct virtio_pci_modern_device *mdev,
+ u16 vector);
+void vp_modern_queue_address(struct virtio_pci_modern_device *mdev,
+ u16 index, u64 desc_addr, u64 driver_addr,
+ u64 device_addr);
+void vp_modern_set_queue_enable(struct virtio_pci_modern_device *mdev,
+ u16 idx, bool enable);
+bool vp_modern_get_queue_enable(struct virtio_pci_modern_device *mdev,
+ u16 idx);
+void vp_modern_set_queue_size(struct virtio_pci_modern_device *mdev,
+ u16 idx, u16 size);
+u16 vp_modern_get_queue_size(struct virtio_pci_modern_device *mdev,
+ u16 idx);
+u16 vp_modern_get_num_queues(struct virtio_pci_modern_device *mdev);
+u16 vp_modern_get_queue_notify_off(struct virtio_pci_modern_device *mdev,
+ u16 idx);
+void __iomem *vp_modern_map_capability(struct virtio_pci_modern_device *mdev,
int off,
+ size_t minlen,
+ u32 align,
+ u32 start, u32 size,
+ size_t *len);
+int vp_modern_probe(struct virtio_pci_modern_device *mdev);
+void vp_modern_remove(struct virtio_pci_modern_device *mdev);
+#endif
--
2.25.1
Michael S. Tsirkin
2021-Feb-05 15:34 UTC
[PATCH V3 16/19] virtio-pci: introduce modern device module
On Mon, Jan 04, 2021 at 02:55:00PM +0800, Jason Wang wrote:> Signed-off-by: Jason Wang <jasowang at redhat.com>I don't exactly get why we need to split the modern driver out, and it can confuse people who are used to be seeing virtio-pci. The vdpa thing so far looks like a development tool, why do we care that it depends on a bit of extra code?> --- > drivers/virtio/Kconfig | 10 +- > drivers/virtio/Makefile | 1 + > drivers/virtio/virtio_pci_common.h | 27 +- > drivers/virtio/virtio_pci_modern.c | 617 ------------------------- > drivers/virtio/virtio_pci_modern_dev.c | 599 ++++++++++++++++++++++++ > include/linux/virtio_pci_modern.h | 111 +++++ > 6 files changed, 721 insertions(+), 644 deletions(-) > create mode 100644 drivers/virtio/virtio_pci_modern_dev.c > create mode 100644 include/linux/virtio_pci_modern.h > > diff --git a/drivers/virtio/Kconfig b/drivers/virtio/Kconfig > index 7b41130d3f35..6b9b81f4b8c2 100644 > --- a/drivers/virtio/Kconfig > +++ b/drivers/virtio/Kconfig > @@ -12,6 +12,14 @@ config ARCH_HAS_RESTRICTED_VIRTIO_MEMORY_ACCESS > This option is selected if the architecture may need to enforce > VIRTIO_F_ACCESS_PLATFORM > > +config VIRTIO_PCI_MODERN > + tristate "Modern Virtio PCI Device" > + depends on PCI > + help > + Modern PCI device implementation. This module implements the > + basic probe and control for devices which are based on modern > + PCI device with possible vendor specific extensions. > + > menuconfig VIRTIO_MENU > bool "Virtio drivers" > default y > @@ -20,7 +28,7 @@ if VIRTIO_MENU > > config VIRTIO_PCI > tristate "PCI driver for virtio devices" > - depends on PCI > + depends on VIRTIO_PCI_MODERN > select VIRTIO > help > This driver provides support for virtio based paravirtual device > diff --git a/drivers/virtio/Makefile b/drivers/virtio/Makefile > index 591e6f72aa54..f097578aaa8f 100644 > --- a/drivers/virtio/Makefile > +++ b/drivers/virtio/Makefile > @@ -1,5 +1,6 @@ > # SPDX-License-Identifier: GPL-2.0 > obj-$(CONFIG_VIRTIO) += virtio.o virtio_ring.o > +obj-$(CONFIG_VIRTIO_PCI_MODERN) += virtio_pci_modern_dev.o > obj-$(CONFIG_VIRTIO_MMIO) += virtio_mmio.o > obj-$(CONFIG_VIRTIO_PCI) += virtio_pci.o > virtio_pci-y := virtio_pci_modern.o virtio_pci_common.o > diff --git a/drivers/virtio/virtio_pci_common.h b/drivers/virtio/virtio_pci_common.h > index f35ff5b6b467..beec047a8f8d 100644 > --- a/drivers/virtio/virtio_pci_common.h > +++ b/drivers/virtio/virtio_pci_common.h > @@ -25,6 +25,7 @@ > #include <linux/virtio_config.h> > #include <linux/virtio_ring.h> > #include <linux/virtio_pci.h> > +#include <linux/virtio_pci_modern.h> > #include <linux/highmem.h> > #include <linux/spinlock.h> > > @@ -39,32 +40,6 @@ struct virtio_pci_vq_info { > unsigned msix_vector; > }; > > -struct virtio_pci_modern_device { > - struct pci_dev *pci_dev; > - > - struct virtio_pci_common_cfg __iomem *common; > - /* Device-specific data (non-legacy mode) */ > - void __iomem *device; > - /* Base of vq notifications (non-legacy mode). */ > - void __iomem *notify_base; > - /* Where to read and clear interrupt */ > - u8 __iomem *isr; > - > - /* So we can sanity-check accesses. */ > - size_t notify_len; > - size_t device_len; > - > - /* Capability for when we need to map notifications per-vq. */ > - int notify_map_cap; > - > - /* Multiply queue_notify_off by this value. (non-legacy mode). */ > - u32 notify_offset_multiplier; > - > - int modern_bars; > - > - struct virtio_device_id id; > -}; > - > /* Our device structure */ > struct virtio_pci_device { > struct virtio_device vdev; > diff --git a/drivers/virtio/virtio_pci_modern.c b/drivers/virtio/virtio_pci_modern.c > index a5e3a5e40323..fbd4ebc00eb6 100644 > --- a/drivers/virtio/virtio_pci_modern.c > +++ b/drivers/virtio/virtio_pci_modern.c > @@ -19,158 +19,6 @@ > #define VIRTIO_RING_NO_LEGACY > #include "virtio_pci_common.h" > > -/* > - * Type-safe wrappers for io accesses. > - * Use these to enforce at compile time the following spec requirement: > - * > - * The driver MUST access each field using the ?natural? access > - * method, i.e. 32-bit accesses for 32-bit fields, 16-bit accesses > - * for 16-bit fields and 8-bit accesses for 8-bit fields. > - */ > -static inline u8 vp_ioread8(const u8 __iomem *addr) > -{ > - return ioread8(addr); > -} > -static inline u16 vp_ioread16 (const __le16 __iomem *addr) > -{ > - return ioread16(addr); > -} > - > -static inline u32 vp_ioread32(const __le32 __iomem *addr) > -{ > - return ioread32(addr); > -} > - > -static inline void vp_iowrite8(u8 value, u8 __iomem *addr) > -{ > - iowrite8(value, addr); > -} > - > -static inline void vp_iowrite16(u16 value, __le16 __iomem *addr) > -{ > - iowrite16(value, addr); > -} > - > -static inline void vp_iowrite32(u32 value, __le32 __iomem *addr) > -{ > - iowrite32(value, addr); > -} > - > -static void vp_iowrite64_twopart(u64 val, > - __le32 __iomem *lo, __le32 __iomem *hi) > -{ > - vp_iowrite32((u32)val, lo); > - vp_iowrite32(val >> 32, hi); > -} > - > -/* > - * vp_modern_map_capability - map a part of virtio pci capability > - * @mdev: the modern virtio-pci device > - * @off: offset of the capability > - * @minlen: minimal length of the capability > - * @align: align requirement > - * @start: start from the capability > - * @size: map size > - * @len: the length that is actually mapped > - * > - * Returns the io address of for the part of the capability > - */ > -void __iomem *vp_modern_map_capability(struct virtio_pci_modern_device *mdev, int off, > - size_t minlen, > - u32 align, > - u32 start, u32 size, > - size_t *len) > -{ > - struct pci_dev *dev = mdev->pci_dev; > - u8 bar; > - u32 offset, length; > - void __iomem *p; > - > - pci_read_config_byte(dev, off + offsetof(struct virtio_pci_cap, > - bar), > - &bar); > - pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, offset), > - &offset); > - pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, length), > - &length); > - > - if (length <= start) { > - dev_err(&dev->dev, > - "virtio_pci: bad capability len %u (>%u expected)\n", > - length, start); > - return NULL; > - } > - > - if (length - start < minlen) { > - dev_err(&dev->dev, > - "virtio_pci: bad capability len %u (>=%zu expected)\n", > - length, minlen); > - return NULL; > - } > - > - length -= start; > - > - if (start + offset < offset) { > - dev_err(&dev->dev, > - "virtio_pci: map wrap-around %u+%u\n", > - start, offset); > - return NULL; > - } > - > - offset += start; > - > - if (offset & (align - 1)) { > - dev_err(&dev->dev, > - "virtio_pci: offset %u not aligned to %u\n", > - offset, align); > - return NULL; > - } > - > - if (length > size) > - length = size; > - > - if (len) > - *len = length; > - > - if (minlen + offset < minlen || > - minlen + offset > pci_resource_len(dev, bar)) { > - dev_err(&dev->dev, > - "virtio_pci: map virtio %zu@%u " > - "out of range on bar %i length %lu\n", > - minlen, offset, > - bar, (unsigned long)pci_resource_len(dev, bar)); > - return NULL; > - } > - > - p = pci_iomap_range(dev, bar, offset, length); > - if (!p) > - dev_err(&dev->dev, > - "virtio_pci: unable to map virtio %u@%u on bar %i\n", > - length, offset, bar); > - return p; > -} > - > -/* > - * vp_modern_get_features - get features from device > - * @mdev: the modern virtio-pci device > - * > - * Returns the features read from the device > - */ > -static u64 vp_modern_get_features(struct virtio_pci_modern_device *mdev) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - u64 features; > - > - vp_iowrite32(0, &cfg->device_feature_select); > - features = vp_ioread32(&cfg->device_feature); > - vp_iowrite32(1, &cfg->device_feature_select); > - features |= ((u64)vp_ioread32(&cfg->device_feature) << 32); > - > - return features; > -} > - > -/* virtio config->get_features() implementation */ > static u64 vp_get_features(struct virtio_device *vdev) > { > struct virtio_pci_device *vp_dev = to_vp_device(vdev); > @@ -188,149 +36,6 @@ static void vp_transport_features(struct virtio_device *vdev, u64 features) > __virtio_set_bit(vdev, VIRTIO_F_SR_IOV); > } > > -/* > - * vp_modern_set_features - set features to device > - * @mdev: the modern virtio-pci device > - * @features: the features set to device > - */ > -static void vp_modern_set_features(struct virtio_pci_modern_device *mdev, > - u64 features) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - vp_iowrite32(0, &cfg->guest_feature_select); > - vp_iowrite32((u32)features, &cfg->guest_feature); > - vp_iowrite32(1, &cfg->guest_feature_select); > - vp_iowrite32(features >> 32, &cfg->guest_feature); > -} > - > -/* > - * vp_modern_queue_vector - set the MSIX vector for a specific virtqueue > - * @mdev: the modern virtio-pci device > - * @index: queue index > - * @vector: the config vector > - * > - * Returns the config vector read from the device > - */ > -static u16 vp_modern_queue_vector(struct virtio_pci_modern_device *mdev, > - u16 index, u16 vector) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - vp_iowrite16(index, &cfg->queue_select); > - vp_iowrite16(vector, &cfg->queue_msix_vector); > - /* Flush the write out to device */ > - return vp_ioread16(&cfg->queue_msix_vector); > -} > - > -/* > - * vp_modern_queue_address - set the virtqueue address > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * @desc_addr: address of the descriptor area > - * @driver_addr: address of the driver area > - * @device_addr: address of the device area > - */ > -static void vp_modern_queue_address(struct virtio_pci_modern_device *mdev, > - u16 index, u64 desc_addr, u64 driver_addr, > - u64 device_addr) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - vp_iowrite16(index, &cfg->queue_select); > - > - vp_iowrite64_twopart(desc_addr, &cfg->queue_desc_lo, > - &cfg->queue_desc_hi); > - vp_iowrite64_twopart(driver_addr, &cfg->queue_avail_lo, > - &cfg->queue_avail_hi); > - vp_iowrite64_twopart(device_addr, &cfg->queue_used_lo, > - &cfg->queue_used_hi); > -} > - > -/* > - * vp_modern_set_queue_enable - enable a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * @enable: whether the virtqueue is enable or not > - */ > -static void vp_modern_set_queue_enable(struct virtio_pci_modern_device *mdev, > - u16 index, bool enable) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - vp_iowrite16(enable, &mdev->common->queue_enable); > -} > - > -/* > - * vp_modern_get_queue_enable - enable a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * > - * Returns whether a virtqueue is enabled or not > - */ > -static bool vp_modern_get_queue_enable(struct virtio_pci_modern_device *mdev, > - u16 index) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - > - return vp_ioread16(&mdev->common->queue_enable); > -} > - > -/* > - * vp_modern_set_queue_size - set size for a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * @size: the size of the virtqueue > - */ > -static void vp_modern_set_queue_size(struct virtio_pci_modern_device *mdev, > - u16 index, u16 size) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - vp_iowrite16(size, &mdev->common->queue_size); > - > -} > - > -/* > - * vp_modern_get_queue_size - get size for a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * > - * Returns the size of the virtqueue > - */ > -static u16 vp_modern_get_queue_size(struct virtio_pci_modern_device *mdev, > - u16 index) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - > - return vp_ioread16(&mdev->common->queue_size); > - > -} > - > -/* > - * vp_modern_get_num_queues - get the number of virtqueues > - * @mdev: the modern virtio-pci device > - * > - * Returns the number of virtqueues > - */ > -static u16 vp_modern_get_num_queues(struct virtio_pci_modern_device *mdev) > -{ > - return vp_ioread16(&mdev->common->num_queues); > -} > - > -/* > - * vp_modern_get_queue_notify_off - get notification offset for a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * > - * Returns the notification offset for a virtqueue > - */ > -static u16 vp_modern_get_queue_notify_off(struct virtio_pci_modern_device *mdev, > - u16 index) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - > - return vp_ioread16(&mdev->common->queue_notify_off); > -} > - > /* virtio config->finalize_features() implementation */ > static int vp_finalize_features(struct virtio_device *vdev) > { > @@ -429,19 +134,6 @@ static void vp_set(struct virtio_device *vdev, unsigned offset, > } > } > > -/* > - * vp_modern_generation - get the device genreation > - * @mdev: the modern virtio-pci device > - * > - * Returns the genreation read from device > - */ > -static u32 vp_modern_generation(struct virtio_pci_modern_device *mdev) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - return vp_ioread8(&cfg->config_generation); > -} > - > static u32 vp_generation(struct virtio_device *vdev) > { > struct virtio_pci_device *vp_dev = to_vp_device(vdev); > @@ -449,19 +141,6 @@ static u32 vp_generation(struct virtio_device *vdev) > return vp_modern_generation(&vp_dev->mdev); > } > > -/* > - * vp_modern_get_status - get the device status > - * @mdev: the modern virtio-pci device > - * > - * Returns the status read from device > - */ > -static u8 vp_modern_get_status(struct virtio_pci_modern_device *mdev) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - return vp_ioread8(&cfg->device_status); > -} > - > /* config->{get,set}_status() implementations */ > static u8 vp_get_status(struct virtio_device *vdev) > { > @@ -470,19 +149,6 @@ static u8 vp_get_status(struct virtio_device *vdev) > return vp_modern_get_status(&vp_dev->mdev); > } > > -/* > - * vp_modern_set_status - set status to device > - * @mdev: the modern virtio-pci device > - * @status: the status set to device > - */ > -static void vp_modern_set_status(struct virtio_pci_modern_device *mdev, > - u8 status) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - vp_iowrite8(status, &cfg->device_status); > -} > - > static void vp_set_status(struct virtio_device *vdev, u8 status) > { > struct virtio_pci_device *vp_dev = to_vp_device(vdev); > @@ -510,25 +176,6 @@ static void vp_reset(struct virtio_device *vdev) > vp_synchronize_vectors(vdev); > } > > -/* > - * vp_modern_config_vector - set the vector for config interrupt > - * @mdev: the modern virtio-pci device > - * @vector: the config vector > - * > - * Returns the config vector read from the device > - */ > -static u16 vp_modern_config_vector(struct virtio_pci_modern_device *mdev, > - u16 vector) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - /* Setup the vector used for configuration events */ > - vp_iowrite16(vector, &cfg->msix_config); > - /* Verify we had enough resources to assign the vector */ > - /* Will also flush the write out to device */ > - return vp_ioread16(&cfg->msix_config); > -} > - > static u16 vp_config_vector(struct virtio_pci_device *vp_dev, u16 vector) > { > return vp_modern_config_vector(&vp_dev->mdev, vector); > @@ -789,253 +436,6 @@ static const struct virtio_config_ops virtio_pci_config_ops = { > .get_shm_region = vp_get_shm_region, > }; > > -/** > - * virtio_pci_find_capability - walk capabilities to find device info. > - * @dev: the pci device > - * @cfg_type: the VIRTIO_PCI_CAP_* value we seek > - * @ioresource_types: IORESOURCE_MEM and/or IORESOURCE_IO. > - * @bars: the bitmask of BARs > - * > - * Returns offset of the capability, or 0. > - */ > -static inline int virtio_pci_find_capability(struct pci_dev *dev, u8 cfg_type, > - u32 ioresource_types, int *bars) > -{ > - int pos; > - > - for (pos = pci_find_capability(dev, PCI_CAP_ID_VNDR); > - pos > 0; > - pos = pci_find_next_capability(dev, pos, PCI_CAP_ID_VNDR)) { > - u8 type, bar; > - pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap, > - cfg_type), > - &type); > - pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap, > - bar), > - &bar); > - > - /* Ignore structures with reserved BAR values */ > - if (bar > 0x5) > - continue; > - > - if (type == cfg_type) { > - if (pci_resource_len(dev, bar) && > - pci_resource_flags(dev, bar) & ioresource_types) { > - *bars |= (1 << bar); > - return pos; > - } > - } > - } > - return 0; > -} > - > -/* This is part of the ABI. Don't screw with it. */ > -static inline void check_offsets(void) > -{ > - /* Note: disk space was harmed in compilation of this function. */ > - BUILD_BUG_ON(VIRTIO_PCI_CAP_VNDR !> - offsetof(struct virtio_pci_cap, cap_vndr)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_NEXT !> - offsetof(struct virtio_pci_cap, cap_next)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_LEN !> - offsetof(struct virtio_pci_cap, cap_len)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_CFG_TYPE !> - offsetof(struct virtio_pci_cap, cfg_type)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_BAR !> - offsetof(struct virtio_pci_cap, bar)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_OFFSET !> - offsetof(struct virtio_pci_cap, offset)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_LENGTH !> - offsetof(struct virtio_pci_cap, length)); > - BUILD_BUG_ON(VIRTIO_PCI_NOTIFY_CAP_MULT !> - offsetof(struct virtio_pci_notify_cap, > - notify_off_multiplier)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_DFSELECT !> - offsetof(struct virtio_pci_common_cfg, > - device_feature_select)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_DF !> - offsetof(struct virtio_pci_common_cfg, device_feature)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_GFSELECT !> - offsetof(struct virtio_pci_common_cfg, > - guest_feature_select)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_GF !> - offsetof(struct virtio_pci_common_cfg, guest_feature)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_MSIX !> - offsetof(struct virtio_pci_common_cfg, msix_config)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_NUMQ !> - offsetof(struct virtio_pci_common_cfg, num_queues)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_STATUS !> - offsetof(struct virtio_pci_common_cfg, device_status)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_CFGGENERATION !> - offsetof(struct virtio_pci_common_cfg, config_generation)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SELECT !> - offsetof(struct virtio_pci_common_cfg, queue_select)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SIZE !> - offsetof(struct virtio_pci_common_cfg, queue_size)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_MSIX !> - offsetof(struct virtio_pci_common_cfg, queue_msix_vector)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_ENABLE !> - offsetof(struct virtio_pci_common_cfg, queue_enable)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_NOFF !> - offsetof(struct virtio_pci_common_cfg, queue_notify_off)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCLO !> - offsetof(struct virtio_pci_common_cfg, queue_desc_lo)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCHI !> - offsetof(struct virtio_pci_common_cfg, queue_desc_hi)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILLO !> - offsetof(struct virtio_pci_common_cfg, queue_avail_lo)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILHI !> - offsetof(struct virtio_pci_common_cfg, queue_avail_hi)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDLO !> - offsetof(struct virtio_pci_common_cfg, queue_used_lo)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDHI !> - offsetof(struct virtio_pci_common_cfg, queue_used_hi)); > -} > - > -/* > - * vp_modern_probe: probe the modern virtio pci device, note that the > - * caller is required to enable PCI device before calling this function. > - * @mdev: the modern virtio-pci device > - * > - * Return 0 on succeed otherwise fail > - */ > -static int vp_modern_probe(struct virtio_pci_modern_device *mdev) > -{ > - struct pci_dev *pci_dev = mdev->pci_dev; > - int err, common, isr, notify, device; > - u32 notify_length; > - u32 notify_offset; > - > - check_offsets(); > - > - mdev->pci_dev = pci_dev; > - > - /* We only own devices >= 0x1000 and <= 0x107f: leave the rest. */ > - if (pci_dev->device < 0x1000 || pci_dev->device > 0x107f) > - return -ENODEV; > - > - if (pci_dev->device < 0x1040) { > - /* Transitional devices: use the PCI subsystem device id as > - * virtio device id, same as legacy driver always did. > - */ > - mdev->id.device = pci_dev->subsystem_device; > - } else { > - /* Modern devices: simply use PCI device id, but start from 0x1040. */ > - mdev->id.device = pci_dev->device - 0x1040; > - } > - mdev->id.vendor = pci_dev->subsystem_vendor; > - > - /* check for a common config: if not, use legacy mode (bar 0). */ > - common = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_COMMON_CFG, > - IORESOURCE_IO | IORESOURCE_MEM, > - &mdev->modern_bars); > - if (!common) { > - dev_info(&pci_dev->dev, > - "virtio_pci: leaving for legacy driver\n"); > - return -ENODEV; > - } > - > - /* If common is there, these should be too... */ > - isr = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_ISR_CFG, > - IORESOURCE_IO | IORESOURCE_MEM, > - &mdev->modern_bars); > - notify = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_NOTIFY_CFG, > - IORESOURCE_IO | IORESOURCE_MEM, > - &mdev->modern_bars); > - if (!isr || !notify) { > - dev_err(&pci_dev->dev, > - "virtio_pci: missing capabilities %i/%i/%i\n", > - common, isr, notify); > - return -EINVAL; > - } > - > - err = dma_set_mask_and_coherent(&pci_dev->dev, DMA_BIT_MASK(64)); > - if (err) > - err = dma_set_mask_and_coherent(&pci_dev->dev, > - DMA_BIT_MASK(32)); > - if (err) > - dev_warn(&pci_dev->dev, "Failed to enable 64-bit or 32-bit DMA. Trying to continue, but this might not work.\n"); > - > - /* Device capability is only mandatory for devices that have > - * device-specific configuration. > - */ > - device = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_DEVICE_CFG, > - IORESOURCE_IO | IORESOURCE_MEM, > - &mdev->modern_bars); > - > - err = pci_request_selected_regions(pci_dev, mdev->modern_bars, > - "virtio-pci-modern"); > - if (err) > - return err; > - > - err = -EINVAL; > - mdev->common = vp_modern_map_capability(mdev, common, > - sizeof(struct virtio_pci_common_cfg), 4, > - 0, sizeof(struct virtio_pci_common_cfg), > - NULL); > - if (!mdev->common) > - goto err_map_common; > - mdev->isr = vp_modern_map_capability(mdev, isr, sizeof(u8), 1, > - 0, 1, > - NULL); > - if (!mdev->isr) > - goto err_map_isr; > - > - /* Read notify_off_multiplier from config space. */ > - pci_read_config_dword(pci_dev, > - notify + offsetof(struct virtio_pci_notify_cap, > - notify_off_multiplier), > - &mdev->notify_offset_multiplier); > - /* Read notify length and offset from config space. */ > - pci_read_config_dword(pci_dev, > - notify + offsetof(struct virtio_pci_notify_cap, > - cap.length), > - ¬ify_length); > - > - pci_read_config_dword(pci_dev, > - notify + offsetof(struct virtio_pci_notify_cap, > - cap.offset), > - ¬ify_offset); > - > - /* We don't know how many VQs we'll map, ahead of the time. > - * If notify length is small, map it all now. > - * Otherwise, map each VQ individually later. > - */ > - if ((u64)notify_length + (notify_offset % PAGE_SIZE) <= PAGE_SIZE) { > - mdev->notify_base = vp_modern_map_capability(mdev, notify, > - 2, 2, > - 0, notify_length, > - &mdev->notify_len); > - if (!mdev->notify_base) > - goto err_map_notify; > - } else { > - mdev->notify_map_cap = notify; > - } > - > - /* Again, we don't know how much we should map, but PAGE_SIZE > - * is more than enough for all existing devices. > - */ > - if (device) { > - mdev->device = vp_modern_map_capability(mdev, device, 0, 4, > - 0, PAGE_SIZE, > - &mdev->device_len); > - if (!mdev->device) > - goto err_map_device; > - } > - > - return 0; > - > -err_map_device: > - if (mdev->notify_base) > - pci_iounmap(pci_dev, mdev->notify_base); > -err_map_notify: > - pci_iounmap(pci_dev, mdev->isr); > -err_map_isr: > - pci_iounmap(pci_dev, mdev->common); > -err_map_common: > - return err; > -} > - > /* the PCI probing function */ > int virtio_pci_modern_probe(struct virtio_pci_device *vp_dev) > { > @@ -1063,23 +463,6 @@ int virtio_pci_modern_probe(struct virtio_pci_device *vp_dev) > return 0; > } > > -/* > - * vp_modern_probe: remove and cleanup the modern virtio pci device > - * @mdev: the modern virtio-pci device > - */ > -static void vp_modern_remove(struct virtio_pci_modern_device *mdev) > -{ > - struct pci_dev *pci_dev = mdev->pci_dev; > - > - if (mdev->device) > - pci_iounmap(pci_dev, mdev->device); > - if (mdev->notify_base) > - pci_iounmap(pci_dev, mdev->notify_base); > - pci_iounmap(pci_dev, mdev->isr); > - pci_iounmap(pci_dev, mdev->common); > - pci_release_selected_regions(pci_dev, mdev->modern_bars); > -} > - > void virtio_pci_modern_remove(struct virtio_pci_device *vp_dev) > { > struct virtio_pci_modern_device *mdev = &vp_dev->mdev; > diff --git a/drivers/virtio/virtio_pci_modern_dev.c b/drivers/virtio/virtio_pci_modern_dev.c > new file mode 100644 > index 000000000000..cbd667496bb1 > --- /dev/null > +++ b/drivers/virtio/virtio_pci_modern_dev.c > @@ -0,0 +1,599 @@ > +// SPDX-License-Identifier: GPL-2.0-or-later > + > +#include <linux/virtio_pci_modern.h> > +#include <linux/module.h> > +#include <linux/pci.h> > + > +/* > + * vp_modern_map_capability - map a part of virtio pci capability > + * @mdev: the modern virtio-pci device > + * @off: offset of the capability > + * @minlen: minimal length of the capability > + * @align: align requirement > + * @start: start from the capability > + * @size: map size > + * @len: the length that is actually mapped > + * > + * Returns the io address of for the part of the capability > + */ > +void __iomem *vp_modern_map_capability(struct virtio_pci_modern_device *mdev, int off, > + size_t minlen, > + u32 align, > + u32 start, u32 size, > + size_t *len) > +{ > + struct pci_dev *dev = mdev->pci_dev; > + u8 bar; > + u32 offset, length; > + void __iomem *p; > + > + pci_read_config_byte(dev, off + offsetof(struct virtio_pci_cap, > + bar), > + &bar); > + pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, offset), > + &offset); > + pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, length), > + &length); > + > + if (length <= start) { > + dev_err(&dev->dev, > + "virtio_pci: bad capability len %u (>%u expected)\n", > + length, start); > + return NULL; > + } > + > + if (length - start < minlen) { > + dev_err(&dev->dev, > + "virtio_pci: bad capability len %u (>=%zu expected)\n", > + length, minlen); > + return NULL; > + } > + > + length -= start; > + > + if (start + offset < offset) { > + dev_err(&dev->dev, > + "virtio_pci: map wrap-around %u+%u\n", > + start, offset); > + return NULL; > + } > + > + offset += start; > + > + if (offset & (align - 1)) { > + dev_err(&dev->dev, > + "virtio_pci: offset %u not aligned to %u\n", > + offset, align); > + return NULL; > + } > + > + if (length > size) > + length = size; > + > + if (len) > + *len = length; > + > + if (minlen + offset < minlen || > + minlen + offset > pci_resource_len(dev, bar)) { > + dev_err(&dev->dev, > + "virtio_pci: map virtio %zu@%u " > + "out of range on bar %i length %lu\n", > + minlen, offset, > + bar, (unsigned long)pci_resource_len(dev, bar)); > + return NULL; > + } > + > + p = pci_iomap_range(dev, bar, offset, length); > + if (!p) > + dev_err(&dev->dev, > + "virtio_pci: unable to map virtio %u@%u on bar %i\n", > + length, offset, bar); > + return p; > +} > +EXPORT_SYMBOL_GPL(vp_modern_map_capability); > + > +/** > + * virtio_pci_find_capability - walk capabilities to find device info. > + * @dev: the pci device > + * @cfg_type: the VIRTIO_PCI_CAP_* value we seek > + * @ioresource_types: IORESOURCE_MEM and/or IORESOURCE_IO. > + * @bars: the bitmask of BARs > + * > + * Returns offset of the capability, or 0. > + */ > +static inline int virtio_pci_find_capability(struct pci_dev *dev, u8 cfg_type, > + u32 ioresource_types, int *bars) > +{ > + int pos; > + > + for (pos = pci_find_capability(dev, PCI_CAP_ID_VNDR); > + pos > 0; > + pos = pci_find_next_capability(dev, pos, PCI_CAP_ID_VNDR)) { > + u8 type, bar; > + pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap, > + cfg_type), > + &type); > + pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap, > + bar), > + &bar); > + > + /* Ignore structures with reserved BAR values */ > + if (bar > 0x5) > + continue; > + > + if (type == cfg_type) { > + if (pci_resource_len(dev, bar) && > + pci_resource_flags(dev, bar) & ioresource_types) { > + *bars |= (1 << bar); > + return pos; > + } > + } > + } > + return 0; > +} > + > +/* This is part of the ABI. Don't screw with it. */ > +static inline void check_offsets(void) > +{ > + /* Note: disk space was harmed in compilation of this function. */ > + BUILD_BUG_ON(VIRTIO_PCI_CAP_VNDR !> + offsetof(struct virtio_pci_cap, cap_vndr)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_NEXT !> + offsetof(struct virtio_pci_cap, cap_next)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_LEN !> + offsetof(struct virtio_pci_cap, cap_len)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_CFG_TYPE !> + offsetof(struct virtio_pci_cap, cfg_type)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_BAR !> + offsetof(struct virtio_pci_cap, bar)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_OFFSET !> + offsetof(struct virtio_pci_cap, offset)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_LENGTH !> + offsetof(struct virtio_pci_cap, length)); > + BUILD_BUG_ON(VIRTIO_PCI_NOTIFY_CAP_MULT !> + offsetof(struct virtio_pci_notify_cap, > + notify_off_multiplier)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_DFSELECT !> + offsetof(struct virtio_pci_common_cfg, > + device_feature_select)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_DF !> + offsetof(struct virtio_pci_common_cfg, device_feature)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_GFSELECT !> + offsetof(struct virtio_pci_common_cfg, > + guest_feature_select)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_GF !> + offsetof(struct virtio_pci_common_cfg, guest_feature)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_MSIX !> + offsetof(struct virtio_pci_common_cfg, msix_config)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_NUMQ !> + offsetof(struct virtio_pci_common_cfg, num_queues)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_STATUS !> + offsetof(struct virtio_pci_common_cfg, device_status)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_CFGGENERATION !> + offsetof(struct virtio_pci_common_cfg, config_generation)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SELECT !> + offsetof(struct virtio_pci_common_cfg, queue_select)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SIZE !> + offsetof(struct virtio_pci_common_cfg, queue_size)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_MSIX !> + offsetof(struct virtio_pci_common_cfg, queue_msix_vector)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_ENABLE !> + offsetof(struct virtio_pci_common_cfg, queue_enable)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_NOFF !> + offsetof(struct virtio_pci_common_cfg, queue_notify_off)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCLO !> + offsetof(struct virtio_pci_common_cfg, queue_desc_lo)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCHI !> + offsetof(struct virtio_pci_common_cfg, queue_desc_hi)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILLO !> + offsetof(struct virtio_pci_common_cfg, queue_avail_lo)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILHI !> + offsetof(struct virtio_pci_common_cfg, queue_avail_hi)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDLO !> + offsetof(struct virtio_pci_common_cfg, queue_used_lo)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDHI !> + offsetof(struct virtio_pci_common_cfg, queue_used_hi)); > +} > + > +/* > + * vp_modern_probe: probe the modern virtio pci device, note that the > + * caller is required to enable PCI device before calling this function. > + * @mdev: the modern virtio-pci device > + * > + * Return 0 on succeed otherwise fail > + */ > +int vp_modern_probe(struct virtio_pci_modern_device *mdev) > +{ > + struct pci_dev *pci_dev = mdev->pci_dev; > + int err, common, isr, notify, device; > + u32 notify_length; > + u32 notify_offset; > + > + check_offsets(); > + > + mdev->pci_dev = pci_dev; > + > + /* We only own devices >= 0x1000 and <= 0x107f: leave the rest. */ > + if (pci_dev->device < 0x1000 || pci_dev->device > 0x107f) > + return -ENODEV; > + > + if (pci_dev->device < 0x1040) { > + /* Transitional devices: use the PCI subsystem device id as > + * virtio device id, same as legacy driver always did. > + */ > + mdev->id.device = pci_dev->subsystem_device; > + } else { > + /* Modern devices: simply use PCI device id, but start from 0x1040. */ > + mdev->id.device = pci_dev->device - 0x1040; > + } > + mdev->id.vendor = pci_dev->subsystem_vendor; > + > + /* check for a common config: if not, use legacy mode (bar 0). */ > + common = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_COMMON_CFG, > + IORESOURCE_IO | IORESOURCE_MEM, > + &mdev->modern_bars); > + if (!common) { > + dev_info(&pci_dev->dev, > + "virtio_pci: leaving for legacy driver\n"); > + return -ENODEV; > + } > + > + /* If common is there, these should be too... */ > + isr = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_ISR_CFG, > + IORESOURCE_IO | IORESOURCE_MEM, > + &mdev->modern_bars); > + notify = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_NOTIFY_CFG, > + IORESOURCE_IO | IORESOURCE_MEM, > + &mdev->modern_bars); > + if (!isr || !notify) { > + dev_err(&pci_dev->dev, > + "virtio_pci: missing capabilities %i/%i/%i\n", > + common, isr, notify); > + return -EINVAL; > + } > + > + err = dma_set_mask_and_coherent(&pci_dev->dev, DMA_BIT_MASK(64)); > + if (err) > + err = dma_set_mask_and_coherent(&pci_dev->dev, > + DMA_BIT_MASK(32)); > + if (err) > + dev_warn(&pci_dev->dev, "Failed to enable 64-bit or 32-bit DMA. Trying to continue, but this might not work.\n"); > + > + /* Device capability is only mandatory for devices that have > + * device-specific configuration. > + */ > + device = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_DEVICE_CFG, > + IORESOURCE_IO | IORESOURCE_MEM, > + &mdev->modern_bars); > + > + err = pci_request_selected_regions(pci_dev, mdev->modern_bars, > + "virtio-pci-modern"); > + if (err) > + return err; > + > + err = -EINVAL; > + mdev->common = vp_modern_map_capability(mdev, common, > + sizeof(struct virtio_pci_common_cfg), 4, > + 0, sizeof(struct virtio_pci_common_cfg), > + NULL); > + if (!mdev->common) > + goto err_map_common; > + mdev->isr = vp_modern_map_capability(mdev, isr, sizeof(u8), 1, > + 0, 1, > + NULL); > + if (!mdev->isr) > + goto err_map_isr; > + > + /* Read notify_off_multiplier from config space. */ > + pci_read_config_dword(pci_dev, > + notify + offsetof(struct virtio_pci_notify_cap, > + notify_off_multiplier), > + &mdev->notify_offset_multiplier); > + /* Read notify length and offset from config space. */ > + pci_read_config_dword(pci_dev, > + notify + offsetof(struct virtio_pci_notify_cap, > + cap.length), > + ¬ify_length); > + > + pci_read_config_dword(pci_dev, > + notify + offsetof(struct virtio_pci_notify_cap, > + cap.offset), > + ¬ify_offset); > + > + /* We don't know how many VQs we'll map, ahead of the time. > + * If notify length is small, map it all now. > + * Otherwise, map each VQ individually later. > + */ > + if ((u64)notify_length + (notify_offset % PAGE_SIZE) <= PAGE_SIZE) { > + mdev->notify_base = vp_modern_map_capability(mdev, notify, > + 2, 2, > + 0, notify_length, > + &mdev->notify_len); > + if (!mdev->notify_base) > + goto err_map_notify; > + } else { > + mdev->notify_map_cap = notify; > + } > + > + /* Again, we don't know how much we should map, but PAGE_SIZE > + * is more than enough for all existing devices. > + */ > + if (device) { > + mdev->device = vp_modern_map_capability(mdev, device, 0, 4, > + 0, PAGE_SIZE, > + &mdev->device_len); > + if (!mdev->device) > + goto err_map_device; > + } > + > + return 0; > + > +err_map_device: > + if (mdev->notify_base) > + pci_iounmap(pci_dev, mdev->notify_base); > +err_map_notify: > + pci_iounmap(pci_dev, mdev->isr); > +err_map_isr: > + pci_iounmap(pci_dev, mdev->common); > +err_map_common: > + return err; > +} > +EXPORT_SYMBOL_GPL(vp_modern_probe); > + > +/* > + * vp_modern_probe: remove and cleanup the modern virtio pci device > + * @mdev: the modern virtio-pci device > + */ > +void vp_modern_remove(struct virtio_pci_modern_device *mdev) > +{ > + struct pci_dev *pci_dev = mdev->pci_dev; > + > + if (mdev->device) > + pci_iounmap(pci_dev, mdev->device); > + if (mdev->notify_base) > + pci_iounmap(pci_dev, mdev->notify_base); > + pci_iounmap(pci_dev, mdev->isr); > + pci_iounmap(pci_dev, mdev->common); > + pci_release_selected_regions(pci_dev, mdev->modern_bars); > +} > +EXPORT_SYMBOL_GPL(vp_modern_remove); > + > +/* > + * vp_modern_get_features - get features from device > + * @mdev: the modern virtio-pci device > + * > + * Returns the features read from the device > + */ > +u64 vp_modern_get_features(struct virtio_pci_modern_device *mdev) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + u64 features; > + > + vp_iowrite32(0, &cfg->device_feature_select); > + features = vp_ioread32(&cfg->device_feature); > + vp_iowrite32(1, &cfg->device_feature_select); > + features |= ((u64)vp_ioread32(&cfg->device_feature) << 32); > + > + return features; > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_features); > + > +/* > + * vp_modern_set_features - set features to device > + * @mdev: the modern virtio-pci device > + * @features: the features set to device > + */ > +void vp_modern_set_features(struct virtio_pci_modern_device *mdev, > + u64 features) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + vp_iowrite32(0, &cfg->guest_feature_select); > + vp_iowrite32((u32)features, &cfg->guest_feature); > + vp_iowrite32(1, &cfg->guest_feature_select); > + vp_iowrite32(features >> 32, &cfg->guest_feature); > +} > +EXPORT_SYMBOL_GPL(vp_modern_set_features); > + > +/* > + * vp_modern_generation - get the device genreation > + * @mdev: the modern virtio-pci device > + * > + * Returns the genreation read from device > + */ > +u32 vp_modern_generation(struct virtio_pci_modern_device *mdev) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + return vp_ioread8(&cfg->config_generation); > +} > +EXPORT_SYMBOL_GPL(vp_modern_generation); > + > +/* > + * vp_modern_get_status - get the device status > + * @mdev: the modern virtio-pci device > + * > + * Returns the status read from device > + */ > +u8 vp_modern_get_status(struct virtio_pci_modern_device *mdev) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + return vp_ioread8(&cfg->device_status); > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_status); > + > +/* > + * vp_modern_set_status - set status to device > + * @mdev: the modern virtio-pci device > + * @status: the status set to device > + */ > +void vp_modern_set_status(struct virtio_pci_modern_device *mdev, > + u8 status) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + vp_iowrite8(status, &cfg->device_status); > +} > +EXPORT_SYMBOL_GPL(vp_modern_set_status); > + > +/* > + * vp_modern_queue_vector - set the MSIX vector for a specific virtqueue > + * @mdev: the modern virtio-pci device > + * @index: queue index > + * @vector: the config vector > + * > + * Returns the config vector read from the device > + */ > +u16 vp_modern_queue_vector(struct virtio_pci_modern_device *mdev, > + u16 index, u16 vector) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + vp_iowrite16(index, &cfg->queue_select); > + vp_iowrite16(vector, &cfg->queue_msix_vector); > + /* Flush the write out to device */ > + return vp_ioread16(&cfg->queue_msix_vector); > +} > +EXPORT_SYMBOL_GPL(vp_modern_queue_vector); > + > +/* > + * vp_modern_config_vector - set the vector for config interrupt > + * @mdev: the modern virtio-pci device > + * @vector: the config vector > + * > + * Returns the config vector read from the device > + */ > +u16 vp_modern_config_vector(struct virtio_pci_modern_device *mdev, > + u16 vector) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + /* Setup the vector used for configuration events */ > + vp_iowrite16(vector, &cfg->msix_config); > + /* Verify we had enough resources to assign the vector */ > + /* Will also flush the write out to device */ > + return vp_ioread16(&cfg->msix_config); > +} > +EXPORT_SYMBOL_GPL(vp_modern_config_vector); > + > +/* > + * vp_modern_queue_address - set the virtqueue address > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * @desc_addr: address of the descriptor area > + * @driver_addr: address of the driver area > + * @device_addr: address of the device area > + */ > +void vp_modern_queue_address(struct virtio_pci_modern_device *mdev, > + u16 index, u64 desc_addr, u64 driver_addr, > + u64 device_addr) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + vp_iowrite16(index, &cfg->queue_select); > + > + vp_iowrite64_twopart(desc_addr, &cfg->queue_desc_lo, > + &cfg->queue_desc_hi); > + vp_iowrite64_twopart(driver_addr, &cfg->queue_avail_lo, > + &cfg->queue_avail_hi); > + vp_iowrite64_twopart(device_addr, &cfg->queue_used_lo, > + &cfg->queue_used_hi); > +} > +EXPORT_SYMBOL_GPL(vp_modern_queue_address); > + > +/* > + * vp_modern_set_queue_enable - enable a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * @enable: whether the virtqueue is enable or not > + */ > +void vp_modern_set_queue_enable(struct virtio_pci_modern_device *mdev, > + u16 index, bool enable) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + vp_iowrite16(enable, &mdev->common->queue_enable); > +} > +EXPORT_SYMBOL_GPL(vp_modern_set_queue_enable); > + > +/* > + * vp_modern_get_queue_enable - enable a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * > + * Returns whether a virtqueue is enabled or not > + */ > +bool vp_modern_get_queue_enable(struct virtio_pci_modern_device *mdev, > + u16 index) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + > + return vp_ioread16(&mdev->common->queue_enable); > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_queue_enable); > + > +/* > + * vp_modern_set_queue_size - set size for a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * @size: the size of the virtqueue > + */ > +void vp_modern_set_queue_size(struct virtio_pci_modern_device *mdev, > + u16 index, u16 size) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + vp_iowrite16(size, &mdev->common->queue_size); > + > +} > +EXPORT_SYMBOL_GPL(vp_modern_set_queue_size); > + > +/* > + * vp_modern_get_queue_size - get size for a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * > + * Returns the size of the virtqueue > + */ > +u16 vp_modern_get_queue_size(struct virtio_pci_modern_device *mdev, > + u16 index) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + > + return vp_ioread16(&mdev->common->queue_size); > + > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_queue_size); > + > +/* > + * vp_modern_get_num_queues - get the number of virtqueues > + * @mdev: the modern virtio-pci device > + * > + * Returns the number of virtqueues > + */ > +u16 vp_modern_get_num_queues(struct virtio_pci_modern_device *mdev) > +{ > + return vp_ioread16(&mdev->common->num_queues); > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_num_queues); > + > +/* > + * vp_modern_get_queue_notify_off - get notification offset for a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * > + * Returns the notification offset for a virtqueue > + */ > +u16 vp_modern_get_queue_notify_off(struct virtio_pci_modern_device *mdev, > + u16 index) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + > + return vp_ioread16(&mdev->common->queue_notify_off); > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_queue_notify_off); > + > +MODULE_VERSION("0.1"); > +MODULE_DESCRIPTION("Modern Virtio PCI Device"); > +MODULE_AUTHOR("Jason Wang <jasowang at redhat.com>"); > +MODULE_LICENSE("GPL"); > diff --git a/include/linux/virtio_pci_modern.h b/include/linux/virtio_pci_modern.h > new file mode 100644 > index 000000000000..f26acbeec965 > --- /dev/null > +++ b/include/linux/virtio_pci_modern.h > @@ -0,0 +1,111 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +#ifndef _LINUX_VIRTIO_PCI_MODERN_H > +#define _LINUX_VIRTIO_PCI_MODERN_H > + > +#include <linux/pci.h> > +#include <linux/virtio_pci.h> > + > +struct virtio_pci_modern_device { > + struct pci_dev *pci_dev; > + > + struct virtio_pci_common_cfg __iomem *common; > + /* Device-specific data (non-legacy mode) */ > + void __iomem *device; > + /* Base of vq notifications (non-legacy mode). */ > + void __iomem *notify_base; > + /* Where to read and clear interrupt */ > + u8 __iomem *isr; > + > + /* So we can sanity-check accesses. */ > + size_t notify_len; > + size_t device_len; > + > + /* Capability for when we need to map notifications per-vq. */ > + int notify_map_cap; > + > + /* Multiply queue_notify_off by this value. (non-legacy mode). */ > + u32 notify_offset_multiplier; > + > + int modern_bars; > + > + struct virtio_device_id id; > +}; > + > +/* > + * Type-safe wrappers for io accesses. > + * Use these to enforce at compile time the following spec requirement: > + * > + * The driver MUST access each field using the ?natural? access > + * method, i.e. 32-bit accesses for 32-bit fields, 16-bit accesses > + * for 16-bit fields and 8-bit accesses for 8-bit fields. > + */ > +static inline u8 vp_ioread8(const u8 __iomem *addr) > +{ > + return ioread8(addr); > +} > +static inline u16 vp_ioread16 (const __le16 __iomem *addr) > +{ > + return ioread16(addr); > +} > + > +static inline u32 vp_ioread32(const __le32 __iomem *addr) > +{ > + return ioread32(addr); > +} > + > +static inline void vp_iowrite8(u8 value, u8 __iomem *addr) > +{ > + iowrite8(value, addr); > +} > + > +static inline void vp_iowrite16(u16 value, __le16 __iomem *addr) > +{ > + iowrite16(value, addr); > +} > + > +static inline void vp_iowrite32(u32 value, __le32 __iomem *addr) > +{ > + iowrite32(value, addr); > +} > + > +static inline void vp_iowrite64_twopart(u64 val, > + __le32 __iomem *lo, > + __le32 __iomem *hi) > +{ > + vp_iowrite32((u32)val, lo); > + vp_iowrite32(val >> 32, hi); > +} > + > +u64 vp_modern_get_features(struct virtio_pci_modern_device *mdev); > +void vp_modern_set_features(struct virtio_pci_modern_device *mdev, > + u64 features); > +u32 vp_modern_generation(struct virtio_pci_modern_device *mdev); > +u8 vp_modern_get_status(struct virtio_pci_modern_device *mdev); > +void vp_modern_set_status(struct virtio_pci_modern_device *mdev, > + u8 status); > +u16 vp_modern_queue_vector(struct virtio_pci_modern_device *mdev, > + u16 idx, u16 vector); > +u16 vp_modern_config_vector(struct virtio_pci_modern_device *mdev, > + u16 vector); > +void vp_modern_queue_address(struct virtio_pci_modern_device *mdev, > + u16 index, u64 desc_addr, u64 driver_addr, > + u64 device_addr); > +void vp_modern_set_queue_enable(struct virtio_pci_modern_device *mdev, > + u16 idx, bool enable); > +bool vp_modern_get_queue_enable(struct virtio_pci_modern_device *mdev, > + u16 idx); > +void vp_modern_set_queue_size(struct virtio_pci_modern_device *mdev, > + u16 idx, u16 size); > +u16 vp_modern_get_queue_size(struct virtio_pci_modern_device *mdev, > + u16 idx); > +u16 vp_modern_get_num_queues(struct virtio_pci_modern_device *mdev); > +u16 vp_modern_get_queue_notify_off(struct virtio_pci_modern_device *mdev, > + u16 idx); > +void __iomem *vp_modern_map_capability(struct virtio_pci_modern_device *mdev, int off, > + size_t minlen, > + u32 align, > + u32 start, u32 size, > + size_t *len); > +int vp_modern_probe(struct virtio_pci_modern_device *mdev); > +void vp_modern_remove(struct virtio_pci_modern_device *mdev); > +#endif > -- > 2.25.1
Guenter Roeck
2021-Feb-12 20:14 UTC
[PATCH V3 16/19] virtio-pci: introduce modern device module
On Mon, Jan 04, 2021 at 02:55:00PM +0800, Jason Wang wrote:> Signed-off-by: Jason Wang <jasowang at redhat.com> > Reported-by: Naresh Kamboju <naresh.kamboju at linaro.org>I don't really see the point of having to enable VIRTIO_PCI_MODERN because otherwise VIRTIO_PCI no longer works. If VIRTIO_PCI now requires VIRTIO_PCI_MODERN, maybe it should select it instead of depending on it ? Alternatively, you could just drop the new configuration flag entirely and build virtio_pci_modern_dev.o with VIRTIO_PCI. One doesn't work without the other anyway, after all. Guenter> --- > drivers/virtio/Kconfig | 10 +- > drivers/virtio/Makefile | 1 + > drivers/virtio/virtio_pci_common.h | 27 +- > drivers/virtio/virtio_pci_modern.c | 617 ------------------------- > drivers/virtio/virtio_pci_modern_dev.c | 599 ++++++++++++++++++++++++ > include/linux/virtio_pci_modern.h | 111 +++++ > 6 files changed, 721 insertions(+), 644 deletions(-) > create mode 100644 drivers/virtio/virtio_pci_modern_dev.c > create mode 100644 include/linux/virtio_pci_modern.h > > diff --git a/drivers/virtio/Kconfig b/drivers/virtio/Kconfig > index 7b41130d3f35..6b9b81f4b8c2 100644 > --- a/drivers/virtio/Kconfig > +++ b/drivers/virtio/Kconfig > @@ -12,6 +12,14 @@ config ARCH_HAS_RESTRICTED_VIRTIO_MEMORY_ACCESS > This option is selected if the architecture may need to enforce > VIRTIO_F_ACCESS_PLATFORM > > +config VIRTIO_PCI_MODERN > + tristate "Modern Virtio PCI Device" > + depends on PCI > + help > + Modern PCI device implementation. This module implements the > + basic probe and control for devices which are based on modern > + PCI device with possible vendor specific extensions. > + > menuconfig VIRTIO_MENU > bool "Virtio drivers" > default y > @@ -20,7 +28,7 @@ if VIRTIO_MENU > > config VIRTIO_PCI > tristate "PCI driver for virtio devices" > - depends on PCI > + depends on VIRTIO_PCI_MODERN > select VIRTIO > help > This driver provides support for virtio based paravirtual device > diff --git a/drivers/virtio/Makefile b/drivers/virtio/Makefile > index 591e6f72aa54..f097578aaa8f 100644 > --- a/drivers/virtio/Makefile > +++ b/drivers/virtio/Makefile > @@ -1,5 +1,6 @@ > # SPDX-License-Identifier: GPL-2.0 > obj-$(CONFIG_VIRTIO) += virtio.o virtio_ring.o > +obj-$(CONFIG_VIRTIO_PCI_MODERN) += virtio_pci_modern_dev.o > obj-$(CONFIG_VIRTIO_MMIO) += virtio_mmio.o > obj-$(CONFIG_VIRTIO_PCI) += virtio_pci.o > virtio_pci-y := virtio_pci_modern.o virtio_pci_common.o > diff --git a/drivers/virtio/virtio_pci_common.h b/drivers/virtio/virtio_pci_common.h > index f35ff5b6b467..beec047a8f8d 100644 > --- a/drivers/virtio/virtio_pci_common.h > +++ b/drivers/virtio/virtio_pci_common.h > @@ -25,6 +25,7 @@ > #include <linux/virtio_config.h> > #include <linux/virtio_ring.h> > #include <linux/virtio_pci.h> > +#include <linux/virtio_pci_modern.h> > #include <linux/highmem.h> > #include <linux/spinlock.h> > > @@ -39,32 +40,6 @@ struct virtio_pci_vq_info { > unsigned msix_vector; > }; > > -struct virtio_pci_modern_device { > - struct pci_dev *pci_dev; > - > - struct virtio_pci_common_cfg __iomem *common; > - /* Device-specific data (non-legacy mode) */ > - void __iomem *device; > - /* Base of vq notifications (non-legacy mode). */ > - void __iomem *notify_base; > - /* Where to read and clear interrupt */ > - u8 __iomem *isr; > - > - /* So we can sanity-check accesses. */ > - size_t notify_len; > - size_t device_len; > - > - /* Capability for when we need to map notifications per-vq. */ > - int notify_map_cap; > - > - /* Multiply queue_notify_off by this value. (non-legacy mode). */ > - u32 notify_offset_multiplier; > - > - int modern_bars; > - > - struct virtio_device_id id; > -}; > - > /* Our device structure */ > struct virtio_pci_device { > struct virtio_device vdev; > diff --git a/drivers/virtio/virtio_pci_modern.c b/drivers/virtio/virtio_pci_modern.c > index a5e3a5e40323..fbd4ebc00eb6 100644 > --- a/drivers/virtio/virtio_pci_modern.c > +++ b/drivers/virtio/virtio_pci_modern.c > @@ -19,158 +19,6 @@ > #define VIRTIO_RING_NO_LEGACY > #include "virtio_pci_common.h" > > -/* > - * Type-safe wrappers for io accesses. > - * Use these to enforce at compile time the following spec requirement: > - * > - * The driver MUST access each field using the ?natural? access > - * method, i.e. 32-bit accesses for 32-bit fields, 16-bit accesses > - * for 16-bit fields and 8-bit accesses for 8-bit fields. > - */ > -static inline u8 vp_ioread8(const u8 __iomem *addr) > -{ > - return ioread8(addr); > -} > -static inline u16 vp_ioread16 (const __le16 __iomem *addr) > -{ > - return ioread16(addr); > -} > - > -static inline u32 vp_ioread32(const __le32 __iomem *addr) > -{ > - return ioread32(addr); > -} > - > -static inline void vp_iowrite8(u8 value, u8 __iomem *addr) > -{ > - iowrite8(value, addr); > -} > - > -static inline void vp_iowrite16(u16 value, __le16 __iomem *addr) > -{ > - iowrite16(value, addr); > -} > - > -static inline void vp_iowrite32(u32 value, __le32 __iomem *addr) > -{ > - iowrite32(value, addr); > -} > - > -static void vp_iowrite64_twopart(u64 val, > - __le32 __iomem *lo, __le32 __iomem *hi) > -{ > - vp_iowrite32((u32)val, lo); > - vp_iowrite32(val >> 32, hi); > -} > - > -/* > - * vp_modern_map_capability - map a part of virtio pci capability > - * @mdev: the modern virtio-pci device > - * @off: offset of the capability > - * @minlen: minimal length of the capability > - * @align: align requirement > - * @start: start from the capability > - * @size: map size > - * @len: the length that is actually mapped > - * > - * Returns the io address of for the part of the capability > - */ > -void __iomem *vp_modern_map_capability(struct virtio_pci_modern_device *mdev, int off, > - size_t minlen, > - u32 align, > - u32 start, u32 size, > - size_t *len) > -{ > - struct pci_dev *dev = mdev->pci_dev; > - u8 bar; > - u32 offset, length; > - void __iomem *p; > - > - pci_read_config_byte(dev, off + offsetof(struct virtio_pci_cap, > - bar), > - &bar); > - pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, offset), > - &offset); > - pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, length), > - &length); > - > - if (length <= start) { > - dev_err(&dev->dev, > - "virtio_pci: bad capability len %u (>%u expected)\n", > - length, start); > - return NULL; > - } > - > - if (length - start < minlen) { > - dev_err(&dev->dev, > - "virtio_pci: bad capability len %u (>=%zu expected)\n", > - length, minlen); > - return NULL; > - } > - > - length -= start; > - > - if (start + offset < offset) { > - dev_err(&dev->dev, > - "virtio_pci: map wrap-around %u+%u\n", > - start, offset); > - return NULL; > - } > - > - offset += start; > - > - if (offset & (align - 1)) { > - dev_err(&dev->dev, > - "virtio_pci: offset %u not aligned to %u\n", > - offset, align); > - return NULL; > - } > - > - if (length > size) > - length = size; > - > - if (len) > - *len = length; > - > - if (minlen + offset < minlen || > - minlen + offset > pci_resource_len(dev, bar)) { > - dev_err(&dev->dev, > - "virtio_pci: map virtio %zu@%u " > - "out of range on bar %i length %lu\n", > - minlen, offset, > - bar, (unsigned long)pci_resource_len(dev, bar)); > - return NULL; > - } > - > - p = pci_iomap_range(dev, bar, offset, length); > - if (!p) > - dev_err(&dev->dev, > - "virtio_pci: unable to map virtio %u@%u on bar %i\n", > - length, offset, bar); > - return p; > -} > - > -/* > - * vp_modern_get_features - get features from device > - * @mdev: the modern virtio-pci device > - * > - * Returns the features read from the device > - */ > -static u64 vp_modern_get_features(struct virtio_pci_modern_device *mdev) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - u64 features; > - > - vp_iowrite32(0, &cfg->device_feature_select); > - features = vp_ioread32(&cfg->device_feature); > - vp_iowrite32(1, &cfg->device_feature_select); > - features |= ((u64)vp_ioread32(&cfg->device_feature) << 32); > - > - return features; > -} > - > -/* virtio config->get_features() implementation */ > static u64 vp_get_features(struct virtio_device *vdev) > { > struct virtio_pci_device *vp_dev = to_vp_device(vdev); > @@ -188,149 +36,6 @@ static void vp_transport_features(struct virtio_device *vdev, u64 features) > __virtio_set_bit(vdev, VIRTIO_F_SR_IOV); > } > > -/* > - * vp_modern_set_features - set features to device > - * @mdev: the modern virtio-pci device > - * @features: the features set to device > - */ > -static void vp_modern_set_features(struct virtio_pci_modern_device *mdev, > - u64 features) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - vp_iowrite32(0, &cfg->guest_feature_select); > - vp_iowrite32((u32)features, &cfg->guest_feature); > - vp_iowrite32(1, &cfg->guest_feature_select); > - vp_iowrite32(features >> 32, &cfg->guest_feature); > -} > - > -/* > - * vp_modern_queue_vector - set the MSIX vector for a specific virtqueue > - * @mdev: the modern virtio-pci device > - * @index: queue index > - * @vector: the config vector > - * > - * Returns the config vector read from the device > - */ > -static u16 vp_modern_queue_vector(struct virtio_pci_modern_device *mdev, > - u16 index, u16 vector) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - vp_iowrite16(index, &cfg->queue_select); > - vp_iowrite16(vector, &cfg->queue_msix_vector); > - /* Flush the write out to device */ > - return vp_ioread16(&cfg->queue_msix_vector); > -} > - > -/* > - * vp_modern_queue_address - set the virtqueue address > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * @desc_addr: address of the descriptor area > - * @driver_addr: address of the driver area > - * @device_addr: address of the device area > - */ > -static void vp_modern_queue_address(struct virtio_pci_modern_device *mdev, > - u16 index, u64 desc_addr, u64 driver_addr, > - u64 device_addr) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - vp_iowrite16(index, &cfg->queue_select); > - > - vp_iowrite64_twopart(desc_addr, &cfg->queue_desc_lo, > - &cfg->queue_desc_hi); > - vp_iowrite64_twopart(driver_addr, &cfg->queue_avail_lo, > - &cfg->queue_avail_hi); > - vp_iowrite64_twopart(device_addr, &cfg->queue_used_lo, > - &cfg->queue_used_hi); > -} > - > -/* > - * vp_modern_set_queue_enable - enable a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * @enable: whether the virtqueue is enable or not > - */ > -static void vp_modern_set_queue_enable(struct virtio_pci_modern_device *mdev, > - u16 index, bool enable) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - vp_iowrite16(enable, &mdev->common->queue_enable); > -} > - > -/* > - * vp_modern_get_queue_enable - enable a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * > - * Returns whether a virtqueue is enabled or not > - */ > -static bool vp_modern_get_queue_enable(struct virtio_pci_modern_device *mdev, > - u16 index) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - > - return vp_ioread16(&mdev->common->queue_enable); > -} > - > -/* > - * vp_modern_set_queue_size - set size for a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * @size: the size of the virtqueue > - */ > -static void vp_modern_set_queue_size(struct virtio_pci_modern_device *mdev, > - u16 index, u16 size) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - vp_iowrite16(size, &mdev->common->queue_size); > - > -} > - > -/* > - * vp_modern_get_queue_size - get size for a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * > - * Returns the size of the virtqueue > - */ > -static u16 vp_modern_get_queue_size(struct virtio_pci_modern_device *mdev, > - u16 index) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - > - return vp_ioread16(&mdev->common->queue_size); > - > -} > - > -/* > - * vp_modern_get_num_queues - get the number of virtqueues > - * @mdev: the modern virtio-pci device > - * > - * Returns the number of virtqueues > - */ > -static u16 vp_modern_get_num_queues(struct virtio_pci_modern_device *mdev) > -{ > - return vp_ioread16(&mdev->common->num_queues); > -} > - > -/* > - * vp_modern_get_queue_notify_off - get notification offset for a virtqueue > - * @mdev: the modern virtio-pci device > - * @index: the queue index > - * > - * Returns the notification offset for a virtqueue > - */ > -static u16 vp_modern_get_queue_notify_off(struct virtio_pci_modern_device *mdev, > - u16 index) > -{ > - vp_iowrite16(index, &mdev->common->queue_select); > - > - return vp_ioread16(&mdev->common->queue_notify_off); > -} > - > /* virtio config->finalize_features() implementation */ > static int vp_finalize_features(struct virtio_device *vdev) > { > @@ -429,19 +134,6 @@ static void vp_set(struct virtio_device *vdev, unsigned offset, > } > } > > -/* > - * vp_modern_generation - get the device genreation > - * @mdev: the modern virtio-pci device > - * > - * Returns the genreation read from device > - */ > -static u32 vp_modern_generation(struct virtio_pci_modern_device *mdev) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - return vp_ioread8(&cfg->config_generation); > -} > - > static u32 vp_generation(struct virtio_device *vdev) > { > struct virtio_pci_device *vp_dev = to_vp_device(vdev); > @@ -449,19 +141,6 @@ static u32 vp_generation(struct virtio_device *vdev) > return vp_modern_generation(&vp_dev->mdev); > } > > -/* > - * vp_modern_get_status - get the device status > - * @mdev: the modern virtio-pci device > - * > - * Returns the status read from device > - */ > -static u8 vp_modern_get_status(struct virtio_pci_modern_device *mdev) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - return vp_ioread8(&cfg->device_status); > -} > - > /* config->{get,set}_status() implementations */ > static u8 vp_get_status(struct virtio_device *vdev) > { > @@ -470,19 +149,6 @@ static u8 vp_get_status(struct virtio_device *vdev) > return vp_modern_get_status(&vp_dev->mdev); > } > > -/* > - * vp_modern_set_status - set status to device > - * @mdev: the modern virtio-pci device > - * @status: the status set to device > - */ > -static void vp_modern_set_status(struct virtio_pci_modern_device *mdev, > - u8 status) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - vp_iowrite8(status, &cfg->device_status); > -} > - > static void vp_set_status(struct virtio_device *vdev, u8 status) > { > struct virtio_pci_device *vp_dev = to_vp_device(vdev); > @@ -510,25 +176,6 @@ static void vp_reset(struct virtio_device *vdev) > vp_synchronize_vectors(vdev); > } > > -/* > - * vp_modern_config_vector - set the vector for config interrupt > - * @mdev: the modern virtio-pci device > - * @vector: the config vector > - * > - * Returns the config vector read from the device > - */ > -static u16 vp_modern_config_vector(struct virtio_pci_modern_device *mdev, > - u16 vector) > -{ > - struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > - > - /* Setup the vector used for configuration events */ > - vp_iowrite16(vector, &cfg->msix_config); > - /* Verify we had enough resources to assign the vector */ > - /* Will also flush the write out to device */ > - return vp_ioread16(&cfg->msix_config); > -} > - > static u16 vp_config_vector(struct virtio_pci_device *vp_dev, u16 vector) > { > return vp_modern_config_vector(&vp_dev->mdev, vector); > @@ -789,253 +436,6 @@ static const struct virtio_config_ops virtio_pci_config_ops = { > .get_shm_region = vp_get_shm_region, > }; > > -/** > - * virtio_pci_find_capability - walk capabilities to find device info. > - * @dev: the pci device > - * @cfg_type: the VIRTIO_PCI_CAP_* value we seek > - * @ioresource_types: IORESOURCE_MEM and/or IORESOURCE_IO. > - * @bars: the bitmask of BARs > - * > - * Returns offset of the capability, or 0. > - */ > -static inline int virtio_pci_find_capability(struct pci_dev *dev, u8 cfg_type, > - u32 ioresource_types, int *bars) > -{ > - int pos; > - > - for (pos = pci_find_capability(dev, PCI_CAP_ID_VNDR); > - pos > 0; > - pos = pci_find_next_capability(dev, pos, PCI_CAP_ID_VNDR)) { > - u8 type, bar; > - pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap, > - cfg_type), > - &type); > - pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap, > - bar), > - &bar); > - > - /* Ignore structures with reserved BAR values */ > - if (bar > 0x5) > - continue; > - > - if (type == cfg_type) { > - if (pci_resource_len(dev, bar) && > - pci_resource_flags(dev, bar) & ioresource_types) { > - *bars |= (1 << bar); > - return pos; > - } > - } > - } > - return 0; > -} > - > -/* This is part of the ABI. Don't screw with it. */ > -static inline void check_offsets(void) > -{ > - /* Note: disk space was harmed in compilation of this function. */ > - BUILD_BUG_ON(VIRTIO_PCI_CAP_VNDR !> - offsetof(struct virtio_pci_cap, cap_vndr)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_NEXT !> - offsetof(struct virtio_pci_cap, cap_next)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_LEN !> - offsetof(struct virtio_pci_cap, cap_len)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_CFG_TYPE !> - offsetof(struct virtio_pci_cap, cfg_type)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_BAR !> - offsetof(struct virtio_pci_cap, bar)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_OFFSET !> - offsetof(struct virtio_pci_cap, offset)); > - BUILD_BUG_ON(VIRTIO_PCI_CAP_LENGTH !> - offsetof(struct virtio_pci_cap, length)); > - BUILD_BUG_ON(VIRTIO_PCI_NOTIFY_CAP_MULT !> - offsetof(struct virtio_pci_notify_cap, > - notify_off_multiplier)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_DFSELECT !> - offsetof(struct virtio_pci_common_cfg, > - device_feature_select)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_DF !> - offsetof(struct virtio_pci_common_cfg, device_feature)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_GFSELECT !> - offsetof(struct virtio_pci_common_cfg, > - guest_feature_select)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_GF !> - offsetof(struct virtio_pci_common_cfg, guest_feature)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_MSIX !> - offsetof(struct virtio_pci_common_cfg, msix_config)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_NUMQ !> - offsetof(struct virtio_pci_common_cfg, num_queues)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_STATUS !> - offsetof(struct virtio_pci_common_cfg, device_status)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_CFGGENERATION !> - offsetof(struct virtio_pci_common_cfg, config_generation)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SELECT !> - offsetof(struct virtio_pci_common_cfg, queue_select)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SIZE !> - offsetof(struct virtio_pci_common_cfg, queue_size)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_MSIX !> - offsetof(struct virtio_pci_common_cfg, queue_msix_vector)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_ENABLE !> - offsetof(struct virtio_pci_common_cfg, queue_enable)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_NOFF !> - offsetof(struct virtio_pci_common_cfg, queue_notify_off)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCLO !> - offsetof(struct virtio_pci_common_cfg, queue_desc_lo)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCHI !> - offsetof(struct virtio_pci_common_cfg, queue_desc_hi)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILLO !> - offsetof(struct virtio_pci_common_cfg, queue_avail_lo)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILHI !> - offsetof(struct virtio_pci_common_cfg, queue_avail_hi)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDLO !> - offsetof(struct virtio_pci_common_cfg, queue_used_lo)); > - BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDHI !> - offsetof(struct virtio_pci_common_cfg, queue_used_hi)); > -} > - > -/* > - * vp_modern_probe: probe the modern virtio pci device, note that the > - * caller is required to enable PCI device before calling this function. > - * @mdev: the modern virtio-pci device > - * > - * Return 0 on succeed otherwise fail > - */ > -static int vp_modern_probe(struct virtio_pci_modern_device *mdev) > -{ > - struct pci_dev *pci_dev = mdev->pci_dev; > - int err, common, isr, notify, device; > - u32 notify_length; > - u32 notify_offset; > - > - check_offsets(); > - > - mdev->pci_dev = pci_dev; > - > - /* We only own devices >= 0x1000 and <= 0x107f: leave the rest. */ > - if (pci_dev->device < 0x1000 || pci_dev->device > 0x107f) > - return -ENODEV; > - > - if (pci_dev->device < 0x1040) { > - /* Transitional devices: use the PCI subsystem device id as > - * virtio device id, same as legacy driver always did. > - */ > - mdev->id.device = pci_dev->subsystem_device; > - } else { > - /* Modern devices: simply use PCI device id, but start from 0x1040. */ > - mdev->id.device = pci_dev->device - 0x1040; > - } > - mdev->id.vendor = pci_dev->subsystem_vendor; > - > - /* check for a common config: if not, use legacy mode (bar 0). */ > - common = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_COMMON_CFG, > - IORESOURCE_IO | IORESOURCE_MEM, > - &mdev->modern_bars); > - if (!common) { > - dev_info(&pci_dev->dev, > - "virtio_pci: leaving for legacy driver\n"); > - return -ENODEV; > - } > - > - /* If common is there, these should be too... */ > - isr = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_ISR_CFG, > - IORESOURCE_IO | IORESOURCE_MEM, > - &mdev->modern_bars); > - notify = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_NOTIFY_CFG, > - IORESOURCE_IO | IORESOURCE_MEM, > - &mdev->modern_bars); > - if (!isr || !notify) { > - dev_err(&pci_dev->dev, > - "virtio_pci: missing capabilities %i/%i/%i\n", > - common, isr, notify); > - return -EINVAL; > - } > - > - err = dma_set_mask_and_coherent(&pci_dev->dev, DMA_BIT_MASK(64)); > - if (err) > - err = dma_set_mask_and_coherent(&pci_dev->dev, > - DMA_BIT_MASK(32)); > - if (err) > - dev_warn(&pci_dev->dev, "Failed to enable 64-bit or 32-bit DMA. Trying to continue, but this might not work.\n"); > - > - /* Device capability is only mandatory for devices that have > - * device-specific configuration. > - */ > - device = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_DEVICE_CFG, > - IORESOURCE_IO | IORESOURCE_MEM, > - &mdev->modern_bars); > - > - err = pci_request_selected_regions(pci_dev, mdev->modern_bars, > - "virtio-pci-modern"); > - if (err) > - return err; > - > - err = -EINVAL; > - mdev->common = vp_modern_map_capability(mdev, common, > - sizeof(struct virtio_pci_common_cfg), 4, > - 0, sizeof(struct virtio_pci_common_cfg), > - NULL); > - if (!mdev->common) > - goto err_map_common; > - mdev->isr = vp_modern_map_capability(mdev, isr, sizeof(u8), 1, > - 0, 1, > - NULL); > - if (!mdev->isr) > - goto err_map_isr; > - > - /* Read notify_off_multiplier from config space. */ > - pci_read_config_dword(pci_dev, > - notify + offsetof(struct virtio_pci_notify_cap, > - notify_off_multiplier), > - &mdev->notify_offset_multiplier); > - /* Read notify length and offset from config space. */ > - pci_read_config_dword(pci_dev, > - notify + offsetof(struct virtio_pci_notify_cap, > - cap.length), > - ¬ify_length); > - > - pci_read_config_dword(pci_dev, > - notify + offsetof(struct virtio_pci_notify_cap, > - cap.offset), > - ¬ify_offset); > - > - /* We don't know how many VQs we'll map, ahead of the time. > - * If notify length is small, map it all now. > - * Otherwise, map each VQ individually later. > - */ > - if ((u64)notify_length + (notify_offset % PAGE_SIZE) <= PAGE_SIZE) { > - mdev->notify_base = vp_modern_map_capability(mdev, notify, > - 2, 2, > - 0, notify_length, > - &mdev->notify_len); > - if (!mdev->notify_base) > - goto err_map_notify; > - } else { > - mdev->notify_map_cap = notify; > - } > - > - /* Again, we don't know how much we should map, but PAGE_SIZE > - * is more than enough for all existing devices. > - */ > - if (device) { > - mdev->device = vp_modern_map_capability(mdev, device, 0, 4, > - 0, PAGE_SIZE, > - &mdev->device_len); > - if (!mdev->device) > - goto err_map_device; > - } > - > - return 0; > - > -err_map_device: > - if (mdev->notify_base) > - pci_iounmap(pci_dev, mdev->notify_base); > -err_map_notify: > - pci_iounmap(pci_dev, mdev->isr); > -err_map_isr: > - pci_iounmap(pci_dev, mdev->common); > -err_map_common: > - return err; > -} > - > /* the PCI probing function */ > int virtio_pci_modern_probe(struct virtio_pci_device *vp_dev) > { > @@ -1063,23 +463,6 @@ int virtio_pci_modern_probe(struct virtio_pci_device *vp_dev) > return 0; > } > > -/* > - * vp_modern_probe: remove and cleanup the modern virtio pci device > - * @mdev: the modern virtio-pci device > - */ > -static void vp_modern_remove(struct virtio_pci_modern_device *mdev) > -{ > - struct pci_dev *pci_dev = mdev->pci_dev; > - > - if (mdev->device) > - pci_iounmap(pci_dev, mdev->device); > - if (mdev->notify_base) > - pci_iounmap(pci_dev, mdev->notify_base); > - pci_iounmap(pci_dev, mdev->isr); > - pci_iounmap(pci_dev, mdev->common); > - pci_release_selected_regions(pci_dev, mdev->modern_bars); > -} > - > void virtio_pci_modern_remove(struct virtio_pci_device *vp_dev) > { > struct virtio_pci_modern_device *mdev = &vp_dev->mdev; > diff --git a/drivers/virtio/virtio_pci_modern_dev.c b/drivers/virtio/virtio_pci_modern_dev.c > new file mode 100644 > index 000000000000..cbd667496bb1 > --- /dev/null > +++ b/drivers/virtio/virtio_pci_modern_dev.c > @@ -0,0 +1,599 @@ > +// SPDX-License-Identifier: GPL-2.0-or-later > + > +#include <linux/virtio_pci_modern.h> > +#include <linux/module.h> > +#include <linux/pci.h> > + > +/* > + * vp_modern_map_capability - map a part of virtio pci capability > + * @mdev: the modern virtio-pci device > + * @off: offset of the capability > + * @minlen: minimal length of the capability > + * @align: align requirement > + * @start: start from the capability > + * @size: map size > + * @len: the length that is actually mapped > + * > + * Returns the io address of for the part of the capability > + */ > +void __iomem *vp_modern_map_capability(struct virtio_pci_modern_device *mdev, int off, > + size_t minlen, > + u32 align, > + u32 start, u32 size, > + size_t *len) > +{ > + struct pci_dev *dev = mdev->pci_dev; > + u8 bar; > + u32 offset, length; > + void __iomem *p; > + > + pci_read_config_byte(dev, off + offsetof(struct virtio_pci_cap, > + bar), > + &bar); > + pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, offset), > + &offset); > + pci_read_config_dword(dev, off + offsetof(struct virtio_pci_cap, length), > + &length); > + > + if (length <= start) { > + dev_err(&dev->dev, > + "virtio_pci: bad capability len %u (>%u expected)\n", > + length, start); > + return NULL; > + } > + > + if (length - start < minlen) { > + dev_err(&dev->dev, > + "virtio_pci: bad capability len %u (>=%zu expected)\n", > + length, minlen); > + return NULL; > + } > + > + length -= start; > + > + if (start + offset < offset) { > + dev_err(&dev->dev, > + "virtio_pci: map wrap-around %u+%u\n", > + start, offset); > + return NULL; > + } > + > + offset += start; > + > + if (offset & (align - 1)) { > + dev_err(&dev->dev, > + "virtio_pci: offset %u not aligned to %u\n", > + offset, align); > + return NULL; > + } > + > + if (length > size) > + length = size; > + > + if (len) > + *len = length; > + > + if (minlen + offset < minlen || > + minlen + offset > pci_resource_len(dev, bar)) { > + dev_err(&dev->dev, > + "virtio_pci: map virtio %zu@%u " > + "out of range on bar %i length %lu\n", > + minlen, offset, > + bar, (unsigned long)pci_resource_len(dev, bar)); > + return NULL; > + } > + > + p = pci_iomap_range(dev, bar, offset, length); > + if (!p) > + dev_err(&dev->dev, > + "virtio_pci: unable to map virtio %u@%u on bar %i\n", > + length, offset, bar); > + return p; > +} > +EXPORT_SYMBOL_GPL(vp_modern_map_capability); > + > +/** > + * virtio_pci_find_capability - walk capabilities to find device info. > + * @dev: the pci device > + * @cfg_type: the VIRTIO_PCI_CAP_* value we seek > + * @ioresource_types: IORESOURCE_MEM and/or IORESOURCE_IO. > + * @bars: the bitmask of BARs > + * > + * Returns offset of the capability, or 0. > + */ > +static inline int virtio_pci_find_capability(struct pci_dev *dev, u8 cfg_type, > + u32 ioresource_types, int *bars) > +{ > + int pos; > + > + for (pos = pci_find_capability(dev, PCI_CAP_ID_VNDR); > + pos > 0; > + pos = pci_find_next_capability(dev, pos, PCI_CAP_ID_VNDR)) { > + u8 type, bar; > + pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap, > + cfg_type), > + &type); > + pci_read_config_byte(dev, pos + offsetof(struct virtio_pci_cap, > + bar), > + &bar); > + > + /* Ignore structures with reserved BAR values */ > + if (bar > 0x5) > + continue; > + > + if (type == cfg_type) { > + if (pci_resource_len(dev, bar) && > + pci_resource_flags(dev, bar) & ioresource_types) { > + *bars |= (1 << bar); > + return pos; > + } > + } > + } > + return 0; > +} > + > +/* This is part of the ABI. Don't screw with it. */ > +static inline void check_offsets(void) > +{ > + /* Note: disk space was harmed in compilation of this function. */ > + BUILD_BUG_ON(VIRTIO_PCI_CAP_VNDR !> + offsetof(struct virtio_pci_cap, cap_vndr)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_NEXT !> + offsetof(struct virtio_pci_cap, cap_next)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_LEN !> + offsetof(struct virtio_pci_cap, cap_len)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_CFG_TYPE !> + offsetof(struct virtio_pci_cap, cfg_type)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_BAR !> + offsetof(struct virtio_pci_cap, bar)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_OFFSET !> + offsetof(struct virtio_pci_cap, offset)); > + BUILD_BUG_ON(VIRTIO_PCI_CAP_LENGTH !> + offsetof(struct virtio_pci_cap, length)); > + BUILD_BUG_ON(VIRTIO_PCI_NOTIFY_CAP_MULT !> + offsetof(struct virtio_pci_notify_cap, > + notify_off_multiplier)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_DFSELECT !> + offsetof(struct virtio_pci_common_cfg, > + device_feature_select)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_DF !> + offsetof(struct virtio_pci_common_cfg, device_feature)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_GFSELECT !> + offsetof(struct virtio_pci_common_cfg, > + guest_feature_select)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_GF !> + offsetof(struct virtio_pci_common_cfg, guest_feature)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_MSIX !> + offsetof(struct virtio_pci_common_cfg, msix_config)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_NUMQ !> + offsetof(struct virtio_pci_common_cfg, num_queues)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_STATUS !> + offsetof(struct virtio_pci_common_cfg, device_status)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_CFGGENERATION !> + offsetof(struct virtio_pci_common_cfg, config_generation)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SELECT !> + offsetof(struct virtio_pci_common_cfg, queue_select)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_SIZE !> + offsetof(struct virtio_pci_common_cfg, queue_size)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_MSIX !> + offsetof(struct virtio_pci_common_cfg, queue_msix_vector)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_ENABLE !> + offsetof(struct virtio_pci_common_cfg, queue_enable)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_NOFF !> + offsetof(struct virtio_pci_common_cfg, queue_notify_off)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCLO !> + offsetof(struct virtio_pci_common_cfg, queue_desc_lo)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_DESCHI !> + offsetof(struct virtio_pci_common_cfg, queue_desc_hi)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILLO !> + offsetof(struct virtio_pci_common_cfg, queue_avail_lo)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_AVAILHI !> + offsetof(struct virtio_pci_common_cfg, queue_avail_hi)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDLO !> + offsetof(struct virtio_pci_common_cfg, queue_used_lo)); > + BUILD_BUG_ON(VIRTIO_PCI_COMMON_Q_USEDHI !> + offsetof(struct virtio_pci_common_cfg, queue_used_hi)); > +} > + > +/* > + * vp_modern_probe: probe the modern virtio pci device, note that the > + * caller is required to enable PCI device before calling this function. > + * @mdev: the modern virtio-pci device > + * > + * Return 0 on succeed otherwise fail > + */ > +int vp_modern_probe(struct virtio_pci_modern_device *mdev) > +{ > + struct pci_dev *pci_dev = mdev->pci_dev; > + int err, common, isr, notify, device; > + u32 notify_length; > + u32 notify_offset; > + > + check_offsets(); > + > + mdev->pci_dev = pci_dev; > + > + /* We only own devices >= 0x1000 and <= 0x107f: leave the rest. */ > + if (pci_dev->device < 0x1000 || pci_dev->device > 0x107f) > + return -ENODEV; > + > + if (pci_dev->device < 0x1040) { > + /* Transitional devices: use the PCI subsystem device id as > + * virtio device id, same as legacy driver always did. > + */ > + mdev->id.device = pci_dev->subsystem_device; > + } else { > + /* Modern devices: simply use PCI device id, but start from 0x1040. */ > + mdev->id.device = pci_dev->device - 0x1040; > + } > + mdev->id.vendor = pci_dev->subsystem_vendor; > + > + /* check for a common config: if not, use legacy mode (bar 0). */ > + common = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_COMMON_CFG, > + IORESOURCE_IO | IORESOURCE_MEM, > + &mdev->modern_bars); > + if (!common) { > + dev_info(&pci_dev->dev, > + "virtio_pci: leaving for legacy driver\n"); > + return -ENODEV; > + } > + > + /* If common is there, these should be too... */ > + isr = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_ISR_CFG, > + IORESOURCE_IO | IORESOURCE_MEM, > + &mdev->modern_bars); > + notify = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_NOTIFY_CFG, > + IORESOURCE_IO | IORESOURCE_MEM, > + &mdev->modern_bars); > + if (!isr || !notify) { > + dev_err(&pci_dev->dev, > + "virtio_pci: missing capabilities %i/%i/%i\n", > + common, isr, notify); > + return -EINVAL; > + } > + > + err = dma_set_mask_and_coherent(&pci_dev->dev, DMA_BIT_MASK(64)); > + if (err) > + err = dma_set_mask_and_coherent(&pci_dev->dev, > + DMA_BIT_MASK(32)); > + if (err) > + dev_warn(&pci_dev->dev, "Failed to enable 64-bit or 32-bit DMA. Trying to continue, but this might not work.\n"); > + > + /* Device capability is only mandatory for devices that have > + * device-specific configuration. > + */ > + device = virtio_pci_find_capability(pci_dev, VIRTIO_PCI_CAP_DEVICE_CFG, > + IORESOURCE_IO | IORESOURCE_MEM, > + &mdev->modern_bars); > + > + err = pci_request_selected_regions(pci_dev, mdev->modern_bars, > + "virtio-pci-modern"); > + if (err) > + return err; > + > + err = -EINVAL; > + mdev->common = vp_modern_map_capability(mdev, common, > + sizeof(struct virtio_pci_common_cfg), 4, > + 0, sizeof(struct virtio_pci_common_cfg), > + NULL); > + if (!mdev->common) > + goto err_map_common; > + mdev->isr = vp_modern_map_capability(mdev, isr, sizeof(u8), 1, > + 0, 1, > + NULL); > + if (!mdev->isr) > + goto err_map_isr; > + > + /* Read notify_off_multiplier from config space. */ > + pci_read_config_dword(pci_dev, > + notify + offsetof(struct virtio_pci_notify_cap, > + notify_off_multiplier), > + &mdev->notify_offset_multiplier); > + /* Read notify length and offset from config space. */ > + pci_read_config_dword(pci_dev, > + notify + offsetof(struct virtio_pci_notify_cap, > + cap.length), > + ¬ify_length); > + > + pci_read_config_dword(pci_dev, > + notify + offsetof(struct virtio_pci_notify_cap, > + cap.offset), > + ¬ify_offset); > + > + /* We don't know how many VQs we'll map, ahead of the time. > + * If notify length is small, map it all now. > + * Otherwise, map each VQ individually later. > + */ > + if ((u64)notify_length + (notify_offset % PAGE_SIZE) <= PAGE_SIZE) { > + mdev->notify_base = vp_modern_map_capability(mdev, notify, > + 2, 2, > + 0, notify_length, > + &mdev->notify_len); > + if (!mdev->notify_base) > + goto err_map_notify; > + } else { > + mdev->notify_map_cap = notify; > + } > + > + /* Again, we don't know how much we should map, but PAGE_SIZE > + * is more than enough for all existing devices. > + */ > + if (device) { > + mdev->device = vp_modern_map_capability(mdev, device, 0, 4, > + 0, PAGE_SIZE, > + &mdev->device_len); > + if (!mdev->device) > + goto err_map_device; > + } > + > + return 0; > + > +err_map_device: > + if (mdev->notify_base) > + pci_iounmap(pci_dev, mdev->notify_base); > +err_map_notify: > + pci_iounmap(pci_dev, mdev->isr); > +err_map_isr: > + pci_iounmap(pci_dev, mdev->common); > +err_map_common: > + return err; > +} > +EXPORT_SYMBOL_GPL(vp_modern_probe); > + > +/* > + * vp_modern_probe: remove and cleanup the modern virtio pci device > + * @mdev: the modern virtio-pci device > + */ > +void vp_modern_remove(struct virtio_pci_modern_device *mdev) > +{ > + struct pci_dev *pci_dev = mdev->pci_dev; > + > + if (mdev->device) > + pci_iounmap(pci_dev, mdev->device); > + if (mdev->notify_base) > + pci_iounmap(pci_dev, mdev->notify_base); > + pci_iounmap(pci_dev, mdev->isr); > + pci_iounmap(pci_dev, mdev->common); > + pci_release_selected_regions(pci_dev, mdev->modern_bars); > +} > +EXPORT_SYMBOL_GPL(vp_modern_remove); > + > +/* > + * vp_modern_get_features - get features from device > + * @mdev: the modern virtio-pci device > + * > + * Returns the features read from the device > + */ > +u64 vp_modern_get_features(struct virtio_pci_modern_device *mdev) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + u64 features; > + > + vp_iowrite32(0, &cfg->device_feature_select); > + features = vp_ioread32(&cfg->device_feature); > + vp_iowrite32(1, &cfg->device_feature_select); > + features |= ((u64)vp_ioread32(&cfg->device_feature) << 32); > + > + return features; > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_features); > + > +/* > + * vp_modern_set_features - set features to device > + * @mdev: the modern virtio-pci device > + * @features: the features set to device > + */ > +void vp_modern_set_features(struct virtio_pci_modern_device *mdev, > + u64 features) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + vp_iowrite32(0, &cfg->guest_feature_select); > + vp_iowrite32((u32)features, &cfg->guest_feature); > + vp_iowrite32(1, &cfg->guest_feature_select); > + vp_iowrite32(features >> 32, &cfg->guest_feature); > +} > +EXPORT_SYMBOL_GPL(vp_modern_set_features); > + > +/* > + * vp_modern_generation - get the device genreation > + * @mdev: the modern virtio-pci device > + * > + * Returns the genreation read from device > + */ > +u32 vp_modern_generation(struct virtio_pci_modern_device *mdev) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + return vp_ioread8(&cfg->config_generation); > +} > +EXPORT_SYMBOL_GPL(vp_modern_generation); > + > +/* > + * vp_modern_get_status - get the device status > + * @mdev: the modern virtio-pci device > + * > + * Returns the status read from device > + */ > +u8 vp_modern_get_status(struct virtio_pci_modern_device *mdev) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + return vp_ioread8(&cfg->device_status); > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_status); > + > +/* > + * vp_modern_set_status - set status to device > + * @mdev: the modern virtio-pci device > + * @status: the status set to device > + */ > +void vp_modern_set_status(struct virtio_pci_modern_device *mdev, > + u8 status) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + vp_iowrite8(status, &cfg->device_status); > +} > +EXPORT_SYMBOL_GPL(vp_modern_set_status); > + > +/* > + * vp_modern_queue_vector - set the MSIX vector for a specific virtqueue > + * @mdev: the modern virtio-pci device > + * @index: queue index > + * @vector: the config vector > + * > + * Returns the config vector read from the device > + */ > +u16 vp_modern_queue_vector(struct virtio_pci_modern_device *mdev, > + u16 index, u16 vector) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + vp_iowrite16(index, &cfg->queue_select); > + vp_iowrite16(vector, &cfg->queue_msix_vector); > + /* Flush the write out to device */ > + return vp_ioread16(&cfg->queue_msix_vector); > +} > +EXPORT_SYMBOL_GPL(vp_modern_queue_vector); > + > +/* > + * vp_modern_config_vector - set the vector for config interrupt > + * @mdev: the modern virtio-pci device > + * @vector: the config vector > + * > + * Returns the config vector read from the device > + */ > +u16 vp_modern_config_vector(struct virtio_pci_modern_device *mdev, > + u16 vector) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + /* Setup the vector used for configuration events */ > + vp_iowrite16(vector, &cfg->msix_config); > + /* Verify we had enough resources to assign the vector */ > + /* Will also flush the write out to device */ > + return vp_ioread16(&cfg->msix_config); > +} > +EXPORT_SYMBOL_GPL(vp_modern_config_vector); > + > +/* > + * vp_modern_queue_address - set the virtqueue address > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * @desc_addr: address of the descriptor area > + * @driver_addr: address of the driver area > + * @device_addr: address of the device area > + */ > +void vp_modern_queue_address(struct virtio_pci_modern_device *mdev, > + u16 index, u64 desc_addr, u64 driver_addr, > + u64 device_addr) > +{ > + struct virtio_pci_common_cfg __iomem *cfg = mdev->common; > + > + vp_iowrite16(index, &cfg->queue_select); > + > + vp_iowrite64_twopart(desc_addr, &cfg->queue_desc_lo, > + &cfg->queue_desc_hi); > + vp_iowrite64_twopart(driver_addr, &cfg->queue_avail_lo, > + &cfg->queue_avail_hi); > + vp_iowrite64_twopart(device_addr, &cfg->queue_used_lo, > + &cfg->queue_used_hi); > +} > +EXPORT_SYMBOL_GPL(vp_modern_queue_address); > + > +/* > + * vp_modern_set_queue_enable - enable a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * @enable: whether the virtqueue is enable or not > + */ > +void vp_modern_set_queue_enable(struct virtio_pci_modern_device *mdev, > + u16 index, bool enable) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + vp_iowrite16(enable, &mdev->common->queue_enable); > +} > +EXPORT_SYMBOL_GPL(vp_modern_set_queue_enable); > + > +/* > + * vp_modern_get_queue_enable - enable a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * > + * Returns whether a virtqueue is enabled or not > + */ > +bool vp_modern_get_queue_enable(struct virtio_pci_modern_device *mdev, > + u16 index) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + > + return vp_ioread16(&mdev->common->queue_enable); > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_queue_enable); > + > +/* > + * vp_modern_set_queue_size - set size for a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * @size: the size of the virtqueue > + */ > +void vp_modern_set_queue_size(struct virtio_pci_modern_device *mdev, > + u16 index, u16 size) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + vp_iowrite16(size, &mdev->common->queue_size); > + > +} > +EXPORT_SYMBOL_GPL(vp_modern_set_queue_size); > + > +/* > + * vp_modern_get_queue_size - get size for a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * > + * Returns the size of the virtqueue > + */ > +u16 vp_modern_get_queue_size(struct virtio_pci_modern_device *mdev, > + u16 index) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + > + return vp_ioread16(&mdev->common->queue_size); > + > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_queue_size); > + > +/* > + * vp_modern_get_num_queues - get the number of virtqueues > + * @mdev: the modern virtio-pci device > + * > + * Returns the number of virtqueues > + */ > +u16 vp_modern_get_num_queues(struct virtio_pci_modern_device *mdev) > +{ > + return vp_ioread16(&mdev->common->num_queues); > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_num_queues); > + > +/* > + * vp_modern_get_queue_notify_off - get notification offset for a virtqueue > + * @mdev: the modern virtio-pci device > + * @index: the queue index > + * > + * Returns the notification offset for a virtqueue > + */ > +u16 vp_modern_get_queue_notify_off(struct virtio_pci_modern_device *mdev, > + u16 index) > +{ > + vp_iowrite16(index, &mdev->common->queue_select); > + > + return vp_ioread16(&mdev->common->queue_notify_off); > +} > +EXPORT_SYMBOL_GPL(vp_modern_get_queue_notify_off); > + > +MODULE_VERSION("0.1"); > +MODULE_DESCRIPTION("Modern Virtio PCI Device"); > +MODULE_AUTHOR("Jason Wang <jasowang at redhat.com>"); > +MODULE_LICENSE("GPL"); > diff --git a/include/linux/virtio_pci_modern.h b/include/linux/virtio_pci_modern.h > new file mode 100644 > index 000000000000..f26acbeec965 > --- /dev/null > +++ b/include/linux/virtio_pci_modern.h > @@ -0,0 +1,111 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +#ifndef _LINUX_VIRTIO_PCI_MODERN_H > +#define _LINUX_VIRTIO_PCI_MODERN_H > + > +#include <linux/pci.h> > +#include <linux/virtio_pci.h> > + > +struct virtio_pci_modern_device { > + struct pci_dev *pci_dev; > + > + struct virtio_pci_common_cfg __iomem *common; > + /* Device-specific data (non-legacy mode) */ > + void __iomem *device; > + /* Base of vq notifications (non-legacy mode). */ > + void __iomem *notify_base; > + /* Where to read and clear interrupt */ > + u8 __iomem *isr; > + > + /* So we can sanity-check accesses. */ > + size_t notify_len; > + size_t device_len; > + > + /* Capability for when we need to map notifications per-vq. */ > + int notify_map_cap; > + > + /* Multiply queue_notify_off by this value. (non-legacy mode). */ > + u32 notify_offset_multiplier; > + > + int modern_bars; > + > + struct virtio_device_id id; > +}; > + > +/* > + * Type-safe wrappers for io accesses. > + * Use these to enforce at compile time the following spec requirement: > + * > + * The driver MUST access each field using the ?natural? access > + * method, i.e. 32-bit accesses for 32-bit fields, 16-bit accesses > + * for 16-bit fields and 8-bit accesses for 8-bit fields. > + */ > +static inline u8 vp_ioread8(const u8 __iomem *addr) > +{ > + return ioread8(addr); > +} > +static inline u16 vp_ioread16 (const __le16 __iomem *addr) > +{ > + return ioread16(addr); > +} > + > +static inline u32 vp_ioread32(const __le32 __iomem *addr) > +{ > + return ioread32(addr); > +} > + > +static inline void vp_iowrite8(u8 value, u8 __iomem *addr) > +{ > + iowrite8(value, addr); > +} > + > +static inline void vp_iowrite16(u16 value, __le16 __iomem *addr) > +{ > + iowrite16(value, addr); > +} > + > +static inline void vp_iowrite32(u32 value, __le32 __iomem *addr) > +{ > + iowrite32(value, addr); > +} > + > +static inline void vp_iowrite64_twopart(u64 val, > + __le32 __iomem *lo, > + __le32 __iomem *hi) > +{ > + vp_iowrite32((u32)val, lo); > + vp_iowrite32(val >> 32, hi); > +} > + > +u64 vp_modern_get_features(struct virtio_pci_modern_device *mdev); > +void vp_modern_set_features(struct virtio_pci_modern_device *mdev, > + u64 features); > +u32 vp_modern_generation(struct virtio_pci_modern_device *mdev); > +u8 vp_modern_get_status(struct virtio_pci_modern_device *mdev); > +void vp_modern_set_status(struct virtio_pci_modern_device *mdev, > + u8 status); > +u16 vp_modern_queue_vector(struct virtio_pci_modern_device *mdev, > + u16 idx, u16 vector); > +u16 vp_modern_config_vector(struct virtio_pci_modern_device *mdev, > + u16 vector); > +void vp_modern_queue_address(struct virtio_pci_modern_device *mdev, > + u16 index, u64 desc_addr, u64 driver_addr, > + u64 device_addr); > +void vp_modern_set_queue_enable(struct virtio_pci_modern_device *mdev, > + u16 idx, bool enable); > +bool vp_modern_get_queue_enable(struct virtio_pci_modern_device *mdev, > + u16 idx); > +void vp_modern_set_queue_size(struct virtio_pci_modern_device *mdev, > + u16 idx, u16 size); > +u16 vp_modern_get_queue_size(struct virtio_pci_modern_device *mdev, > + u16 idx); > +u16 vp_modern_get_num_queues(struct virtio_pci_modern_device *mdev); > +u16 vp_modern_get_queue_notify_off(struct virtio_pci_modern_device *mdev, > + u16 idx); > +void __iomem *vp_modern_map_capability(struct virtio_pci_modern_device *mdev, int off, > + size_t minlen, > + u32 align, > + u32 start, u32 size, > + size_t *len); > +int vp_modern_probe(struct virtio_pci_modern_device *mdev); > +void vp_modern_remove(struct virtio_pci_modern_device *mdev); > +#endif