mirror of https://gitee.com/openkylin/linux.git
Merge branches 'v5.10/vfio/bardirty', 'v5.10/vfio/dma_avail', 'v5.10/vfio/misc', 'v5.10/vfio/no-cmd-mem' and 'v5.10/vfio/yan_zhao_fixes' into v5.10/vfio/next
This commit is contained in:
commit
3de066f8f8
|
@ -197,9 +197,10 @@ void pcibios_bus_add_device(struct pci_dev *pdev)
|
||||||
* With pdev->no_vf_scan the common PCI probing code does not
|
* With pdev->no_vf_scan the common PCI probing code does not
|
||||||
* perform PF/VF linking.
|
* perform PF/VF linking.
|
||||||
*/
|
*/
|
||||||
if (zdev->vfn)
|
if (zdev->vfn) {
|
||||||
zpci_bus_setup_virtfn(zdev->zbus, pdev, zdev->vfn);
|
zpci_bus_setup_virtfn(zdev->zbus, pdev, zdev->vfn);
|
||||||
|
pdev->no_command_memory = 1;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static int zpci_bus_add_device(struct zpci_bus *zbus, struct zpci_dev *zdev)
|
static int zpci_bus_add_device(struct zpci_bus *zbus, struct zpci_dev *zdev)
|
||||||
|
|
|
@ -180,6 +180,7 @@ int pci_iov_add_virtfn(struct pci_dev *dev, int id)
|
||||||
virtfn->device = iov->vf_device;
|
virtfn->device = iov->vf_device;
|
||||||
virtfn->is_virtfn = 1;
|
virtfn->is_virtfn = 1;
|
||||||
virtfn->physfn = pci_dev_get(dev);
|
virtfn->physfn = pci_dev_get(dev);
|
||||||
|
virtfn->no_command_memory = 1;
|
||||||
|
|
||||||
if (id == 0)
|
if (id == 0)
|
||||||
pci_read_vf_config_common(virtfn);
|
pci_read_vf_config_common(virtfn);
|
||||||
|
|
|
@ -1862,7 +1862,6 @@ static const struct vfio_device_ops vfio_pci_ops = {
|
||||||
|
|
||||||
static int vfio_pci_reflck_attach(struct vfio_pci_device *vdev);
|
static int vfio_pci_reflck_attach(struct vfio_pci_device *vdev);
|
||||||
static void vfio_pci_reflck_put(struct vfio_pci_reflck *reflck);
|
static void vfio_pci_reflck_put(struct vfio_pci_reflck *reflck);
|
||||||
static struct pci_driver vfio_pci_driver;
|
|
||||||
|
|
||||||
static int vfio_pci_bus_notifier(struct notifier_block *nb,
|
static int vfio_pci_bus_notifier(struct notifier_block *nb,
|
||||||
unsigned long action, void *data)
|
unsigned long action, void *data)
|
||||||
|
|
|
@ -406,7 +406,7 @@ bool __vfio_pci_memory_enabled(struct vfio_pci_device *vdev)
|
||||||
* PF SR-IOV capability, there's therefore no need to trigger
|
* PF SR-IOV capability, there's therefore no need to trigger
|
||||||
* faults based on the virtual value.
|
* faults based on the virtual value.
|
||||||
*/
|
*/
|
||||||
return pdev->is_virtfn || (cmd & PCI_COMMAND_MEMORY);
|
return pdev->no_command_memory || (cmd & PCI_COMMAND_MEMORY);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -523,8 +523,8 @@ static int vfio_basic_config_read(struct vfio_pci_device *vdev, int pos,
|
||||||
|
|
||||||
count = vfio_default_config_read(vdev, pos, count, perm, offset, val);
|
count = vfio_default_config_read(vdev, pos, count, perm, offset, val);
|
||||||
|
|
||||||
/* Mask in virtual memory enable for SR-IOV devices */
|
/* Mask in virtual memory enable */
|
||||||
if (offset == PCI_COMMAND && vdev->pdev->is_virtfn) {
|
if (offset == PCI_COMMAND && vdev->pdev->no_command_memory) {
|
||||||
u16 cmd = le16_to_cpu(*(__le16 *)&vdev->vconfig[PCI_COMMAND]);
|
u16 cmd = le16_to_cpu(*(__le16 *)&vdev->vconfig[PCI_COMMAND]);
|
||||||
u32 tmp_val = le32_to_cpu(*val);
|
u32 tmp_val = le32_to_cpu(*val);
|
||||||
|
|
||||||
|
@ -592,9 +592,11 @@ static int vfio_basic_config_write(struct vfio_pci_device *vdev, int pos,
|
||||||
* shows it disabled (phys_mem/io, then the device has
|
* shows it disabled (phys_mem/io, then the device has
|
||||||
* undergone some kind of backdoor reset and needs to be
|
* undergone some kind of backdoor reset and needs to be
|
||||||
* restored before we allow it to enable the bars.
|
* restored before we allow it to enable the bars.
|
||||||
* SR-IOV devices will trigger this, but we catch them later
|
* SR-IOV devices will trigger this - for mem enable let's
|
||||||
|
* catch this now and for io enable it will be caught later
|
||||||
*/
|
*/
|
||||||
if ((new_mem && virt_mem && !phys_mem) ||
|
if ((new_mem && virt_mem && !phys_mem &&
|
||||||
|
!pdev->no_command_memory) ||
|
||||||
(new_io && virt_io && !phys_io) ||
|
(new_io && virt_io && !phys_io) ||
|
||||||
vfio_need_bar_restore(vdev))
|
vfio_need_bar_restore(vdev))
|
||||||
vfio_bar_restore(vdev);
|
vfio_bar_restore(vdev);
|
||||||
|
@ -1737,12 +1739,14 @@ int vfio_config_init(struct vfio_pci_device *vdev)
|
||||||
vconfig[PCI_INTERRUPT_PIN]);
|
vconfig[PCI_INTERRUPT_PIN]);
|
||||||
|
|
||||||
vconfig[PCI_INTERRUPT_PIN] = 0; /* Gratuitous for good VFs */
|
vconfig[PCI_INTERRUPT_PIN] = 0; /* Gratuitous for good VFs */
|
||||||
|
}
|
||||||
|
if (pdev->no_command_memory) {
|
||||||
/*
|
/*
|
||||||
* VFs do no implement the memory enable bit of the COMMAND
|
* VFs and devices that set pdev->no_command_memory do not
|
||||||
* register therefore we'll not have it set in our initial
|
* implement the memory enable bit of the COMMAND register
|
||||||
* copy of config space after pci_enable_device(). For
|
* therefore we'll not have it set in our initial copy of
|
||||||
* consistency with PFs, set the virtual enable bit here.
|
* config space after pci_enable_device(). For consistency
|
||||||
|
* with PFs, set the virtual enable bit here.
|
||||||
*/
|
*/
|
||||||
*(__le16 *)&vconfig[PCI_COMMAND] |=
|
*(__le16 *)&vconfig[PCI_COMMAND] |=
|
||||||
cpu_to_le16(PCI_COMMAND_MEMORY);
|
cpu_to_le16(PCI_COMMAND_MEMORY);
|
||||||
|
|
|
@ -1949,8 +1949,10 @@ int vfio_pin_pages(struct device *dev, unsigned long *user_pfn, int npage,
|
||||||
if (!group)
|
if (!group)
|
||||||
return -ENODEV;
|
return -ENODEV;
|
||||||
|
|
||||||
if (group->dev_counter > 1)
|
if (group->dev_counter > 1) {
|
||||||
return -EINVAL;
|
ret = -EINVAL;
|
||||||
|
goto err_pin_pages;
|
||||||
|
}
|
||||||
|
|
||||||
ret = vfio_group_add_container_user(group);
|
ret = vfio_group_add_container_user(group);
|
||||||
if (ret)
|
if (ret)
|
||||||
|
@ -2051,6 +2053,9 @@ int vfio_group_pin_pages(struct vfio_group *group,
|
||||||
if (!group || !user_iova_pfn || !phys_pfn || !npage)
|
if (!group || !user_iova_pfn || !phys_pfn || !npage)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
|
if (group->dev_counter > 1)
|
||||||
|
return -EINVAL;
|
||||||
|
|
||||||
if (npage > VFIO_PIN_PAGES_MAX_ENTRIES)
|
if (npage > VFIO_PIN_PAGES_MAX_ENTRIES)
|
||||||
return -E2BIG;
|
return -E2BIG;
|
||||||
|
|
||||||
|
|
|
@ -2609,6 +2609,20 @@ static int vfio_iommu_migration_build_caps(struct vfio_iommu *iommu,
|
||||||
return vfio_info_add_capability(caps, &cap_mig.header, sizeof(cap_mig));
|
return vfio_info_add_capability(caps, &cap_mig.header, sizeof(cap_mig));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int vfio_iommu_dma_avail_build_caps(struct vfio_iommu *iommu,
|
||||||
|
struct vfio_info_cap *caps)
|
||||||
|
{
|
||||||
|
struct vfio_iommu_type1_info_dma_avail cap_dma_avail;
|
||||||
|
|
||||||
|
cap_dma_avail.header.id = VFIO_IOMMU_TYPE1_INFO_DMA_AVAIL;
|
||||||
|
cap_dma_avail.header.version = 1;
|
||||||
|
|
||||||
|
cap_dma_avail.avail = iommu->dma_avail;
|
||||||
|
|
||||||
|
return vfio_info_add_capability(caps, &cap_dma_avail.header,
|
||||||
|
sizeof(cap_dma_avail));
|
||||||
|
}
|
||||||
|
|
||||||
static int vfio_iommu_type1_get_info(struct vfio_iommu *iommu,
|
static int vfio_iommu_type1_get_info(struct vfio_iommu *iommu,
|
||||||
unsigned long arg)
|
unsigned long arg)
|
||||||
{
|
{
|
||||||
|
@ -2641,6 +2655,9 @@ static int vfio_iommu_type1_get_info(struct vfio_iommu *iommu,
|
||||||
|
|
||||||
ret = vfio_iommu_migration_build_caps(iommu, &caps);
|
ret = vfio_iommu_migration_build_caps(iommu, &caps);
|
||||||
|
|
||||||
|
if (!ret)
|
||||||
|
ret = vfio_iommu_dma_avail_build_caps(iommu, &caps);
|
||||||
|
|
||||||
if (!ret)
|
if (!ret)
|
||||||
ret = vfio_iommu_iova_build_caps(iommu, &caps);
|
ret = vfio_iommu_iova_build_caps(iommu, &caps);
|
||||||
|
|
||||||
|
@ -2933,7 +2950,8 @@ static int vfio_iommu_type1_dma_rw_chunk(struct vfio_iommu *iommu,
|
||||||
* size
|
* size
|
||||||
*/
|
*/
|
||||||
bitmap_set(dma->bitmap, offset >> pgshift,
|
bitmap_set(dma->bitmap, offset >> pgshift,
|
||||||
*copied >> pgshift);
|
((offset + *copied - 1) >> pgshift) -
|
||||||
|
(offset >> pgshift) + 1);
|
||||||
}
|
}
|
||||||
} else
|
} else
|
||||||
*copied = copy_from_user(data, (void __user *)vaddr,
|
*copied = copy_from_user(data, (void __user *)vaddr,
|
||||||
|
|
|
@ -445,6 +445,7 @@ struct pci_dev {
|
||||||
unsigned int is_probed:1; /* Device probing in progress */
|
unsigned int is_probed:1; /* Device probing in progress */
|
||||||
unsigned int link_active_reporting:1;/* Device capable of reporting link active */
|
unsigned int link_active_reporting:1;/* Device capable of reporting link active */
|
||||||
unsigned int no_vf_scan:1; /* Don't scan for VFs after IOV enablement */
|
unsigned int no_vf_scan:1; /* Don't scan for VFs after IOV enablement */
|
||||||
|
unsigned int no_command_memory:1; /* No PCI_COMMAND_MEMORY */
|
||||||
pci_dev_flags_t dev_flags;
|
pci_dev_flags_t dev_flags;
|
||||||
atomic_t enable_cnt; /* pci_enable_device has been called */
|
atomic_t enable_cnt; /* pci_enable_device has been called */
|
||||||
|
|
||||||
|
|
|
@ -462,7 +462,7 @@ struct vfio_region_gfx_edid {
|
||||||
* 5. Resumed
|
* 5. Resumed
|
||||||
* |--------->|
|
* |--------->|
|
||||||
*
|
*
|
||||||
* 0. Default state of VFIO device is _RUNNNG when the user application starts.
|
* 0. Default state of VFIO device is _RUNNING when the user application starts.
|
||||||
* 1. During normal shutdown of the user application, the user application may
|
* 1. During normal shutdown of the user application, the user application may
|
||||||
* optionally change the VFIO device state from _RUNNING to _STOP. This
|
* optionally change the VFIO device state from _RUNNING to _STOP. This
|
||||||
* transition is optional. The vendor driver must support this transition but
|
* transition is optional. The vendor driver must support this transition but
|
||||||
|
@ -1039,6 +1039,21 @@ struct vfio_iommu_type1_info_cap_migration {
|
||||||
__u64 max_dirty_bitmap_size; /* in bytes */
|
__u64 max_dirty_bitmap_size; /* in bytes */
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/*
|
||||||
|
* The DMA available capability allows to report the current number of
|
||||||
|
* simultaneously outstanding DMA mappings that are allowed.
|
||||||
|
*
|
||||||
|
* The structure below defines version 1 of this capability.
|
||||||
|
*
|
||||||
|
* avail: specifies the current number of outstanding DMA mappings allowed.
|
||||||
|
*/
|
||||||
|
#define VFIO_IOMMU_TYPE1_INFO_DMA_AVAIL 3
|
||||||
|
|
||||||
|
struct vfio_iommu_type1_info_dma_avail {
|
||||||
|
struct vfio_info_cap_header header;
|
||||||
|
__u32 avail;
|
||||||
|
};
|
||||||
|
|
||||||
#define VFIO_IOMMU_GET_INFO _IO(VFIO_TYPE, VFIO_BASE + 12)
|
#define VFIO_IOMMU_GET_INFO _IO(VFIO_TYPE, VFIO_BASE + 12)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
Loading…
Reference in New Issue