mirror of https://github.com/xemu-project/xemu.git
vfio/container: Move pgsizes and dma_max_mappings to base container
No functional change intended. Signed-off-by: Eric Auger <eric.auger@redhat.com> Signed-off-by: Yi Liu <yi.l.liu@intel.com> Signed-off-by: Yi Sun <yi.y.sun@linux.intel.com> Signed-off-by: Zhenzhong Duan <zhenzhong.duan@intel.com> Reviewed-by: Cédric Le Goater <clg@redhat.com> Signed-off-by: Cédric Le Goater <clg@redhat.com>
This commit is contained in:
parent
e1cac6b203
commit
7ab1cb74ff
|
@ -401,6 +401,7 @@ static int vfio_ram_discard_notify_populate(RamDiscardListener *rdl,
|
||||||
static void vfio_register_ram_discard_listener(VFIOContainer *container,
|
static void vfio_register_ram_discard_listener(VFIOContainer *container,
|
||||||
MemoryRegionSection *section)
|
MemoryRegionSection *section)
|
||||||
{
|
{
|
||||||
|
VFIOContainerBase *bcontainer = &container->bcontainer;
|
||||||
RamDiscardManager *rdm = memory_region_get_ram_discard_manager(section->mr);
|
RamDiscardManager *rdm = memory_region_get_ram_discard_manager(section->mr);
|
||||||
VFIORamDiscardListener *vrdl;
|
VFIORamDiscardListener *vrdl;
|
||||||
|
|
||||||
|
@ -419,8 +420,8 @@ static void vfio_register_ram_discard_listener(VFIOContainer *container,
|
||||||
section->mr);
|
section->mr);
|
||||||
|
|
||||||
g_assert(vrdl->granularity && is_power_of_2(vrdl->granularity));
|
g_assert(vrdl->granularity && is_power_of_2(vrdl->granularity));
|
||||||
g_assert(container->pgsizes &&
|
g_assert(bcontainer->pgsizes &&
|
||||||
vrdl->granularity >= 1ULL << ctz64(container->pgsizes));
|
vrdl->granularity >= 1ULL << ctz64(bcontainer->pgsizes));
|
||||||
|
|
||||||
ram_discard_listener_init(&vrdl->listener,
|
ram_discard_listener_init(&vrdl->listener,
|
||||||
vfio_ram_discard_notify_populate,
|
vfio_ram_discard_notify_populate,
|
||||||
|
@ -441,7 +442,7 @@ static void vfio_register_ram_discard_listener(VFIOContainer *container,
|
||||||
* number of sections in the address space we could have over time,
|
* number of sections in the address space we could have over time,
|
||||||
* also consuming DMA mappings.
|
* also consuming DMA mappings.
|
||||||
*/
|
*/
|
||||||
if (container->dma_max_mappings) {
|
if (bcontainer->dma_max_mappings) {
|
||||||
unsigned int vrdl_count = 0, vrdl_mappings = 0, max_memslots = 512;
|
unsigned int vrdl_count = 0, vrdl_mappings = 0, max_memslots = 512;
|
||||||
|
|
||||||
#ifdef CONFIG_KVM
|
#ifdef CONFIG_KVM
|
||||||
|
@ -462,11 +463,11 @@ static void vfio_register_ram_discard_listener(VFIOContainer *container,
|
||||||
}
|
}
|
||||||
|
|
||||||
if (vrdl_mappings + max_memslots - vrdl_count >
|
if (vrdl_mappings + max_memslots - vrdl_count >
|
||||||
container->dma_max_mappings) {
|
bcontainer->dma_max_mappings) {
|
||||||
warn_report("%s: possibly running out of DMA mappings. E.g., try"
|
warn_report("%s: possibly running out of DMA mappings. E.g., try"
|
||||||
" increasing the 'block-size' of virtio-mem devies."
|
" increasing the 'block-size' of virtio-mem devies."
|
||||||
" Maximum possible DMA mappings: %d, Maximum possible"
|
" Maximum possible DMA mappings: %d, Maximum possible"
|
||||||
" memslots: %d", __func__, container->dma_max_mappings,
|
" memslots: %d", __func__, bcontainer->dma_max_mappings,
|
||||||
max_memslots);
|
max_memslots);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -626,7 +627,7 @@ static void vfio_listener_region_add(MemoryListener *listener,
|
||||||
iommu_idx);
|
iommu_idx);
|
||||||
|
|
||||||
ret = memory_region_iommu_set_page_size_mask(giommu->iommu_mr,
|
ret = memory_region_iommu_set_page_size_mask(giommu->iommu_mr,
|
||||||
container->pgsizes,
|
bcontainer->pgsizes,
|
||||||
&err);
|
&err);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
g_free(giommu);
|
g_free(giommu);
|
||||||
|
@ -675,7 +676,7 @@ static void vfio_listener_region_add(MemoryListener *listener,
|
||||||
llsize = int128_sub(llend, int128_make64(iova));
|
llsize = int128_sub(llend, int128_make64(iova));
|
||||||
|
|
||||||
if (memory_region_is_ram_device(section->mr)) {
|
if (memory_region_is_ram_device(section->mr)) {
|
||||||
hwaddr pgmask = (1ULL << ctz64(container->pgsizes)) - 1;
|
hwaddr pgmask = (1ULL << ctz64(bcontainer->pgsizes)) - 1;
|
||||||
|
|
||||||
if ((iova & pgmask) || (int128_get64(llsize) & pgmask)) {
|
if ((iova & pgmask) || (int128_get64(llsize) & pgmask)) {
|
||||||
trace_vfio_listener_region_add_no_dma_map(
|
trace_vfio_listener_region_add_no_dma_map(
|
||||||
|
@ -777,7 +778,7 @@ static void vfio_listener_region_del(MemoryListener *listener,
|
||||||
if (memory_region_is_ram_device(section->mr)) {
|
if (memory_region_is_ram_device(section->mr)) {
|
||||||
hwaddr pgmask;
|
hwaddr pgmask;
|
||||||
|
|
||||||
pgmask = (1ULL << ctz64(container->pgsizes)) - 1;
|
pgmask = (1ULL << ctz64(bcontainer->pgsizes)) - 1;
|
||||||
try_unmap = !((iova & pgmask) || (int128_get64(llsize) & pgmask));
|
try_unmap = !((iova & pgmask) || (int128_get64(llsize) & pgmask));
|
||||||
} else if (memory_region_has_ram_discard_manager(section->mr)) {
|
} else if (memory_region_has_ram_discard_manager(section->mr)) {
|
||||||
vfio_unregister_ram_discard_listener(container, section);
|
vfio_unregister_ram_discard_listener(container, section);
|
||||||
|
|
|
@ -52,6 +52,7 @@ void vfio_container_init(VFIOContainerBase *bcontainer, VFIOAddressSpace *space,
|
||||||
bcontainer->ops = ops;
|
bcontainer->ops = ops;
|
||||||
bcontainer->space = space;
|
bcontainer->space = space;
|
||||||
bcontainer->dirty_pages_supported = false;
|
bcontainer->dirty_pages_supported = false;
|
||||||
|
bcontainer->dma_max_mappings = 0;
|
||||||
QLIST_INIT(&bcontainer->giommu_list);
|
QLIST_INIT(&bcontainer->giommu_list);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -154,7 +154,7 @@ static int vfio_legacy_dma_unmap(VFIOContainerBase *bcontainer, hwaddr iova,
|
||||||
if (errno == EINVAL && unmap.size && !(unmap.iova + unmap.size) &&
|
if (errno == EINVAL && unmap.size && !(unmap.iova + unmap.size) &&
|
||||||
container->iommu_type == VFIO_TYPE1v2_IOMMU) {
|
container->iommu_type == VFIO_TYPE1v2_IOMMU) {
|
||||||
trace_vfio_legacy_dma_unmap_overflow_workaround();
|
trace_vfio_legacy_dma_unmap_overflow_workaround();
|
||||||
unmap.size -= 1ULL << ctz64(container->pgsizes);
|
unmap.size -= 1ULL << ctz64(bcontainer->pgsizes);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
error_report("VFIO_UNMAP_DMA failed: %s", strerror(errno));
|
error_report("VFIO_UNMAP_DMA failed: %s", strerror(errno));
|
||||||
|
@ -559,7 +559,6 @@ static int vfio_connect_container(VFIOGroup *group, AddressSpace *as,
|
||||||
container = g_malloc0(sizeof(*container));
|
container = g_malloc0(sizeof(*container));
|
||||||
container->fd = fd;
|
container->fd = fd;
|
||||||
container->error = NULL;
|
container->error = NULL;
|
||||||
container->dma_max_mappings = 0;
|
|
||||||
container->iova_ranges = NULL;
|
container->iova_ranges = NULL;
|
||||||
QLIST_INIT(&container->vrdl_list);
|
QLIST_INIT(&container->vrdl_list);
|
||||||
bcontainer = &container->bcontainer;
|
bcontainer = &container->bcontainer;
|
||||||
|
@ -589,13 +588,13 @@ static int vfio_connect_container(VFIOGroup *group, AddressSpace *as,
|
||||||
}
|
}
|
||||||
|
|
||||||
if (info->flags & VFIO_IOMMU_INFO_PGSIZES) {
|
if (info->flags & VFIO_IOMMU_INFO_PGSIZES) {
|
||||||
container->pgsizes = info->iova_pgsizes;
|
bcontainer->pgsizes = info->iova_pgsizes;
|
||||||
} else {
|
} else {
|
||||||
container->pgsizes = qemu_real_host_page_size();
|
bcontainer->pgsizes = qemu_real_host_page_size();
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!vfio_get_info_dma_avail(info, &container->dma_max_mappings)) {
|
if (!vfio_get_info_dma_avail(info, &bcontainer->dma_max_mappings)) {
|
||||||
container->dma_max_mappings = 65535;
|
bcontainer->dma_max_mappings = 65535;
|
||||||
}
|
}
|
||||||
|
|
||||||
vfio_get_info_iova_range(info, container);
|
vfio_get_info_iova_range(info, container);
|
||||||
|
|
|
@ -226,6 +226,7 @@ static int vfio_spapr_create_window(VFIOContainer *container,
|
||||||
hwaddr *pgsize)
|
hwaddr *pgsize)
|
||||||
{
|
{
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
VFIOContainerBase *bcontainer = &container->bcontainer;
|
||||||
IOMMUMemoryRegion *iommu_mr = IOMMU_MEMORY_REGION(section->mr);
|
IOMMUMemoryRegion *iommu_mr = IOMMU_MEMORY_REGION(section->mr);
|
||||||
uint64_t pagesize = memory_region_iommu_get_min_page_size(iommu_mr), pgmask;
|
uint64_t pagesize = memory_region_iommu_get_min_page_size(iommu_mr), pgmask;
|
||||||
unsigned entries, bits_total, bits_per_level, max_levels;
|
unsigned entries, bits_total, bits_per_level, max_levels;
|
||||||
|
@ -239,13 +240,13 @@ static int vfio_spapr_create_window(VFIOContainer *container,
|
||||||
if (pagesize > rampagesize) {
|
if (pagesize > rampagesize) {
|
||||||
pagesize = rampagesize;
|
pagesize = rampagesize;
|
||||||
}
|
}
|
||||||
pgmask = container->pgsizes & (pagesize | (pagesize - 1));
|
pgmask = bcontainer->pgsizes & (pagesize | (pagesize - 1));
|
||||||
pagesize = pgmask ? (1ULL << (63 - clz64(pgmask))) : 0;
|
pagesize = pgmask ? (1ULL << (63 - clz64(pgmask))) : 0;
|
||||||
if (!pagesize) {
|
if (!pagesize) {
|
||||||
error_report("Host doesn't support page size 0x%"PRIx64
|
error_report("Host doesn't support page size 0x%"PRIx64
|
||||||
", the supported mask is 0x%lx",
|
", the supported mask is 0x%lx",
|
||||||
memory_region_iommu_get_min_page_size(iommu_mr),
|
memory_region_iommu_get_min_page_size(iommu_mr),
|
||||||
container->pgsizes);
|
bcontainer->pgsizes);
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -421,6 +422,7 @@ void vfio_container_del_section_window(VFIOContainer *container,
|
||||||
|
|
||||||
int vfio_spapr_container_init(VFIOContainer *container, Error **errp)
|
int vfio_spapr_container_init(VFIOContainer *container, Error **errp)
|
||||||
{
|
{
|
||||||
|
VFIOContainerBase *bcontainer = &container->bcontainer;
|
||||||
struct vfio_iommu_spapr_tce_info info;
|
struct vfio_iommu_spapr_tce_info info;
|
||||||
bool v2 = container->iommu_type == VFIO_SPAPR_TCE_v2_IOMMU;
|
bool v2 = container->iommu_type == VFIO_SPAPR_TCE_v2_IOMMU;
|
||||||
int ret, fd = container->fd;
|
int ret, fd = container->fd;
|
||||||
|
@ -461,7 +463,7 @@ int vfio_spapr_container_init(VFIOContainer *container, Error **errp)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (v2) {
|
if (v2) {
|
||||||
container->pgsizes = info.ddw.pgsizes;
|
bcontainer->pgsizes = info.ddw.pgsizes;
|
||||||
/*
|
/*
|
||||||
* There is a default window in just created container.
|
* There is a default window in just created container.
|
||||||
* To make region_add/del simpler, we better remove this
|
* To make region_add/del simpler, we better remove this
|
||||||
|
@ -476,7 +478,7 @@ int vfio_spapr_container_init(VFIOContainer *container, Error **errp)
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
/* The default table uses 4K pages */
|
/* The default table uses 4K pages */
|
||||||
container->pgsizes = 0x1000;
|
bcontainer->pgsizes = 0x1000;
|
||||||
vfio_host_win_add(container, info.dma32_window_start,
|
vfio_host_win_add(container, info.dma32_window_start,
|
||||||
info.dma32_window_start +
|
info.dma32_window_start +
|
||||||
info.dma32_window_size - 1,
|
info.dma32_window_size - 1,
|
||||||
|
|
|
@ -85,8 +85,6 @@ typedef struct VFIOContainer {
|
||||||
bool initialized;
|
bool initialized;
|
||||||
uint64_t dirty_pgsizes;
|
uint64_t dirty_pgsizes;
|
||||||
uint64_t max_dirty_bitmap_size;
|
uint64_t max_dirty_bitmap_size;
|
||||||
unsigned long pgsizes;
|
|
||||||
unsigned int dma_max_mappings;
|
|
||||||
QLIST_HEAD(, VFIOHostDMAWindow) hostwin_list;
|
QLIST_HEAD(, VFIOHostDMAWindow) hostwin_list;
|
||||||
QLIST_HEAD(, VFIOGroup) group_list;
|
QLIST_HEAD(, VFIOGroup) group_list;
|
||||||
QLIST_HEAD(, VFIORamDiscardListener) vrdl_list;
|
QLIST_HEAD(, VFIORamDiscardListener) vrdl_list;
|
||||||
|
|
|
@ -36,6 +36,8 @@ typedef struct VFIOAddressSpace {
|
||||||
typedef struct VFIOContainerBase {
|
typedef struct VFIOContainerBase {
|
||||||
const VFIOIOMMUOps *ops;
|
const VFIOIOMMUOps *ops;
|
||||||
VFIOAddressSpace *space;
|
VFIOAddressSpace *space;
|
||||||
|
unsigned long pgsizes;
|
||||||
|
unsigned int dma_max_mappings;
|
||||||
bool dirty_pages_supported;
|
bool dirty_pages_supported;
|
||||||
QLIST_HEAD(, VFIOGuestIOMMU) giommu_list;
|
QLIST_HEAD(, VFIOGuestIOMMU) giommu_list;
|
||||||
QLIST_ENTRY(VFIOContainerBase) next;
|
QLIST_ENTRY(VFIOContainerBase) next;
|
||||||
|
|
Loading…
Reference in New Issue