|
@@ -37,7 +37,7 @@
|
|
|
|
|
|
static MSIMessage msix_get_message(PCIDevice *dev, unsigned vector)
|
|
|
{
|
|
|
- uint8_t *table_entry = dev->msix_table_page + vector * PCI_MSIX_ENTRY_SIZE;
|
|
|
+ uint8_t *table_entry = dev->msix_table + vector * PCI_MSIX_ENTRY_SIZE;
|
|
|
MSIMessage msg;
|
|
|
|
|
|
msg.address = pci_get_quad(table_entry + PCI_MSIX_ENTRY_LOWER_ADDR);
|
|
@@ -93,7 +93,7 @@ static uint8_t msix_pending_mask(int vector)
|
|
|
|
|
|
static uint8_t *msix_pending_byte(PCIDevice *dev, int vector)
|
|
|
{
|
|
|
- return dev->msix_table_page + MSIX_PAGE_PENDING + vector / 8;
|
|
|
+ return dev->msix_pba + vector / 8;
|
|
|
}
|
|
|
|
|
|
static int msix_is_pending(PCIDevice *dev, int vector)
|
|
@@ -114,7 +114,7 @@ static void msix_clr_pending(PCIDevice *dev, int vector)
|
|
|
static bool msix_vector_masked(PCIDevice *dev, int vector, bool fmask)
|
|
|
{
|
|
|
unsigned offset = vector * PCI_MSIX_ENTRY_SIZE + PCI_MSIX_ENTRY_VECTOR_CTRL;
|
|
|
- return fmask || dev->msix_table_page[offset] & PCI_MSIX_ENTRY_CTRL_MASKBIT;
|
|
|
+ return fmask || dev->msix_table[offset] & PCI_MSIX_ENTRY_CTRL_MASKBIT;
|
|
|
}
|
|
|
|
|
|
static bool msix_is_masked(PCIDevice *dev, int vector)
|
|
@@ -193,37 +193,47 @@ void msix_write_config(PCIDevice *dev, uint32_t addr,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static uint64_t msix_mmio_read(void *opaque, target_phys_addr_t addr,
|
|
|
- unsigned size)
|
|
|
+static uint64_t msix_table_mmio_read(void *opaque, target_phys_addr_t addr,
|
|
|
+ unsigned size)
|
|
|
{
|
|
|
PCIDevice *dev = opaque;
|
|
|
- unsigned int offset = addr & (MSIX_PAGE_SIZE - 1) & ~0x3;
|
|
|
- void *page = dev->msix_table_page;
|
|
|
|
|
|
- return pci_get_long(page + offset);
|
|
|
+ return pci_get_long(dev->msix_table + addr);
|
|
|
}
|
|
|
|
|
|
-static void msix_mmio_write(void *opaque, target_phys_addr_t addr,
|
|
|
- uint64_t val, unsigned size)
|
|
|
+static void msix_table_mmio_write(void *opaque, target_phys_addr_t addr,
|
|
|
+ uint64_t val, unsigned size)
|
|
|
{
|
|
|
PCIDevice *dev = opaque;
|
|
|
- unsigned int offset = addr & (MSIX_PAGE_SIZE - 1) & ~0x3;
|
|
|
- int vector = offset / PCI_MSIX_ENTRY_SIZE;
|
|
|
+ int vector = addr / PCI_MSIX_ENTRY_SIZE;
|
|
|
bool was_masked;
|
|
|
|
|
|
- /* MSI-X page includes a read-only PBA and a writeable Vector Control. */
|
|
|
- if (vector >= dev->msix_entries_nr) {
|
|
|
- return;
|
|
|
- }
|
|
|
-
|
|
|
was_masked = msix_is_masked(dev, vector);
|
|
|
- pci_set_long(dev->msix_table_page + offset, val);
|
|
|
+ pci_set_long(dev->msix_table + addr, val);
|
|
|
msix_handle_mask_update(dev, vector, was_masked);
|
|
|
}
|
|
|
|
|
|
-static const MemoryRegionOps msix_mmio_ops = {
|
|
|
- .read = msix_mmio_read,
|
|
|
- .write = msix_mmio_write,
|
|
|
+static const MemoryRegionOps msix_table_mmio_ops = {
|
|
|
+ .read = msix_table_mmio_read,
|
|
|
+ .write = msix_table_mmio_write,
|
|
|
+ /* TODO: MSIX should be LITTLE_ENDIAN. */
|
|
|
+ .endianness = DEVICE_NATIVE_ENDIAN,
|
|
|
+ .valid = {
|
|
|
+ .min_access_size = 4,
|
|
|
+ .max_access_size = 4,
|
|
|
+ },
|
|
|
+};
|
|
|
+
|
|
|
+static uint64_t msix_pba_mmio_read(void *opaque, target_phys_addr_t addr,
|
|
|
+ unsigned size)
|
|
|
+{
|
|
|
+ PCIDevice *dev = opaque;
|
|
|
+
|
|
|
+ return pci_get_long(dev->msix_pba + addr);
|
|
|
+}
|
|
|
+
|
|
|
+static const MemoryRegionOps msix_pba_mmio_ops = {
|
|
|
+ .read = msix_pba_mmio_read,
|
|
|
/* TODO: MSIX should be LITTLE_ENDIAN. */
|
|
|
.endianness = DEVICE_NATIVE_ENDIAN,
|
|
|
.valid = {
|
|
@@ -236,11 +246,14 @@ static void msix_mmio_setup(PCIDevice *d, MemoryRegion *bar)
|
|
|
{
|
|
|
uint8_t *config = d->config + d->msix_cap;
|
|
|
uint32_t table = pci_get_long(config + PCI_MSIX_TABLE);
|
|
|
- uint32_t offset = table & ~(MSIX_PAGE_SIZE - 1);
|
|
|
+ uint32_t table_offset = table & ~PCI_MSIX_FLAGS_BIRMASK;
|
|
|
+ uint32_t pba = pci_get_long(config + PCI_MSIX_PBA);
|
|
|
+ uint32_t pba_offset = pba & ~PCI_MSIX_FLAGS_BIRMASK;
|
|
|
/* TODO: for assigned devices, we'll want to make it possible to map
|
|
|
* pending bits separately in case they are in a separate bar. */
|
|
|
|
|
|
- memory_region_add_subregion(bar, offset, &d->msix_mmio);
|
|
|
+ memory_region_add_subregion(bar, table_offset, &d->msix_table_mmio);
|
|
|
+ memory_region_add_subregion(bar, pba_offset, &d->msix_pba_mmio);
|
|
|
}
|
|
|
|
|
|
static void msix_mask_all(struct PCIDevice *dev, unsigned nentries)
|
|
@@ -252,7 +265,7 @@ static void msix_mask_all(struct PCIDevice *dev, unsigned nentries)
|
|
|
vector * PCI_MSIX_ENTRY_SIZE + PCI_MSIX_ENTRY_VECTOR_CTRL;
|
|
|
bool was_masked = msix_is_masked(dev, vector);
|
|
|
|
|
|
- dev->msix_table_page[offset] |= PCI_MSIX_ENTRY_CTRL_MASKBIT;
|
|
|
+ dev->msix_table[offset] |= PCI_MSIX_ENTRY_CTRL_MASKBIT;
|
|
|
msix_handle_mask_update(dev, vector, was_masked);
|
|
|
}
|
|
|
}
|
|
@@ -264,6 +277,7 @@ int msix_init(struct PCIDevice *dev, unsigned short nentries,
|
|
|
unsigned bar_nr, unsigned bar_size)
|
|
|
{
|
|
|
int ret;
|
|
|
+ unsigned table_size, pba_size;
|
|
|
|
|
|
/* Nothing to do if MSI is not supported by interrupt controller */
|
|
|
if (!msi_supported) {
|
|
@@ -272,14 +286,20 @@ int msix_init(struct PCIDevice *dev, unsigned short nentries,
|
|
|
if (nentries > MSIX_MAX_ENTRIES)
|
|
|
return -EINVAL;
|
|
|
|
|
|
+ table_size = nentries * PCI_MSIX_ENTRY_SIZE;
|
|
|
+ pba_size = QEMU_ALIGN_UP(nentries, 64) / 8;
|
|
|
+
|
|
|
dev->msix_entry_used = g_malloc0(MSIX_MAX_ENTRIES *
|
|
|
sizeof *dev->msix_entry_used);
|
|
|
|
|
|
- dev->msix_table_page = g_malloc0(MSIX_PAGE_SIZE);
|
|
|
+ dev->msix_table = g_malloc0(table_size);
|
|
|
+ dev->msix_pba = g_malloc0(pba_size);
|
|
|
msix_mask_all(dev, nentries);
|
|
|
|
|
|
- memory_region_init_io(&dev->msix_mmio, &msix_mmio_ops, dev,
|
|
|
- "msix", MSIX_PAGE_SIZE);
|
|
|
+ memory_region_init_io(&dev->msix_table_mmio, &msix_table_mmio_ops, dev,
|
|
|
+ "msix-table", table_size);
|
|
|
+ memory_region_init_io(&dev->msix_pba_mmio, &msix_pba_mmio_ops, dev,
|
|
|
+ "msix-pba", pba_size);
|
|
|
|
|
|
dev->msix_entries_nr = nentries;
|
|
|
ret = msix_add_config(dev, nentries, bar_nr, bar_size);
|
|
@@ -292,9 +312,12 @@ int msix_init(struct PCIDevice *dev, unsigned short nentries,
|
|
|
|
|
|
err_config:
|
|
|
dev->msix_entries_nr = 0;
|
|
|
- memory_region_destroy(&dev->msix_mmio);
|
|
|
- g_free(dev->msix_table_page);
|
|
|
- dev->msix_table_page = NULL;
|
|
|
+ memory_region_destroy(&dev->msix_pba_mmio);
|
|
|
+ g_free(dev->msix_pba);
|
|
|
+ dev->msix_pba = NULL;
|
|
|
+ memory_region_destroy(&dev->msix_table_mmio);
|
|
|
+ g_free(dev->msix_table);
|
|
|
+ dev->msix_table = NULL;
|
|
|
g_free(dev->msix_entry_used);
|
|
|
dev->msix_entry_used = NULL;
|
|
|
return ret;
|
|
@@ -359,10 +382,14 @@ int msix_uninit(PCIDevice *dev, MemoryRegion *bar)
|
|
|
dev->msix_cap = 0;
|
|
|
msix_free_irq_entries(dev);
|
|
|
dev->msix_entries_nr = 0;
|
|
|
- memory_region_del_subregion(bar, &dev->msix_mmio);
|
|
|
- memory_region_destroy(&dev->msix_mmio);
|
|
|
- g_free(dev->msix_table_page);
|
|
|
- dev->msix_table_page = NULL;
|
|
|
+ memory_region_del_subregion(bar, &dev->msix_pba_mmio);
|
|
|
+ memory_region_destroy(&dev->msix_pba_mmio);
|
|
|
+ g_free(dev->msix_pba);
|
|
|
+ dev->msix_pba = NULL;
|
|
|
+ memory_region_del_subregion(bar, &dev->msix_table_mmio);
|
|
|
+ memory_region_destroy(&dev->msix_table_mmio);
|
|
|
+ g_free(dev->msix_table);
|
|
|
+ dev->msix_table = NULL;
|
|
|
g_free(dev->msix_entry_used);
|
|
|
dev->msix_entry_used = NULL;
|
|
|
dev->cap_present &= ~QEMU_PCI_CAP_MSIX;
|
|
@@ -385,8 +412,8 @@ void msix_save(PCIDevice *dev, QEMUFile *f)
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
- qemu_put_buffer(f, dev->msix_table_page, n * PCI_MSIX_ENTRY_SIZE);
|
|
|
- qemu_put_buffer(f, dev->msix_table_page + MSIX_PAGE_PENDING, (n + 7) / 8);
|
|
|
+ qemu_put_buffer(f, dev->msix_table, n * PCI_MSIX_ENTRY_SIZE);
|
|
|
+ qemu_put_buffer(f, dev->msix_pba, (n + 7) / 8);
|
|
|
}
|
|
|
|
|
|
/* Should be called after restoring the config space. */
|
|
@@ -400,8 +427,8 @@ void msix_load(PCIDevice *dev, QEMUFile *f)
|
|
|
}
|
|
|
|
|
|
msix_free_irq_entries(dev);
|
|
|
- qemu_get_buffer(f, dev->msix_table_page, n * PCI_MSIX_ENTRY_SIZE);
|
|
|
- qemu_get_buffer(f, dev->msix_table_page + MSIX_PAGE_PENDING, (n + 7) / 8);
|
|
|
+ qemu_get_buffer(f, dev->msix_table, n * PCI_MSIX_ENTRY_SIZE);
|
|
|
+ qemu_get_buffer(f, dev->msix_pba, (n + 7) / 8);
|
|
|
msix_update_function_masked(dev);
|
|
|
|
|
|
for (vector = 0; vector < n; vector++) {
|
|
@@ -448,7 +475,8 @@ void msix_reset(PCIDevice *dev)
|
|
|
msix_free_irq_entries(dev);
|
|
|
dev->config[dev->msix_cap + MSIX_CONTROL_OFFSET] &=
|
|
|
~dev->wmask[dev->msix_cap + MSIX_CONTROL_OFFSET];
|
|
|
- memset(dev->msix_table_page, 0, MSIX_PAGE_SIZE);
|
|
|
+ memset(dev->msix_table, 0, dev->msix_entries_nr * PCI_MSIX_ENTRY_SIZE);
|
|
|
+ memset(dev->msix_pba, 0, QEMU_ALIGN_UP(dev->msix_entries_nr, 64) / 8);
|
|
|
msix_mask_all(dev, dev->msix_entries_nr);
|
|
|
}
|
|
|
|