mirror of https://gitlab.com/qemu-project/qemu
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
2950 lines
90 KiB
C
2950 lines
90 KiB
C
/*
|
|
* QEMU PCI bus manager
|
|
*
|
|
* Copyright (c) 2004 Fabrice Bellard
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to deal
|
|
* in the Software without restriction, including without limitation the rights
|
|
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
* copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
* THE SOFTWARE.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/datadir.h"
|
|
#include "qemu/units.h"
|
|
#include "hw/irq.h"
|
|
#include "hw/pci/pci.h"
|
|
#include "hw/pci/pci_bridge.h"
|
|
#include "hw/pci/pci_bus.h"
|
|
#include "hw/pci/pci_host.h"
|
|
#include "hw/qdev-properties.h"
|
|
#include "hw/qdev-properties-system.h"
|
|
#include "migration/qemu-file-types.h"
|
|
#include "migration/vmstate.h"
|
|
#include "net/net.h"
|
|
#include "sysemu/numa.h"
|
|
#include "sysemu/runstate.h"
|
|
#include "sysemu/sysemu.h"
|
|
#include "hw/loader.h"
|
|
#include "qemu/error-report.h"
|
|
#include "qemu/range.h"
|
|
#include "trace.h"
|
|
#include "hw/pci/msi.h"
|
|
#include "hw/pci/msix.h"
|
|
#include "hw/hotplug.h"
|
|
#include "hw/boards.h"
|
|
#include "qapi/error.h"
|
|
#include "qemu/cutils.h"
|
|
#include "pci-internal.h"
|
|
|
|
#include "hw/xen/xen.h"
|
|
#include "hw/i386/kvm/xen_evtchn.h"
|
|
|
|
//#define DEBUG_PCI
|
|
#ifdef DEBUG_PCI
|
|
# define PCI_DPRINTF(format, ...) printf(format, ## __VA_ARGS__)
|
|
#else
|
|
# define PCI_DPRINTF(format, ...) do { } while (0)
|
|
#endif
|
|
|
|
bool pci_available = true;
|
|
|
|
static char *pcibus_get_dev_path(DeviceState *dev);
|
|
static char *pcibus_get_fw_dev_path(DeviceState *dev);
|
|
static void pcibus_reset_hold(Object *obj, ResetType type);
|
|
static bool pcie_has_upstream_port(PCIDevice *dev);
|
|
|
|
static void prop_pci_busnr_get(Object *obj, Visitor *v, const char *name,
|
|
void *opaque, Error **errp)
|
|
{
|
|
uint8_t busnr = pci_dev_bus_num(PCI_DEVICE(obj));
|
|
|
|
visit_type_uint8(v, name, &busnr, errp);
|
|
}
|
|
|
|
static const PropertyInfo prop_pci_busnr = {
|
|
.name = "busnr",
|
|
.get = prop_pci_busnr_get,
|
|
};
|
|
|
|
static Property pci_props[] = {
|
|
DEFINE_PROP_PCI_DEVFN("addr", PCIDevice, devfn, -1),
|
|
DEFINE_PROP_STRING("romfile", PCIDevice, romfile),
|
|
DEFINE_PROP_UINT32("romsize", PCIDevice, romsize, UINT32_MAX),
|
|
DEFINE_PROP_UINT32("rombar", PCIDevice, rom_bar, 1),
|
|
DEFINE_PROP_BIT("multifunction", PCIDevice, cap_present,
|
|
QEMU_PCI_CAP_MULTIFUNCTION_BITNR, false),
|
|
DEFINE_PROP_BIT("x-pcie-lnksta-dllla", PCIDevice, cap_present,
|
|
QEMU_PCIE_LNKSTA_DLLLA_BITNR, true),
|
|
DEFINE_PROP_BIT("x-pcie-extcap-init", PCIDevice, cap_present,
|
|
QEMU_PCIE_EXTCAP_INIT_BITNR, true),
|
|
DEFINE_PROP_STRING("failover_pair_id", PCIDevice,
|
|
failover_pair_id),
|
|
DEFINE_PROP_UINT32("acpi-index", PCIDevice, acpi_index, 0),
|
|
DEFINE_PROP_BIT("x-pcie-err-unc-mask", PCIDevice, cap_present,
|
|
QEMU_PCIE_ERR_UNC_MASK_BITNR, true),
|
|
DEFINE_PROP_BIT("x-pcie-ari-nextfn-1", PCIDevice, cap_present,
|
|
QEMU_PCIE_ARI_NEXTFN_1_BITNR, false),
|
|
DEFINE_PROP_SIZE32("x-max-bounce-buffer-size", PCIDevice,
|
|
max_bounce_buffer_size, DEFAULT_MAX_BOUNCE_BUFFER_SIZE),
|
|
DEFINE_PROP_BIT("x-pcie-ext-tag", PCIDevice, cap_present,
|
|
QEMU_PCIE_EXT_TAG_BITNR, true),
|
|
{ .name = "busnr", .info = &prop_pci_busnr },
|
|
DEFINE_PROP_END_OF_LIST()
|
|
};
|
|
|
|
static const VMStateDescription vmstate_pcibus = {
|
|
.name = "PCIBUS",
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.fields = (const VMStateField[]) {
|
|
VMSTATE_INT32_EQUAL(nirq, PCIBus, NULL),
|
|
VMSTATE_VARRAY_INT32(irq_count, PCIBus,
|
|
nirq, 0, vmstate_info_int32,
|
|
int32_t),
|
|
VMSTATE_END_OF_LIST()
|
|
}
|
|
};
|
|
|
|
static gint g_cmp_uint32(gconstpointer a, gconstpointer b, gpointer user_data)
|
|
{
|
|
return a - b;
|
|
}
|
|
|
|
static GSequence *pci_acpi_index_list(void)
|
|
{
|
|
static GSequence *used_acpi_index_list;
|
|
|
|
if (!used_acpi_index_list) {
|
|
used_acpi_index_list = g_sequence_new(NULL);
|
|
}
|
|
return used_acpi_index_list;
|
|
}
|
|
|
|
static void pci_init_bus_master(PCIDevice *pci_dev)
|
|
{
|
|
AddressSpace *dma_as = pci_device_iommu_address_space(pci_dev);
|
|
|
|
memory_region_init_alias(&pci_dev->bus_master_enable_region,
|
|
OBJECT(pci_dev), "bus master",
|
|
dma_as->root, 0, memory_region_size(dma_as->root));
|
|
memory_region_set_enabled(&pci_dev->bus_master_enable_region, false);
|
|
memory_region_add_subregion(&pci_dev->bus_master_container_region, 0,
|
|
&pci_dev->bus_master_enable_region);
|
|
}
|
|
|
|
static void pcibus_machine_done(Notifier *notifier, void *data)
|
|
{
|
|
PCIBus *bus = container_of(notifier, PCIBus, machine_done);
|
|
int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(bus->devices); ++i) {
|
|
if (bus->devices[i]) {
|
|
pci_init_bus_master(bus->devices[i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void pci_bus_realize(BusState *qbus, Error **errp)
|
|
{
|
|
PCIBus *bus = PCI_BUS(qbus);
|
|
|
|
bus->machine_done.notify = pcibus_machine_done;
|
|
qemu_add_machine_init_done_notifier(&bus->machine_done);
|
|
|
|
vmstate_register_any(NULL, &vmstate_pcibus, bus);
|
|
}
|
|
|
|
static void pcie_bus_realize(BusState *qbus, Error **errp)
|
|
{
|
|
PCIBus *bus = PCI_BUS(qbus);
|
|
Error *local_err = NULL;
|
|
|
|
pci_bus_realize(qbus, &local_err);
|
|
if (local_err) {
|
|
error_propagate(errp, local_err);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* A PCI-E bus can support extended config space if it's the root
|
|
* bus, or if the bus/bridge above it does as well
|
|
*/
|
|
if (pci_bus_is_root(bus)) {
|
|
bus->flags |= PCI_BUS_EXTENDED_CONFIG_SPACE;
|
|
} else {
|
|
PCIBus *parent_bus = pci_get_bus(bus->parent_dev);
|
|
|
|
if (pci_bus_allows_extended_config_space(parent_bus)) {
|
|
bus->flags |= PCI_BUS_EXTENDED_CONFIG_SPACE;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void pci_bus_unrealize(BusState *qbus)
|
|
{
|
|
PCIBus *bus = PCI_BUS(qbus);
|
|
|
|
qemu_remove_machine_init_done_notifier(&bus->machine_done);
|
|
|
|
vmstate_unregister(NULL, &vmstate_pcibus, bus);
|
|
}
|
|
|
|
static int pcibus_num(PCIBus *bus)
|
|
{
|
|
if (pci_bus_is_root(bus)) {
|
|
return 0; /* pci host bridge */
|
|
}
|
|
return bus->parent_dev->config[PCI_SECONDARY_BUS];
|
|
}
|
|
|
|
static uint16_t pcibus_numa_node(PCIBus *bus)
|
|
{
|
|
return NUMA_NODE_UNASSIGNED;
|
|
}
|
|
|
|
static void pci_bus_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
BusClass *k = BUS_CLASS(klass);
|
|
PCIBusClass *pbc = PCI_BUS_CLASS(klass);
|
|
ResettableClass *rc = RESETTABLE_CLASS(klass);
|
|
|
|
k->print_dev = pcibus_dev_print;
|
|
k->get_dev_path = pcibus_get_dev_path;
|
|
k->get_fw_dev_path = pcibus_get_fw_dev_path;
|
|
k->realize = pci_bus_realize;
|
|
k->unrealize = pci_bus_unrealize;
|
|
|
|
rc->phases.hold = pcibus_reset_hold;
|
|
|
|
pbc->bus_num = pcibus_num;
|
|
pbc->numa_node = pcibus_numa_node;
|
|
}
|
|
|
|
static const TypeInfo pci_bus_info = {
|
|
.name = TYPE_PCI_BUS,
|
|
.parent = TYPE_BUS,
|
|
.instance_size = sizeof(PCIBus),
|
|
.class_size = sizeof(PCIBusClass),
|
|
.class_init = pci_bus_class_init,
|
|
};
|
|
|
|
static const TypeInfo cxl_interface_info = {
|
|
.name = INTERFACE_CXL_DEVICE,
|
|
.parent = TYPE_INTERFACE,
|
|
};
|
|
|
|
static const TypeInfo pcie_interface_info = {
|
|
.name = INTERFACE_PCIE_DEVICE,
|
|
.parent = TYPE_INTERFACE,
|
|
};
|
|
|
|
static const TypeInfo conventional_pci_interface_info = {
|
|
.name = INTERFACE_CONVENTIONAL_PCI_DEVICE,
|
|
.parent = TYPE_INTERFACE,
|
|
};
|
|
|
|
static void pcie_bus_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
BusClass *k = BUS_CLASS(klass);
|
|
|
|
k->realize = pcie_bus_realize;
|
|
}
|
|
|
|
static const TypeInfo pcie_bus_info = {
|
|
.name = TYPE_PCIE_BUS,
|
|
.parent = TYPE_PCI_BUS,
|
|
.class_init = pcie_bus_class_init,
|
|
};
|
|
|
|
static const TypeInfo cxl_bus_info = {
|
|
.name = TYPE_CXL_BUS,
|
|
.parent = TYPE_PCIE_BUS,
|
|
.class_init = pcie_bus_class_init,
|
|
};
|
|
|
|
static void pci_update_mappings(PCIDevice *d);
|
|
static void pci_irq_handler(void *opaque, int irq_num, int level);
|
|
static void pci_add_option_rom(PCIDevice *pdev, bool is_default_rom, Error **);
|
|
static void pci_del_option_rom(PCIDevice *pdev);
|
|
|
|
static uint16_t pci_default_sub_vendor_id = PCI_SUBVENDOR_ID_REDHAT_QUMRANET;
|
|
static uint16_t pci_default_sub_device_id = PCI_SUBDEVICE_ID_QEMU;
|
|
|
|
PCIHostStateList pci_host_bridges;
|
|
|
|
int pci_bar(PCIDevice *d, int reg)
|
|
{
|
|
uint8_t type;
|
|
|
|
/* PCIe virtual functions do not have their own BARs */
|
|
assert(!pci_is_vf(d));
|
|
|
|
if (reg != PCI_ROM_SLOT)
|
|
return PCI_BASE_ADDRESS_0 + reg * 4;
|
|
|
|
type = d->config[PCI_HEADER_TYPE] & ~PCI_HEADER_TYPE_MULTI_FUNCTION;
|
|
return type == PCI_HEADER_TYPE_BRIDGE ? PCI_ROM_ADDRESS1 : PCI_ROM_ADDRESS;
|
|
}
|
|
|
|
static inline int pci_irq_state(PCIDevice *d, int irq_num)
|
|
{
|
|
return (d->irq_state >> irq_num) & 0x1;
|
|
}
|
|
|
|
static inline void pci_set_irq_state(PCIDevice *d, int irq_num, int level)
|
|
{
|
|
d->irq_state &= ~(0x1 << irq_num);
|
|
d->irq_state |= level << irq_num;
|
|
}
|
|
|
|
static void pci_bus_change_irq_level(PCIBus *bus, int irq_num, int change)
|
|
{
|
|
assert(irq_num >= 0);
|
|
assert(irq_num < bus->nirq);
|
|
bus->irq_count[irq_num] += change;
|
|
bus->set_irq(bus->irq_opaque, irq_num, bus->irq_count[irq_num] != 0);
|
|
}
|
|
|
|
static void pci_change_irq_level(PCIDevice *pci_dev, int irq_num, int change)
|
|
{
|
|
PCIBus *bus;
|
|
for (;;) {
|
|
int dev_irq = irq_num;
|
|
bus = pci_get_bus(pci_dev);
|
|
assert(bus->map_irq);
|
|
irq_num = bus->map_irq(pci_dev, irq_num);
|
|
trace_pci_route_irq(dev_irq, DEVICE(pci_dev)->canonical_path, irq_num,
|
|
pci_bus_is_root(bus) ? "root-complex"
|
|
: DEVICE(bus->parent_dev)->canonical_path);
|
|
if (bus->set_irq)
|
|
break;
|
|
pci_dev = bus->parent_dev;
|
|
}
|
|
pci_bus_change_irq_level(bus, irq_num, change);
|
|
}
|
|
|
|
int pci_bus_get_irq_level(PCIBus *bus, int irq_num)
|
|
{
|
|
assert(irq_num >= 0);
|
|
assert(irq_num < bus->nirq);
|
|
return !!bus->irq_count[irq_num];
|
|
}
|
|
|
|
/* Update interrupt status bit in config space on interrupt
|
|
* state change. */
|
|
static void pci_update_irq_status(PCIDevice *dev)
|
|
{
|
|
if (dev->irq_state) {
|
|
dev->config[PCI_STATUS] |= PCI_STATUS_INTERRUPT;
|
|
} else {
|
|
dev->config[PCI_STATUS] &= ~PCI_STATUS_INTERRUPT;
|
|
}
|
|
}
|
|
|
|
void pci_device_deassert_intx(PCIDevice *dev)
|
|
{
|
|
int i;
|
|
for (i = 0; i < PCI_NUM_PINS; ++i) {
|
|
pci_irq_handler(dev, i, 0);
|
|
}
|
|
}
|
|
|
|
static void pci_msi_trigger(PCIDevice *dev, MSIMessage msg)
|
|
{
|
|
MemTxAttrs attrs = {};
|
|
|
|
/*
|
|
* Xen uses the high bits of the address to contain some of the bits
|
|
* of the PIRQ#. Therefore we can't just send the write cycle and
|
|
* trust that it's caught by the APIC at 0xfee00000 because the
|
|
* target of the write might be e.g. 0x0x1000fee46000 for PIRQ#4166.
|
|
* So we intercept the delivery here instead of in kvm_send_msi().
|
|
*/
|
|
if (xen_mode == XEN_EMULATE &&
|
|
xen_evtchn_deliver_pirq_msi(msg.address, msg.data)) {
|
|
return;
|
|
}
|
|
attrs.requester_id = pci_requester_id(dev);
|
|
address_space_stl_le(&dev->bus_master_as, msg.address, msg.data,
|
|
attrs, NULL);
|
|
}
|
|
|
|
static void pci_reset_regions(PCIDevice *dev)
|
|
{
|
|
int r;
|
|
if (pci_is_vf(dev)) {
|
|
return;
|
|
}
|
|
|
|
for (r = 0; r < PCI_NUM_REGIONS; ++r) {
|
|
PCIIORegion *region = &dev->io_regions[r];
|
|
if (!region->size) {
|
|
continue;
|
|
}
|
|
|
|
if (!(region->type & PCI_BASE_ADDRESS_SPACE_IO) &&
|
|
region->type & PCI_BASE_ADDRESS_MEM_TYPE_64) {
|
|
pci_set_quad(dev->config + pci_bar(dev, r), region->type);
|
|
} else {
|
|
pci_set_long(dev->config + pci_bar(dev, r), region->type);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void pci_do_device_reset(PCIDevice *dev)
|
|
{
|
|
pci_device_deassert_intx(dev);
|
|
assert(dev->irq_state == 0);
|
|
|
|
/* Clear all writable bits */
|
|
pci_word_test_and_clear_mask(dev->config + PCI_COMMAND,
|
|
pci_get_word(dev->wmask + PCI_COMMAND) |
|
|
pci_get_word(dev->w1cmask + PCI_COMMAND));
|
|
pci_word_test_and_clear_mask(dev->config + PCI_STATUS,
|
|
pci_get_word(dev->wmask + PCI_STATUS) |
|
|
pci_get_word(dev->w1cmask + PCI_STATUS));
|
|
/* Some devices make bits of PCI_INTERRUPT_LINE read only */
|
|
pci_byte_test_and_clear_mask(dev->config + PCI_INTERRUPT_LINE,
|
|
pci_get_word(dev->wmask + PCI_INTERRUPT_LINE) |
|
|
pci_get_word(dev->w1cmask + PCI_INTERRUPT_LINE));
|
|
dev->config[PCI_CACHE_LINE_SIZE] = 0x0;
|
|
pci_reset_regions(dev);
|
|
pci_update_mappings(dev);
|
|
|
|
msi_reset(dev);
|
|
msix_reset(dev);
|
|
pcie_sriov_pf_reset(dev);
|
|
}
|
|
|
|
/*
|
|
* This function is called on #RST and FLR.
|
|
* FLR if PCI_EXP_DEVCTL_BCR_FLR is set
|
|
*/
|
|
void pci_device_reset(PCIDevice *dev)
|
|
{
|
|
device_cold_reset(&dev->qdev);
|
|
pci_do_device_reset(dev);
|
|
}
|
|
|
|
/*
|
|
* Trigger pci bus reset under a given bus.
|
|
* Called via bus_cold_reset on RST# assert, after the devices
|
|
* have been reset device_cold_reset-ed already.
|
|
*/
|
|
static void pcibus_reset_hold(Object *obj, ResetType type)
|
|
{
|
|
PCIBus *bus = PCI_BUS(obj);
|
|
int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(bus->devices); ++i) {
|
|
if (bus->devices[i]) {
|
|
pci_do_device_reset(bus->devices[i]);
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < bus->nirq; i++) {
|
|
assert(bus->irq_count[i] == 0);
|
|
}
|
|
}
|
|
|
|
static void pci_host_bus_register(DeviceState *host)
|
|
{
|
|
PCIHostState *host_bridge = PCI_HOST_BRIDGE(host);
|
|
|
|
QLIST_INSERT_HEAD(&pci_host_bridges, host_bridge, next);
|
|
}
|
|
|
|
static void pci_host_bus_unregister(DeviceState *host)
|
|
{
|
|
PCIHostState *host_bridge = PCI_HOST_BRIDGE(host);
|
|
|
|
QLIST_REMOVE(host_bridge, next);
|
|
}
|
|
|
|
PCIBus *pci_device_root_bus(const PCIDevice *d)
|
|
{
|
|
PCIBus *bus = pci_get_bus(d);
|
|
|
|
while (!pci_bus_is_root(bus)) {
|
|
d = bus->parent_dev;
|
|
assert(d != NULL);
|
|
|
|
bus = pci_get_bus(d);
|
|
}
|
|
|
|
return bus;
|
|
}
|
|
|
|
const char *pci_root_bus_path(PCIDevice *dev)
|
|
{
|
|
PCIBus *rootbus = pci_device_root_bus(dev);
|
|
PCIHostState *host_bridge = PCI_HOST_BRIDGE(rootbus->qbus.parent);
|
|
PCIHostBridgeClass *hc = PCI_HOST_BRIDGE_GET_CLASS(host_bridge);
|
|
|
|
assert(host_bridge->bus == rootbus);
|
|
|
|
if (hc->root_bus_path) {
|
|
return (*hc->root_bus_path)(host_bridge, rootbus);
|
|
}
|
|
|
|
return rootbus->qbus.name;
|
|
}
|
|
|
|
bool pci_bus_bypass_iommu(PCIBus *bus)
|
|
{
|
|
PCIBus *rootbus = bus;
|
|
PCIHostState *host_bridge;
|
|
|
|
if (!pci_bus_is_root(bus)) {
|
|
rootbus = pci_device_root_bus(bus->parent_dev);
|
|
}
|
|
|
|
host_bridge = PCI_HOST_BRIDGE(rootbus->qbus.parent);
|
|
|
|
assert(host_bridge->bus == rootbus);
|
|
|
|
return host_bridge->bypass_iommu;
|
|
}
|
|
|
|
static void pci_root_bus_internal_init(PCIBus *bus, DeviceState *parent,
|
|
MemoryRegion *mem, MemoryRegion *io,
|
|
uint8_t devfn_min)
|
|
{
|
|
assert(PCI_FUNC(devfn_min) == 0);
|
|
bus->devfn_min = devfn_min;
|
|
bus->slot_reserved_mask = 0x0;
|
|
bus->address_space_mem = mem;
|
|
bus->address_space_io = io;
|
|
bus->flags |= PCI_BUS_IS_ROOT;
|
|
|
|
/* host bridge */
|
|
QLIST_INIT(&bus->child);
|
|
|
|
pci_host_bus_register(parent);
|
|
}
|
|
|
|
static void pci_bus_uninit(PCIBus *bus)
|
|
{
|
|
pci_host_bus_unregister(BUS(bus)->parent);
|
|
}
|
|
|
|
bool pci_bus_is_express(const PCIBus *bus)
|
|
{
|
|
return object_dynamic_cast(OBJECT(bus), TYPE_PCIE_BUS);
|
|
}
|
|
|
|
void pci_root_bus_init(PCIBus *bus, size_t bus_size, DeviceState *parent,
|
|
const char *name,
|
|
MemoryRegion *mem, MemoryRegion *io,
|
|
uint8_t devfn_min, const char *typename)
|
|
{
|
|
qbus_init(bus, bus_size, typename, parent, name);
|
|
pci_root_bus_internal_init(bus, parent, mem, io, devfn_min);
|
|
}
|
|
|
|
PCIBus *pci_root_bus_new(DeviceState *parent, const char *name,
|
|
MemoryRegion *mem, MemoryRegion *io,
|
|
uint8_t devfn_min, const char *typename)
|
|
{
|
|
PCIBus *bus;
|
|
|
|
bus = PCI_BUS(qbus_new(typename, parent, name));
|
|
pci_root_bus_internal_init(bus, parent, mem, io, devfn_min);
|
|
return bus;
|
|
}
|
|
|
|
void pci_root_bus_cleanup(PCIBus *bus)
|
|
{
|
|
pci_bus_uninit(bus);
|
|
/* the caller of the unplug hotplug handler will delete this device */
|
|
qbus_unrealize(BUS(bus));
|
|
}
|
|
|
|
void pci_bus_irqs(PCIBus *bus, pci_set_irq_fn set_irq,
|
|
void *irq_opaque, int nirq)
|
|
{
|
|
bus->set_irq = set_irq;
|
|
bus->irq_opaque = irq_opaque;
|
|
bus->nirq = nirq;
|
|
g_free(bus->irq_count);
|
|
bus->irq_count = g_malloc0(nirq * sizeof(bus->irq_count[0]));
|
|
}
|
|
|
|
void pci_bus_map_irqs(PCIBus *bus, pci_map_irq_fn map_irq)
|
|
{
|
|
bus->map_irq = map_irq;
|
|
}
|
|
|
|
void pci_bus_irqs_cleanup(PCIBus *bus)
|
|
{
|
|
bus->set_irq = NULL;
|
|
bus->map_irq = NULL;
|
|
bus->irq_opaque = NULL;
|
|
bus->nirq = 0;
|
|
g_free(bus->irq_count);
|
|
bus->irq_count = NULL;
|
|
}
|
|
|
|
PCIBus *pci_register_root_bus(DeviceState *parent, const char *name,
|
|
pci_set_irq_fn set_irq, pci_map_irq_fn map_irq,
|
|
void *irq_opaque,
|
|
MemoryRegion *mem, MemoryRegion *io,
|
|
uint8_t devfn_min, int nirq,
|
|
const char *typename)
|
|
{
|
|
PCIBus *bus;
|
|
|
|
bus = pci_root_bus_new(parent, name, mem, io, devfn_min, typename);
|
|
pci_bus_irqs(bus, set_irq, irq_opaque, nirq);
|
|
pci_bus_map_irqs(bus, map_irq);
|
|
return bus;
|
|
}
|
|
|
|
void pci_unregister_root_bus(PCIBus *bus)
|
|
{
|
|
pci_bus_irqs_cleanup(bus);
|
|
pci_root_bus_cleanup(bus);
|
|
}
|
|
|
|
int pci_bus_num(PCIBus *s)
|
|
{
|
|
return PCI_BUS_GET_CLASS(s)->bus_num(s);
|
|
}
|
|
|
|
/* Returns the min and max bus numbers of a PCI bus hierarchy */
|
|
void pci_bus_range(PCIBus *bus, int *min_bus, int *max_bus)
|
|
{
|
|
int i;
|
|
*min_bus = *max_bus = pci_bus_num(bus);
|
|
|
|
for (i = 0; i < ARRAY_SIZE(bus->devices); ++i) {
|
|
PCIDevice *dev = bus->devices[i];
|
|
|
|
if (dev && IS_PCI_BRIDGE(dev)) {
|
|
*min_bus = MIN(*min_bus, dev->config[PCI_SECONDARY_BUS]);
|
|
*max_bus = MAX(*max_bus, dev->config[PCI_SUBORDINATE_BUS]);
|
|
}
|
|
}
|
|
}
|
|
|
|
int pci_bus_numa_node(PCIBus *bus)
|
|
{
|
|
return PCI_BUS_GET_CLASS(bus)->numa_node(bus);
|
|
}
|
|
|
|
static int get_pci_config_device(QEMUFile *f, void *pv, size_t size,
|
|
const VMStateField *field)
|
|
{
|
|
PCIDevice *s = container_of(pv, PCIDevice, config);
|
|
uint8_t *config;
|
|
int i;
|
|
|
|
assert(size == pci_config_size(s));
|
|
config = g_malloc(size);
|
|
|
|
qemu_get_buffer(f, config, size);
|
|
for (i = 0; i < size; ++i) {
|
|
if ((config[i] ^ s->config[i]) &
|
|
s->cmask[i] & ~s->wmask[i] & ~s->w1cmask[i]) {
|
|
error_report("%s: Bad config data: i=0x%x read: %x device: %x "
|
|
"cmask: %x wmask: %x w1cmask:%x", __func__,
|
|
i, config[i], s->config[i],
|
|
s->cmask[i], s->wmask[i], s->w1cmask[i]);
|
|
g_free(config);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
memcpy(s->config, config, size);
|
|
|
|
pci_update_mappings(s);
|
|
if (IS_PCI_BRIDGE(s)) {
|
|
pci_bridge_update_mappings(PCI_BRIDGE(s));
|
|
}
|
|
|
|
memory_region_set_enabled(&s->bus_master_enable_region,
|
|
pci_get_word(s->config + PCI_COMMAND)
|
|
& PCI_COMMAND_MASTER);
|
|
|
|
g_free(config);
|
|
return 0;
|
|
}
|
|
|
|
/* just put buffer */
|
|
static int put_pci_config_device(QEMUFile *f, void *pv, size_t size,
|
|
const VMStateField *field, JSONWriter *vmdesc)
|
|
{
|
|
const uint8_t **v = pv;
|
|
assert(size == pci_config_size(container_of(pv, PCIDevice, config)));
|
|
qemu_put_buffer(f, *v, size);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const VMStateInfo vmstate_info_pci_config = {
|
|
.name = "pci config",
|
|
.get = get_pci_config_device,
|
|
.put = put_pci_config_device,
|
|
};
|
|
|
|
static int get_pci_irq_state(QEMUFile *f, void *pv, size_t size,
|
|
const VMStateField *field)
|
|
{
|
|
PCIDevice *s = container_of(pv, PCIDevice, irq_state);
|
|
uint32_t irq_state[PCI_NUM_PINS];
|
|
int i;
|
|
for (i = 0; i < PCI_NUM_PINS; ++i) {
|
|
irq_state[i] = qemu_get_be32(f);
|
|
if (irq_state[i] != 0x1 && irq_state[i] != 0) {
|
|
fprintf(stderr, "irq state %d: must be 0 or 1.\n",
|
|
irq_state[i]);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < PCI_NUM_PINS; ++i) {
|
|
pci_set_irq_state(s, i, irq_state[i]);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int put_pci_irq_state(QEMUFile *f, void *pv, size_t size,
|
|
const VMStateField *field, JSONWriter *vmdesc)
|
|
{
|
|
int i;
|
|
PCIDevice *s = container_of(pv, PCIDevice, irq_state);
|
|
|
|
for (i = 0; i < PCI_NUM_PINS; ++i) {
|
|
qemu_put_be32(f, pci_irq_state(s, i));
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const VMStateInfo vmstate_info_pci_irq_state = {
|
|
.name = "pci irq state",
|
|
.get = get_pci_irq_state,
|
|
.put = put_pci_irq_state,
|
|
};
|
|
|
|
static bool migrate_is_pcie(void *opaque, int version_id)
|
|
{
|
|
return pci_is_express((PCIDevice *)opaque);
|
|
}
|
|
|
|
static bool migrate_is_not_pcie(void *opaque, int version_id)
|
|
{
|
|
return !pci_is_express((PCIDevice *)opaque);
|
|
}
|
|
|
|
const VMStateDescription vmstate_pci_device = {
|
|
.name = "PCIDevice",
|
|
.version_id = 2,
|
|
.minimum_version_id = 1,
|
|
.fields = (const VMStateField[]) {
|
|
VMSTATE_INT32_POSITIVE_LE(version_id, PCIDevice),
|
|
VMSTATE_BUFFER_UNSAFE_INFO_TEST(config, PCIDevice,
|
|
migrate_is_not_pcie,
|
|
0, vmstate_info_pci_config,
|
|
PCI_CONFIG_SPACE_SIZE),
|
|
VMSTATE_BUFFER_UNSAFE_INFO_TEST(config, PCIDevice,
|
|
migrate_is_pcie,
|
|
0, vmstate_info_pci_config,
|
|
PCIE_CONFIG_SPACE_SIZE),
|
|
VMSTATE_BUFFER_UNSAFE_INFO(irq_state, PCIDevice, 2,
|
|
vmstate_info_pci_irq_state,
|
|
PCI_NUM_PINS * sizeof(int32_t)),
|
|
VMSTATE_END_OF_LIST()
|
|
}
|
|
};
|
|
|
|
|
|
void pci_device_save(PCIDevice *s, QEMUFile *f)
|
|
{
|
|
/* Clear interrupt status bit: it is implicit
|
|
* in irq_state which we are saving.
|
|
* This makes us compatible with old devices
|
|
* which never set or clear this bit. */
|
|
s->config[PCI_STATUS] &= ~PCI_STATUS_INTERRUPT;
|
|
vmstate_save_state(f, &vmstate_pci_device, s, NULL);
|
|
/* Restore the interrupt status bit. */
|
|
pci_update_irq_status(s);
|
|
}
|
|
|
|
int pci_device_load(PCIDevice *s, QEMUFile *f)
|
|
{
|
|
int ret;
|
|
ret = vmstate_load_state(f, &vmstate_pci_device, s, s->version_id);
|
|
/* Restore the interrupt status bit. */
|
|
pci_update_irq_status(s);
|
|
return ret;
|
|
}
|
|
|
|
static void pci_set_default_subsystem_id(PCIDevice *pci_dev)
|
|
{
|
|
pci_set_word(pci_dev->config + PCI_SUBSYSTEM_VENDOR_ID,
|
|
pci_default_sub_vendor_id);
|
|
pci_set_word(pci_dev->config + PCI_SUBSYSTEM_ID,
|
|
pci_default_sub_device_id);
|
|
}
|
|
|
|
/*
|
|
* Parse [[<domain>:]<bus>:]<slot>, return -1 on error if funcp == NULL
|
|
* [[<domain>:]<bus>:]<slot>.<func>, return -1 on error
|
|
*/
|
|
static int pci_parse_devaddr(const char *addr, int *domp, int *busp,
|
|
unsigned int *slotp, unsigned int *funcp)
|
|
{
|
|
const char *p;
|
|
char *e;
|
|
unsigned long val;
|
|
unsigned long dom = 0, bus = 0;
|
|
unsigned int slot = 0;
|
|
unsigned int func = 0;
|
|
|
|
p = addr;
|
|
val = strtoul(p, &e, 16);
|
|
if (e == p)
|
|
return -1;
|
|
if (*e == ':') {
|
|
bus = val;
|
|
p = e + 1;
|
|
val = strtoul(p, &e, 16);
|
|
if (e == p)
|
|
return -1;
|
|
if (*e == ':') {
|
|
dom = bus;
|
|
bus = val;
|
|
p = e + 1;
|
|
val = strtoul(p, &e, 16);
|
|
if (e == p)
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
slot = val;
|
|
|
|
if (funcp != NULL) {
|
|
if (*e != '.')
|
|
return -1;
|
|
|
|
p = e + 1;
|
|
val = strtoul(p, &e, 16);
|
|
if (e == p)
|
|
return -1;
|
|
|
|
func = val;
|
|
}
|
|
|
|
/* if funcp == NULL func is 0 */
|
|
if (dom > 0xffff || bus > 0xff || slot > 0x1f || func > 7)
|
|
return -1;
|
|
|
|
if (*e)
|
|
return -1;
|
|
|
|
*domp = dom;
|
|
*busp = bus;
|
|
*slotp = slot;
|
|
if (funcp != NULL)
|
|
*funcp = func;
|
|
return 0;
|
|
}
|
|
|
|
static void pci_init_cmask(PCIDevice *dev)
|
|
{
|
|
pci_set_word(dev->cmask + PCI_VENDOR_ID, 0xffff);
|
|
pci_set_word(dev->cmask + PCI_DEVICE_ID, 0xffff);
|
|
dev->cmask[PCI_STATUS] = PCI_STATUS_CAP_LIST;
|
|
dev->cmask[PCI_REVISION_ID] = 0xff;
|
|
dev->cmask[PCI_CLASS_PROG] = 0xff;
|
|
pci_set_word(dev->cmask + PCI_CLASS_DEVICE, 0xffff);
|
|
dev->cmask[PCI_HEADER_TYPE] = 0xff;
|
|
dev->cmask[PCI_CAPABILITY_LIST] = 0xff;
|
|
}
|
|
|
|
static void pci_init_wmask(PCIDevice *dev)
|
|
{
|
|
int config_size = pci_config_size(dev);
|
|
|
|
dev->wmask[PCI_CACHE_LINE_SIZE] = 0xff;
|
|
dev->wmask[PCI_INTERRUPT_LINE] = 0xff;
|
|
pci_set_word(dev->wmask + PCI_COMMAND,
|
|
PCI_COMMAND_IO | PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER |
|
|
PCI_COMMAND_INTX_DISABLE);
|
|
pci_word_test_and_set_mask(dev->wmask + PCI_COMMAND, PCI_COMMAND_SERR);
|
|
|
|
memset(dev->wmask + PCI_CONFIG_HEADER_SIZE, 0xff,
|
|
config_size - PCI_CONFIG_HEADER_SIZE);
|
|
}
|
|
|
|
static void pci_init_w1cmask(PCIDevice *dev)
|
|
{
|
|
/*
|
|
* Note: It's okay to set w1cmask even for readonly bits as
|
|
* long as their value is hardwired to 0.
|
|
*/
|
|
pci_set_word(dev->w1cmask + PCI_STATUS,
|
|
PCI_STATUS_PARITY | PCI_STATUS_SIG_TARGET_ABORT |
|
|
PCI_STATUS_REC_TARGET_ABORT | PCI_STATUS_REC_MASTER_ABORT |
|
|
PCI_STATUS_SIG_SYSTEM_ERROR | PCI_STATUS_DETECTED_PARITY);
|
|
}
|
|
|
|
static void pci_init_mask_bridge(PCIDevice *d)
|
|
{
|
|
/* PCI_PRIMARY_BUS, PCI_SECONDARY_BUS, PCI_SUBORDINATE_BUS and
|
|
PCI_SEC_LATENCY_TIMER */
|
|
memset(d->wmask + PCI_PRIMARY_BUS, 0xff, 4);
|
|
|
|
/* base and limit */
|
|
d->wmask[PCI_IO_BASE] = PCI_IO_RANGE_MASK & 0xff;
|
|
d->wmask[PCI_IO_LIMIT] = PCI_IO_RANGE_MASK & 0xff;
|
|
pci_set_word(d->wmask + PCI_MEMORY_BASE,
|
|
PCI_MEMORY_RANGE_MASK & 0xffff);
|
|
pci_set_word(d->wmask + PCI_MEMORY_LIMIT,
|
|
PCI_MEMORY_RANGE_MASK & 0xffff);
|
|
pci_set_word(d->wmask + PCI_PREF_MEMORY_BASE,
|
|
PCI_PREF_RANGE_MASK & 0xffff);
|
|
pci_set_word(d->wmask + PCI_PREF_MEMORY_LIMIT,
|
|
PCI_PREF_RANGE_MASK & 0xffff);
|
|
|
|
/* PCI_PREF_BASE_UPPER32 and PCI_PREF_LIMIT_UPPER32 */
|
|
memset(d->wmask + PCI_PREF_BASE_UPPER32, 0xff, 8);
|
|
|
|
/* Supported memory and i/o types */
|
|
d->config[PCI_IO_BASE] |= PCI_IO_RANGE_TYPE_16;
|
|
d->config[PCI_IO_LIMIT] |= PCI_IO_RANGE_TYPE_16;
|
|
pci_word_test_and_set_mask(d->config + PCI_PREF_MEMORY_BASE,
|
|
PCI_PREF_RANGE_TYPE_64);
|
|
pci_word_test_and_set_mask(d->config + PCI_PREF_MEMORY_LIMIT,
|
|
PCI_PREF_RANGE_TYPE_64);
|
|
|
|
/*
|
|
* TODO: Bridges default to 10-bit VGA decoding but we currently only
|
|
* implement 16-bit decoding (no alias support).
|
|
*/
|
|
pci_set_word(d->wmask + PCI_BRIDGE_CONTROL,
|
|
PCI_BRIDGE_CTL_PARITY |
|
|
PCI_BRIDGE_CTL_SERR |
|
|
PCI_BRIDGE_CTL_ISA |
|
|
PCI_BRIDGE_CTL_VGA |
|
|
PCI_BRIDGE_CTL_VGA_16BIT |
|
|
PCI_BRIDGE_CTL_MASTER_ABORT |
|
|
PCI_BRIDGE_CTL_BUS_RESET |
|
|
PCI_BRIDGE_CTL_FAST_BACK |
|
|
PCI_BRIDGE_CTL_DISCARD |
|
|
PCI_BRIDGE_CTL_SEC_DISCARD |
|
|
PCI_BRIDGE_CTL_DISCARD_SERR);
|
|
/* Below does not do anything as we never set this bit, put here for
|
|
* completeness. */
|
|
pci_set_word(d->w1cmask + PCI_BRIDGE_CONTROL,
|
|
PCI_BRIDGE_CTL_DISCARD_STATUS);
|
|
d->cmask[PCI_IO_BASE] |= PCI_IO_RANGE_TYPE_MASK;
|
|
d->cmask[PCI_IO_LIMIT] |= PCI_IO_RANGE_TYPE_MASK;
|
|
pci_word_test_and_set_mask(d->cmask + PCI_PREF_MEMORY_BASE,
|
|
PCI_PREF_RANGE_TYPE_MASK);
|
|
pci_word_test_and_set_mask(d->cmask + PCI_PREF_MEMORY_LIMIT,
|
|
PCI_PREF_RANGE_TYPE_MASK);
|
|
}
|
|
|
|
static void pci_init_multifunction(PCIBus *bus, PCIDevice *dev, Error **errp)
|
|
{
|
|
uint8_t slot = PCI_SLOT(dev->devfn);
|
|
uint8_t func;
|
|
|
|
if (dev->cap_present & QEMU_PCI_CAP_MULTIFUNCTION) {
|
|
dev->config[PCI_HEADER_TYPE] |= PCI_HEADER_TYPE_MULTI_FUNCTION;
|
|
}
|
|
|
|
/*
|
|
* With SR/IOV and ARI, a device at function 0 need not be a multifunction
|
|
* device, as it may just be a VF that ended up with function 0 in
|
|
* the legacy PCI interpretation. Avoid failing in such cases:
|
|
*/
|
|
if (pci_is_vf(dev) &&
|
|
dev->exp.sriov_vf.pf->cap_present & QEMU_PCI_CAP_MULTIFUNCTION) {
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* multifunction bit is interpreted in two ways as follows.
|
|
* - all functions must set the bit to 1.
|
|
* Example: Intel X53
|
|
* - function 0 must set the bit, but the rest function (> 0)
|
|
* is allowed to leave the bit to 0.
|
|
* Example: PIIX3(also in qemu), PIIX4(also in qemu), ICH10,
|
|
*
|
|
* So OS (at least Linux) checks the bit of only function 0,
|
|
* and doesn't see the bit of function > 0.
|
|
*
|
|
* The below check allows both interpretation.
|
|
*/
|
|
if (PCI_FUNC(dev->devfn)) {
|
|
PCIDevice *f0 = bus->devices[PCI_DEVFN(slot, 0)];
|
|
if (f0 && !(f0->cap_present & QEMU_PCI_CAP_MULTIFUNCTION)) {
|
|
/* function 0 should set multifunction bit */
|
|
error_setg(errp, "PCI: single function device can't be populated "
|
|
"in function %x.%x", slot, PCI_FUNC(dev->devfn));
|
|
return;
|
|
}
|
|
return;
|
|
}
|
|
|
|
if (dev->cap_present & QEMU_PCI_CAP_MULTIFUNCTION) {
|
|
return;
|
|
}
|
|
/* function 0 indicates single function, so function > 0 must be NULL */
|
|
for (func = 1; func < PCI_FUNC_MAX; ++func) {
|
|
if (bus->devices[PCI_DEVFN(slot, func)]) {
|
|
error_setg(errp, "PCI: %x.0 indicates single function, "
|
|
"but %x.%x is already populated.",
|
|
slot, slot, func);
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void pci_config_alloc(PCIDevice *pci_dev)
|
|
{
|
|
int config_size = pci_config_size(pci_dev);
|
|
|
|
pci_dev->config = g_malloc0(config_size);
|
|
pci_dev->cmask = g_malloc0(config_size);
|
|
pci_dev->wmask = g_malloc0(config_size);
|
|
pci_dev->w1cmask = g_malloc0(config_size);
|
|
pci_dev->used = g_malloc0(config_size);
|
|
}
|
|
|
|
static void pci_config_free(PCIDevice *pci_dev)
|
|
{
|
|
g_free(pci_dev->config);
|
|
g_free(pci_dev->cmask);
|
|
g_free(pci_dev->wmask);
|
|
g_free(pci_dev->w1cmask);
|
|
g_free(pci_dev->used);
|
|
}
|
|
|
|
static void do_pci_unregister_device(PCIDevice *pci_dev)
|
|
{
|
|
pci_get_bus(pci_dev)->devices[pci_dev->devfn] = NULL;
|
|
pci_config_free(pci_dev);
|
|
|
|
if (xen_mode == XEN_EMULATE) {
|
|
xen_evtchn_remove_pci_device(pci_dev);
|
|
}
|
|
if (memory_region_is_mapped(&pci_dev->bus_master_enable_region)) {
|
|
memory_region_del_subregion(&pci_dev->bus_master_container_region,
|
|
&pci_dev->bus_master_enable_region);
|
|
}
|
|
address_space_destroy(&pci_dev->bus_master_as);
|
|
}
|
|
|
|
/* Extract PCIReqIDCache into BDF format */
|
|
static uint16_t pci_req_id_cache_extract(PCIReqIDCache *cache)
|
|
{
|
|
uint8_t bus_n;
|
|
uint16_t result;
|
|
|
|
switch (cache->type) {
|
|
case PCI_REQ_ID_BDF:
|
|
result = pci_get_bdf(cache->dev);
|
|
break;
|
|
case PCI_REQ_ID_SECONDARY_BUS:
|
|
bus_n = pci_dev_bus_num(cache->dev);
|
|
result = PCI_BUILD_BDF(bus_n, 0);
|
|
break;
|
|
default:
|
|
error_report("Invalid PCI requester ID cache type: %d",
|
|
cache->type);
|
|
exit(1);
|
|
break;
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
/* Parse bridges up to the root complex and return requester ID
|
|
* cache for specific device. For full PCIe topology, the cache
|
|
* result would be exactly the same as getting BDF of the device.
|
|
* However, several tricks are required when system mixed up with
|
|
* legacy PCI devices and PCIe-to-PCI bridges.
|
|
*
|
|
* Here we cache the proxy device (and type) not requester ID since
|
|
* bus number might change from time to time.
|
|
*/
|
|
static PCIReqIDCache pci_req_id_cache_get(PCIDevice *dev)
|
|
{
|
|
PCIDevice *parent;
|
|
PCIReqIDCache cache = {
|
|
.dev = dev,
|
|
.type = PCI_REQ_ID_BDF,
|
|
};
|
|
|
|
while (!pci_bus_is_root(pci_get_bus(dev))) {
|
|
/* We are under PCI/PCIe bridges */
|
|
parent = pci_get_bus(dev)->parent_dev;
|
|
if (pci_is_express(parent)) {
|
|
if (pcie_cap_get_type(parent) == PCI_EXP_TYPE_PCI_BRIDGE) {
|
|
/* When we pass through PCIe-to-PCI/PCIX bridges, we
|
|
* override the requester ID using secondary bus
|
|
* number of parent bridge with zeroed devfn
|
|
* (pcie-to-pci bridge spec chap 2.3). */
|
|
cache.type = PCI_REQ_ID_SECONDARY_BUS;
|
|
cache.dev = dev;
|
|
}
|
|
} else {
|
|
/* Legacy PCI, override requester ID with the bridge's
|
|
* BDF upstream. When the root complex connects to
|
|
* legacy PCI devices (including buses), it can only
|
|
* obtain requester ID info from directly attached
|
|
* devices. If devices are attached under bridges, only
|
|
* the requester ID of the bridge that is directly
|
|
* attached to the root complex can be recognized. */
|
|
cache.type = PCI_REQ_ID_BDF;
|
|
cache.dev = parent;
|
|
}
|
|
dev = parent;
|
|
}
|
|
|
|
return cache;
|
|
}
|
|
|
|
uint16_t pci_requester_id(PCIDevice *dev)
|
|
{
|
|
return pci_req_id_cache_extract(&dev->requester_id_cache);
|
|
}
|
|
|
|
static bool pci_bus_devfn_available(PCIBus *bus, int devfn)
|
|
{
|
|
return !(bus->devices[devfn]);
|
|
}
|
|
|
|
static bool pci_bus_devfn_reserved(PCIBus *bus, int devfn)
|
|
{
|
|
return bus->slot_reserved_mask & (1UL << PCI_SLOT(devfn));
|
|
}
|
|
|
|
uint32_t pci_bus_get_slot_reserved_mask(PCIBus *bus)
|
|
{
|
|
return bus->slot_reserved_mask;
|
|
}
|
|
|
|
void pci_bus_set_slot_reserved_mask(PCIBus *bus, uint32_t mask)
|
|
{
|
|
bus->slot_reserved_mask |= mask;
|
|
}
|
|
|
|
void pci_bus_clear_slot_reserved_mask(PCIBus *bus, uint32_t mask)
|
|
{
|
|
bus->slot_reserved_mask &= ~mask;
|
|
}
|
|
|
|
/* -1 for devfn means auto assign */
|
|
static PCIDevice *do_pci_register_device(PCIDevice *pci_dev,
|
|
const char *name, int devfn,
|
|
Error **errp)
|
|
{
|
|
PCIDeviceClass *pc = PCI_DEVICE_GET_CLASS(pci_dev);
|
|
PCIConfigReadFunc *config_read = pc->config_read;
|
|
PCIConfigWriteFunc *config_write = pc->config_write;
|
|
Error *local_err = NULL;
|
|
DeviceState *dev = DEVICE(pci_dev);
|
|
PCIBus *bus = pci_get_bus(pci_dev);
|
|
bool is_bridge = IS_PCI_BRIDGE(pci_dev);
|
|
|
|
/* Only pci bridges can be attached to extra PCI root buses */
|
|
if (pci_bus_is_root(bus) && bus->parent_dev && !is_bridge) {
|
|
error_setg(errp,
|
|
"PCI: Only PCI/PCIe bridges can be plugged into %s",
|
|
bus->parent_dev->name);
|
|
return NULL;
|
|
}
|
|
|
|
if (devfn < 0) {
|
|
for(devfn = bus->devfn_min ; devfn < ARRAY_SIZE(bus->devices);
|
|
devfn += PCI_FUNC_MAX) {
|
|
if (pci_bus_devfn_available(bus, devfn) &&
|
|
!pci_bus_devfn_reserved(bus, devfn)) {
|
|
goto found;
|
|
}
|
|
}
|
|
error_setg(errp, "PCI: no slot/function available for %s, all in use "
|
|
"or reserved", name);
|
|
return NULL;
|
|
found: ;
|
|
} else if (pci_bus_devfn_reserved(bus, devfn)) {
|
|
error_setg(errp, "PCI: slot %d function %d not available for %s,"
|
|
" reserved",
|
|
PCI_SLOT(devfn), PCI_FUNC(devfn), name);
|
|
return NULL;
|
|
} else if (!pci_bus_devfn_available(bus, devfn)) {
|
|
error_setg(errp, "PCI: slot %d function %d not available for %s,"
|
|
" in use by %s,id=%s",
|
|
PCI_SLOT(devfn), PCI_FUNC(devfn), name,
|
|
bus->devices[devfn]->name, bus->devices[devfn]->qdev.id);
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* Populating function 0 triggers a scan from the guest that
|
|
* exposes other non-zero functions. Hence we need to ensure that
|
|
* function 0 wasn't added yet.
|
|
*/
|
|
if (dev->hotplugged && !pci_is_vf(pci_dev) &&
|
|
pci_get_function_0(pci_dev)) {
|
|
error_setg(errp, "PCI: slot %d function 0 already occupied by %s,"
|
|
" new func %s cannot be exposed to guest.",
|
|
PCI_SLOT(pci_get_function_0(pci_dev)->devfn),
|
|
pci_get_function_0(pci_dev)->name,
|
|
name);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
pci_dev->devfn = devfn;
|
|
pci_dev->requester_id_cache = pci_req_id_cache_get(pci_dev);
|
|
pstrcpy(pci_dev->name, sizeof(pci_dev->name), name);
|
|
|
|
memory_region_init(&pci_dev->bus_master_container_region, OBJECT(pci_dev),
|
|
"bus master container", UINT64_MAX);
|
|
address_space_init(&pci_dev->bus_master_as,
|
|
&pci_dev->bus_master_container_region, pci_dev->name);
|
|
pci_dev->bus_master_as.max_bounce_buffer_size =
|
|
pci_dev->max_bounce_buffer_size;
|
|
|
|
if (phase_check(PHASE_MACHINE_READY)) {
|
|
pci_init_bus_master(pci_dev);
|
|
}
|
|
pci_dev->irq_state = 0;
|
|
pci_config_alloc(pci_dev);
|
|
|
|
pci_config_set_vendor_id(pci_dev->config, pc->vendor_id);
|
|
pci_config_set_device_id(pci_dev->config, pc->device_id);
|
|
pci_config_set_revision(pci_dev->config, pc->revision);
|
|
pci_config_set_class(pci_dev->config, pc->class_id);
|
|
|
|
if (!is_bridge) {
|
|
if (pc->subsystem_vendor_id || pc->subsystem_id) {
|
|
pci_set_word(pci_dev->config + PCI_SUBSYSTEM_VENDOR_ID,
|
|
pc->subsystem_vendor_id);
|
|
pci_set_word(pci_dev->config + PCI_SUBSYSTEM_ID,
|
|
pc->subsystem_id);
|
|
} else {
|
|
pci_set_default_subsystem_id(pci_dev);
|
|
}
|
|
} else {
|
|
/* subsystem_vendor_id/subsystem_id are only for header type 0 */
|
|
assert(!pc->subsystem_vendor_id);
|
|
assert(!pc->subsystem_id);
|
|
}
|
|
pci_init_cmask(pci_dev);
|
|
pci_init_wmask(pci_dev);
|
|
pci_init_w1cmask(pci_dev);
|
|
if (is_bridge) {
|
|
pci_init_mask_bridge(pci_dev);
|
|
}
|
|
pci_init_multifunction(bus, pci_dev, &local_err);
|
|
if (local_err) {
|
|
error_propagate(errp, local_err);
|
|
do_pci_unregister_device(pci_dev);
|
|
return NULL;
|
|
}
|
|
|
|
if (!config_read)
|
|
config_read = pci_default_read_config;
|
|
if (!config_write)
|
|
config_write = pci_default_write_config;
|
|
pci_dev->config_read = config_read;
|
|
pci_dev->config_write = config_write;
|
|
bus->devices[devfn] = pci_dev;
|
|
pci_dev->version_id = 2; /* Current pci device vmstate version */
|
|
return pci_dev;
|
|
}
|
|
|
|
static void pci_unregister_io_regions(PCIDevice *pci_dev)
|
|
{
|
|
PCIIORegion *r;
|
|
int i;
|
|
|
|
for(i = 0; i < PCI_NUM_REGIONS; i++) {
|
|
r = &pci_dev->io_regions[i];
|
|
if (!r->size || r->addr == PCI_BAR_UNMAPPED)
|
|
continue;
|
|
memory_region_del_subregion(r->address_space, r->memory);
|
|
}
|
|
|
|
pci_unregister_vga(pci_dev);
|
|
}
|
|
|
|
static void pci_qdev_unrealize(DeviceState *dev)
|
|
{
|
|
PCIDevice *pci_dev = PCI_DEVICE(dev);
|
|
PCIDeviceClass *pc = PCI_DEVICE_GET_CLASS(pci_dev);
|
|
|
|
pci_unregister_io_regions(pci_dev);
|
|
pci_del_option_rom(pci_dev);
|
|
|
|
if (pc->exit) {
|
|
pc->exit(pci_dev);
|
|
}
|
|
|
|
pci_device_deassert_intx(pci_dev);
|
|
do_pci_unregister_device(pci_dev);
|
|
|
|
pci_dev->msi_trigger = NULL;
|
|
|
|
/*
|
|
* clean up acpi-index so it could reused by another device
|
|
*/
|
|
if (pci_dev->acpi_index) {
|
|
GSequence *used_indexes = pci_acpi_index_list();
|
|
|
|
g_sequence_remove(g_sequence_lookup(used_indexes,
|
|
GINT_TO_POINTER(pci_dev->acpi_index),
|
|
g_cmp_uint32, NULL));
|
|
}
|
|
}
|
|
|
|
void pci_register_bar(PCIDevice *pci_dev, int region_num,
|
|
uint8_t type, MemoryRegion *memory)
|
|
{
|
|
PCIIORegion *r;
|
|
uint32_t addr; /* offset in pci config space */
|
|
uint64_t wmask;
|
|
pcibus_t size = memory_region_size(memory);
|
|
uint8_t hdr_type;
|
|
|
|
assert(!pci_is_vf(pci_dev)); /* VFs must use pcie_sriov_vf_register_bar */
|
|
assert(region_num >= 0);
|
|
assert(region_num < PCI_NUM_REGIONS);
|
|
assert(is_power_of_2(size));
|
|
|
|
/* A PCI bridge device (with Type 1 header) may only have at most 2 BARs */
|
|
hdr_type =
|
|
pci_dev->config[PCI_HEADER_TYPE] & ~PCI_HEADER_TYPE_MULTI_FUNCTION;
|
|
assert(hdr_type != PCI_HEADER_TYPE_BRIDGE || region_num < 2);
|
|
|
|
r = &pci_dev->io_regions[region_num];
|
|
r->addr = PCI_BAR_UNMAPPED;
|
|
r->size = size;
|
|
r->type = type;
|
|
r->memory = memory;
|
|
r->address_space = type & PCI_BASE_ADDRESS_SPACE_IO
|
|
? pci_get_bus(pci_dev)->address_space_io
|
|
: pci_get_bus(pci_dev)->address_space_mem;
|
|
|
|
wmask = ~(size - 1);
|
|
if (region_num == PCI_ROM_SLOT) {
|
|
/* ROM enable bit is writable */
|
|
wmask |= PCI_ROM_ADDRESS_ENABLE;
|
|
}
|
|
|
|
addr = pci_bar(pci_dev, region_num);
|
|
pci_set_long(pci_dev->config + addr, type);
|
|
|
|
if (!(r->type & PCI_BASE_ADDRESS_SPACE_IO) &&
|
|
r->type & PCI_BASE_ADDRESS_MEM_TYPE_64) {
|
|
pci_set_quad(pci_dev->wmask + addr, wmask);
|
|
pci_set_quad(pci_dev->cmask + addr, ~0ULL);
|
|
} else {
|
|
pci_set_long(pci_dev->wmask + addr, wmask & 0xffffffff);
|
|
pci_set_long(pci_dev->cmask + addr, 0xffffffff);
|
|
}
|
|
}
|
|
|
|
static void pci_update_vga(PCIDevice *pci_dev)
|
|
{
|
|
uint16_t cmd;
|
|
|
|
if (!pci_dev->has_vga) {
|
|
return;
|
|
}
|
|
|
|
cmd = pci_get_word(pci_dev->config + PCI_COMMAND);
|
|
|
|
memory_region_set_enabled(pci_dev->vga_regions[QEMU_PCI_VGA_MEM],
|
|
cmd & PCI_COMMAND_MEMORY);
|
|
memory_region_set_enabled(pci_dev->vga_regions[QEMU_PCI_VGA_IO_LO],
|
|
cmd & PCI_COMMAND_IO);
|
|
memory_region_set_enabled(pci_dev->vga_regions[QEMU_PCI_VGA_IO_HI],
|
|
cmd & PCI_COMMAND_IO);
|
|
}
|
|
|
|
void pci_register_vga(PCIDevice *pci_dev, MemoryRegion *mem,
|
|
MemoryRegion *io_lo, MemoryRegion *io_hi)
|
|
{
|
|
PCIBus *bus = pci_get_bus(pci_dev);
|
|
|
|
assert(!pci_dev->has_vga);
|
|
|
|
assert(memory_region_size(mem) == QEMU_PCI_VGA_MEM_SIZE);
|
|
pci_dev->vga_regions[QEMU_PCI_VGA_MEM] = mem;
|
|
memory_region_add_subregion_overlap(bus->address_space_mem,
|
|
QEMU_PCI_VGA_MEM_BASE, mem, 1);
|
|
|
|
assert(memory_region_size(io_lo) == QEMU_PCI_VGA_IO_LO_SIZE);
|
|
pci_dev->vga_regions[QEMU_PCI_VGA_IO_LO] = io_lo;
|
|
memory_region_add_subregion_overlap(bus->address_space_io,
|
|
QEMU_PCI_VGA_IO_LO_BASE, io_lo, 1);
|
|
|
|
assert(memory_region_size(io_hi) == QEMU_PCI_VGA_IO_HI_SIZE);
|
|
pci_dev->vga_regions[QEMU_PCI_VGA_IO_HI] = io_hi;
|
|
memory_region_add_subregion_overlap(bus->address_space_io,
|
|
QEMU_PCI_VGA_IO_HI_BASE, io_hi, 1);
|
|
pci_dev->has_vga = true;
|
|
|
|
pci_update_vga(pci_dev);
|
|
}
|
|
|
|
void pci_unregister_vga(PCIDevice *pci_dev)
|
|
{
|
|
PCIBus *bus = pci_get_bus(pci_dev);
|
|
|
|
if (!pci_dev->has_vga) {
|
|
return;
|
|
}
|
|
|
|
memory_region_del_subregion(bus->address_space_mem,
|
|
pci_dev->vga_regions[QEMU_PCI_VGA_MEM]);
|
|
memory_region_del_subregion(bus->address_space_io,
|
|
pci_dev->vga_regions[QEMU_PCI_VGA_IO_LO]);
|
|
memory_region_del_subregion(bus->address_space_io,
|
|
pci_dev->vga_regions[QEMU_PCI_VGA_IO_HI]);
|
|
pci_dev->has_vga = false;
|
|
}
|
|
|
|
pcibus_t pci_get_bar_addr(PCIDevice *pci_dev, int region_num)
|
|
{
|
|
return pci_dev->io_regions[region_num].addr;
|
|
}
|
|
|
|
static pcibus_t pci_config_get_bar_addr(PCIDevice *d, int reg,
|
|
uint8_t type, pcibus_t size)
|
|
{
|
|
pcibus_t new_addr;
|
|
if (!pci_is_vf(d)) {
|
|
int bar = pci_bar(d, reg);
|
|
if (type & PCI_BASE_ADDRESS_MEM_TYPE_64) {
|
|
new_addr = pci_get_quad(d->config + bar);
|
|
} else {
|
|
new_addr = pci_get_long(d->config + bar);
|
|
}
|
|
} else {
|
|
PCIDevice *pf = d->exp.sriov_vf.pf;
|
|
uint16_t sriov_cap = pf->exp.sriov_cap;
|
|
int bar = sriov_cap + PCI_SRIOV_BAR + reg * 4;
|
|
uint16_t vf_offset =
|
|
pci_get_word(pf->config + sriov_cap + PCI_SRIOV_VF_OFFSET);
|
|
uint16_t vf_stride =
|
|
pci_get_word(pf->config + sriov_cap + PCI_SRIOV_VF_STRIDE);
|
|
uint32_t vf_num = (d->devfn - (pf->devfn + vf_offset)) / vf_stride;
|
|
|
|
if (type & PCI_BASE_ADDRESS_MEM_TYPE_64) {
|
|
new_addr = pci_get_quad(pf->config + bar);
|
|
} else {
|
|
new_addr = pci_get_long(pf->config + bar);
|
|
}
|
|
new_addr += vf_num * size;
|
|
}
|
|
/* The ROM slot has a specific enable bit, keep it intact */
|
|
if (reg != PCI_ROM_SLOT) {
|
|
new_addr &= ~(size - 1);
|
|
}
|
|
return new_addr;
|
|
}
|
|
|
|
pcibus_t pci_bar_address(PCIDevice *d,
|
|
int reg, uint8_t type, pcibus_t size)
|
|
{
|
|
pcibus_t new_addr, last_addr;
|
|
uint16_t cmd = pci_get_word(d->config + PCI_COMMAND);
|
|
MachineClass *mc = MACHINE_GET_CLASS(qdev_get_machine());
|
|
bool allow_0_address = mc->pci_allow_0_address;
|
|
|
|
if (type & PCI_BASE_ADDRESS_SPACE_IO) {
|
|
if (!(cmd & PCI_COMMAND_IO)) {
|
|
return PCI_BAR_UNMAPPED;
|
|
}
|
|
new_addr = pci_config_get_bar_addr(d, reg, type, size);
|
|
last_addr = new_addr + size - 1;
|
|
/* Check if 32 bit BAR wraps around explicitly.
|
|
* TODO: make priorities correct and remove this work around.
|
|
*/
|
|
if (last_addr <= new_addr || last_addr >= UINT32_MAX ||
|
|
(!allow_0_address && new_addr == 0)) {
|
|
return PCI_BAR_UNMAPPED;
|
|
}
|
|
return new_addr;
|
|
}
|
|
|
|
if (!(cmd & PCI_COMMAND_MEMORY)) {
|
|
return PCI_BAR_UNMAPPED;
|
|
}
|
|
new_addr = pci_config_get_bar_addr(d, reg, type, size);
|
|
/* the ROM slot has a specific enable bit */
|
|
if (reg == PCI_ROM_SLOT && !(new_addr & PCI_ROM_ADDRESS_ENABLE)) {
|
|
return PCI_BAR_UNMAPPED;
|
|
}
|
|
new_addr &= ~(size - 1);
|
|
last_addr = new_addr + size - 1;
|
|
/* NOTE: we do not support wrapping */
|
|
/* XXX: as we cannot support really dynamic
|
|
mappings, we handle specific values as invalid
|
|
mappings. */
|
|
if (last_addr <= new_addr || last_addr == PCI_BAR_UNMAPPED ||
|
|
(!allow_0_address && new_addr == 0)) {
|
|
return PCI_BAR_UNMAPPED;
|
|
}
|
|
|
|
/* Now pcibus_t is 64bit.
|
|
* Check if 32 bit BAR wraps around explicitly.
|
|
* Without this, PC ide doesn't work well.
|
|
* TODO: remove this work around.
|
|
*/
|
|
if (!(type & PCI_BASE_ADDRESS_MEM_TYPE_64) && last_addr >= UINT32_MAX) {
|
|
return PCI_BAR_UNMAPPED;
|
|
}
|
|
|
|
/*
|
|
* OS is allowed to set BAR beyond its addressable
|
|
* bits. For example, 32 bit OS can set 64bit bar
|
|
* to >4G. Check it. TODO: we might need to support
|
|
* it in the future for e.g. PAE.
|
|
*/
|
|
if (last_addr >= HWADDR_MAX) {
|
|
return PCI_BAR_UNMAPPED;
|
|
}
|
|
|
|
return new_addr;
|
|
}
|
|
|
|
static void pci_update_mappings(PCIDevice *d)
|
|
{
|
|
PCIIORegion *r;
|
|
int i;
|
|
pcibus_t new_addr;
|
|
|
|
for(i = 0; i < PCI_NUM_REGIONS; i++) {
|
|
r = &d->io_regions[i];
|
|
|
|
/* this region isn't registered */
|
|
if (!r->size)
|
|
continue;
|
|
|
|
new_addr = pci_bar_address(d, i, r->type, r->size);
|
|
if (!d->has_power) {
|
|
new_addr = PCI_BAR_UNMAPPED;
|
|
}
|
|
|
|
/* This bar isn't changed */
|
|
if (new_addr == r->addr)
|
|
continue;
|
|
|
|
/* now do the real mapping */
|
|
if (r->addr != PCI_BAR_UNMAPPED) {
|
|
trace_pci_update_mappings_del(d->name, pci_dev_bus_num(d),
|
|
PCI_SLOT(d->devfn),
|
|
PCI_FUNC(d->devfn),
|
|
i, r->addr, r->size);
|
|
memory_region_del_subregion(r->address_space, r->memory);
|
|
}
|
|
r->addr = new_addr;
|
|
if (r->addr != PCI_BAR_UNMAPPED) {
|
|
trace_pci_update_mappings_add(d->name, pci_dev_bus_num(d),
|
|
PCI_SLOT(d->devfn),
|
|
PCI_FUNC(d->devfn),
|
|
i, r->addr, r->size);
|
|
memory_region_add_subregion_overlap(r->address_space,
|
|
r->addr, r->memory, 1);
|
|
}
|
|
}
|
|
|
|
pci_update_vga(d);
|
|
}
|
|
|
|
static inline int pci_irq_disabled(PCIDevice *d)
|
|
{
|
|
return pci_get_word(d->config + PCI_COMMAND) & PCI_COMMAND_INTX_DISABLE;
|
|
}
|
|
|
|
/* Called after interrupt disabled field update in config space,
|
|
* assert/deassert interrupts if necessary.
|
|
* Gets original interrupt disable bit value (before update). */
|
|
static void pci_update_irq_disabled(PCIDevice *d, int was_irq_disabled)
|
|
{
|
|
int i, disabled = pci_irq_disabled(d);
|
|
if (disabled == was_irq_disabled)
|
|
return;
|
|
for (i = 0; i < PCI_NUM_PINS; ++i) {
|
|
int state = pci_irq_state(d, i);
|
|
pci_change_irq_level(d, i, disabled ? -state : state);
|
|
}
|
|
}
|
|
|
|
uint32_t pci_default_read_config(PCIDevice *d,
|
|
uint32_t address, int len)
|
|
{
|
|
uint32_t val = 0;
|
|
|
|
assert(address + len <= pci_config_size(d));
|
|
|
|
if (pci_is_express_downstream_port(d) &&
|
|
ranges_overlap(address, len, d->exp.exp_cap + PCI_EXP_LNKSTA, 2)) {
|
|
pcie_sync_bridge_lnk(d);
|
|
}
|
|
memcpy(&val, d->config + address, len);
|
|
return le32_to_cpu(val);
|
|
}
|
|
|
|
void pci_default_write_config(PCIDevice *d, uint32_t addr, uint32_t val_in, int l)
|
|
{
|
|
int i, was_irq_disabled = pci_irq_disabled(d);
|
|
uint32_t val = val_in;
|
|
|
|
assert(addr + l <= pci_config_size(d));
|
|
|
|
for (i = 0; i < l; val >>= 8, ++i) {
|
|
uint8_t wmask = d->wmask[addr + i];
|
|
uint8_t w1cmask = d->w1cmask[addr + i];
|
|
assert(!(wmask & w1cmask));
|
|
d->config[addr + i] = (d->config[addr + i] & ~wmask) | (val & wmask);
|
|
d->config[addr + i] &= ~(val & w1cmask); /* W1C: Write 1 to Clear */
|
|
}
|
|
if (ranges_overlap(addr, l, PCI_BASE_ADDRESS_0, 24) ||
|
|
ranges_overlap(addr, l, PCI_ROM_ADDRESS, 4) ||
|
|
ranges_overlap(addr, l, PCI_ROM_ADDRESS1, 4) ||
|
|
range_covers_byte(addr, l, PCI_COMMAND))
|
|
pci_update_mappings(d);
|
|
|
|
if (ranges_overlap(addr, l, PCI_COMMAND, 2)) {
|
|
pci_update_irq_disabled(d, was_irq_disabled);
|
|
memory_region_set_enabled(&d->bus_master_enable_region,
|
|
(pci_get_word(d->config + PCI_COMMAND)
|
|
& PCI_COMMAND_MASTER) && d->has_power);
|
|
}
|
|
|
|
msi_write_config(d, addr, val_in, l);
|
|
msix_write_config(d, addr, val_in, l);
|
|
pcie_sriov_config_write(d, addr, val_in, l);
|
|
}
|
|
|
|
/***********************************************************/
|
|
/* generic PCI irq support */
|
|
|
|
/* 0 <= irq_num <= 3. level must be 0 or 1 */
|
|
static void pci_irq_handler(void *opaque, int irq_num, int level)
|
|
{
|
|
PCIDevice *pci_dev = opaque;
|
|
int change;
|
|
|
|
assert(0 <= irq_num && irq_num < PCI_NUM_PINS);
|
|
assert(level == 0 || level == 1);
|
|
change = level - pci_irq_state(pci_dev, irq_num);
|
|
if (!change)
|
|
return;
|
|
|
|
pci_set_irq_state(pci_dev, irq_num, level);
|
|
pci_update_irq_status(pci_dev);
|
|
if (pci_irq_disabled(pci_dev))
|
|
return;
|
|
pci_change_irq_level(pci_dev, irq_num, change);
|
|
}
|
|
|
|
qemu_irq pci_allocate_irq(PCIDevice *pci_dev)
|
|
{
|
|
int intx = pci_intx(pci_dev);
|
|
assert(0 <= intx && intx < PCI_NUM_PINS);
|
|
|
|
return qemu_allocate_irq(pci_irq_handler, pci_dev, intx);
|
|
}
|
|
|
|
void pci_set_irq(PCIDevice *pci_dev, int level)
|
|
{
|
|
int intx = pci_intx(pci_dev);
|
|
pci_irq_handler(pci_dev, intx, level);
|
|
}
|
|
|
|
/* Special hooks used by device assignment */
|
|
void pci_bus_set_route_irq_fn(PCIBus *bus, pci_route_irq_fn route_intx_to_irq)
|
|
{
|
|
assert(pci_bus_is_root(bus));
|
|
bus->route_intx_to_irq = route_intx_to_irq;
|
|
}
|
|
|
|
PCIINTxRoute pci_device_route_intx_to_irq(PCIDevice *dev, int pin)
|
|
{
|
|
PCIBus *bus;
|
|
|
|
do {
|
|
int dev_irq = pin;
|
|
bus = pci_get_bus(dev);
|
|
pin = bus->map_irq(dev, pin);
|
|
trace_pci_route_irq(dev_irq, DEVICE(dev)->canonical_path, pin,
|
|
pci_bus_is_root(bus) ? "root-complex"
|
|
: DEVICE(bus->parent_dev)->canonical_path);
|
|
dev = bus->parent_dev;
|
|
} while (dev);
|
|
|
|
if (!bus->route_intx_to_irq) {
|
|
error_report("PCI: Bug - unimplemented PCI INTx routing (%s)",
|
|
object_get_typename(OBJECT(bus->qbus.parent)));
|
|
return (PCIINTxRoute) { PCI_INTX_DISABLED, -1 };
|
|
}
|
|
|
|
return bus->route_intx_to_irq(bus->irq_opaque, pin);
|
|
}
|
|
|
|
bool pci_intx_route_changed(PCIINTxRoute *old, PCIINTxRoute *new)
|
|
{
|
|
return old->mode != new->mode || old->irq != new->irq;
|
|
}
|
|
|
|
void pci_bus_fire_intx_routing_notifier(PCIBus *bus)
|
|
{
|
|
PCIDevice *dev;
|
|
PCIBus *sec;
|
|
int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(bus->devices); ++i) {
|
|
dev = bus->devices[i];
|
|
if (dev && dev->intx_routing_notifier) {
|
|
dev->intx_routing_notifier(dev);
|
|
}
|
|
}
|
|
|
|
QLIST_FOREACH(sec, &bus->child, sibling) {
|
|
pci_bus_fire_intx_routing_notifier(sec);
|
|
}
|
|
}
|
|
|
|
void pci_device_set_intx_routing_notifier(PCIDevice *dev,
|
|
PCIINTxRoutingNotifier notifier)
|
|
{
|
|
dev->intx_routing_notifier = notifier;
|
|
}
|
|
|
|
/*
|
|
* PCI-to-PCI bridge specification
|
|
* 9.1: Interrupt routing. Table 9-1
|
|
*
|
|
* the PCI Express Base Specification, Revision 2.1
|
|
* 2.2.8.1: INTx interrupt signaling - Rules
|
|
* the Implementation Note
|
|
* Table 2-20
|
|
*/
|
|
/*
|
|
* 0 <= pin <= 3 0 = INTA, 1 = INTB, 2 = INTC, 3 = INTD
|
|
* 0-origin unlike PCI interrupt pin register.
|
|
*/
|
|
int pci_swizzle_map_irq_fn(PCIDevice *pci_dev, int pin)
|
|
{
|
|
return pci_swizzle(PCI_SLOT(pci_dev->devfn), pin);
|
|
}
|
|
|
|
/***********************************************************/
|
|
/* monitor info on PCI */
|
|
|
|
static const pci_class_desc pci_class_descriptions[] =
|
|
{
|
|
{ 0x0001, "VGA controller", "display"},
|
|
{ 0x0100, "SCSI controller", "scsi"},
|
|
{ 0x0101, "IDE controller", "ide"},
|
|
{ 0x0102, "Floppy controller", "fdc"},
|
|
{ 0x0103, "IPI controller", "ipi"},
|
|
{ 0x0104, "RAID controller", "raid"},
|
|
{ 0x0106, "SATA controller"},
|
|
{ 0x0107, "SAS controller"},
|
|
{ 0x0180, "Storage controller"},
|
|
{ 0x0200, "Ethernet controller", "ethernet"},
|
|
{ 0x0201, "Token Ring controller", "token-ring"},
|
|
{ 0x0202, "FDDI controller", "fddi"},
|
|
{ 0x0203, "ATM controller", "atm"},
|
|
{ 0x0280, "Network controller"},
|
|
{ 0x0300, "VGA controller", "display", 0x00ff},
|
|
{ 0x0301, "XGA controller"},
|
|
{ 0x0302, "3D controller"},
|
|
{ 0x0380, "Display controller"},
|
|
{ 0x0400, "Video controller", "video"},
|
|
{ 0x0401, "Audio controller", "sound"},
|
|
{ 0x0402, "Phone"},
|
|
{ 0x0403, "Audio controller", "sound"},
|
|
{ 0x0480, "Multimedia controller"},
|
|
{ 0x0500, "RAM controller", "memory"},
|
|
{ 0x0501, "Flash controller", "flash"},
|
|
{ 0x0580, "Memory controller"},
|
|
{ 0x0600, "Host bridge", "host"},
|
|
{ 0x0601, "ISA bridge", "isa"},
|
|
{ 0x0602, "EISA bridge", "eisa"},
|
|
{ 0x0603, "MC bridge", "mca"},
|
|
{ 0x0604, "PCI bridge", "pci-bridge"},
|
|
{ 0x0605, "PCMCIA bridge", "pcmcia"},
|
|
{ 0x0606, "NUBUS bridge", "nubus"},
|
|
{ 0x0607, "CARDBUS bridge", "cardbus"},
|
|
{ 0x0608, "RACEWAY bridge"},
|
|
{ 0x0680, "Bridge"},
|
|
{ 0x0700, "Serial port", "serial"},
|
|
{ 0x0701, "Parallel port", "parallel"},
|
|
{ 0x0800, "Interrupt controller", "interrupt-controller"},
|
|
{ 0x0801, "DMA controller", "dma-controller"},
|
|
{ 0x0802, "Timer", "timer"},
|
|
{ 0x0803, "RTC", "rtc"},
|
|
{ 0x0900, "Keyboard", "keyboard"},
|
|
{ 0x0901, "Pen", "pen"},
|
|
{ 0x0902, "Mouse", "mouse"},
|
|
{ 0x0A00, "Dock station", "dock", 0x00ff},
|
|
{ 0x0B00, "i386 cpu", "cpu", 0x00ff},
|
|
{ 0x0c00, "Firewire controller", "firewire"},
|
|
{ 0x0c01, "Access bus controller", "access-bus"},
|
|
{ 0x0c02, "SSA controller", "ssa"},
|
|
{ 0x0c03, "USB controller", "usb"},
|
|
{ 0x0c04, "Fibre channel controller", "fibre-channel"},
|
|
{ 0x0c05, "SMBus"},
|
|
{ 0, NULL}
|
|
};
|
|
|
|
void pci_for_each_device_under_bus_reverse(PCIBus *bus,
|
|
pci_bus_dev_fn fn,
|
|
void *opaque)
|
|
{
|
|
PCIDevice *d;
|
|
int devfn;
|
|
|
|
for (devfn = 0; devfn < ARRAY_SIZE(bus->devices); devfn++) {
|
|
d = bus->devices[ARRAY_SIZE(bus->devices) - 1 - devfn];
|
|
if (d) {
|
|
fn(bus, d, opaque);
|
|
}
|
|
}
|
|
}
|
|
|
|
void pci_for_each_device_reverse(PCIBus *bus, int bus_num,
|
|
pci_bus_dev_fn fn, void *opaque)
|
|
{
|
|
bus = pci_find_bus_nr(bus, bus_num);
|
|
|
|
if (bus) {
|
|
pci_for_each_device_under_bus_reverse(bus, fn, opaque);
|
|
}
|
|
}
|
|
|
|
void pci_for_each_device_under_bus(PCIBus *bus,
|
|
pci_bus_dev_fn fn, void *opaque)
|
|
{
|
|
PCIDevice *d;
|
|
int devfn;
|
|
|
|
for(devfn = 0; devfn < ARRAY_SIZE(bus->devices); devfn++) {
|
|
d = bus->devices[devfn];
|
|
if (d) {
|
|
fn(bus, d, opaque);
|
|
}
|
|
}
|
|
}
|
|
|
|
void pci_for_each_device(PCIBus *bus, int bus_num,
|
|
pci_bus_dev_fn fn, void *opaque)
|
|
{
|
|
bus = pci_find_bus_nr(bus, bus_num);
|
|
|
|
if (bus) {
|
|
pci_for_each_device_under_bus(bus, fn, opaque);
|
|
}
|
|
}
|
|
|
|
const pci_class_desc *get_class_desc(int class)
|
|
{
|
|
const pci_class_desc *desc;
|
|
|
|
desc = pci_class_descriptions;
|
|
while (desc->desc && class != desc->class) {
|
|
desc++;
|
|
}
|
|
|
|
return desc;
|
|
}
|
|
|
|
void pci_init_nic_devices(PCIBus *bus, const char *default_model)
|
|
{
|
|
qemu_create_nic_bus_devices(&bus->qbus, TYPE_PCI_DEVICE, default_model,
|
|
"virtio", "virtio-net-pci");
|
|
}
|
|
|
|
bool pci_init_nic_in_slot(PCIBus *rootbus, const char *model,
|
|
const char *alias, const char *devaddr)
|
|
{
|
|
NICInfo *nd = qemu_find_nic_info(model, true, alias);
|
|
int dom, busnr, devfn;
|
|
PCIDevice *pci_dev;
|
|
unsigned slot;
|
|
PCIBus *bus;
|
|
|
|
if (!nd) {
|
|
return false;
|
|
}
|
|
|
|
if (!devaddr || pci_parse_devaddr(devaddr, &dom, &busnr, &slot, NULL) < 0) {
|
|
error_report("Invalid PCI device address %s for device %s",
|
|
devaddr, model);
|
|
exit(1);
|
|
}
|
|
|
|
if (dom != 0) {
|
|
error_report("No support for non-zero PCI domains");
|
|
exit(1);
|
|
}
|
|
|
|
devfn = PCI_DEVFN(slot, 0);
|
|
|
|
bus = pci_find_bus_nr(rootbus, busnr);
|
|
if (!bus) {
|
|
error_report("Invalid PCI device address %s for device %s",
|
|
devaddr, model);
|
|
exit(1);
|
|
}
|
|
|
|
pci_dev = pci_new(devfn, model);
|
|
qdev_set_nic_properties(&pci_dev->qdev, nd);
|
|
pci_realize_and_unref(pci_dev, bus, &error_fatal);
|
|
return true;
|
|
}
|
|
|
|
PCIDevice *pci_vga_init(PCIBus *bus)
|
|
{
|
|
vga_interface_created = true;
|
|
switch (vga_interface_type) {
|
|
case VGA_CIRRUS:
|
|
return pci_create_simple(bus, -1, "cirrus-vga");
|
|
case VGA_QXL:
|
|
return pci_create_simple(bus, -1, "qxl-vga");
|
|
case VGA_STD:
|
|
return pci_create_simple(bus, -1, "VGA");
|
|
case VGA_VMWARE:
|
|
return pci_create_simple(bus, -1, "vmware-svga");
|
|
case VGA_VIRTIO:
|
|
return pci_create_simple(bus, -1, "virtio-vga");
|
|
case VGA_NONE:
|
|
default: /* Other non-PCI types. Checking for unsupported types is already
|
|
done in vl.c. */
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
/* Whether a given bus number is in range of the secondary
|
|
* bus of the given bridge device. */
|
|
static bool pci_secondary_bus_in_range(PCIDevice *dev, int bus_num)
|
|
{
|
|
return !(pci_get_word(dev->config + PCI_BRIDGE_CONTROL) &
|
|
PCI_BRIDGE_CTL_BUS_RESET) /* Don't walk the bus if it's reset. */ &&
|
|
dev->config[PCI_SECONDARY_BUS] <= bus_num &&
|
|
bus_num <= dev->config[PCI_SUBORDINATE_BUS];
|
|
}
|
|
|
|
/* Whether a given bus number is in a range of a root bus */
|
|
static bool pci_root_bus_in_range(PCIBus *bus, int bus_num)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(bus->devices); ++i) {
|
|
PCIDevice *dev = bus->devices[i];
|
|
|
|
if (dev && IS_PCI_BRIDGE(dev)) {
|
|
if (pci_secondary_bus_in_range(dev, bus_num)) {
|
|
return true;
|
|
}
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
PCIBus *pci_find_bus_nr(PCIBus *bus, int bus_num)
|
|
{
|
|
PCIBus *sec;
|
|
|
|
if (!bus) {
|
|
return NULL;
|
|
}
|
|
|
|
if (pci_bus_num(bus) == bus_num) {
|
|
return bus;
|
|
}
|
|
|
|
/* Consider all bus numbers in range for the host pci bridge. */
|
|
if (!pci_bus_is_root(bus) &&
|
|
!pci_secondary_bus_in_range(bus->parent_dev, bus_num)) {
|
|
return NULL;
|
|
}
|
|
|
|
/* try child bus */
|
|
for (; bus; bus = sec) {
|
|
QLIST_FOREACH(sec, &bus->child, sibling) {
|
|
if (pci_bus_num(sec) == bus_num) {
|
|
return sec;
|
|
}
|
|
/* PXB buses assumed to be children of bus 0 */
|
|
if (pci_bus_is_root(sec)) {
|
|
if (pci_root_bus_in_range(sec, bus_num)) {
|
|
break;
|
|
}
|
|
} else {
|
|
if (pci_secondary_bus_in_range(sec->parent_dev, bus_num)) {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
void pci_for_each_bus_depth_first(PCIBus *bus, pci_bus_ret_fn begin,
|
|
pci_bus_fn end, void *parent_state)
|
|
{
|
|
PCIBus *sec;
|
|
void *state;
|
|
|
|
if (!bus) {
|
|
return;
|
|
}
|
|
|
|
if (begin) {
|
|
state = begin(bus, parent_state);
|
|
} else {
|
|
state = parent_state;
|
|
}
|
|
|
|
QLIST_FOREACH(sec, &bus->child, sibling) {
|
|
pci_for_each_bus_depth_first(sec, begin, end, state);
|
|
}
|
|
|
|
if (end) {
|
|
end(bus, state);
|
|
}
|
|
}
|
|
|
|
|
|
PCIDevice *pci_find_device(PCIBus *bus, int bus_num, uint8_t devfn)
|
|
{
|
|
bus = pci_find_bus_nr(bus, bus_num);
|
|
|
|
if (!bus)
|
|
return NULL;
|
|
|
|
return bus->devices[devfn];
|
|
}
|
|
|
|
#define ONBOARD_INDEX_MAX (16 * 1024 - 1)
|
|
|
|
static void pci_qdev_realize(DeviceState *qdev, Error **errp)
|
|
{
|
|
PCIDevice *pci_dev = (PCIDevice *)qdev;
|
|
PCIDeviceClass *pc = PCI_DEVICE_GET_CLASS(pci_dev);
|
|
ObjectClass *klass = OBJECT_CLASS(pc);
|
|
Error *local_err = NULL;
|
|
bool is_default_rom;
|
|
uint16_t class_id;
|
|
|
|
/*
|
|
* capped by systemd (see: udev-builtin-net_id.c)
|
|
* as it's the only known user honor it to avoid users
|
|
* misconfigure QEMU and then wonder why acpi-index doesn't work
|
|
*/
|
|
if (pci_dev->acpi_index > ONBOARD_INDEX_MAX) {
|
|
error_setg(errp, "acpi-index should be less or equal to %u",
|
|
ONBOARD_INDEX_MAX);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* make sure that acpi-index is unique across all present PCI devices
|
|
*/
|
|
if (pci_dev->acpi_index) {
|
|
GSequence *used_indexes = pci_acpi_index_list();
|
|
|
|
if (g_sequence_lookup(used_indexes,
|
|
GINT_TO_POINTER(pci_dev->acpi_index),
|
|
g_cmp_uint32, NULL)) {
|
|
error_setg(errp, "a PCI device with acpi-index = %" PRIu32
|
|
" already exist", pci_dev->acpi_index);
|
|
return;
|
|
}
|
|
g_sequence_insert_sorted(used_indexes,
|
|
GINT_TO_POINTER(pci_dev->acpi_index),
|
|
g_cmp_uint32, NULL);
|
|
}
|
|
|
|
if (pci_dev->romsize != UINT32_MAX && !is_power_of_2(pci_dev->romsize)) {
|
|
error_setg(errp, "ROM size %u is not a power of two", pci_dev->romsize);
|
|
return;
|
|
}
|
|
|
|
/* initialize cap_present for pci_is_express() and pci_config_size(),
|
|
* Note that hybrid PCIs are not set automatically and need to manage
|
|
* QEMU_PCI_CAP_EXPRESS manually */
|
|
if (object_class_dynamic_cast(klass, INTERFACE_PCIE_DEVICE) &&
|
|
!object_class_dynamic_cast(klass, INTERFACE_CONVENTIONAL_PCI_DEVICE)) {
|
|
pci_dev->cap_present |= QEMU_PCI_CAP_EXPRESS;
|
|
}
|
|
|
|
if (object_class_dynamic_cast(klass, INTERFACE_CXL_DEVICE)) {
|
|
pci_dev->cap_present |= QEMU_PCIE_CAP_CXL;
|
|
}
|
|
|
|
pci_dev = do_pci_register_device(pci_dev,
|
|
object_get_typename(OBJECT(qdev)),
|
|
pci_dev->devfn, errp);
|
|
if (pci_dev == NULL)
|
|
return;
|
|
|
|
if (pc->realize) {
|
|
pc->realize(pci_dev, &local_err);
|
|
if (local_err) {
|
|
error_propagate(errp, local_err);
|
|
do_pci_unregister_device(pci_dev);
|
|
return;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* A PCIe Downstream Port that do not have ARI Forwarding enabled must
|
|
* associate only Device 0 with the device attached to the bus
|
|
* representing the Link from the Port (PCIe base spec rev 4.0 ver 0.3,
|
|
* sec 7.3.1).
|
|
* With ARI, PCI_SLOT() can return non-zero value as the traditional
|
|
* 5-bit Device Number and 3-bit Function Number fields in its associated
|
|
* Routing IDs, Requester IDs and Completer IDs are interpreted as a
|
|
* single 8-bit Function Number. Hence, ignore ARI capable devices.
|
|
*/
|
|
if (pci_is_express(pci_dev) &&
|
|
!pcie_find_capability(pci_dev, PCI_EXT_CAP_ID_ARI) &&
|
|
pcie_has_upstream_port(pci_dev) &&
|
|
PCI_SLOT(pci_dev->devfn)) {
|
|
warn_report("PCI: slot %d is not valid for %s,"
|
|
" parent device only allows plugging into slot 0.",
|
|
PCI_SLOT(pci_dev->devfn), pci_dev->name);
|
|
}
|
|
|
|
if (pci_dev->failover_pair_id) {
|
|
if (!pci_bus_is_express(pci_get_bus(pci_dev))) {
|
|
error_setg(errp, "failover primary device must be on "
|
|
"PCIExpress bus");
|
|
pci_qdev_unrealize(DEVICE(pci_dev));
|
|
return;
|
|
}
|
|
class_id = pci_get_word(pci_dev->config + PCI_CLASS_DEVICE);
|
|
if (class_id != PCI_CLASS_NETWORK_ETHERNET) {
|
|
error_setg(errp, "failover primary device is not an "
|
|
"Ethernet device");
|
|
pci_qdev_unrealize(DEVICE(pci_dev));
|
|
return;
|
|
}
|
|
if ((pci_dev->cap_present & QEMU_PCI_CAP_MULTIFUNCTION)
|
|
|| (PCI_FUNC(pci_dev->devfn) != 0)) {
|
|
error_setg(errp, "failover: primary device must be in its own "
|
|
"PCI slot");
|
|
pci_qdev_unrealize(DEVICE(pci_dev));
|
|
return;
|
|
}
|
|
qdev->allow_unplug_during_migration = true;
|
|
}
|
|
|
|
/* rom loading */
|
|
is_default_rom = false;
|
|
if (pci_dev->romfile == NULL && pc->romfile != NULL) {
|
|
pci_dev->romfile = g_strdup(pc->romfile);
|
|
is_default_rom = true;
|
|
}
|
|
|
|
pci_add_option_rom(pci_dev, is_default_rom, &local_err);
|
|
if (local_err) {
|
|
error_propagate(errp, local_err);
|
|
pci_qdev_unrealize(DEVICE(pci_dev));
|
|
return;
|
|
}
|
|
|
|
pci_set_power(pci_dev, true);
|
|
|
|
pci_dev->msi_trigger = pci_msi_trigger;
|
|
}
|
|
|
|
static PCIDevice *pci_new_internal(int devfn, bool multifunction,
|
|
const char *name)
|
|
{
|
|
DeviceState *dev;
|
|
|
|
dev = qdev_new(name);
|
|
qdev_prop_set_int32(dev, "addr", devfn);
|
|
qdev_prop_set_bit(dev, "multifunction", multifunction);
|
|
return PCI_DEVICE(dev);
|
|
}
|
|
|
|
PCIDevice *pci_new_multifunction(int devfn, const char *name)
|
|
{
|
|
return pci_new_internal(devfn, true, name);
|
|
}
|
|
|
|
PCIDevice *pci_new(int devfn, const char *name)
|
|
{
|
|
return pci_new_internal(devfn, false, name);
|
|
}
|
|
|
|
bool pci_realize_and_unref(PCIDevice *dev, PCIBus *bus, Error **errp)
|
|
{
|
|
return qdev_realize_and_unref(&dev->qdev, &bus->qbus, errp);
|
|
}
|
|
|
|
PCIDevice *pci_create_simple_multifunction(PCIBus *bus, int devfn,
|
|
const char *name)
|
|
{
|
|
PCIDevice *dev = pci_new_multifunction(devfn, name);
|
|
pci_realize_and_unref(dev, bus, &error_fatal);
|
|
return dev;
|
|
}
|
|
|
|
PCIDevice *pci_create_simple(PCIBus *bus, int devfn, const char *name)
|
|
{
|
|
PCIDevice *dev = pci_new(devfn, name);
|
|
pci_realize_and_unref(dev, bus, &error_fatal);
|
|
return dev;
|
|
}
|
|
|
|
static uint8_t pci_find_space(PCIDevice *pdev, uint8_t size)
|
|
{
|
|
int offset = PCI_CONFIG_HEADER_SIZE;
|
|
int i;
|
|
for (i = PCI_CONFIG_HEADER_SIZE; i < PCI_CONFIG_SPACE_SIZE; ++i) {
|
|
if (pdev->used[i])
|
|
offset = i + 1;
|
|
else if (i - offset + 1 == size)
|
|
return offset;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static uint8_t pci_find_capability_list(PCIDevice *pdev, uint8_t cap_id,
|
|
uint8_t *prev_p)
|
|
{
|
|
uint8_t next, prev;
|
|
|
|
if (!(pdev->config[PCI_STATUS] & PCI_STATUS_CAP_LIST))
|
|
return 0;
|
|
|
|
for (prev = PCI_CAPABILITY_LIST; (next = pdev->config[prev]);
|
|
prev = next + PCI_CAP_LIST_NEXT)
|
|
if (pdev->config[next + PCI_CAP_LIST_ID] == cap_id)
|
|
break;
|
|
|
|
if (prev_p)
|
|
*prev_p = prev;
|
|
return next;
|
|
}
|
|
|
|
static uint8_t pci_find_capability_at_offset(PCIDevice *pdev, uint8_t offset)
|
|
{
|
|
uint8_t next, prev, found = 0;
|
|
|
|
if (!(pdev->used[offset])) {
|
|
return 0;
|
|
}
|
|
|
|
assert(pdev->config[PCI_STATUS] & PCI_STATUS_CAP_LIST);
|
|
|
|
for (prev = PCI_CAPABILITY_LIST; (next = pdev->config[prev]);
|
|
prev = next + PCI_CAP_LIST_NEXT) {
|
|
if (next <= offset && next > found) {
|
|
found = next;
|
|
}
|
|
}
|
|
return found;
|
|
}
|
|
|
|
/* Patch the PCI vendor and device ids in a PCI rom image if necessary.
|
|
This is needed for an option rom which is used for more than one device. */
|
|
static void pci_patch_ids(PCIDevice *pdev, uint8_t *ptr, uint32_t size)
|
|
{
|
|
uint16_t vendor_id;
|
|
uint16_t device_id;
|
|
uint16_t rom_vendor_id;
|
|
uint16_t rom_device_id;
|
|
uint16_t rom_magic;
|
|
uint16_t pcir_offset;
|
|
uint8_t checksum;
|
|
|
|
/* Words in rom data are little endian (like in PCI configuration),
|
|
so they can be read / written with pci_get_word / pci_set_word. */
|
|
|
|
/* Only a valid rom will be patched. */
|
|
rom_magic = pci_get_word(ptr);
|
|
if (rom_magic != 0xaa55) {
|
|
PCI_DPRINTF("Bad ROM magic %04x\n", rom_magic);
|
|
return;
|
|
}
|
|
pcir_offset = pci_get_word(ptr + 0x18);
|
|
if (pcir_offset + 8 >= size || memcmp(ptr + pcir_offset, "PCIR", 4)) {
|
|
PCI_DPRINTF("Bad PCIR offset 0x%x or signature\n", pcir_offset);
|
|
return;
|
|
}
|
|
|
|
vendor_id = pci_get_word(pdev->config + PCI_VENDOR_ID);
|
|
device_id = pci_get_word(pdev->config + PCI_DEVICE_ID);
|
|
rom_vendor_id = pci_get_word(ptr + pcir_offset + 4);
|
|
rom_device_id = pci_get_word(ptr + pcir_offset + 6);
|
|
|
|
PCI_DPRINTF("%s: ROM id %04x%04x / PCI id %04x%04x\n", pdev->romfile,
|
|
vendor_id, device_id, rom_vendor_id, rom_device_id);
|
|
|
|
checksum = ptr[6];
|
|
|
|
if (vendor_id != rom_vendor_id) {
|
|
/* Patch vendor id and checksum (at offset 6 for etherboot roms). */
|
|
checksum += (uint8_t)rom_vendor_id + (uint8_t)(rom_vendor_id >> 8);
|
|
checksum -= (uint8_t)vendor_id + (uint8_t)(vendor_id >> 8);
|
|
PCI_DPRINTF("ROM checksum %02x / %02x\n", ptr[6], checksum);
|
|
ptr[6] = checksum;
|
|
pci_set_word(ptr + pcir_offset + 4, vendor_id);
|
|
}
|
|
|
|
if (device_id != rom_device_id) {
|
|
/* Patch device id and checksum (at offset 6 for etherboot roms). */
|
|
checksum += (uint8_t)rom_device_id + (uint8_t)(rom_device_id >> 8);
|
|
checksum -= (uint8_t)device_id + (uint8_t)(device_id >> 8);
|
|
PCI_DPRINTF("ROM checksum %02x / %02x\n", ptr[6], checksum);
|
|
ptr[6] = checksum;
|
|
pci_set_word(ptr + pcir_offset + 6, device_id);
|
|
}
|
|
}
|
|
|
|
/* Add an option rom for the device */
|
|
static void pci_add_option_rom(PCIDevice *pdev, bool is_default_rom,
|
|
Error **errp)
|
|
{
|
|
int64_t size = 0;
|
|
g_autofree char *path = NULL;
|
|
char name[32];
|
|
const VMStateDescription *vmsd;
|
|
|
|
/*
|
|
* In case of incoming migration ROM will come with migration stream, no
|
|
* reason to load the file. Neither we want to fail if local ROM file
|
|
* mismatches with specified romsize.
|
|
*/
|
|
bool load_file = !runstate_check(RUN_STATE_INMIGRATE);
|
|
|
|
if (!pdev->romfile || !strlen(pdev->romfile)) {
|
|
return;
|
|
}
|
|
|
|
if (!pdev->rom_bar) {
|
|
/*
|
|
* Load rom via fw_cfg instead of creating a rom bar,
|
|
* for 0.11 compatibility.
|
|
*/
|
|
int class = pci_get_word(pdev->config + PCI_CLASS_DEVICE);
|
|
|
|
/*
|
|
* Hot-plugged devices can't use the option ROM
|
|
* if the rom bar is disabled.
|
|
*/
|
|
if (DEVICE(pdev)->hotplugged) {
|
|
error_setg(errp, "Hot-plugged device without ROM bar"
|
|
" can't have an option ROM");
|
|
return;
|
|
}
|
|
|
|
if (class == 0x0300) {
|
|
rom_add_vga(pdev->romfile);
|
|
} else {
|
|
rom_add_option(pdev->romfile, -1);
|
|
}
|
|
return;
|
|
}
|
|
|
|
if (load_file || pdev->romsize == UINT32_MAX) {
|
|
path = qemu_find_file(QEMU_FILE_TYPE_BIOS, pdev->romfile);
|
|
if (path == NULL) {
|
|
path = g_strdup(pdev->romfile);
|
|
}
|
|
|
|
size = get_image_size(path);
|
|
if (size < 0) {
|
|
error_setg(errp, "failed to find romfile \"%s\"", pdev->romfile);
|
|
return;
|
|
} else if (size == 0) {
|
|
error_setg(errp, "romfile \"%s\" is empty", pdev->romfile);
|
|
return;
|
|
} else if (size > 2 * GiB) {
|
|
error_setg(errp,
|
|
"romfile \"%s\" too large (size cannot exceed 2 GiB)",
|
|
pdev->romfile);
|
|
return;
|
|
}
|
|
if (pdev->romsize != UINT_MAX) {
|
|
if (size > pdev->romsize) {
|
|
error_setg(errp, "romfile \"%s\" (%u bytes) "
|
|
"is too large for ROM size %u",
|
|
pdev->romfile, (uint32_t)size, pdev->romsize);
|
|
return;
|
|
}
|
|
} else {
|
|
pdev->romsize = pow2ceil(size);
|
|
}
|
|
}
|
|
|
|
vmsd = qdev_get_vmsd(DEVICE(pdev));
|
|
snprintf(name, sizeof(name), "%s.rom",
|
|
vmsd ? vmsd->name : object_get_typename(OBJECT(pdev)));
|
|
|
|
pdev->has_rom = true;
|
|
memory_region_init_rom(&pdev->rom, OBJECT(pdev), name, pdev->romsize,
|
|
&error_fatal);
|
|
|
|
if (load_file) {
|
|
void *ptr = memory_region_get_ram_ptr(&pdev->rom);
|
|
|
|
if (load_image_size(path, ptr, size) < 0) {
|
|
error_setg(errp, "failed to load romfile \"%s\"", pdev->romfile);
|
|
return;
|
|
}
|
|
|
|
if (is_default_rom) {
|
|
/* Only the default rom images will be patched (if needed). */
|
|
pci_patch_ids(pdev, ptr, size);
|
|
}
|
|
}
|
|
|
|
pci_register_bar(pdev, PCI_ROM_SLOT, 0, &pdev->rom);
|
|
}
|
|
|
|
static void pci_del_option_rom(PCIDevice *pdev)
|
|
{
|
|
if (!pdev->has_rom)
|
|
return;
|
|
|
|
vmstate_unregister_ram(&pdev->rom, &pdev->qdev);
|
|
pdev->has_rom = false;
|
|
}
|
|
|
|
/*
|
|
* On success, pci_add_capability() returns a positive value
|
|
* that the offset of the pci capability.
|
|
* On failure, it sets an error and returns a negative error
|
|
* code.
|
|
*/
|
|
int pci_add_capability(PCIDevice *pdev, uint8_t cap_id,
|
|
uint8_t offset, uint8_t size,
|
|
Error **errp)
|
|
{
|
|
uint8_t *config;
|
|
int i, overlapping_cap;
|
|
|
|
if (!offset) {
|
|
offset = pci_find_space(pdev, size);
|
|
/* out of PCI config space is programming error */
|
|
assert(offset);
|
|
} else {
|
|
/* Verify that capabilities don't overlap. Note: device assignment
|
|
* depends on this check to verify that the device is not broken.
|
|
* Should never trigger for emulated devices, but it's helpful
|
|
* for debugging these. */
|
|
for (i = offset; i < offset + size; i++) {
|
|
overlapping_cap = pci_find_capability_at_offset(pdev, i);
|
|
if (overlapping_cap) {
|
|
error_setg(errp, "%s:%02x:%02x.%x "
|
|
"Attempt to add PCI capability %x at offset "
|
|
"%x overlaps existing capability %x at offset %x",
|
|
pci_root_bus_path(pdev), pci_dev_bus_num(pdev),
|
|
PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn),
|
|
cap_id, offset, overlapping_cap, i);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
}
|
|
|
|
config = pdev->config + offset;
|
|
config[PCI_CAP_LIST_ID] = cap_id;
|
|
config[PCI_CAP_LIST_NEXT] = pdev->config[PCI_CAPABILITY_LIST];
|
|
pdev->config[PCI_CAPABILITY_LIST] = offset;
|
|
pdev->config[PCI_STATUS] |= PCI_STATUS_CAP_LIST;
|
|
memset(pdev->used + offset, 0xFF, QEMU_ALIGN_UP(size, 4));
|
|
/* Make capability read-only by default */
|
|
memset(pdev->wmask + offset, 0, size);
|
|
/* Check capability by default */
|
|
memset(pdev->cmask + offset, 0xFF, size);
|
|
return offset;
|
|
}
|
|
|
|
/* Unlink capability from the pci config space. */
|
|
void pci_del_capability(PCIDevice *pdev, uint8_t cap_id, uint8_t size)
|
|
{
|
|
uint8_t prev, offset = pci_find_capability_list(pdev, cap_id, &prev);
|
|
if (!offset)
|
|
return;
|
|
pdev->config[prev] = pdev->config[offset + PCI_CAP_LIST_NEXT];
|
|
/* Make capability writable again */
|
|
memset(pdev->wmask + offset, 0xff, size);
|
|
memset(pdev->w1cmask + offset, 0, size);
|
|
/* Clear cmask as device-specific registers can't be checked */
|
|
memset(pdev->cmask + offset, 0, size);
|
|
memset(pdev->used + offset, 0, QEMU_ALIGN_UP(size, 4));
|
|
|
|
if (!pdev->config[PCI_CAPABILITY_LIST])
|
|
pdev->config[PCI_STATUS] &= ~PCI_STATUS_CAP_LIST;
|
|
}
|
|
|
|
uint8_t pci_find_capability(PCIDevice *pdev, uint8_t cap_id)
|
|
{
|
|
return pci_find_capability_list(pdev, cap_id, NULL);
|
|
}
|
|
|
|
static char *pci_dev_fw_name(DeviceState *dev, char *buf, int len)
|
|
{
|
|
PCIDevice *d = (PCIDevice *)dev;
|
|
const char *name = NULL;
|
|
const pci_class_desc *desc = pci_class_descriptions;
|
|
int class = pci_get_word(d->config + PCI_CLASS_DEVICE);
|
|
|
|
while (desc->desc &&
|
|
(class & ~desc->fw_ign_bits) !=
|
|
(desc->class & ~desc->fw_ign_bits)) {
|
|
desc++;
|
|
}
|
|
|
|
if (desc->desc) {
|
|
name = desc->fw_name;
|
|
}
|
|
|
|
if (name) {
|
|
pstrcpy(buf, len, name);
|
|
} else {
|
|
snprintf(buf, len, "pci%04x,%04x",
|
|
pci_get_word(d->config + PCI_VENDOR_ID),
|
|
pci_get_word(d->config + PCI_DEVICE_ID));
|
|
}
|
|
|
|
return buf;
|
|
}
|
|
|
|
static char *pcibus_get_fw_dev_path(DeviceState *dev)
|
|
{
|
|
PCIDevice *d = (PCIDevice *)dev;
|
|
char name[33];
|
|
int has_func = !!PCI_FUNC(d->devfn);
|
|
|
|
return g_strdup_printf("%s@%x%s%.*x",
|
|
pci_dev_fw_name(dev, name, sizeof(name)),
|
|
PCI_SLOT(d->devfn),
|
|
has_func ? "," : "",
|
|
has_func,
|
|
PCI_FUNC(d->devfn));
|
|
}
|
|
|
|
static char *pcibus_get_dev_path(DeviceState *dev)
|
|
{
|
|
PCIDevice *d = container_of(dev, PCIDevice, qdev);
|
|
PCIDevice *t;
|
|
int slot_depth;
|
|
/* Path format: Domain:00:Slot.Function:Slot.Function....:Slot.Function.
|
|
* 00 is added here to make this format compatible with
|
|
* domain:Bus:Slot.Func for systems without nested PCI bridges.
|
|
* Slot.Function list specifies the slot and function numbers for all
|
|
* devices on the path from root to the specific device. */
|
|
const char *root_bus_path;
|
|
int root_bus_len;
|
|
char slot[] = ":SS.F";
|
|
int slot_len = sizeof slot - 1 /* For '\0' */;
|
|
int path_len;
|
|
char *path, *p;
|
|
int s;
|
|
|
|
root_bus_path = pci_root_bus_path(d);
|
|
root_bus_len = strlen(root_bus_path);
|
|
|
|
/* Calculate # of slots on path between device and root. */;
|
|
slot_depth = 0;
|
|
for (t = d; t; t = pci_get_bus(t)->parent_dev) {
|
|
++slot_depth;
|
|
}
|
|
|
|
path_len = root_bus_len + slot_len * slot_depth;
|
|
|
|
/* Allocate memory, fill in the terminating null byte. */
|
|
path = g_malloc(path_len + 1 /* For '\0' */);
|
|
path[path_len] = '\0';
|
|
|
|
memcpy(path, root_bus_path, root_bus_len);
|
|
|
|
/* Fill in slot numbers. We walk up from device to root, so need to print
|
|
* them in the reverse order, last to first. */
|
|
p = path + path_len;
|
|
for (t = d; t; t = pci_get_bus(t)->parent_dev) {
|
|
p -= slot_len;
|
|
s = snprintf(slot, sizeof slot, ":%02x.%x",
|
|
PCI_SLOT(t->devfn), PCI_FUNC(t->devfn));
|
|
assert(s == slot_len);
|
|
memcpy(p, slot, slot_len);
|
|
}
|
|
|
|
return path;
|
|
}
|
|
|
|
static int pci_qdev_find_recursive(PCIBus *bus,
|
|
const char *id, PCIDevice **pdev)
|
|
{
|
|
DeviceState *qdev = qdev_find_recursive(&bus->qbus, id);
|
|
if (!qdev) {
|
|
return -ENODEV;
|
|
}
|
|
|
|
/* roughly check if given qdev is pci device */
|
|
if (object_dynamic_cast(OBJECT(qdev), TYPE_PCI_DEVICE)) {
|
|
*pdev = PCI_DEVICE(qdev);
|
|
return 0;
|
|
}
|
|
return -EINVAL;
|
|
}
|
|
|
|
int pci_qdev_find_device(const char *id, PCIDevice **pdev)
|
|
{
|
|
PCIHostState *host_bridge;
|
|
int rc = -ENODEV;
|
|
|
|
QLIST_FOREACH(host_bridge, &pci_host_bridges, next) {
|
|
int tmp = pci_qdev_find_recursive(host_bridge->bus, id, pdev);
|
|
if (!tmp) {
|
|
rc = 0;
|
|
break;
|
|
}
|
|
if (tmp != -ENODEV) {
|
|
rc = tmp;
|
|
}
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
MemoryRegion *pci_address_space(PCIDevice *dev)
|
|
{
|
|
return pci_get_bus(dev)->address_space_mem;
|
|
}
|
|
|
|
MemoryRegion *pci_address_space_io(PCIDevice *dev)
|
|
{
|
|
return pci_get_bus(dev)->address_space_io;
|
|
}
|
|
|
|
static void pci_device_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *k = DEVICE_CLASS(klass);
|
|
|
|
k->realize = pci_qdev_realize;
|
|
k->unrealize = pci_qdev_unrealize;
|
|
k->bus_type = TYPE_PCI_BUS;
|
|
device_class_set_props(k, pci_props);
|
|
object_class_property_set_description(
|
|
klass, "x-max-bounce-buffer-size",
|
|
"Maximum buffer size allocated for bounce buffers used for mapped "
|
|
"access to indirect DMA memory");
|
|
}
|
|
|
|
static void pci_device_class_base_init(ObjectClass *klass, void *data)
|
|
{
|
|
if (!object_class_is_abstract(klass)) {
|
|
ObjectClass *conventional =
|
|
object_class_dynamic_cast(klass, INTERFACE_CONVENTIONAL_PCI_DEVICE);
|
|
ObjectClass *pcie =
|
|
object_class_dynamic_cast(klass, INTERFACE_PCIE_DEVICE);
|
|
ObjectClass *cxl =
|
|
object_class_dynamic_cast(klass, INTERFACE_CXL_DEVICE);
|
|
assert(conventional || pcie || cxl);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Get IOMMU root bus, aliased bus and devfn of a PCI device
|
|
*
|
|
* IOMMU root bus is needed by all call sites to call into iommu_ops.
|
|
* For call sites which don't need aliased BDF, passing NULL to
|
|
* aliased_[bus|devfn] is allowed.
|
|
*
|
|
* @piommu_bus: return root #PCIBus backed by an IOMMU for the PCI device.
|
|
*
|
|
* @aliased_bus: return aliased #PCIBus of the PCI device, optional.
|
|
*
|
|
* @aliased_devfn: return aliased devfn of the PCI device, optional.
|
|
*/
|
|
static void pci_device_get_iommu_bus_devfn(PCIDevice *dev,
|
|
PCIBus **piommu_bus,
|
|
PCIBus **aliased_bus,
|
|
int *aliased_devfn)
|
|
{
|
|
PCIBus *bus = pci_get_bus(dev);
|
|
PCIBus *iommu_bus = bus;
|
|
int devfn = dev->devfn;
|
|
|
|
while (iommu_bus && !iommu_bus->iommu_ops && iommu_bus->parent_dev) {
|
|
PCIBus *parent_bus = pci_get_bus(iommu_bus->parent_dev);
|
|
|
|
/*
|
|
* The requester ID of the provided device may be aliased, as seen from
|
|
* the IOMMU, due to topology limitations. The IOMMU relies on a
|
|
* requester ID to provide a unique AddressSpace for devices, but
|
|
* conventional PCI buses pre-date such concepts. Instead, the PCIe-
|
|
* to-PCI bridge creates and accepts transactions on behalf of down-
|
|
* stream devices. When doing so, all downstream devices are masked
|
|
* (aliased) behind a single requester ID. The requester ID used
|
|
* depends on the format of the bridge devices. Proper PCIe-to-PCI
|
|
* bridges, with a PCIe capability indicating such, follow the
|
|
* guidelines of chapter 2.3 of the PCIe-to-PCI/X bridge specification,
|
|
* where the bridge uses the seconary bus as the bridge portion of the
|
|
* requester ID and devfn of 00.0. For other bridges, typically those
|
|
* found on the root complex such as the dmi-to-pci-bridge, we follow
|
|
* the convention of typical bare-metal hardware, which uses the
|
|
* requester ID of the bridge itself. There are device specific
|
|
* exceptions to these rules, but these are the defaults that the
|
|
* Linux kernel uses when determining DMA aliases itself and believed
|
|
* to be true for the bare metal equivalents of the devices emulated
|
|
* in QEMU.
|
|
*/
|
|
if (!pci_bus_is_express(iommu_bus)) {
|
|
PCIDevice *parent = iommu_bus->parent_dev;
|
|
|
|
if (pci_is_express(parent) &&
|
|
pcie_cap_get_type(parent) == PCI_EXP_TYPE_PCI_BRIDGE) {
|
|
devfn = PCI_DEVFN(0, 0);
|
|
bus = iommu_bus;
|
|
} else {
|
|
devfn = parent->devfn;
|
|
bus = parent_bus;
|
|
}
|
|
}
|
|
|
|
iommu_bus = parent_bus;
|
|
}
|
|
|
|
assert(0 <= devfn && devfn < PCI_DEVFN_MAX);
|
|
assert(iommu_bus);
|
|
|
|
if (pci_bus_bypass_iommu(bus) || !iommu_bus->iommu_ops) {
|
|
iommu_bus = NULL;
|
|
}
|
|
|
|
*piommu_bus = iommu_bus;
|
|
|
|
if (aliased_bus) {
|
|
*aliased_bus = bus;
|
|
}
|
|
|
|
if (aliased_devfn) {
|
|
*aliased_devfn = devfn;
|
|
}
|
|
}
|
|
|
|
AddressSpace *pci_device_iommu_address_space(PCIDevice *dev)
|
|
{
|
|
PCIBus *bus;
|
|
PCIBus *iommu_bus;
|
|
int devfn;
|
|
|
|
pci_device_get_iommu_bus_devfn(dev, &iommu_bus, &bus, &devfn);
|
|
if (iommu_bus) {
|
|
return iommu_bus->iommu_ops->get_address_space(bus,
|
|
iommu_bus->iommu_opaque, devfn);
|
|
}
|
|
return &address_space_memory;
|
|
}
|
|
|
|
bool pci_device_set_iommu_device(PCIDevice *dev, HostIOMMUDevice *hiod,
|
|
Error **errp)
|
|
{
|
|
PCIBus *iommu_bus, *aliased_bus;
|
|
int aliased_devfn;
|
|
|
|
/* set_iommu_device requires device's direct BDF instead of aliased BDF */
|
|
pci_device_get_iommu_bus_devfn(dev, &iommu_bus,
|
|
&aliased_bus, &aliased_devfn);
|
|
if (iommu_bus && iommu_bus->iommu_ops->set_iommu_device) {
|
|
hiod->aliased_bus = aliased_bus;
|
|
hiod->aliased_devfn = aliased_devfn;
|
|
return iommu_bus->iommu_ops->set_iommu_device(pci_get_bus(dev),
|
|
iommu_bus->iommu_opaque,
|
|
dev->devfn, hiod, errp);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
void pci_device_unset_iommu_device(PCIDevice *dev)
|
|
{
|
|
PCIBus *iommu_bus;
|
|
|
|
pci_device_get_iommu_bus_devfn(dev, &iommu_bus, NULL, NULL);
|
|
if (iommu_bus && iommu_bus->iommu_ops->unset_iommu_device) {
|
|
return iommu_bus->iommu_ops->unset_iommu_device(pci_get_bus(dev),
|
|
iommu_bus->iommu_opaque,
|
|
dev->devfn);
|
|
}
|
|
}
|
|
|
|
void pci_setup_iommu(PCIBus *bus, const PCIIOMMUOps *ops, void *opaque)
|
|
{
|
|
/*
|
|
* If called, pci_setup_iommu() should provide a minimum set of
|
|
* useful callbacks for the bus.
|
|
*/
|
|
assert(ops);
|
|
assert(ops->get_address_space);
|
|
|
|
bus->iommu_ops = ops;
|
|
bus->iommu_opaque = opaque;
|
|
}
|
|
|
|
static void pci_dev_get_w64(PCIBus *b, PCIDevice *dev, void *opaque)
|
|
{
|
|
Range *range = opaque;
|
|
uint16_t cmd = pci_get_word(dev->config + PCI_COMMAND);
|
|
int i;
|
|
|
|
if (!(cmd & PCI_COMMAND_MEMORY)) {
|
|
return;
|
|
}
|
|
|
|
if (IS_PCI_BRIDGE(dev)) {
|
|
pcibus_t base = pci_bridge_get_base(dev, PCI_BASE_ADDRESS_MEM_PREFETCH);
|
|
pcibus_t limit = pci_bridge_get_limit(dev, PCI_BASE_ADDRESS_MEM_PREFETCH);
|
|
|
|
base = MAX(base, 0x1ULL << 32);
|
|
|
|
if (limit >= base) {
|
|
Range pref_range;
|
|
range_set_bounds(&pref_range, base, limit);
|
|
range_extend(range, &pref_range);
|
|
}
|
|
}
|
|
for (i = 0; i < PCI_NUM_REGIONS; ++i) {
|
|
PCIIORegion *r = &dev->io_regions[i];
|
|
pcibus_t lob, upb;
|
|
Range region_range;
|
|
|
|
if (!r->size ||
|
|
(r->type & PCI_BASE_ADDRESS_SPACE_IO) ||
|
|
!(r->type & PCI_BASE_ADDRESS_MEM_TYPE_64)) {
|
|
continue;
|
|
}
|
|
|
|
lob = pci_bar_address(dev, i, r->type, r->size);
|
|
upb = lob + r->size - 1;
|
|
if (lob == PCI_BAR_UNMAPPED) {
|
|
continue;
|
|
}
|
|
|
|
lob = MAX(lob, 0x1ULL << 32);
|
|
|
|
if (upb >= lob) {
|
|
range_set_bounds(®ion_range, lob, upb);
|
|
range_extend(range, ®ion_range);
|
|
}
|
|
}
|
|
}
|
|
|
|
void pci_bus_get_w64_range(PCIBus *bus, Range *range)
|
|
{
|
|
range_make_empty(range);
|
|
pci_for_each_device_under_bus(bus, pci_dev_get_w64, range);
|
|
}
|
|
|
|
static bool pcie_has_upstream_port(PCIDevice *dev)
|
|
{
|
|
PCIDevice *parent_dev = pci_bridge_get_device(pci_get_bus(dev));
|
|
|
|
/* Device associated with an upstream port.
|
|
* As there are several types of these, it's easier to check the
|
|
* parent device: upstream ports are always connected to
|
|
* root or downstream ports.
|
|
*/
|
|
return parent_dev &&
|
|
pci_is_express(parent_dev) &&
|
|
parent_dev->exp.exp_cap &&
|
|
(pcie_cap_get_type(parent_dev) == PCI_EXP_TYPE_ROOT_PORT ||
|
|
pcie_cap_get_type(parent_dev) == PCI_EXP_TYPE_DOWNSTREAM);
|
|
}
|
|
|
|
PCIDevice *pci_get_function_0(PCIDevice *pci_dev)
|
|
{
|
|
PCIBus *bus = pci_get_bus(pci_dev);
|
|
|
|
if(pcie_has_upstream_port(pci_dev)) {
|
|
/* With an upstream PCIe port, we only support 1 device at slot 0 */
|
|
return bus->devices[0];
|
|
} else {
|
|
/* Other bus types might support multiple devices at slots 0-31 */
|
|
return bus->devices[PCI_DEVFN(PCI_SLOT(pci_dev->devfn), 0)];
|
|
}
|
|
}
|
|
|
|
MSIMessage pci_get_msi_message(PCIDevice *dev, int vector)
|
|
{
|
|
MSIMessage msg;
|
|
if (msix_enabled(dev)) {
|
|
msg = msix_get_message(dev, vector);
|
|
} else if (msi_enabled(dev)) {
|
|
msg = msi_get_message(dev, vector);
|
|
} else {
|
|
/* Should never happen */
|
|
error_report("%s: unknown interrupt type", __func__);
|
|
abort();
|
|
}
|
|
return msg;
|
|
}
|
|
|
|
void pci_set_power(PCIDevice *d, bool state)
|
|
{
|
|
if (d->has_power == state) {
|
|
return;
|
|
}
|
|
|
|
d->has_power = state;
|
|
pci_update_mappings(d);
|
|
memory_region_set_enabled(&d->bus_master_enable_region,
|
|
(pci_get_word(d->config + PCI_COMMAND)
|
|
& PCI_COMMAND_MASTER) && d->has_power);
|
|
if (!d->has_power) {
|
|
pci_device_reset(d);
|
|
}
|
|
}
|
|
|
|
static const TypeInfo pci_device_type_info = {
|
|
.name = TYPE_PCI_DEVICE,
|
|
.parent = TYPE_DEVICE,
|
|
.instance_size = sizeof(PCIDevice),
|
|
.abstract = true,
|
|
.class_size = sizeof(PCIDeviceClass),
|
|
.class_init = pci_device_class_init,
|
|
.class_base_init = pci_device_class_base_init,
|
|
};
|
|
|
|
static void pci_register_types(void)
|
|
{
|
|
type_register_static(&pci_bus_info);
|
|
type_register_static(&pcie_bus_info);
|
|
type_register_static(&cxl_bus_info);
|
|
type_register_static(&conventional_pci_interface_info);
|
|
type_register_static(&cxl_interface_info);
|
|
type_register_static(&pcie_interface_info);
|
|
type_register_static(&pci_device_type_info);
|
|
}
|
|
|
|
type_init(pci_register_types)
|