qemu

FORK: QEMU emulator
git clone https://git.neptards.moe/neptards/qemu.git
Log | Files | Refs | Submodules | LICENSE

smmu-common.c (17468B)


      1 /*
      2  * Copyright (C) 2014-2016 Broadcom Corporation
      3  * Copyright (c) 2017 Red Hat, Inc.
      4  * Written by Prem Mallappa, Eric Auger
      5  *
      6  * This program is free software; you can redistribute it and/or modify
      7  * it under the terms of the GNU General Public License version 2 as
      8  * published by the Free Software Foundation.
      9  *
     10  * This program is distributed in the hope that it will be useful,
     11  * but WITHOUT ANY WARRANTY; without even the implied warranty of
     12  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
     13  * GNU General Public License for more details.
     14  *
     15  * Author: Prem Mallappa <pmallapp@broadcom.com>
     16  *
     17  */
     18 
     19 #include "qemu/osdep.h"
     20 #include "trace.h"
     21 #include "exec/target_page.h"
     22 #include "hw/core/cpu.h"
     23 #include "hw/qdev-properties.h"
     24 #include "qapi/error.h"
     25 #include "qemu/jhash.h"
     26 #include "qemu/module.h"
     27 
     28 #include "qemu/error-report.h"
     29 #include "hw/arm/smmu-common.h"
     30 #include "smmu-internal.h"
     31 
     32 /* IOTLB Management */
     33 
     34 static guint smmu_iotlb_key_hash(gconstpointer v)
     35 {
     36     SMMUIOTLBKey *key = (SMMUIOTLBKey *)v;
     37     uint32_t a, b, c;
     38 
     39     /* Jenkins hash */
     40     a = b = c = JHASH_INITVAL + sizeof(*key);
     41     a += key->asid + key->level + key->tg;
     42     b += extract64(key->iova, 0, 32);
     43     c += extract64(key->iova, 32, 32);
     44 
     45     __jhash_mix(a, b, c);
     46     __jhash_final(a, b, c);
     47 
     48     return c;
     49 }
     50 
     51 static gboolean smmu_iotlb_key_equal(gconstpointer v1, gconstpointer v2)
     52 {
     53     SMMUIOTLBKey *k1 = (SMMUIOTLBKey *)v1, *k2 = (SMMUIOTLBKey *)v2;
     54 
     55     return (k1->asid == k2->asid) && (k1->iova == k2->iova) &&
     56            (k1->level == k2->level) && (k1->tg == k2->tg);
     57 }
     58 
     59 SMMUIOTLBKey smmu_get_iotlb_key(uint16_t asid, uint64_t iova,
     60                                 uint8_t tg, uint8_t level)
     61 {
     62     SMMUIOTLBKey key = {.asid = asid, .iova = iova, .tg = tg, .level = level};
     63 
     64     return key;
     65 }
     66 
     67 SMMUTLBEntry *smmu_iotlb_lookup(SMMUState *bs, SMMUTransCfg *cfg,
     68                                 SMMUTransTableInfo *tt, hwaddr iova)
     69 {
     70     uint8_t tg = (tt->granule_sz - 10) / 2;
     71     uint8_t inputsize = 64 - tt->tsz;
     72     uint8_t stride = tt->granule_sz - 3;
     73     uint8_t level = 4 - (inputsize - 4) / stride;
     74     SMMUTLBEntry *entry = NULL;
     75 
     76     while (level <= 3) {
     77         uint64_t subpage_size = 1ULL << level_shift(level, tt->granule_sz);
     78         uint64_t mask = subpage_size - 1;
     79         SMMUIOTLBKey key;
     80 
     81         key = smmu_get_iotlb_key(cfg->asid, iova & ~mask, tg, level);
     82         entry = g_hash_table_lookup(bs->iotlb, &key);
     83         if (entry) {
     84             break;
     85         }
     86         level++;
     87     }
     88 
     89     if (entry) {
     90         cfg->iotlb_hits++;
     91         trace_smmu_iotlb_lookup_hit(cfg->asid, iova,
     92                                     cfg->iotlb_hits, cfg->iotlb_misses,
     93                                     100 * cfg->iotlb_hits /
     94                                     (cfg->iotlb_hits + cfg->iotlb_misses));
     95     } else {
     96         cfg->iotlb_misses++;
     97         trace_smmu_iotlb_lookup_miss(cfg->asid, iova,
     98                                      cfg->iotlb_hits, cfg->iotlb_misses,
     99                                      100 * cfg->iotlb_hits /
    100                                      (cfg->iotlb_hits + cfg->iotlb_misses));
    101     }
    102     return entry;
    103 }
    104 
    105 void smmu_iotlb_insert(SMMUState *bs, SMMUTransCfg *cfg, SMMUTLBEntry *new)
    106 {
    107     SMMUIOTLBKey *key = g_new0(SMMUIOTLBKey, 1);
    108     uint8_t tg = (new->granule - 10) / 2;
    109 
    110     if (g_hash_table_size(bs->iotlb) >= SMMU_IOTLB_MAX_SIZE) {
    111         smmu_iotlb_inv_all(bs);
    112     }
    113 
    114     *key = smmu_get_iotlb_key(cfg->asid, new->entry.iova, tg, new->level);
    115     trace_smmu_iotlb_insert(cfg->asid, new->entry.iova, tg, new->level);
    116     g_hash_table_insert(bs->iotlb, key, new);
    117 }
    118 
    119 inline void smmu_iotlb_inv_all(SMMUState *s)
    120 {
    121     trace_smmu_iotlb_inv_all();
    122     g_hash_table_remove_all(s->iotlb);
    123 }
    124 
    125 static gboolean smmu_hash_remove_by_asid(gpointer key, gpointer value,
    126                                          gpointer user_data)
    127 {
    128     uint16_t asid = *(uint16_t *)user_data;
    129     SMMUIOTLBKey *iotlb_key = (SMMUIOTLBKey *)key;
    130 
    131     return SMMU_IOTLB_ASID(*iotlb_key) == asid;
    132 }
    133 
    134 static gboolean smmu_hash_remove_by_asid_iova(gpointer key, gpointer value,
    135                                               gpointer user_data)
    136 {
    137     SMMUTLBEntry *iter = (SMMUTLBEntry *)value;
    138     IOMMUTLBEntry *entry = &iter->entry;
    139     SMMUIOTLBPageInvInfo *info = (SMMUIOTLBPageInvInfo *)user_data;
    140     SMMUIOTLBKey iotlb_key = *(SMMUIOTLBKey *)key;
    141 
    142     if (info->asid >= 0 && info->asid != SMMU_IOTLB_ASID(iotlb_key)) {
    143         return false;
    144     }
    145     return ((info->iova & ~entry->addr_mask) == entry->iova) ||
    146            ((entry->iova & ~info->mask) == info->iova);
    147 }
    148 
    149 inline void
    150 smmu_iotlb_inv_iova(SMMUState *s, int asid, dma_addr_t iova,
    151                     uint8_t tg, uint64_t num_pages, uint8_t ttl)
    152 {
    153     /* if tg is not set we use 4KB range invalidation */
    154     uint8_t granule = tg ? tg * 2 + 10 : 12;
    155 
    156     if (ttl && (num_pages == 1) && (asid >= 0)) {
    157         SMMUIOTLBKey key = smmu_get_iotlb_key(asid, iova, tg, ttl);
    158 
    159         if (g_hash_table_remove(s->iotlb, &key)) {
    160             return;
    161         }
    162         /*
    163          * if the entry is not found, let's see if it does not
    164          * belong to a larger IOTLB entry
    165          */
    166     }
    167 
    168     SMMUIOTLBPageInvInfo info = {
    169         .asid = asid, .iova = iova,
    170         .mask = (num_pages * 1 << granule) - 1};
    171 
    172     g_hash_table_foreach_remove(s->iotlb,
    173                                 smmu_hash_remove_by_asid_iova,
    174                                 &info);
    175 }
    176 
    177 inline void smmu_iotlb_inv_asid(SMMUState *s, uint16_t asid)
    178 {
    179     trace_smmu_iotlb_inv_asid(asid);
    180     g_hash_table_foreach_remove(s->iotlb, smmu_hash_remove_by_asid, &asid);
    181 }
    182 
    183 /* VMSAv8-64 Translation */
    184 
    185 /**
    186  * get_pte - Get the content of a page table entry located at
    187  * @base_addr[@index]
    188  */
    189 static int get_pte(dma_addr_t baseaddr, uint32_t index, uint64_t *pte,
    190                    SMMUPTWEventInfo *info)
    191 {
    192     int ret;
    193     dma_addr_t addr = baseaddr + index * sizeof(*pte);
    194 
    195     /* TODO: guarantee 64-bit single-copy atomicity */
    196     ret = dma_memory_read(&address_space_memory, addr, pte, sizeof(*pte),
    197                           MEMTXATTRS_UNSPECIFIED);
    198 
    199     if (ret != MEMTX_OK) {
    200         info->type = SMMU_PTW_ERR_WALK_EABT;
    201         info->addr = addr;
    202         return -EINVAL;
    203     }
    204     trace_smmu_get_pte(baseaddr, index, addr, *pte);
    205     return 0;
    206 }
    207 
    208 /* VMSAv8-64 Translation Table Format Descriptor Decoding */
    209 
    210 /**
    211  * get_page_pte_address - returns the L3 descriptor output address,
    212  * ie. the page frame
    213  * ARM ARM spec: Figure D4-17 VMSAv8-64 level 3 descriptor format
    214  */
    215 static inline hwaddr get_page_pte_address(uint64_t pte, int granule_sz)
    216 {
    217     return PTE_ADDRESS(pte, granule_sz);
    218 }
    219 
    220 /**
    221  * get_table_pte_address - return table descriptor output address,
    222  * ie. address of next level table
    223  * ARM ARM Figure D4-16 VMSAv8-64 level0, level1, and level 2 descriptor formats
    224  */
    225 static inline hwaddr get_table_pte_address(uint64_t pte, int granule_sz)
    226 {
    227     return PTE_ADDRESS(pte, granule_sz);
    228 }
    229 
    230 /**
    231  * get_block_pte_address - return block descriptor output address and block size
    232  * ARM ARM Figure D4-16 VMSAv8-64 level0, level1, and level 2 descriptor formats
    233  */
    234 static inline hwaddr get_block_pte_address(uint64_t pte, int level,
    235                                            int granule_sz, uint64_t *bsz)
    236 {
    237     int n = level_shift(level, granule_sz);
    238 
    239     *bsz = 1ULL << n;
    240     return PTE_ADDRESS(pte, n);
    241 }
    242 
    243 SMMUTransTableInfo *select_tt(SMMUTransCfg *cfg, dma_addr_t iova)
    244 {
    245     bool tbi = extract64(iova, 55, 1) ? TBI1(cfg->tbi) : TBI0(cfg->tbi);
    246     uint8_t tbi_byte = tbi * 8;
    247 
    248     if (cfg->tt[0].tsz &&
    249         !extract64(iova, 64 - cfg->tt[0].tsz, cfg->tt[0].tsz - tbi_byte)) {
    250         /* there is a ttbr0 region and we are in it (high bits all zero) */
    251         return &cfg->tt[0];
    252     } else if (cfg->tt[1].tsz &&
    253            !extract64(iova, 64 - cfg->tt[1].tsz, cfg->tt[1].tsz - tbi_byte)) {
    254         /* there is a ttbr1 region and we are in it (high bits all one) */
    255         return &cfg->tt[1];
    256     } else if (!cfg->tt[0].tsz) {
    257         /* ttbr0 region is "everything not in the ttbr1 region" */
    258         return &cfg->tt[0];
    259     } else if (!cfg->tt[1].tsz) {
    260         /* ttbr1 region is "everything not in the ttbr0 region" */
    261         return &cfg->tt[1];
    262     }
    263     /* in the gap between the two regions, this is a Translation fault */
    264     return NULL;
    265 }
    266 
    267 /**
    268  * smmu_ptw_64 - VMSAv8-64 Walk of the page tables for a given IOVA
    269  * @cfg: translation config
    270  * @iova: iova to translate
    271  * @perm: access type
    272  * @tlbe: SMMUTLBEntry (out)
    273  * @info: handle to an error info
    274  *
    275  * Return 0 on success, < 0 on error. In case of error, @info is filled
    276  * and tlbe->perm is set to IOMMU_NONE.
    277  * Upon success, @tlbe is filled with translated_addr and entry
    278  * permission rights.
    279  */
    280 static int smmu_ptw_64(SMMUTransCfg *cfg,
    281                        dma_addr_t iova, IOMMUAccessFlags perm,
    282                        SMMUTLBEntry *tlbe, SMMUPTWEventInfo *info)
    283 {
    284     dma_addr_t baseaddr, indexmask;
    285     int stage = cfg->stage;
    286     SMMUTransTableInfo *tt = select_tt(cfg, iova);
    287     uint8_t level, granule_sz, inputsize, stride;
    288 
    289     if (!tt || tt->disabled) {
    290         info->type = SMMU_PTW_ERR_TRANSLATION;
    291         goto error;
    292     }
    293 
    294     granule_sz = tt->granule_sz;
    295     stride = granule_sz - 3;
    296     inputsize = 64 - tt->tsz;
    297     level = 4 - (inputsize - 4) / stride;
    298     indexmask = (1ULL << (inputsize - (stride * (4 - level)))) - 1;
    299     baseaddr = extract64(tt->ttb, 0, 48);
    300     baseaddr &= ~indexmask;
    301 
    302     while (level <= 3) {
    303         uint64_t subpage_size = 1ULL << level_shift(level, granule_sz);
    304         uint64_t mask = subpage_size - 1;
    305         uint32_t offset = iova_level_offset(iova, inputsize, level, granule_sz);
    306         uint64_t pte, gpa;
    307         dma_addr_t pte_addr = baseaddr + offset * sizeof(pte);
    308         uint8_t ap;
    309 
    310         if (get_pte(baseaddr, offset, &pte, info)) {
    311                 goto error;
    312         }
    313         trace_smmu_ptw_level(level, iova, subpage_size,
    314                              baseaddr, offset, pte);
    315 
    316         if (is_invalid_pte(pte) || is_reserved_pte(pte, level)) {
    317             trace_smmu_ptw_invalid_pte(stage, level, baseaddr,
    318                                        pte_addr, offset, pte);
    319             break;
    320         }
    321 
    322         if (is_table_pte(pte, level)) {
    323             ap = PTE_APTABLE(pte);
    324 
    325             if (is_permission_fault(ap, perm) && !tt->had) {
    326                 info->type = SMMU_PTW_ERR_PERMISSION;
    327                 goto error;
    328             }
    329             baseaddr = get_table_pte_address(pte, granule_sz);
    330             level++;
    331             continue;
    332         } else if (is_page_pte(pte, level)) {
    333             gpa = get_page_pte_address(pte, granule_sz);
    334             trace_smmu_ptw_page_pte(stage, level, iova,
    335                                     baseaddr, pte_addr, pte, gpa);
    336         } else {
    337             uint64_t block_size;
    338 
    339             gpa = get_block_pte_address(pte, level, granule_sz,
    340                                         &block_size);
    341             trace_smmu_ptw_block_pte(stage, level, baseaddr,
    342                                      pte_addr, pte, iova, gpa,
    343                                      block_size >> 20);
    344         }
    345         ap = PTE_AP(pte);
    346         if (is_permission_fault(ap, perm)) {
    347             info->type = SMMU_PTW_ERR_PERMISSION;
    348             goto error;
    349         }
    350 
    351         tlbe->entry.translated_addr = gpa;
    352         tlbe->entry.iova = iova & ~mask;
    353         tlbe->entry.addr_mask = mask;
    354         tlbe->entry.perm = PTE_AP_TO_PERM(ap);
    355         tlbe->level = level;
    356         tlbe->granule = granule_sz;
    357         return 0;
    358     }
    359     info->type = SMMU_PTW_ERR_TRANSLATION;
    360 
    361 error:
    362     tlbe->entry.perm = IOMMU_NONE;
    363     return -EINVAL;
    364 }
    365 
    366 /**
    367  * smmu_ptw - Walk the page tables for an IOVA, according to @cfg
    368  *
    369  * @cfg: translation configuration
    370  * @iova: iova to translate
    371  * @perm: tentative access type
    372  * @tlbe: returned entry
    373  * @info: ptw event handle
    374  *
    375  * return 0 on success
    376  */
    377 inline int smmu_ptw(SMMUTransCfg *cfg, dma_addr_t iova, IOMMUAccessFlags perm,
    378                     SMMUTLBEntry *tlbe, SMMUPTWEventInfo *info)
    379 {
    380     if (!cfg->aa64) {
    381         /*
    382          * This code path is not entered as we check this while decoding
    383          * the configuration data in the derived SMMU model.
    384          */
    385         g_assert_not_reached();
    386     }
    387 
    388     return smmu_ptw_64(cfg, iova, perm, tlbe, info);
    389 }
    390 
    391 /**
    392  * The bus number is used for lookup when SID based invalidation occurs.
    393  * In that case we lazily populate the SMMUPciBus array from the bus hash
    394  * table. At the time the SMMUPciBus is created (smmu_find_add_as), the bus
    395  * numbers may not be always initialized yet.
    396  */
    397 SMMUPciBus *smmu_find_smmu_pcibus(SMMUState *s, uint8_t bus_num)
    398 {
    399     SMMUPciBus *smmu_pci_bus = s->smmu_pcibus_by_bus_num[bus_num];
    400     GHashTableIter iter;
    401 
    402     if (smmu_pci_bus) {
    403         return smmu_pci_bus;
    404     }
    405 
    406     g_hash_table_iter_init(&iter, s->smmu_pcibus_by_busptr);
    407     while (g_hash_table_iter_next(&iter, NULL, (void **)&smmu_pci_bus)) {
    408         if (pci_bus_num(smmu_pci_bus->bus) == bus_num) {
    409             s->smmu_pcibus_by_bus_num[bus_num] = smmu_pci_bus;
    410             return smmu_pci_bus;
    411         }
    412     }
    413 
    414     return NULL;
    415 }
    416 
    417 static AddressSpace *smmu_find_add_as(PCIBus *bus, void *opaque, int devfn)
    418 {
    419     SMMUState *s = opaque;
    420     SMMUPciBus *sbus = g_hash_table_lookup(s->smmu_pcibus_by_busptr, bus);
    421     SMMUDevice *sdev;
    422     static unsigned int index;
    423 
    424     if (!sbus) {
    425         sbus = g_malloc0(sizeof(SMMUPciBus) +
    426                          sizeof(SMMUDevice *) * SMMU_PCI_DEVFN_MAX);
    427         sbus->bus = bus;
    428         g_hash_table_insert(s->smmu_pcibus_by_busptr, bus, sbus);
    429     }
    430 
    431     sdev = sbus->pbdev[devfn];
    432     if (!sdev) {
    433         char *name = g_strdup_printf("%s-%d-%d", s->mrtypename, devfn, index++);
    434 
    435         sdev = sbus->pbdev[devfn] = g_new0(SMMUDevice, 1);
    436 
    437         sdev->smmu = s;
    438         sdev->bus = bus;
    439         sdev->devfn = devfn;
    440 
    441         memory_region_init_iommu(&sdev->iommu, sizeof(sdev->iommu),
    442                                  s->mrtypename,
    443                                  OBJECT(s), name, 1ULL << SMMU_MAX_VA_BITS);
    444         address_space_init(&sdev->as,
    445                            MEMORY_REGION(&sdev->iommu), name);
    446         trace_smmu_add_mr(name);
    447         g_free(name);
    448     }
    449 
    450     return &sdev->as;
    451 }
    452 
    453 IOMMUMemoryRegion *smmu_iommu_mr(SMMUState *s, uint32_t sid)
    454 {
    455     uint8_t bus_n, devfn;
    456     SMMUPciBus *smmu_bus;
    457     SMMUDevice *smmu;
    458 
    459     bus_n = PCI_BUS_NUM(sid);
    460     smmu_bus = smmu_find_smmu_pcibus(s, bus_n);
    461     if (smmu_bus) {
    462         devfn = SMMU_PCI_DEVFN(sid);
    463         smmu = smmu_bus->pbdev[devfn];
    464         if (smmu) {
    465             return &smmu->iommu;
    466         }
    467     }
    468     return NULL;
    469 }
    470 
    471 /* Unmap the whole notifier's range */
    472 static void smmu_unmap_notifier_range(IOMMUNotifier *n)
    473 {
    474     IOMMUTLBEvent event;
    475 
    476     event.type = IOMMU_NOTIFIER_UNMAP;
    477     event.entry.target_as = &address_space_memory;
    478     event.entry.iova = n->start;
    479     event.entry.perm = IOMMU_NONE;
    480     event.entry.addr_mask = n->end - n->start;
    481 
    482     memory_region_notify_iommu_one(n, &event);
    483 }
    484 
    485 /* Unmap all notifiers attached to @mr */
    486 inline void smmu_inv_notifiers_mr(IOMMUMemoryRegion *mr)
    487 {
    488     IOMMUNotifier *n;
    489 
    490     trace_smmu_inv_notifiers_mr(mr->parent_obj.name);
    491     IOMMU_NOTIFIER_FOREACH(n, mr) {
    492         smmu_unmap_notifier_range(n);
    493     }
    494 }
    495 
    496 /* Unmap all notifiers of all mr's */
    497 void smmu_inv_notifiers_all(SMMUState *s)
    498 {
    499     SMMUDevice *sdev;
    500 
    501     QLIST_FOREACH(sdev, &s->devices_with_notifiers, next) {
    502         smmu_inv_notifiers_mr(&sdev->iommu);
    503     }
    504 }
    505 
    506 static void smmu_base_realize(DeviceState *dev, Error **errp)
    507 {
    508     SMMUState *s = ARM_SMMU(dev);
    509     SMMUBaseClass *sbc = ARM_SMMU_GET_CLASS(dev);
    510     Error *local_err = NULL;
    511 
    512     sbc->parent_realize(dev, &local_err);
    513     if (local_err) {
    514         error_propagate(errp, local_err);
    515         return;
    516     }
    517     s->configs = g_hash_table_new_full(NULL, NULL, NULL, g_free);
    518     s->iotlb = g_hash_table_new_full(smmu_iotlb_key_hash, smmu_iotlb_key_equal,
    519                                      g_free, g_free);
    520     s->smmu_pcibus_by_busptr = g_hash_table_new(NULL, NULL);
    521 
    522     if (s->primary_bus) {
    523         pci_setup_iommu(s->primary_bus, smmu_find_add_as, s);
    524     } else {
    525         error_setg(errp, "SMMU is not attached to any PCI bus!");
    526     }
    527 }
    528 
    529 static void smmu_base_reset(DeviceState *dev)
    530 {
    531     SMMUState *s = ARM_SMMU(dev);
    532 
    533     g_hash_table_remove_all(s->configs);
    534     g_hash_table_remove_all(s->iotlb);
    535 }
    536 
    537 static Property smmu_dev_properties[] = {
    538     DEFINE_PROP_UINT8("bus_num", SMMUState, bus_num, 0),
    539     DEFINE_PROP_LINK("primary-bus", SMMUState, primary_bus, "PCI", PCIBus *),
    540     DEFINE_PROP_END_OF_LIST(),
    541 };
    542 
    543 static void smmu_base_class_init(ObjectClass *klass, void *data)
    544 {
    545     DeviceClass *dc = DEVICE_CLASS(klass);
    546     SMMUBaseClass *sbc = ARM_SMMU_CLASS(klass);
    547 
    548     device_class_set_props(dc, smmu_dev_properties);
    549     device_class_set_parent_realize(dc, smmu_base_realize,
    550                                     &sbc->parent_realize);
    551     dc->reset = smmu_base_reset;
    552 }
    553 
    554 static const TypeInfo smmu_base_info = {
    555     .name          = TYPE_ARM_SMMU,
    556     .parent        = TYPE_SYS_BUS_DEVICE,
    557     .instance_size = sizeof(SMMUState),
    558     .class_data    = NULL,
    559     .class_size    = sizeof(SMMUBaseClass),
    560     .class_init    = smmu_base_class_init,
    561     .abstract      = true,
    562 };
    563 
    564 static void smmu_base_register_types(void)
    565 {
    566     type_register_static(&smmu_base_info);
    567 }
    568 
    569 type_init(smmu_base_register_types)
    570