Spaces:
Runtime error
Runtime error
File size: 9,617 Bytes
8df6da4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 |
/*
* Intel IOMMU APIs
*
* Copyright (C) 2016 Red Hat, Inc.
*
* Authors:
* Peter Xu <peterx@redhat.com>,
*
* This work is licensed under the terms of the GNU LGPL, version 2 or
* later.
*/
#include "intel-iommu.h"
#include "libcflat.h"
#include "pci.h"
#include "atomic.h"
/*
* VT-d in QEMU currently only support 39 bits address width, which is
* 3-level translation.
*/
#define VTD_PAGE_LEVEL 3
#define VTD_CE_AW_39BIT 0x1
typedef uint64_t vtd_pte_t;
struct vtd_root_entry {
/* Quad 1 */
uint64_t present:1;
uint64_t __reserved:11;
uint64_t context_table_p:52;
/* Quad 2 */
uint64_t __reserved_2;
} __attribute__ ((packed));
typedef struct vtd_root_entry vtd_re_t;
struct vtd_context_entry {
/* Quad 1 */
uint64_t present:1;
uint64_t disable_fault_report:1;
uint64_t trans_type:2;
uint64_t __reserved:8;
uint64_t slptptr:52;
/* Quad 2 */
uint64_t addr_width:3;
uint64_t __ignore:4;
uint64_t __reserved_2:1;
uint64_t domain_id:16;
uint64_t __reserved_3:40;
} __attribute__ ((packed));
typedef struct vtd_context_entry vtd_ce_t;
struct vtd_irte {
uint32_t present:1;
uint32_t fault_disable:1; /* Fault Processing Disable */
uint32_t dest_mode:1; /* Destination Mode */
uint32_t redir_hint:1; /* Redirection Hint */
uint32_t trigger_mode:1; /* Trigger Mode */
uint32_t delivery_mode:3; /* Delivery Mode */
uint32_t __avail:4; /* Available spaces for software */
uint32_t __reserved_0:3; /* Reserved 0 */
uint32_t irte_mode:1; /* IRTE Mode */
uint32_t vector:8; /* Interrupt Vector */
uint32_t __reserved_1:8; /* Reserved 1 */
uint32_t dest_id; /* Destination ID */
uint16_t source_id:16; /* Source-ID */
uint64_t sid_q:2; /* Source-ID Qualifier */
uint64_t sid_vtype:2; /* Source-ID Validation Type */
uint64_t __reserved_2:44; /* Reserved 2 */
} __attribute__ ((packed));
typedef struct vtd_irte vtd_irte_t;
#define VTD_RTA_MASK (PAGE_MASK)
#define VTD_IRTA_MASK (PAGE_MASK)
void *vtd_reg_base;
static uint64_t vtd_root_table(void)
{
/* No extend root table support yet */
return vtd_readq(DMAR_RTADDR_REG) & VTD_RTA_MASK;
}
static uint64_t vtd_ir_table(void)
{
return vtd_readq(DMAR_IRTA_REG) & VTD_IRTA_MASK;
}
static void vtd_gcmd_or(uint32_t cmd)
{
uint32_t status;
/* We only allow set one bit for each time */
assert(is_power_of_2(cmd));
status = vtd_readl(DMAR_GSTS_REG);
vtd_writel(DMAR_GCMD_REG, status | cmd);
if (cmd & VTD_GCMD_ONE_SHOT_BITS) {
/* One-shot bits are taking effect immediately */
return;
}
/* Make sure IOMMU handled our command request */
while (!(vtd_readl(DMAR_GSTS_REG) & cmd))
cpu_relax();
}
static void vtd_dump_init_info(void)
{
uint32_t version;
version = vtd_readl(DMAR_VER_REG);
/* Major version >= 1 */
assert(((version >> 3) & 0xf) >= 1);
printf("VT-d version: %#x\n", version);
printf(" cap: %#018lx\n", vtd_readq(DMAR_CAP_REG));
printf(" ecap: %#018lx\n", vtd_readq(DMAR_ECAP_REG));
}
static void vtd_setup_root_table(void)
{
void *root = alloc_page();
memset(root, 0, PAGE_SIZE);
vtd_writeq(DMAR_RTADDR_REG, virt_to_phys(root));
vtd_gcmd_or(VTD_GCMD_ROOT);
printf("DMAR table address: %#018lx\n", vtd_root_table());
}
static void vtd_setup_ir_table(void)
{
void *root = alloc_page();
memset(root, 0, PAGE_SIZE);
/* 0xf stands for table size (2^(0xf+1) == 65536) */
vtd_writeq(DMAR_IRTA_REG, virt_to_phys(root) | 0xf);
vtd_gcmd_or(VTD_GCMD_IR_TABLE);
printf("IR table address: %#018lx\n", vtd_ir_table());
}
static void vtd_install_pte(vtd_pte_t *root, iova_t iova,
phys_addr_t pa, int level_target)
{
int level;
unsigned int offset;
void *page;
for (level = VTD_PAGE_LEVEL; level > level_target; level--) {
offset = PGDIR_OFFSET(iova, level);
if (!(root[offset] & VTD_PTE_RW)) {
page = alloc_page();
memset(page, 0, PAGE_SIZE);
root[offset] = virt_to_phys(page) | VTD_PTE_RW;
}
root = (uint64_t *)(phys_to_virt(root[offset] &
VTD_PTE_ADDR));
}
offset = PGDIR_OFFSET(iova, level);
root[offset] = pa | VTD_PTE_RW;
if (level != 1) {
/* This is huge page */
root[offset] |= VTD_PTE_HUGE;
}
}
/**
* vtd_map_range: setup IO address mapping for specific memory range
*
* @sid: source ID of the device to setup
* @iova: start IO virtual address
* @pa: start physical address
* @size: size of the mapping area
*/
void vtd_map_range(uint16_t sid, iova_t iova, phys_addr_t pa, size_t size)
{
uint8_t bus_n, devfn;
void *slptptr;
vtd_ce_t *ce;
vtd_re_t *re = phys_to_virt(vtd_root_table());
assert(IS_ALIGNED(iova, SZ_4K));
assert(IS_ALIGNED(pa, SZ_4K));
assert(IS_ALIGNED(size, SZ_4K));
bus_n = PCI_BDF_GET_BUS(sid);
devfn = PCI_BDF_GET_DEVFN(sid);
/* Point to the correct root entry */
re += bus_n;
if (!re->present) {
ce = alloc_page();
memset(ce, 0, PAGE_SIZE);
memset(re, 0, sizeof(*re));
re->context_table_p = virt_to_phys(ce) >> VTD_PAGE_SHIFT;
re->present = 1;
printf("allocated vt-d root entry for PCI bus %d\n",
bus_n);
} else
ce = phys_to_virt(re->context_table_p << VTD_PAGE_SHIFT);
/* Point to the correct context entry */
ce += devfn;
if (!ce->present) {
slptptr = alloc_page();
memset(slptptr, 0, PAGE_SIZE);
memset(ce, 0, sizeof(*ce));
/* To make it simple, domain ID is the same as SID */
ce->domain_id = sid;
/* We only test 39 bits width case (3-level paging) */
ce->addr_width = VTD_CE_AW_39BIT;
ce->slptptr = virt_to_phys(slptptr) >> VTD_PAGE_SHIFT;
ce->trans_type = VTD_CONTEXT_TT_MULTI_LEVEL;
ce->present = 1;
/* No error reporting yet */
ce->disable_fault_report = 1;
printf("allocated vt-d context entry for devfn %#x\n",
devfn);
} else
slptptr = phys_to_virt(ce->slptptr << VTD_PAGE_SHIFT);
while (size) {
/* TODO: currently we only map 4K pages (level = 1) */
printf("map 4K page IOVA %#lx to %#lx (sid=%#06x)\n",
iova, pa, sid);
vtd_install_pte(slptptr, iova, pa, 1);
size -= VTD_PAGE_SIZE;
iova += VTD_PAGE_SIZE;
pa += VTD_PAGE_SIZE;
}
}
static uint16_t vtd_intr_index_alloc(void)
{
static volatile int index_ctr = 0;
int ctr;
assert(index_ctr < 65535);
ctr = atomic_inc_fetch(&index_ctr);
printf("INTR: alloc IRTE index %d\n", ctr);
return ctr;
}
static void vtd_setup_irte(struct pci_dev *dev, vtd_irte_t *irte,
int vector, int dest_id, trigger_mode_t trigger)
{
assert(sizeof(vtd_irte_t) == 16);
memset(irte, 0, sizeof(*irte));
irte->fault_disable = 1;
irte->dest_mode = 0; /* physical */
irte->trigger_mode = trigger;
irte->delivery_mode = 0; /* fixed */
irte->irte_mode = 0; /* remapped */
irte->vector = vector;
irte->dest_id = dest_id;
irte->source_id = dev->bdf;
irte->sid_q = 0;
irte->sid_vtype = 1; /* full-sid verify */
irte->present = 1;
}
struct vtd_msi_addr {
uint32_t __dont_care:2;
uint32_t handle_15:1; /* handle[15] */
uint32_t shv:1;
uint32_t interrupt_format:1;
uint32_t handle_0_14:15; /* handle[0:14] */
uint32_t head:12; /* 0xfee */
uint32_t addr_hi; /* not used except with x2apic */
} __attribute__ ((packed));
typedef struct vtd_msi_addr vtd_msi_addr_t;
struct vtd_msi_data {
uint16_t __reserved;
uint16_t subhandle;
} __attribute__ ((packed));
typedef struct vtd_msi_data vtd_msi_data_t;
struct vtd_ioapic_entry {
uint64_t vector:8;
uint64_t __zeros:3;
uint64_t index_15:1;
uint64_t delivery_status:1;
uint64_t polarity:1;
uint64_t remote_irr:1;
uint64_t trigger_mode:1;
uint64_t mask:1;
uint64_t __zeros_2:31;
uint64_t interrupt_format:1;
uint64_t index_0_14:15;
} __attribute__ ((packed));
typedef struct vtd_ioapic_entry vtd_ioapic_entry_t;
/**
* vtd_setup_msi - setup MSI message for a device
*
* @dev: PCI device to setup MSI
* @vector: interrupt vector
* @dest_id: destination processor
*/
bool vtd_setup_msi(struct pci_dev *dev, int vector, int dest_id)
{
vtd_msi_data_t msi_data = {};
vtd_msi_addr_t msi_addr = {};
vtd_irte_t *irte = phys_to_virt(vtd_ir_table());
uint16_t index = vtd_intr_index_alloc();
assert(sizeof(vtd_msi_addr_t) == 8);
assert(sizeof(vtd_msi_data_t) == 4);
/* Use edge irq as default */
vtd_setup_irte(dev, irte + index, vector,
dest_id, TRIGGER_EDGE);
msi_addr.handle_15 = index >> 15 & 1;
msi_addr.shv = 0;
msi_addr.interrupt_format = 1;
msi_addr.handle_0_14 = index & 0x7fff;
msi_addr.head = 0xfee;
msi_data.subhandle = 0;
printf("%s: msi_addr=%#" PRIx64 ", msi_data=%#x\n", __func__,
*(uint64_t *)&msi_addr, *(uint32_t *)&msi_data);
return pci_setup_msi(dev, *(uint64_t *)&msi_addr,
*(uint32_t *)&msi_data);
}
void vtd_setup_ioapic_irq(struct pci_dev *dev, int vector,
int dest_id, trigger_mode_t trigger)
{
vtd_ioapic_entry_t entry = {};
vtd_irte_t *irte = phys_to_virt(vtd_ir_table());
ioapic_redir_entry_t *entry_2 = (ioapic_redir_entry_t *)&entry;
uint16_t index = vtd_intr_index_alloc();
uint8_t line;
assert(dev);
assert(sizeof(vtd_ioapic_entry_t) == 8);
vtd_setup_irte(dev, irte + index, vector,
dest_id, trigger);
entry.vector = vector;
entry.trigger_mode = trigger;
entry.index_15 = (index >> 15) & 1;
entry.interrupt_format = 1;
entry.index_0_14 = index & 0x7fff;
line = pci_intx_line(dev);
ioapic_write_redir(line, *entry_2);
}
void vtd_init(void)
{
setup_vm();
smp_init();
vtd_reg_base = ioremap(Q35_HOST_BRIDGE_IOMMU_ADDR, PAGE_SIZE);
vtd_dump_init_info();
vtd_gcmd_or(VTD_GCMD_QI); /* Enable QI */
vtd_setup_root_table();
vtd_setup_ir_table();
vtd_gcmd_or(VTD_GCMD_DMAR); /* Enable DMAR */
vtd_gcmd_or(VTD_GCMD_IR); /* Enable IR */
}
|