Commit | Line | Data |
---|---|---|
8d283c35 | 1 | /* |
5d0d7156 | 2 | * Copyright (C) 2007-2010 Advanced Micro Devices, Inc. |
8d283c35 JR |
3 | * Author: Joerg Roedel <joerg.roedel@amd.com> |
4 | * Leo Duran <leo.duran@amd.com> | |
5 | * | |
6 | * This program is free software; you can redistribute it and/or modify it | |
7 | * under the terms of the GNU General Public License version 2 as published | |
8 | * by the Free Software Foundation. | |
9 | * | |
10 | * This program is distributed in the hope that it will be useful, | |
11 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
12 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
13 | * GNU General Public License for more details. | |
14 | * | |
15 | * You should have received a copy of the GNU General Public License | |
16 | * along with this program; if not, write to the Free Software | |
17 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA | |
18 | */ | |
19 | ||
1965aae3 PA |
20 | #ifndef _ASM_X86_AMD_IOMMU_TYPES_H |
21 | #define _ASM_X86_AMD_IOMMU_TYPES_H | |
8d283c35 JR |
22 | |
23 | #include <linux/types.h> | |
5d214fe6 | 24 | #include <linux/mutex.h> |
8d283c35 JR |
25 | #include <linux/list.h> |
26 | #include <linux/spinlock.h> | |
c5081cd7 | 27 | #include <linux/pci.h> |
4b180d97 | 28 | #include <linux/irqreturn.h> |
8d283c35 | 29 | |
bb52777e JR |
30 | /* |
31 | * Maximum number of IOMMUs supported | |
32 | */ | |
33 | #define MAX_IOMMUS 32 | |
34 | ||
8d283c35 JR |
35 | /* |
36 | * some size calculation constants | |
37 | */ | |
83f5aac1 | 38 | #define DEV_TABLE_ENTRY_SIZE 32 |
8d283c35 JR |
39 | #define ALIAS_TABLE_ENTRY_SIZE 2 |
40 | #define RLOOKUP_TABLE_ENTRY_SIZE (sizeof(void *)) | |
41 | ||
8d283c35 JR |
42 | /* Capability offsets used by the driver */ |
43 | #define MMIO_CAP_HDR_OFFSET 0x00 | |
44 | #define MMIO_RANGE_OFFSET 0x0c | |
a80dc3e0 | 45 | #define MMIO_MISC_OFFSET 0x10 |
8d283c35 JR |
46 | |
47 | /* Masks, shifts and macros to parse the device range capability */ | |
48 | #define MMIO_RANGE_LD_MASK 0xff000000 | |
49 | #define MMIO_RANGE_FD_MASK 0x00ff0000 | |
50 | #define MMIO_RANGE_BUS_MASK 0x0000ff00 | |
51 | #define MMIO_RANGE_LD_SHIFT 24 | |
52 | #define MMIO_RANGE_FD_SHIFT 16 | |
53 | #define MMIO_RANGE_BUS_SHIFT 8 | |
54 | #define MMIO_GET_LD(x) (((x) & MMIO_RANGE_LD_MASK) >> MMIO_RANGE_LD_SHIFT) | |
55 | #define MMIO_GET_FD(x) (((x) & MMIO_RANGE_FD_MASK) >> MMIO_RANGE_FD_SHIFT) | |
56 | #define MMIO_GET_BUS(x) (((x) & MMIO_RANGE_BUS_MASK) >> MMIO_RANGE_BUS_SHIFT) | |
a80dc3e0 | 57 | #define MMIO_MSI_NUM(x) ((x) & 0x1f) |
8d283c35 JR |
58 | |
59 | /* Flag masks for the AMD IOMMU exclusion range */ | |
60 | #define MMIO_EXCL_ENABLE_MASK 0x01ULL | |
61 | #define MMIO_EXCL_ALLOW_MASK 0x02ULL | |
62 | ||
63 | /* Used offsets into the MMIO space */ | |
64 | #define MMIO_DEV_TABLE_OFFSET 0x0000 | |
65 | #define MMIO_CMD_BUF_OFFSET 0x0008 | |
66 | #define MMIO_EVT_BUF_OFFSET 0x0010 | |
67 | #define MMIO_CONTROL_OFFSET 0x0018 | |
68 | #define MMIO_EXCL_BASE_OFFSET 0x0020 | |
69 | #define MMIO_EXCL_LIMIT_OFFSET 0x0028 | |
d99ddec3 | 70 | #define MMIO_EXT_FEATURES 0x0030 |
1a29ac01 | 71 | #define MMIO_PPR_LOG_OFFSET 0x0038 |
8d283c35 JR |
72 | #define MMIO_CMD_HEAD_OFFSET 0x2000 |
73 | #define MMIO_CMD_TAIL_OFFSET 0x2008 | |
74 | #define MMIO_EVT_HEAD_OFFSET 0x2010 | |
75 | #define MMIO_EVT_TAIL_OFFSET 0x2018 | |
76 | #define MMIO_STATUS_OFFSET 0x2020 | |
1a29ac01 JR |
77 | #define MMIO_PPR_HEAD_OFFSET 0x2030 |
78 | #define MMIO_PPR_TAIL_OFFSET 0x2038 | |
30861ddc SK |
79 | #define MMIO_CNTR_CONF_OFFSET 0x4000 |
80 | #define MMIO_CNTR_REG_OFFSET 0x40000 | |
81 | #define MMIO_REG_END_OFFSET 0x80000 | |
82 | ||
8d283c35 | 83 | |
d99ddec3 JR |
84 | |
85 | /* Extended Feature Bits */ | |
86 | #define FEATURE_PREFETCH (1ULL<<0) | |
87 | #define FEATURE_PPR (1ULL<<1) | |
88 | #define FEATURE_X2APIC (1ULL<<2) | |
89 | #define FEATURE_NX (1ULL<<3) | |
90 | #define FEATURE_GT (1ULL<<4) | |
91 | #define FEATURE_IA (1ULL<<6) | |
92 | #define FEATURE_GA (1ULL<<7) | |
93 | #define FEATURE_HE (1ULL<<8) | |
94 | #define FEATURE_PC (1ULL<<9) | |
95 | ||
62f71abb JR |
96 | #define FEATURE_PASID_SHIFT 32 |
97 | #define FEATURE_PASID_MASK (0x1fULL << FEATURE_PASID_SHIFT) | |
98 | ||
52815b75 JR |
99 | #define FEATURE_GLXVAL_SHIFT 14 |
100 | #define FEATURE_GLXVAL_MASK (0x03ULL << FEATURE_GLXVAL_SHIFT) | |
101 | ||
a919a018 SS |
102 | /* Note: |
103 | * The current driver only support 16-bit PASID. | |
104 | * Currently, hardware only implement upto 16-bit PASID | |
105 | * even though the spec says it could have upto 20 bits. | |
106 | */ | |
107 | #define PASID_MASK 0x0000ffff | |
52815b75 | 108 | |
519c31ba | 109 | /* MMIO status bits */ |
925fe08b | 110 | #define MMIO_STATUS_EVT_INT_MASK (1 << 1) |
72e1dcc4 JR |
111 | #define MMIO_STATUS_COM_WAIT_INT_MASK (1 << 2) |
112 | #define MMIO_STATUS_PPR_INT_MASK (1 << 6) | |
519c31ba | 113 | |
90008ee4 JR |
114 | /* event logging constants */ |
115 | #define EVENT_ENTRY_SIZE 0x10 | |
116 | #define EVENT_TYPE_SHIFT 28 | |
117 | #define EVENT_TYPE_MASK 0xf | |
118 | #define EVENT_TYPE_ILL_DEV 0x1 | |
119 | #define EVENT_TYPE_IO_FAULT 0x2 | |
120 | #define EVENT_TYPE_DEV_TAB_ERR 0x3 | |
121 | #define EVENT_TYPE_PAGE_TAB_ERR 0x4 | |
122 | #define EVENT_TYPE_ILL_CMD 0x5 | |
123 | #define EVENT_TYPE_CMD_HARD_ERR 0x6 | |
124 | #define EVENT_TYPE_IOTLB_INV_TO 0x7 | |
125 | #define EVENT_TYPE_INV_DEV_REQ 0x8 | |
126 | #define EVENT_DEVID_MASK 0xffff | |
127 | #define EVENT_DEVID_SHIFT 0 | |
128 | #define EVENT_DOMID_MASK 0xffff | |
129 | #define EVENT_DOMID_SHIFT 0 | |
130 | #define EVENT_FLAGS_MASK 0xfff | |
131 | #define EVENT_FLAGS_SHIFT 0x10 | |
132 | ||
8d283c35 JR |
133 | /* feature control bits */ |
134 | #define CONTROL_IOMMU_EN 0x00ULL | |
135 | #define CONTROL_HT_TUN_EN 0x01ULL | |
136 | #define CONTROL_EVT_LOG_EN 0x02ULL | |
137 | #define CONTROL_EVT_INT_EN 0x03ULL | |
138 | #define CONTROL_COMWAIT_EN 0x04ULL | |
1456e9d2 | 139 | #define CONTROL_INV_TIMEOUT 0x05ULL |
8d283c35 JR |
140 | #define CONTROL_PASSPW_EN 0x08ULL |
141 | #define CONTROL_RESPASSPW_EN 0x09ULL | |
142 | #define CONTROL_COHERENT_EN 0x0aULL | |
143 | #define CONTROL_ISOC_EN 0x0bULL | |
144 | #define CONTROL_CMDBUF_EN 0x0cULL | |
145 | #define CONTROL_PPFLOG_EN 0x0dULL | |
146 | #define CONTROL_PPFINT_EN 0x0eULL | |
1a29ac01 | 147 | #define CONTROL_PPR_EN 0x0fULL |
cbc33a90 | 148 | #define CONTROL_GT_EN 0x10ULL |
8d283c35 | 149 | |
1456e9d2 JR |
150 | #define CTRL_INV_TO_MASK (7 << CONTROL_INV_TIMEOUT) |
151 | #define CTRL_INV_TO_NONE 0 | |
152 | #define CTRL_INV_TO_1MS 1 | |
153 | #define CTRL_INV_TO_10MS 2 | |
154 | #define CTRL_INV_TO_100MS 3 | |
155 | #define CTRL_INV_TO_1S 4 | |
156 | #define CTRL_INV_TO_10S 5 | |
157 | #define CTRL_INV_TO_100S 6 | |
158 | ||
8d283c35 JR |
159 | /* command specific defines */ |
160 | #define CMD_COMPL_WAIT 0x01 | |
161 | #define CMD_INV_DEV_ENTRY 0x02 | |
cb41ed85 JR |
162 | #define CMD_INV_IOMMU_PAGES 0x03 |
163 | #define CMD_INV_IOTLB_PAGES 0x04 | |
7ef2798d | 164 | #define CMD_INV_IRT 0x05 |
c99afa25 | 165 | #define CMD_COMPLETE_PPR 0x07 |
58fc7f14 | 166 | #define CMD_INV_ALL 0x08 |
8d283c35 JR |
167 | |
168 | #define CMD_COMPL_WAIT_STORE_MASK 0x01 | |
519c31ba | 169 | #define CMD_COMPL_WAIT_INT_MASK 0x02 |
8d283c35 JR |
170 | #define CMD_INV_IOMMU_PAGES_SIZE_MASK 0x01 |
171 | #define CMD_INV_IOMMU_PAGES_PDE_MASK 0x02 | |
22e266c7 | 172 | #define CMD_INV_IOMMU_PAGES_GN_MASK 0x04 |
8d283c35 | 173 | |
c99afa25 JR |
174 | #define PPR_STATUS_MASK 0xf |
175 | #define PPR_STATUS_SHIFT 12 | |
176 | ||
999ba417 JR |
177 | #define CMD_INV_IOMMU_ALL_PAGES_ADDRESS 0x7fffffffffffffffULL |
178 | ||
8d283c35 JR |
179 | /* macros and definitions for device table entries */ |
180 | #define DEV_ENTRY_VALID 0x00 | |
181 | #define DEV_ENTRY_TRANSLATION 0x01 | |
182 | #define DEV_ENTRY_IR 0x3d | |
183 | #define DEV_ENTRY_IW 0x3e | |
9f5f5fb3 | 184 | #define DEV_ENTRY_NO_PAGE_FAULT 0x62 |
8d283c35 JR |
185 | #define DEV_ENTRY_EX 0x67 |
186 | #define DEV_ENTRY_SYSMGT1 0x68 | |
187 | #define DEV_ENTRY_SYSMGT2 0x69 | |
0ea2c422 | 188 | #define DEV_ENTRY_IRQ_TBL_EN 0x80 |
8d283c35 JR |
189 | #define DEV_ENTRY_INIT_PASS 0xb8 |
190 | #define DEV_ENTRY_EINT_PASS 0xb9 | |
191 | #define DEV_ENTRY_NMI_PASS 0xba | |
192 | #define DEV_ENTRY_LINT0_PASS 0xbe | |
193 | #define DEV_ENTRY_LINT1_PASS 0xbf | |
38ddf41b JR |
194 | #define DEV_ENTRY_MODE_MASK 0x07 |
195 | #define DEV_ENTRY_MODE_SHIFT 0x09 | |
8d283c35 | 196 | |
7ef2798d JR |
197 | #define MAX_DEV_TABLE_ENTRIES 0xffff |
198 | ||
8d283c35 JR |
199 | /* constants to configure the command buffer */ |
200 | #define CMD_BUFFER_SIZE 8192 | |
549c90dc | 201 | #define CMD_BUFFER_UNINITIALIZED 1 |
8d283c35 JR |
202 | #define CMD_BUFFER_ENTRIES 512 |
203 | #define MMIO_CMD_SIZE_SHIFT 56 | |
204 | #define MMIO_CMD_SIZE_512 (0x9ULL << MMIO_CMD_SIZE_SHIFT) | |
205 | ||
335503e5 JR |
206 | /* constants for event buffer handling */ |
207 | #define EVT_BUFFER_SIZE 8192 /* 512 entries */ | |
208 | #define EVT_LEN_MASK (0x9ULL << 56) | |
209 | ||
1a29ac01 JR |
210 | /* Constants for PPR Log handling */ |
211 | #define PPR_LOG_ENTRIES 512 | |
212 | #define PPR_LOG_SIZE_SHIFT 56 | |
213 | #define PPR_LOG_SIZE_512 (0x9ULL << PPR_LOG_SIZE_SHIFT) | |
214 | #define PPR_ENTRY_SIZE 16 | |
215 | #define PPR_LOG_SIZE (PPR_ENTRY_SIZE * PPR_LOG_ENTRIES) | |
216 | ||
72e1dcc4 JR |
217 | #define PPR_REQ_TYPE(x) (((x) >> 60) & 0xfULL) |
218 | #define PPR_FLAGS(x) (((x) >> 48) & 0xfffULL) | |
219 | #define PPR_DEVID(x) ((x) & 0xffffULL) | |
220 | #define PPR_TAG(x) (((x) >> 32) & 0x3ffULL) | |
221 | #define PPR_PASID1(x) (((x) >> 16) & 0xffffULL) | |
222 | #define PPR_PASID2(x) (((x) >> 42) & 0xfULL) | |
223 | #define PPR_PASID(x) ((PPR_PASID2(x) << 16) | PPR_PASID1(x)) | |
224 | ||
225 | #define PPR_REQ_FAULT 0x01 | |
226 | ||
0feae533 | 227 | #define PAGE_MODE_NONE 0x00 |
8d283c35 JR |
228 | #define PAGE_MODE_1_LEVEL 0x01 |
229 | #define PAGE_MODE_2_LEVEL 0x02 | |
230 | #define PAGE_MODE_3_LEVEL 0x03 | |
9355a081 JR |
231 | #define PAGE_MODE_4_LEVEL 0x04 |
232 | #define PAGE_MODE_5_LEVEL 0x05 | |
233 | #define PAGE_MODE_6_LEVEL 0x06 | |
8d283c35 | 234 | |
9355a081 JR |
235 | #define PM_LEVEL_SHIFT(x) (12 + ((x) * 9)) |
236 | #define PM_LEVEL_SIZE(x) (((x) < 6) ? \ | |
237 | ((1ULL << PM_LEVEL_SHIFT((x))) - 1): \ | |
238 | (0xffffffffffffffffULL)) | |
239 | #define PM_LEVEL_INDEX(x, a) (((a) >> PM_LEVEL_SHIFT((x))) & 0x1ffULL) | |
50020fb6 JR |
240 | #define PM_LEVEL_ENC(x) (((x) << 9) & 0xe00ULL) |
241 | #define PM_LEVEL_PDE(x, a) ((a) | PM_LEVEL_ENC((x)) | \ | |
242 | IOMMU_PTE_P | IOMMU_PTE_IR | IOMMU_PTE_IW) | |
a6b256b4 | 243 | #define PM_PTE_LEVEL(pte) (((pte) >> 9) & 0x7ULL) |
50020fb6 | 244 | |
abdc5eb3 JR |
245 | #define PM_MAP_4k 0 |
246 | #define PM_ADDR_MASK 0x000ffffffffff000ULL | |
247 | #define PM_MAP_MASK(lvl) (PM_ADDR_MASK & \ | |
248 | (~((1ULL << (12 + ((lvl) * 9))) - 1))) | |
249 | #define PM_ALIGNED(lvl, addr) ((PM_MAP_MASK(lvl) & (addr)) == (addr)) | |
8d283c35 | 250 | |
cbb9d729 JR |
251 | /* |
252 | * Returns the page table level to use for a given page size | |
253 | * Pagesize is expected to be a power-of-two | |
254 | */ | |
255 | #define PAGE_SIZE_LEVEL(pagesize) \ | |
256 | ((__ffs(pagesize) - 12) / 9) | |
257 | /* | |
258 | * Returns the number of ptes to use for a given page size | |
259 | * Pagesize is expected to be a power-of-two | |
260 | */ | |
261 | #define PAGE_SIZE_PTE_COUNT(pagesize) \ | |
262 | (1ULL << ((__ffs(pagesize) - 12) % 9)) | |
263 | ||
264 | /* | |
265 | * Aligns a given io-virtual address to a given page size | |
266 | * Pagesize is expected to be a power-of-two | |
267 | */ | |
268 | #define PAGE_SIZE_ALIGN(address, pagesize) \ | |
269 | ((address) & ~((pagesize) - 1)) | |
270 | /* | |
df805abb | 271 | * Creates an IOMMU PTE for an address and a given pagesize |
cbb9d729 JR |
272 | * The PTE has no permission bits set |
273 | * Pagesize is expected to be a power-of-two larger than 4096 | |
274 | */ | |
275 | #define PAGE_SIZE_PTE(address, pagesize) \ | |
276 | (((address) | ((pagesize) - 1)) & \ | |
277 | (~(pagesize >> 1)) & PM_ADDR_MASK) | |
278 | ||
24cd7723 JR |
279 | /* |
280 | * Takes a PTE value with mode=0x07 and returns the page size it maps | |
281 | */ | |
282 | #define PTE_PAGE_SIZE(pte) \ | |
283 | (1ULL << (1 + ffz(((pte) | 0xfffULL)))) | |
284 | ||
8d283c35 | 285 | #define IOMMU_PTE_P (1ULL << 0) |
38ddf41b | 286 | #define IOMMU_PTE_TV (1ULL << 1) |
8d283c35 JR |
287 | #define IOMMU_PTE_U (1ULL << 59) |
288 | #define IOMMU_PTE_FC (1ULL << 60) | |
289 | #define IOMMU_PTE_IR (1ULL << 61) | |
290 | #define IOMMU_PTE_IW (1ULL << 62) | |
291 | ||
ee6c2868 | 292 | #define DTE_FLAG_IOTLB (0x01UL << 32) |
52815b75 JR |
293 | #define DTE_FLAG_GV (0x01ULL << 55) |
294 | #define DTE_GLX_SHIFT (56) | |
295 | #define DTE_GLX_MASK (3) | |
296 | ||
297 | #define DTE_GCR3_VAL_A(x) (((x) >> 12) & 0x00007ULL) | |
298 | #define DTE_GCR3_VAL_B(x) (((x) >> 15) & 0x0ffffULL) | |
299 | #define DTE_GCR3_VAL_C(x) (((x) >> 31) & 0xfffffULL) | |
300 | ||
301 | #define DTE_GCR3_INDEX_A 0 | |
302 | #define DTE_GCR3_INDEX_B 1 | |
303 | #define DTE_GCR3_INDEX_C 1 | |
304 | ||
305 | #define DTE_GCR3_SHIFT_A 58 | |
306 | #define DTE_GCR3_SHIFT_B 16 | |
307 | #define DTE_GCR3_SHIFT_C 43 | |
308 | ||
b16137b1 | 309 | #define GCR3_VALID 0x01ULL |
fd7b5535 | 310 | |
8d283c35 JR |
311 | #define IOMMU_PAGE_MASK (((1ULL << 52) - 1) & ~0xfffULL) |
312 | #define IOMMU_PTE_PRESENT(pte) ((pte) & IOMMU_PTE_P) | |
313 | #define IOMMU_PTE_PAGE(pte) (phys_to_virt((pte) & IOMMU_PAGE_MASK)) | |
314 | #define IOMMU_PTE_MODE(pte) (((pte) >> 9) & 0x07) | |
315 | ||
316 | #define IOMMU_PROT_MASK 0x03 | |
317 | #define IOMMU_PROT_IR 0x01 | |
318 | #define IOMMU_PROT_IW 0x02 | |
319 | ||
320 | /* IOMMU capabilities */ | |
321 | #define IOMMU_CAP_IOTLB 24 | |
322 | #define IOMMU_CAP_NPCACHE 26 | |
d99ddec3 | 323 | #define IOMMU_CAP_EFR 27 |
8d283c35 JR |
324 | |
325 | #define MAX_DOMAIN_ID 65536 | |
326 | ||
9fdb19d6 JR |
327 | /* Protection domain flags */ |
328 | #define PD_DMA_OPS_MASK (1UL << 0) /* domain used for dma_ops */ | |
e2dc14a2 JR |
329 | #define PD_DEFAULT_MASK (1UL << 1) /* domain is a default dma_ops |
330 | domain for an IOMMU */ | |
0feae533 JR |
331 | #define PD_PASSTHROUGH_MASK (1UL << 2) /* domain has no page |
332 | translation */ | |
52815b75 | 333 | #define PD_IOMMUV2_MASK (1UL << 3) /* domain has gcr3 table */ |
0feae533 | 334 | |
fefda117 JR |
335 | extern bool amd_iommu_dump; |
336 | #define DUMP_printk(format, arg...) \ | |
337 | do { \ | |
338 | if (amd_iommu_dump) \ | |
4c6f40d4 | 339 | printk(KERN_INFO "AMD-Vi: " format, ## arg); \ |
fefda117 | 340 | } while(0); |
9fdb19d6 | 341 | |
318afd41 JR |
342 | /* global flag if IOMMUs cache non-present entries */ |
343 | extern bool amd_iommu_np_cache; | |
60f723b4 JR |
344 | /* Only true if all IOMMUs support device IOTLBs */ |
345 | extern bool amd_iommu_iotlb_sup; | |
318afd41 | 346 | |
05152a04 JR |
347 | #define MAX_IRQS_PER_TABLE 256 |
348 | #define IRQ_TABLE_ALIGNMENT 128 | |
349 | ||
0ea2c422 JR |
350 | struct irq_remap_table { |
351 | spinlock_t lock; | |
352 | unsigned min_index; | |
353 | u32 *table; | |
354 | }; | |
355 | ||
356 | extern struct irq_remap_table **irq_lookup_table; | |
357 | ||
05152a04 JR |
358 | /* Interrupt remapping feature used? */ |
359 | extern bool amd_iommu_irq_remap; | |
360 | ||
361 | /* kmem_cache to get tables with 128 byte alignement */ | |
362 | extern struct kmem_cache *amd_iommu_irq_cache; | |
363 | ||
3bd22172 JR |
364 | /* |
365 | * Make iterating over all IOMMUs easier | |
366 | */ | |
367 | #define for_each_iommu(iommu) \ | |
368 | list_for_each_entry((iommu), &amd_iommu_list, list) | |
369 | #define for_each_iommu_safe(iommu, next) \ | |
370 | list_for_each_entry_safe((iommu), (next), &amd_iommu_list, list) | |
371 | ||
384de729 JR |
372 | #define APERTURE_RANGE_SHIFT 27 /* 128 MB */ |
373 | #define APERTURE_RANGE_SIZE (1ULL << APERTURE_RANGE_SHIFT) | |
374 | #define APERTURE_RANGE_PAGES (APERTURE_RANGE_SIZE >> PAGE_SHIFT) | |
375 | #define APERTURE_MAX_RANGES 32 /* allows 4GB of DMA address space */ | |
376 | #define APERTURE_RANGE_INDEX(a) ((a) >> APERTURE_RANGE_SHIFT) | |
377 | #define APERTURE_PAGE_INDEX(a) (((a) >> 21) & 0x3fULL) | |
9fdb19d6 | 378 | |
72e1dcc4 JR |
379 | |
380 | /* | |
381 | * This struct is used to pass information about | |
382 | * incoming PPR faults around. | |
383 | */ | |
384 | struct amd_iommu_fault { | |
385 | u64 address; /* IO virtual address of the fault*/ | |
386 | u32 pasid; /* Address space identifier */ | |
387 | u16 device_id; /* Originating PCI device id */ | |
388 | u16 tag; /* PPR tag */ | |
389 | u16 flags; /* Fault flags */ | |
390 | ||
391 | }; | |
392 | ||
393 | #define PPR_FAULT_EXEC (1 << 1) | |
394 | #define PPR_FAULT_READ (1 << 2) | |
395 | #define PPR_FAULT_WRITE (1 << 5) | |
396 | #define PPR_FAULT_USER (1 << 6) | |
397 | #define PPR_FAULT_RSVD (1 << 7) | |
398 | #define PPR_FAULT_GN (1 << 8) | |
399 | ||
f3572db8 JR |
400 | struct iommu_domain; |
401 | ||
5694703f JR |
402 | /* |
403 | * This structure contains generic data for IOMMU protection domains | |
404 | * independent of their use. | |
405 | */ | |
8d283c35 | 406 | struct protection_domain { |
aeb26f55 | 407 | struct list_head list; /* for list of all protection domains */ |
7c392cbe | 408 | struct list_head dev_list; /* List of all devices in this domain */ |
9fdb19d6 | 409 | spinlock_t lock; /* mostly used to lock the page table*/ |
5d214fe6 | 410 | struct mutex api_lock; /* protect page tables in the iommu-api path */ |
9fdb19d6 JR |
411 | u16 id; /* the domain id written to the device table */ |
412 | int mode; /* paging mode (0-6 levels) */ | |
413 | u64 *pt_root; /* page table root pointer */ | |
52815b75 JR |
414 | int glx; /* Number of levels for GCR3 table */ |
415 | u64 *gcr3_tbl; /* Guest CR3 table */ | |
9fdb19d6 | 416 | unsigned long flags; /* flags to find out type of domain */ |
04bfdd84 | 417 | bool updated; /* complete domain flush required */ |
863c74eb | 418 | unsigned dev_cnt; /* devices assigned to this domain */ |
c4596114 | 419 | unsigned dev_iommu[MAX_IOMMUS]; /* per-IOMMU reference count */ |
9fdb19d6 | 420 | void *priv; /* private data */ |
f3572db8 JR |
421 | struct iommu_domain *iommu_domain; /* Pointer to generic |
422 | domain structure */ | |
c4596114 | 423 | |
8d283c35 JR |
424 | }; |
425 | ||
657cbb6b JR |
426 | /* |
427 | * This struct contains device specific data for the IOMMU | |
428 | */ | |
429 | struct iommu_dev_data { | |
7c392cbe | 430 | struct list_head list; /* For domain->dev_list */ |
8fa5f802 | 431 | struct list_head dev_data_list; /* For global dev_data_list */ |
71f77580 | 432 | struct iommu_dev_data *alias_data;/* The alias dev_data */ |
657cbb6b | 433 | struct protection_domain *domain; /* Domain the device is bound to */ |
df805abb | 434 | atomic_t bind; /* Domain attach reference count */ |
78bfa9f3 | 435 | struct iommu_group *group; /* IOMMU group for virtual aliases */ |
f62dda66 | 436 | u16 devid; /* PCI Device ID */ |
5abcdba4 JR |
437 | bool iommu_v2; /* Device can make use of IOMMUv2 */ |
438 | bool passthrough; /* Default for device is pt_domain */ | |
ea61cddb JR |
439 | struct { |
440 | bool enabled; | |
441 | int qdep; | |
442 | } ats; /* ATS state */ | |
c99afa25 JR |
443 | bool pri_tlp; /* PASID TLB required for |
444 | PPR completions */ | |
6a113ddc | 445 | u32 errata; /* Bitmap for errata to apply */ |
657cbb6b JR |
446 | }; |
447 | ||
c3239567 JR |
448 | /* |
449 | * For dynamic growth the aperture size is split into ranges of 128MB of | |
450 | * DMA address space each. This struct represents one such range. | |
451 | */ | |
452 | struct aperture_range { | |
453 | ||
454 | /* address allocation bitmap */ | |
455 | unsigned long *bitmap; | |
456 | ||
457 | /* | |
458 | * Array of PTE pages for the aperture. In this array we save all the | |
459 | * leaf pages of the domain page table used for the aperture. This way | |
460 | * we don't need to walk the page table to find a specific PTE. We can | |
461 | * just calculate its address in constant time. | |
462 | */ | |
463 | u64 *pte_pages[64]; | |
384de729 JR |
464 | |
465 | unsigned long offset; | |
c3239567 JR |
466 | }; |
467 | ||
5694703f JR |
468 | /* |
469 | * Data container for a dma_ops specific protection domain | |
470 | */ | |
8d283c35 JR |
471 | struct dma_ops_domain { |
472 | struct list_head list; | |
5694703f JR |
473 | |
474 | /* generic protection domain information */ | |
8d283c35 | 475 | struct protection_domain domain; |
5694703f JR |
476 | |
477 | /* size of the aperture for the mappings */ | |
8d283c35 | 478 | unsigned long aperture_size; |
5694703f JR |
479 | |
480 | /* address we start to search for free addresses */ | |
803b8cb4 | 481 | unsigned long next_address; |
5694703f | 482 | |
c3239567 | 483 | /* address space relevant data */ |
384de729 | 484 | struct aperture_range *aperture[APERTURE_MAX_RANGES]; |
1c655773 JR |
485 | |
486 | /* This will be set to true when TLB needs to be flushed */ | |
487 | bool need_flush; | |
bd60b735 JR |
488 | |
489 | /* | |
490 | * if this is a preallocated domain, keep the device for which it was | |
491 | * preallocated in this variable | |
492 | */ | |
493 | u16 target_dev; | |
8d283c35 JR |
494 | }; |
495 | ||
5694703f JR |
496 | /* |
497 | * Structure where we save information about one hardware AMD IOMMU in the | |
498 | * system. | |
499 | */ | |
8d283c35 JR |
500 | struct amd_iommu { |
501 | struct list_head list; | |
5694703f | 502 | |
bb52777e JR |
503 | /* Index within the IOMMU array */ |
504 | int index; | |
505 | ||
5694703f | 506 | /* locks the accesses to the hardware */ |
8d283c35 JR |
507 | spinlock_t lock; |
508 | ||
3eaf28a1 JR |
509 | /* Pointer to PCI device of this IOMMU */ |
510 | struct pci_dev *dev; | |
511 | ||
c1bf94ec JR |
512 | /* Cache pdev to root device for resume quirks */ |
513 | struct pci_dev *root_pdev; | |
514 | ||
5694703f | 515 | /* physical address of MMIO space */ |
8d283c35 | 516 | u64 mmio_phys; |
30861ddc SK |
517 | |
518 | /* physical end address of MMIO space */ | |
519 | u64 mmio_phys_end; | |
520 | ||
5694703f | 521 | /* virtual address of MMIO space */ |
98f1ad25 | 522 | u8 __iomem *mmio_base; |
5694703f JR |
523 | |
524 | /* capabilities of that IOMMU read from ACPI */ | |
8d283c35 | 525 | u32 cap; |
5694703f | 526 | |
e9bf5197 JR |
527 | /* flags read from acpi table */ |
528 | u8 acpi_flags; | |
529 | ||
d99ddec3 JR |
530 | /* Extended features */ |
531 | u64 features; | |
532 | ||
400a28a0 JR |
533 | /* IOMMUv2 */ |
534 | bool is_iommu_v2; | |
535 | ||
23c742db JR |
536 | /* PCI device id of the IOMMU device */ |
537 | u16 devid; | |
538 | ||
eac9fbc6 RK |
539 | /* |
540 | * Capability pointer. There could be more than one IOMMU per PCI | |
541 | * device function if there are more than one AMD IOMMU capability | |
542 | * pointers. | |
543 | */ | |
544 | u16 cap_ptr; | |
545 | ||
ee893c24 JR |
546 | /* pci domain of this IOMMU */ |
547 | u16 pci_seg; | |
548 | ||
5694703f | 549 | /* first device this IOMMU handles. read from PCI */ |
8d283c35 | 550 | u16 first_device; |
5694703f | 551 | /* last device this IOMMU handles. read from PCI */ |
8d283c35 | 552 | u16 last_device; |
5694703f JR |
553 | |
554 | /* start of exclusion range of that IOMMU */ | |
8d283c35 | 555 | u64 exclusion_start; |
5694703f | 556 | /* length of exclusion range of that IOMMU */ |
8d283c35 JR |
557 | u64 exclusion_length; |
558 | ||
5694703f | 559 | /* command buffer virtual address */ |
8d283c35 | 560 | u8 *cmd_buf; |
5694703f | 561 | /* size of command buffer */ |
8d283c35 JR |
562 | u32 cmd_buf_size; |
563 | ||
335503e5 JR |
564 | /* size of event buffer */ |
565 | u32 evt_buf_size; | |
eac9fbc6 RK |
566 | /* event buffer virtual address */ |
567 | u8 *evt_buf; | |
335503e5 | 568 | |
1a29ac01 JR |
569 | /* Base of the PPR log, if present */ |
570 | u8 *ppr_log; | |
571 | ||
a80dc3e0 JR |
572 | /* true if interrupts for this IOMMU are already enabled */ |
573 | bool int_enabled; | |
574 | ||
eac9fbc6 | 575 | /* if one, we need to send a completion wait command */ |
0cfd7aa9 | 576 | bool need_sync; |
eac9fbc6 | 577 | |
5694703f | 578 | /* default dma_ops domain for that IOMMU */ |
8d283c35 | 579 | struct dma_ops_domain *default_dom; |
4c894f47 JR |
580 | |
581 | /* | |
5bcd757f MG |
582 | * We can't rely on the BIOS to restore all values on reinit, so we |
583 | * need to stash them | |
4c894f47 | 584 | */ |
5bcd757f MG |
585 | |
586 | /* The iommu BAR */ | |
587 | u32 stored_addr_lo; | |
588 | u32 stored_addr_hi; | |
589 | ||
590 | /* | |
591 | * Each iommu has 6 l1s, each of which is documented as having 0x12 | |
592 | * registers | |
593 | */ | |
594 | u32 stored_l1[6][0x12]; | |
595 | ||
596 | /* The l2 indirect registers */ | |
597 | u32 stored_l2[0x83]; | |
30861ddc SK |
598 | |
599 | /* The maximum PC banks and counters/bank (PCSup=1) */ | |
600 | u8 max_banks; | |
601 | u8 max_counters; | |
8d283c35 JR |
602 | }; |
603 | ||
6efed63b JR |
604 | struct devid_map { |
605 | struct list_head list; | |
606 | u8 id; | |
607 | u16 devid; | |
31cff67f | 608 | bool cmd_line; |
6efed63b JR |
609 | }; |
610 | ||
611 | /* Map HPET and IOAPIC ids to the devid used by the IOMMU */ | |
612 | extern struct list_head ioapic_map; | |
613 | extern struct list_head hpet_map; | |
614 | ||
5694703f JR |
615 | /* |
616 | * List with all IOMMUs in the system. This list is not locked because it is | |
617 | * only written and read at driver initialization or suspend time | |
618 | */ | |
8d283c35 JR |
619 | extern struct list_head amd_iommu_list; |
620 | ||
bb52777e JR |
621 | /* |
622 | * Array with pointers to each IOMMU struct | |
623 | * The indices are referenced in the protection domains | |
624 | */ | |
625 | extern struct amd_iommu *amd_iommus[MAX_IOMMUS]; | |
626 | ||
627 | /* Number of IOMMUs present in the system */ | |
628 | extern int amd_iommus_present; | |
629 | ||
aeb26f55 JR |
630 | /* |
631 | * Declarations for the global list of all protection domains | |
632 | */ | |
633 | extern spinlock_t amd_iommu_pd_lock; | |
634 | extern struct list_head amd_iommu_pd_list; | |
635 | ||
5694703f JR |
636 | /* |
637 | * Structure defining one entry in the device table | |
638 | */ | |
8d283c35 | 639 | struct dev_table_entry { |
ee6c2868 | 640 | u64 data[4]; |
8d283c35 JR |
641 | }; |
642 | ||
5694703f JR |
643 | /* |
644 | * One entry for unity mappings parsed out of the ACPI table. | |
645 | */ | |
8d283c35 JR |
646 | struct unity_map_entry { |
647 | struct list_head list; | |
5694703f JR |
648 | |
649 | /* starting device id this entry is used for (including) */ | |
8d283c35 | 650 | u16 devid_start; |
5694703f | 651 | /* end device id this entry is used for (including) */ |
8d283c35 | 652 | u16 devid_end; |
5694703f JR |
653 | |
654 | /* start address to unity map (including) */ | |
8d283c35 | 655 | u64 address_start; |
5694703f | 656 | /* end address to unity map (including) */ |
8d283c35 | 657 | u64 address_end; |
5694703f JR |
658 | |
659 | /* required protection */ | |
8d283c35 JR |
660 | int prot; |
661 | }; | |
662 | ||
5694703f JR |
663 | /* |
664 | * List of all unity mappings. It is not locked because as runtime it is only | |
665 | * read. It is created at ACPI table parsing time. | |
666 | */ | |
8d283c35 JR |
667 | extern struct list_head amd_iommu_unity_map; |
668 | ||
5694703f JR |
669 | /* |
670 | * Data structures for device handling | |
671 | */ | |
672 | ||
673 | /* | |
674 | * Device table used by hardware. Read and write accesses by software are | |
675 | * locked with the amd_iommu_pd_table lock. | |
676 | */ | |
8d283c35 | 677 | extern struct dev_table_entry *amd_iommu_dev_table; |
5694703f JR |
678 | |
679 | /* | |
680 | * Alias table to find requestor ids to device ids. Not locked because only | |
681 | * read on runtime. | |
682 | */ | |
8d283c35 | 683 | extern u16 *amd_iommu_alias_table; |
5694703f JR |
684 | |
685 | /* | |
686 | * Reverse lookup table to find the IOMMU which translates a specific device. | |
687 | */ | |
8d283c35 JR |
688 | extern struct amd_iommu **amd_iommu_rlookup_table; |
689 | ||
5694703f | 690 | /* size of the dma_ops aperture as power of 2 */ |
8d283c35 JR |
691 | extern unsigned amd_iommu_aperture_order; |
692 | ||
5694703f | 693 | /* largest PCI device id we expect translation requests for */ |
8d283c35 JR |
694 | extern u16 amd_iommu_last_bdf; |
695 | ||
5694703f | 696 | /* allocation bitmap for domain ids */ |
8d283c35 JR |
697 | extern unsigned long *amd_iommu_pd_alloc_bitmap; |
698 | ||
afa9fdc2 FT |
699 | /* |
700 | * If true, the addresses will be flushed on unmap time, not when | |
701 | * they are reused | |
702 | */ | |
3775d481 | 703 | extern u32 amd_iommu_unmap_flush; |
afa9fdc2 | 704 | |
a919a018 SS |
705 | /* Smallest max PASID supported by any IOMMU in the system */ |
706 | extern u32 amd_iommu_max_pasid; | |
62f71abb | 707 | |
400a28a0 JR |
708 | extern bool amd_iommu_v2_present; |
709 | ||
5abcdba4 JR |
710 | extern bool amd_iommu_force_isolation; |
711 | ||
52815b75 JR |
712 | /* Max levels of glxval supported */ |
713 | extern int amd_iommu_max_glx_val; | |
714 | ||
98f1ad25 JR |
715 | /* |
716 | * This function flushes all internal caches of | |
717 | * the IOMMU used by this driver. | |
718 | */ | |
719 | extern void iommu_flush_all_caches(struct amd_iommu *iommu); | |
720 | ||
6efed63b JR |
721 | static inline int get_ioapic_devid(int id) |
722 | { | |
723 | struct devid_map *entry; | |
724 | ||
725 | list_for_each_entry(entry, &ioapic_map, list) { | |
726 | if (entry->id == id) | |
727 | return entry->devid; | |
728 | } | |
729 | ||
730 | return -EINVAL; | |
731 | } | |
732 | ||
733 | static inline int get_hpet_devid(int id) | |
734 | { | |
735 | struct devid_map *entry; | |
736 | ||
737 | list_for_each_entry(entry, &hpet_map, list) { | |
738 | if (entry->id == id) | |
739 | return entry->devid; | |
740 | } | |
741 | ||
742 | return -EINVAL; | |
743 | } | |
744 | ||
a9dddbe0 JR |
745 | #ifdef CONFIG_AMD_IOMMU_STATS |
746 | ||
747 | struct __iommu_counter { | |
748 | char *name; | |
749 | struct dentry *dent; | |
750 | u64 value; | |
751 | }; | |
752 | ||
753 | #define DECLARE_STATS_COUNTER(nm) \ | |
754 | static struct __iommu_counter nm = { \ | |
755 | .name = #nm, \ | |
756 | } | |
757 | ||
758 | #define INC_STATS_COUNTER(name) name.value += 1 | |
759 | #define ADD_STATS_COUNTER(name, x) name.value += (x) | |
760 | #define SUB_STATS_COUNTER(name, x) name.value -= (x) | |
761 | ||
762 | #else /* CONFIG_AMD_IOMMU_STATS */ | |
763 | ||
764 | #define DECLARE_STATS_COUNTER(name) | |
765 | #define INC_STATS_COUNTER(name) | |
766 | #define ADD_STATS_COUNTER(name, x) | |
767 | #define SUB_STATS_COUNTER(name, x) | |
768 | ||
769 | #endif /* CONFIG_AMD_IOMMU_STATS */ | |
770 | ||
1965aae3 | 771 | #endif /* _ASM_X86_AMD_IOMMU_TYPES_H */ |