PCI: hisi: Fix deferred probing
[deliverable/linux.git] / arch / arm / kernel / bios32.c
CommitLineData
1da177e4
LT
1/*
2 * linux/arch/arm/kernel/bios32.c
3 *
4 * PCI bios-type initialisation for PCI machines
5 *
6 * Bits taken from various places.
7 */
ecea4ab6 8#include <linux/export.h>
1da177e4
LT
9#include <linux/kernel.h>
10#include <linux/pci.h>
11#include <linux/slab.h>
12#include <linux/init.h>
fced80c7 13#include <linux/io.h>
1da177e4 14
1da177e4 15#include <asm/mach-types.h>
c2794437 16#include <asm/mach/map.h>
1da177e4
LT
17#include <asm/mach/pci.h>
18
19static int debug_pci;
b3a72384
GP
20static resource_size_t (*align_resource)(struct pci_dev *dev,
21 const struct resource *res,
22 resource_size_t start,
23 resource_size_t size,
24 resource_size_t align) = NULL;
1da177e4
LT
25
26/*
5ee09efe 27 * We can't use pci_get_device() here since we are
1da177e4
LT
28 * called from interrupt context.
29 */
30static void pcibios_bus_report_status(struct pci_bus *bus, u_int status_mask, int warn)
31{
32 struct pci_dev *dev;
33
34 list_for_each_entry(dev, &bus->devices, bus_list) {
35 u16 status;
36
37 /*
38 * ignore host bridge - we handle
39 * that separately
40 */
41 if (dev->bus->number == 0 && dev->devfn == 0)
42 continue;
43
44 pci_read_config_word(dev, PCI_STATUS, &status);
45 if (status == 0xffff)
46 continue;
47
48 if ((status & status_mask) == 0)
49 continue;
50
51 /* clear the status errors */
52 pci_write_config_word(dev, PCI_STATUS, status & status_mask);
53
54 if (warn)
55 printk("(%s: %04X) ", pci_name(dev), status);
56 }
57
58 list_for_each_entry(dev, &bus->devices, bus_list)
59 if (dev->subordinate)
60 pcibios_bus_report_status(dev->subordinate, status_mask, warn);
61}
62
63void pcibios_report_status(u_int status_mask, int warn)
64{
f0197e0c 65 struct pci_bus *bus;
1da177e4 66
f0197e0c 67 list_for_each_entry(bus, &pci_root_buses, node)
1da177e4 68 pcibios_bus_report_status(bus, status_mask, warn);
1da177e4
LT
69}
70
71/*
72 * We don't use this to fix the device, but initialisation of it.
73 * It's not the correct use for this, but it works.
74 * Note that the arbiter/ISA bridge appears to be buggy, specifically in
75 * the following area:
76 * 1. park on CPU
77 * 2. ISA bridge ping-pong
78 * 3. ISA bridge master handling of target RETRY
79 *
80 * Bug 3 is responsible for the sound DMA grinding to a halt. We now
81 * live with bug 2.
82 */
351a102d 83static void pci_fixup_83c553(struct pci_dev *dev)
1da177e4
LT
84{
85 /*
86 * Set memory region to start at address 0, and enable IO
87 */
88 pci_write_config_dword(dev, PCI_BASE_ADDRESS_0, PCI_BASE_ADDRESS_SPACE_MEMORY);
89 pci_write_config_word(dev, PCI_COMMAND, PCI_COMMAND_IO);
90
91 dev->resource[0].end -= dev->resource[0].start;
92 dev->resource[0].start = 0;
93
94 /*
95 * All memory requests from ISA to be channelled to PCI
96 */
97 pci_write_config_byte(dev, 0x48, 0xff);
98
99 /*
100 * Enable ping-pong on bus master to ISA bridge transactions.
101 * This improves the sound DMA substantially. The fixed
102 * priority arbiter also helps (see below).
103 */
104 pci_write_config_byte(dev, 0x42, 0x01);
105
106 /*
107 * Enable PCI retry
108 */
109 pci_write_config_byte(dev, 0x40, 0x22);
110
111 /*
112 * We used to set the arbiter to "park on last master" (bit
113 * 1 set), but unfortunately the CyberPro does not park the
114 * bus. We must therefore park on CPU. Unfortunately, this
115 * may trigger yet another bug in the 553.
116 */
117 pci_write_config_byte(dev, 0x83, 0x02);
118
119 /*
120 * Make the ISA DMA request lowest priority, and disable
121 * rotating priorities completely.
122 */
123 pci_write_config_byte(dev, 0x80, 0x11);
124 pci_write_config_byte(dev, 0x81, 0x00);
125
126 /*
127 * Route INTA input to IRQ 11, and set IRQ11 to be level
128 * sensitive.
129 */
130 pci_write_config_word(dev, 0x44, 0xb000);
131 outb(0x08, 0x4d1);
132}
133DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_WINBOND, PCI_DEVICE_ID_WINBOND_83C553, pci_fixup_83c553);
134
351a102d 135static void pci_fixup_unassign(struct pci_dev *dev)
1da177e4
LT
136{
137 dev->resource[0].end -= dev->resource[0].start;
138 dev->resource[0].start = 0;
139}
140DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_WINBOND2, PCI_DEVICE_ID_WINBOND2_89C940F, pci_fixup_unassign);
141
142/*
143 * Prevent the PCI layer from seeing the resources allocated to this device
144 * if it is the host bridge by marking it as such. These resources are of
145 * no consequence to the PCI layer (they are handled elsewhere).
146 */
351a102d 147static void pci_fixup_dec21285(struct pci_dev *dev)
1da177e4
LT
148{
149 int i;
150
151 if (dev->devfn == 0) {
152 dev->class &= 0xff;
153 dev->class |= PCI_CLASS_BRIDGE_HOST << 8;
154 for (i = 0; i < PCI_NUM_RESOURCES; i++) {
155 dev->resource[i].start = 0;
156 dev->resource[i].end = 0;
157 dev->resource[i].flags = 0;
158 }
159 }
160}
161DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_21285, pci_fixup_dec21285);
162
1da177e4
LT
163/*
164 * PCI IDE controllers use non-standard I/O port decoding, respect it.
165 */
351a102d 166static void pci_fixup_ide_bases(struct pci_dev *dev)
1da177e4
LT
167{
168 struct resource *r;
169 int i;
170
171 if ((dev->class >> 8) != PCI_CLASS_STORAGE_IDE)
172 return;
173
174 for (i = 0; i < PCI_NUM_RESOURCES; i++) {
175 r = dev->resource + i;
176 if ((r->start & ~0x80) == 0x374) {
177 r->start |= 2;
178 r->end = r->start;
179 }
180 }
181}
182DECLARE_PCI_FIXUP_HEADER(PCI_ANY_ID, PCI_ANY_ID, pci_fixup_ide_bases);
183
184/*
185 * Put the DEC21142 to sleep
186 */
351a102d 187static void pci_fixup_dec21142(struct pci_dev *dev)
1da177e4
LT
188{
189 pci_write_config_dword(dev, 0x40, 0x80000000);
190}
191DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_21142, pci_fixup_dec21142);
192
193/*
194 * The CY82C693 needs some rather major fixups to ensure that it does
195 * the right thing. Idea from the Alpha people, with a few additions.
196 *
197 * We ensure that the IDE base registers are set to 1f0/3f4 for the
198 * primary bus, and 170/374 for the secondary bus. Also, hide them
199 * from the PCI subsystem view as well so we won't try to perform
200 * our own auto-configuration on them.
201 *
202 * In addition, we ensure that the PCI IDE interrupts are routed to
203 * IRQ 14 and IRQ 15 respectively.
204 *
205 * The above gets us to a point where the IDE on this device is
206 * functional. However, The CY82C693U _does not work_ in bus
207 * master mode without locking the PCI bus solid.
208 */
351a102d 209static void pci_fixup_cy82c693(struct pci_dev *dev)
1da177e4
LT
210{
211 if ((dev->class >> 8) == PCI_CLASS_STORAGE_IDE) {
212 u32 base0, base1;
213
214 if (dev->class & 0x80) { /* primary */
215 base0 = 0x1f0;
216 base1 = 0x3f4;
217 } else { /* secondary */
218 base0 = 0x170;
219 base1 = 0x374;
220 }
221
222 pci_write_config_dword(dev, PCI_BASE_ADDRESS_0,
223 base0 | PCI_BASE_ADDRESS_SPACE_IO);
224 pci_write_config_dword(dev, PCI_BASE_ADDRESS_1,
225 base1 | PCI_BASE_ADDRESS_SPACE_IO);
226
227 dev->resource[0].start = 0;
228 dev->resource[0].end = 0;
229 dev->resource[0].flags = 0;
230
231 dev->resource[1].start = 0;
232 dev->resource[1].end = 0;
233 dev->resource[1].flags = 0;
234 } else if (PCI_FUNC(dev->devfn) == 0) {
235 /*
236 * Setup IDE IRQ routing.
237 */
238 pci_write_config_byte(dev, 0x4b, 14);
239 pci_write_config_byte(dev, 0x4c, 15);
240
241 /*
242 * Disable FREQACK handshake, enable USB.
243 */
244 pci_write_config_byte(dev, 0x4d, 0x41);
245
246 /*
247 * Enable PCI retry, and PCI post-write buffer.
248 */
249 pci_write_config_byte(dev, 0x44, 0x17);
250
251 /*
252 * Enable ISA master and DMA post write buffering.
253 */
254 pci_write_config_byte(dev, 0x45, 0x03);
255 }
256}
257DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_CONTAQ, PCI_DEVICE_ID_CONTAQ_82C693, pci_fixup_cy82c693);
258
351a102d 259static void pci_fixup_it8152(struct pci_dev *dev)
a8fc0789
MR
260{
261 int i;
262 /* fixup for ITE 8152 devices */
263 /* FIXME: add defines for class 0x68000 and 0x80103 */
264 if ((dev->class >> 8) == PCI_CLASS_BRIDGE_HOST ||
265 dev->class == 0x68000 ||
266 dev->class == 0x80103) {
267 for (i = 0; i < PCI_NUM_RESOURCES; i++) {
268 dev->resource[i].start = 0;
269 dev->resource[i].end = 0;
270 dev->resource[i].flags = 0;
271 }
272 }
273}
274DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ITE, PCI_DEVICE_ID_ITE_8152, pci_fixup_it8152);
275
1da177e4
LT
276/*
277 * If the bus contains any of these devices, then we must not turn on
278 * parity checking of any kind. Currently this is CyberPro 20x0 only.
279 */
280static inline int pdev_bad_for_parity(struct pci_dev *dev)
281{
a8fc0789
MR
282 return ((dev->vendor == PCI_VENDOR_ID_INTERG &&
283 (dev->device == PCI_DEVICE_ID_INTERG_2000 ||
284 dev->device == PCI_DEVICE_ID_INTERG_2010)) ||
285 (dev->vendor == PCI_VENDOR_ID_ITE &&
286 dev->device == PCI_DEVICE_ID_ITE_8152));
287
1da177e4
LT
288}
289
1da177e4
LT
290/*
291 * pcibios_fixup_bus - Called after each bus is probed,
292 * but before its children are examined.
293 */
46edfc54 294void pcibios_fixup_bus(struct pci_bus *bus)
1da177e4 295{
1da177e4
LT
296 struct pci_dev *dev;
297 u16 features = PCI_COMMAND_SERR | PCI_COMMAND_PARITY | PCI_COMMAND_FAST_BACK;
298
1da177e4
LT
299 /*
300 * Walk the devices on this bus, working out what we can
301 * and can't support.
302 */
303 list_for_each_entry(dev, &bus->devices, bus_list) {
304 u16 status;
305
1da177e4
LT
306 pci_read_config_word(dev, PCI_STATUS, &status);
307
308 /*
309 * If any device on this bus does not support fast back
310 * to back transfers, then the bus as a whole is not able
311 * to support them. Having fast back to back transfers
312 * on saves us one PCI cycle per transaction.
313 */
314 if (!(status & PCI_STATUS_FAST_BACK))
315 features &= ~PCI_COMMAND_FAST_BACK;
316
317 if (pdev_bad_for_parity(dev))
318 features &= ~(PCI_COMMAND_SERR | PCI_COMMAND_PARITY);
319
320 switch (dev->class >> 8) {
1da177e4
LT
321 case PCI_CLASS_BRIDGE_PCI:
322 pci_read_config_word(dev, PCI_BRIDGE_CONTROL, &status);
323 status |= PCI_BRIDGE_CTL_PARITY|PCI_BRIDGE_CTL_MASTER_ABORT;
324 status &= ~(PCI_BRIDGE_CTL_BUS_RESET|PCI_BRIDGE_CTL_FAST_BACK);
325 pci_write_config_word(dev, PCI_BRIDGE_CONTROL, status);
326 break;
327
328 case PCI_CLASS_BRIDGE_CARDBUS:
329 pci_read_config_word(dev, PCI_CB_BRIDGE_CONTROL, &status);
330 status |= PCI_CB_BRIDGE_CTL_PARITY|PCI_CB_BRIDGE_CTL_MASTER_ABORT;
331 pci_write_config_word(dev, PCI_CB_BRIDGE_CONTROL, status);
332 break;
333 }
334 }
335
336 /*
337 * Now walk the devices again, this time setting them up.
338 */
339 list_for_each_entry(dev, &bus->devices, bus_list) {
340 u16 cmd;
341
342 pci_read_config_word(dev, PCI_COMMAND, &cmd);
343 cmd |= features;
344 pci_write_config_word(dev, PCI_COMMAND, cmd);
345
346 pci_write_config_byte(dev, PCI_CACHE_LINE_SIZE,
347 L1_CACHE_BYTES >> 2);
348 }
349
350 /*
351 * Propagate the flags to the PCI bridge.
352 */
353 if (bus->self && bus->self->hdr_type == PCI_HEADER_TYPE_BRIDGE) {
354 if (features & PCI_COMMAND_FAST_BACK)
355 bus->bridge_ctl |= PCI_BRIDGE_CTL_FAST_BACK;
356 if (features & PCI_COMMAND_PARITY)
357 bus->bridge_ctl |= PCI_BRIDGE_CTL_PARITY;
358 }
359
360 /*
361 * Report what we did for this bus
362 */
4ed89f22 363 pr_info("PCI: bus%d: Fast back to back transfers %sabled\n",
1da177e4
LT
364 bus->number, (features & PCI_COMMAND_FAST_BACK) ? "en" : "dis");
365}
b214bea5 366EXPORT_SYMBOL(pcibios_fixup_bus);
1da177e4 367
1da177e4 368/*
daeb4c0c
RK
369 * Swizzle the device pin each time we cross a bridge. If a platform does
370 * not provide a swizzle function, we perform the standard PCI swizzling.
371 *
372 * The default swizzling walks up the bus tree one level at a time, applying
373 * the standard swizzle function at each step, stopping when it finds the PCI
374 * root bus. This will return the slot number of the bridge device on the
375 * root bus and the interrupt pin on that device which should correspond
376 * with the downstream device interrupt.
377 *
378 * Platforms may override this, in which case the slot and pin returned
379 * depend entirely on the platform code. However, please note that the
380 * PCI standard swizzle is implemented on plug-in cards and Cardbus based
381 * PCI extenders, so it can not be ignored.
1da177e4 382 */
351a102d 383static u8 pcibios_swizzle(struct pci_dev *dev, u8 *pin)
1da177e4
LT
384{
385 struct pci_sys_data *sys = dev->sysdata;
daeb4c0c 386 int slot, oldpin = *pin;
1da177e4
LT
387
388 if (sys->swizzle)
389 slot = sys->swizzle(dev, pin);
daeb4c0c
RK
390 else
391 slot = pci_common_swizzle(dev, pin);
1da177e4
LT
392
393 if (debug_pci)
394 printk("PCI: %s swizzling pin %d => pin %d slot %d\n",
395 pci_name(dev), oldpin, *pin, slot);
396
397 return slot;
398}
399
400/*
401 * Map a slot/pin to an IRQ.
402 */
d5341942 403static int pcibios_map_irq(const struct pci_dev *dev, u8 slot, u8 pin)
1da177e4
LT
404{
405 struct pci_sys_data *sys = dev->sysdata;
406 int irq = -1;
407
408 if (sys->map_irq)
409 irq = sys->map_irq(dev, slot, pin);
410
411 if (debug_pci)
412 printk("PCI: %s mapping slot %d pin %d => irq %d\n",
413 pci_name(dev), slot, pin, irq);
414
415 return irq;
416}
417
276fd3c5 418static int pcibios_init_resources(int busnr, struct pci_sys_data *sys)
3c5d1699
RH
419{
420 int ret;
14d76b68 421 struct resource_entry *window;
3c5d1699
RH
422
423 if (list_empty(&sys->resources)) {
424 pci_add_resource_offset(&sys->resources,
425 &iomem_resource, sys->mem_offset);
426 }
427
14d76b68 428 resource_list_for_each_entry(window, &sys->resources)
3c5d1699
RH
429 if (resource_type(window->res) == IORESOURCE_IO)
430 return 0;
3c5d1699
RH
431
432 sys->io_res.start = (busnr * SZ_64K) ? : pcibios_min_io;
433 sys->io_res.end = (busnr + 1) * SZ_64K - 1;
434 sys->io_res.flags = IORESOURCE_IO;
435 sys->io_res.name = sys->io_res_name;
436 sprintf(sys->io_res_name, "PCI%d I/O", busnr);
437
438 ret = request_resource(&ioport_resource, &sys->io_res);
439 if (ret) {
440 pr_err("PCI: unable to allocate I/O port region (%d)\n", ret);
441 return ret;
442 }
443 pci_add_resource_offset(&sys->resources, &sys->io_res,
444 sys->io_offset);
445
446 return 0;
447}
448
14d86e72
LW
449static void pcibios_init_hw(struct device *parent, struct hw_pci *hw,
450 struct list_head *head)
1da177e4
LT
451{
452 struct pci_sys_data *sys = NULL;
453 int ret;
454 int nr, busnr;
455
456 for (nr = busnr = 0; nr < hw->nr_controllers; nr++) {
d2a02b93 457 sys = kzalloc(sizeof(struct pci_sys_data), GFP_KERNEL);
ada8b675
LP
458 if (WARN(!sys, "PCI: unable to allocate sys data!"))
459 break;
1da177e4 460
1da177e4
LT
461 sys->busnr = busnr;
462 sys->swizzle = hw->swizzle;
463 sys->map_irq = hw->map_irq;
b3a72384 464 align_resource = hw->align_resource;
37d15909 465 INIT_LIST_HEAD(&sys->resources);
1da177e4 466
352af7d4
TR
467 if (hw->private_data)
468 sys->private_data = hw->private_data[nr];
469
1da177e4
LT
470 ret = hw->setup(nr, sys);
471
472 if (ret > 0) {
3c5d1699
RH
473 ret = pcibios_init_resources(nr, sys);
474 if (ret) {
475 kfree(sys);
476 break;
37d15909
BH
477 }
478
c23bfc38
RK
479 if (hw->scan)
480 sys->bus = hw->scan(nr, sys);
481 else
8953aab1
LP
482 sys->bus = pci_scan_root_bus_msi(parent,
483 sys->busnr, hw->ops, sys,
484 &sys->resources, hw->msi_ctrl);
1da177e4 485
ada8b675
LP
486 if (WARN(!sys->bus, "PCI: unable to scan bus!")) {
487 kfree(sys);
488 break;
489 }
1da177e4 490
b918c62e 491 busnr = sys->bus->busn_res.end + 1;
1da177e4 492
90cf2418 493 list_add(&sys->node, head);
1da177e4
LT
494 } else {
495 kfree(sys);
496 if (ret < 0)
497 break;
498 }
499 }
500}
501
14d86e72 502void pci_common_init_dev(struct device *parent, struct hw_pci *hw)
1da177e4
LT
503{
504 struct pci_sys_data *sys;
90cf2418 505 LIST_HEAD(head);
1da177e4 506
6696cbc3 507 pci_add_flags(PCI_REASSIGN_ALL_RSRC);
1da177e4
LT
508 if (hw->preinit)
509 hw->preinit();
14d86e72 510 pcibios_init_hw(parent, hw, &head);
1da177e4
LT
511 if (hw->postinit)
512 hw->postinit();
513
514 pci_fixup_irqs(pcibios_swizzle, pcibios_map_irq);
515
90cf2418 516 list_for_each_entry(sys, &head, node) {
1da177e4
LT
517 struct pci_bus *bus = sys->bus;
518
a4fab044 519 if (!pci_has_flag(PCI_PROBE_ONLY)) {
808b27a5
MK
520 struct pci_bus *child;
521
1da177e4
LT
522 /*
523 * Size the bridge windows.
524 */
525 pci_bus_size_bridges(bus);
526
527 /*
528 * Assign resources.
529 */
530 pci_bus_assign_resources(bus);
1da177e4 531
808b27a5
MK
532 list_for_each_entry(child, &bus->children, node)
533 pcie_bus_configure_settings(child);
534 }
1da177e4
LT
535 /*
536 * Tell drivers about devices found.
537 */
538 pci_bus_add_devices(bus);
539 }
540}
541
168c8619
MS
542#ifndef CONFIG_PCI_HOST_ITE8152
543void pcibios_set_master(struct pci_dev *dev)
544{
545 /* No special bus mastering setup handling */
546}
547#endif
548
1da177e4
LT
549char * __init pcibios_setup(char *str)
550{
551 if (!strcmp(str, "debug")) {
552 debug_pci = 1;
553 return NULL;
554 } else if (!strcmp(str, "firmware")) {
a4fab044 555 pci_add_flags(PCI_PROBE_ONLY);
1da177e4
LT
556 return NULL;
557 }
558 return str;
559}
560
561/*
562 * From arch/i386/kernel/pci-i386.c:
563 *
564 * We need to avoid collisions with `mirrored' VGA ports
565 * and other strange ISA hardware, so we always want the
566 * addresses to be allocated in the 0x000-0x0ff region
567 * modulo 0x400.
568 *
569 * Why? Because some silly external IO cards only decode
570 * the low 10 bits of the IO address. The 0x00-0xff region
571 * is reserved for motherboard devices that decode all 16
572 * bits, so it's ok to allocate at, say, 0x2800-0x28ff,
573 * but we want to try to avoid allocating at 0x2900-0x2bff
574 * which might be mirrored at 0x0100-0x03ff..
575 */
3b7a17fc 576resource_size_t pcibios_align_resource(void *data, const struct resource *res,
b26b2d49 577 resource_size_t size, resource_size_t align)
1da177e4 578{
029baf14 579 struct pci_dev *dev = data;
e31dd6e4 580 resource_size_t start = res->start;
1da177e4
LT
581
582 if (res->flags & IORESOURCE_IO && start & 0x300)
583 start = (start + 0x3ff) & ~0x3ff;
584
b26b2d49
DB
585 start = (start + align - 1) & ~(align - 1);
586
b3a72384
GP
587 if (align_resource)
588 return align_resource(dev, res, start, size, align);
029baf14 589
b26b2d49 590 return start;
1da177e4
LT
591}
592
593/**
594 * pcibios_enable_device - Enable I/O and memory.
595 * @dev: PCI device to be enabled
596 */
597int pcibios_enable_device(struct pci_dev *dev, int mask)
598{
27e8efdb
WD
599 if (pci_has_flag(PCI_PROBE_ONLY))
600 return 0;
1da177e4 601
27e8efdb 602 return pci_enable_resources(dev, mask);
1da177e4
LT
603}
604
605int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma,
606 enum pci_mmap_state mmap_state, int write_combine)
607{
415ae101 608 if (mmap_state == pci_mmap_io)
1da177e4 609 return -EINVAL;
1da177e4
LT
610
611 /*
612 * Mark this as IO
613 */
1da177e4
LT
614 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
615
415ae101 616 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff,
1da177e4
LT
617 vma->vm_end - vma->vm_start,
618 vma->vm_page_prot))
619 return -EAGAIN;
620
621 return 0;
622}
c2794437
RH
623
624void __init pci_map_io_early(unsigned long pfn)
625{
626 struct map_desc pci_io_desc = {
627 .virtual = PCI_IO_VIRT_BASE,
628 .type = MT_DEVICE,
629 .length = SZ_64K,
630 };
631
632 pci_io_desc.pfn = pfn;
633 iotable_init(&pci_io_desc, 1);
634}
This page took 0.694387 seconds and 5 git commands to generate.