1
2
3
4
5
6
7
8
9#include <linux/types.h>
10#include <linux/kernel.h>
11#include <linux/pci.h>
12#include <linux/init.h>
13#include <linux/interrupt.h>
14#include <linux/ide.h>
15#include <linux/dma-mapping.h>
16
17#include <asm/io.h>
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32static int ide_setup_pci_baseregs(struct pci_dev *dev, const char *name)
33{
34 u8 progif = 0;
35
36
37
38
39 if (pci_read_config_byte(dev, PCI_CLASS_PROG, &progif) ||
40 (progif & 5) != 5) {
41 if ((progif & 0xa) != 0xa) {
42 printk(KERN_INFO "%s %s: device not capable of full "
43 "native PCI mode\n", name, pci_name(dev));
44 return -EOPNOTSUPP;
45 }
46 printk(KERN_INFO "%s %s: placing both ports into native PCI "
47 "mode\n", name, pci_name(dev));
48 (void) pci_write_config_byte(dev, PCI_CLASS_PROG, progif|5);
49 if (pci_read_config_byte(dev, PCI_CLASS_PROG, &progif) ||
50 (progif & 5) != 5) {
51 printk(KERN_ERR "%s %s: rewrite of PROGIF failed, "
52 "wanted 0x%04x, got 0x%04x\n",
53 name, pci_name(dev), progif | 5, progif);
54 return -EOPNOTSUPP;
55 }
56 }
57 return 0;
58}
59
60#ifdef CONFIG_BLK_DEV_IDEDMA_PCI
61static int ide_pci_clear_simplex(unsigned long dma_base, const char *name)
62{
63 u8 dma_stat = inb(dma_base + 2);
64
65 outb(dma_stat & 0x60, dma_base + 2);
66 dma_stat = inb(dma_base + 2);
67
68 return (dma_stat & 0x80) ? 1 : 0;
69}
70
71
72
73
74
75
76
77
78
79unsigned long ide_pci_dma_base(ide_hwif_t *hwif, const struct ide_port_info *d)
80{
81 struct pci_dev *dev = to_pci_dev(hwif->dev);
82 unsigned long dma_base = 0;
83
84 if (hwif->host_flags & IDE_HFLAG_MMIO)
85 return hwif->dma_base;
86
87 if (hwif->mate && hwif->mate->dma_base) {
88 dma_base = hwif->mate->dma_base - (hwif->channel ? 0 : 8);
89 } else {
90 u8 baridx = (d->host_flags & IDE_HFLAG_CS5520) ? 2 : 4;
91
92 dma_base = pci_resource_start(dev, baridx);
93
94 if (dma_base == 0) {
95 printk(KERN_ERR "%s %s: DMA base is invalid\n",
96 d->name, pci_name(dev));
97 return 0;
98 }
99 }
100
101 if (hwif->channel)
102 dma_base += 8;
103
104 return dma_base;
105}
106EXPORT_SYMBOL_GPL(ide_pci_dma_base);
107
108int ide_pci_check_simplex(ide_hwif_t *hwif, const struct ide_port_info *d)
109{
110 struct pci_dev *dev = to_pci_dev(hwif->dev);
111 u8 dma_stat;
112
113 if (d->host_flags & (IDE_HFLAG_MMIO | IDE_HFLAG_CS5520))
114 goto out;
115
116 if (d->host_flags & IDE_HFLAG_CLEAR_SIMPLEX) {
117 if (ide_pci_clear_simplex(hwif->dma_base, d->name))
118 printk(KERN_INFO "%s %s: simplex device: DMA forced\n",
119 d->name, pci_name(dev));
120 goto out;
121 }
122
123
124
125
126
127
128
129
130
131
132
133 dma_stat = hwif->tp_ops->read_sff_dma_status(hwif);
134 if ((dma_stat & 0x80) && hwif->mate && hwif->mate->dma_base) {
135 printk(KERN_INFO "%s %s: simplex device: DMA disabled\n",
136 d->name, pci_name(dev));
137 return -1;
138 }
139out:
140 return 0;
141}
142EXPORT_SYMBOL_GPL(ide_pci_check_simplex);
143
144
145
146
147int ide_pci_set_master(struct pci_dev *dev, const char *name)
148{
149 u16 pcicmd;
150
151 pci_read_config_word(dev, PCI_COMMAND, &pcicmd);
152
153 if ((pcicmd & PCI_COMMAND_MASTER) == 0) {
154 pci_set_master(dev);
155
156 if (pci_read_config_word(dev, PCI_COMMAND, &pcicmd) ||
157 (pcicmd & PCI_COMMAND_MASTER) == 0) {
158 printk(KERN_ERR "%s %s: error updating PCICMD\n",
159 name, pci_name(dev));
160 return -EIO;
161 }
162 }
163
164 return 0;
165}
166EXPORT_SYMBOL_GPL(ide_pci_set_master);
167#endif
168
169void ide_setup_pci_noise(struct pci_dev *dev, const struct ide_port_info *d)
170{
171 printk(KERN_INFO "%s %s: IDE controller (0x%04x:0x%04x rev 0x%02x)\n",
172 d->name, pci_name(dev),
173 dev->vendor, dev->device, dev->revision);
174}
175EXPORT_SYMBOL_GPL(ide_setup_pci_noise);
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191static int ide_pci_enable(struct pci_dev *dev, const struct ide_port_info *d)
192{
193 int ret, bars;
194
195 if (pci_enable_device(dev)) {
196 ret = pci_enable_device_io(dev);
197 if (ret < 0) {
198 printk(KERN_WARNING "%s %s: couldn't enable device\n",
199 d->name, pci_name(dev));
200 goto out;
201 }
202 printk(KERN_WARNING "%s %s: BIOS configuration fixed\n",
203 d->name, pci_name(dev));
204 }
205
206
207
208
209
210
211 ret = pci_set_dma_mask(dev, DMA_32BIT_MASK);
212 if (ret < 0) {
213 printk(KERN_ERR "%s %s: can't set DMA mask\n",
214 d->name, pci_name(dev));
215 goto out;
216 }
217
218 if (d->host_flags & IDE_HFLAG_SINGLE)
219 bars = (1 << 2) - 1;
220 else
221 bars = (1 << 4) - 1;
222
223 if ((d->host_flags & IDE_HFLAG_NO_DMA) == 0) {
224 if (d->host_flags & IDE_HFLAG_CS5520)
225 bars |= (1 << 2);
226 else
227 bars |= (1 << 4);
228 }
229
230 ret = pci_request_selected_regions(dev, bars, d->name);
231 if (ret < 0)
232 printk(KERN_ERR "%s %s: can't reserve resources\n",
233 d->name, pci_name(dev));
234out:
235 return ret;
236}
237
238
239
240
241
242
243
244
245
246
247static int ide_pci_configure(struct pci_dev *dev, const struct ide_port_info *d)
248{
249 u16 pcicmd = 0;
250
251
252
253
254
255
256
257 if (ide_setup_pci_baseregs(dev, d->name) ||
258 pci_write_config_word(dev, PCI_COMMAND, pcicmd | PCI_COMMAND_IO)) {
259 printk(KERN_INFO "%s %s: device disabled (BIOS)\n",
260 d->name, pci_name(dev));
261 return -ENODEV;
262 }
263 if (pci_read_config_word(dev, PCI_COMMAND, &pcicmd)) {
264 printk(KERN_ERR "%s %s: error accessing PCI regs\n",
265 d->name, pci_name(dev));
266 return -EIO;
267 }
268 if (!(pcicmd & PCI_COMMAND_IO)) {
269 printk(KERN_ERR "%s %s: unable to enable IDE controller\n",
270 d->name, pci_name(dev));
271 return -ENXIO;
272 }
273 return 0;
274}
275
276
277
278
279
280
281
282
283
284
285
286static int ide_pci_check_iomem(struct pci_dev *dev, const struct ide_port_info *d,
287 int bar)
288{
289 ulong flags = pci_resource_flags(dev, bar);
290
291
292 if (!flags || pci_resource_len(dev, bar) == 0)
293 return 0;
294
295
296 if (flags & IORESOURCE_IO)
297 return 0;
298
299
300 return -EINVAL;
301}
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318static int ide_hw_configure(struct pci_dev *dev, const struct ide_port_info *d,
319 unsigned int port, int irq, hw_regs_t *hw)
320{
321 unsigned long ctl = 0, base = 0;
322
323 if ((d->host_flags & IDE_HFLAG_ISA_PORTS) == 0) {
324 if (ide_pci_check_iomem(dev, d, 2 * port) ||
325 ide_pci_check_iomem(dev, d, 2 * port + 1)) {
326 printk(KERN_ERR "%s %s: I/O baseregs (BIOS) are "
327 "reported as MEM for port %d!\n",
328 d->name, pci_name(dev), port);
329 return -EINVAL;
330 }
331
332 ctl = pci_resource_start(dev, 2*port+1);
333 base = pci_resource_start(dev, 2*port);
334 } else {
335
336 ctl = port ? 0x374 : 0x3f4;
337 base = port ? 0x170 : 0x1f0;
338 }
339
340 if (!base || !ctl) {
341 printk(KERN_ERR "%s %s: bad PCI BARs for port %d, skipping\n",
342 d->name, pci_name(dev), port);
343 return -EINVAL;
344 }
345
346 memset(hw, 0, sizeof(*hw));
347 hw->irq = irq;
348 hw->dev = &dev->dev;
349 hw->chipset = d->chipset ? d->chipset : ide_pci;
350 ide_std_init_ports(hw, base, ctl | 2);
351
352 return 0;
353}
354
355#ifdef CONFIG_BLK_DEV_IDEDMA_PCI
356
357
358
359
360
361
362
363
364
365
366int ide_hwif_setup_dma(ide_hwif_t *hwif, const struct ide_port_info *d)
367{
368 struct pci_dev *dev = to_pci_dev(hwif->dev);
369
370 if ((d->host_flags & IDE_HFLAG_NO_AUTODMA) == 0 ||
371 ((dev->class >> 8) == PCI_CLASS_STORAGE_IDE &&
372 (dev->class & 0x80))) {
373 unsigned long base = ide_pci_dma_base(hwif, d);
374
375 if (base == 0)
376 return -1;
377
378 hwif->dma_base = base;
379
380 if (ide_pci_check_simplex(hwif, d) < 0)
381 return -1;
382
383 if (ide_pci_set_master(dev, d->name) < 0)
384 return -1;
385
386 if (hwif->host_flags & IDE_HFLAG_MMIO)
387 printk(KERN_INFO " %s: MMIO-DMA\n", hwif->name);
388 else
389 printk(KERN_INFO " %s: BM-DMA at 0x%04lx-0x%04lx\n",
390 hwif->name, base, base + 7);
391
392 hwif->extra_base = base + (hwif->channel ? 8 : 16);
393
394 if (ide_allocate_dma_engine(hwif))
395 return -1;
396
397 hwif->dma_ops = &sff_dma_ops;
398 }
399
400 return 0;
401}
402#endif
403
404
405
406
407
408
409
410
411
412
413
414
415static int ide_setup_pci_controller(struct pci_dev *dev,
416 const struct ide_port_info *d, int noisy)
417{
418 int ret;
419 u16 pcicmd;
420
421 if (noisy)
422 ide_setup_pci_noise(dev, d);
423
424 ret = ide_pci_enable(dev, d);
425 if (ret < 0)
426 goto out;
427
428 ret = pci_read_config_word(dev, PCI_COMMAND, &pcicmd);
429 if (ret < 0) {
430 printk(KERN_ERR "%s %s: error accessing PCI regs\n",
431 d->name, pci_name(dev));
432 goto out;
433 }
434 if (!(pcicmd & PCI_COMMAND_IO)) {
435 ret = ide_pci_configure(dev, d);
436 if (ret < 0)
437 goto out;
438 printk(KERN_INFO "%s %s: device enabled (Linux)\n",
439 d->name, pci_name(dev));
440 }
441
442out:
443 return ret;
444}
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463void ide_pci_setup_ports(struct pci_dev *dev, const struct ide_port_info *d,
464 int pciirq, hw_regs_t *hw, hw_regs_t **hws)
465{
466 int channels = (d->host_flags & IDE_HFLAG_SINGLE) ? 1 : 2, port;
467 u8 tmp;
468
469
470
471
472
473 for (port = 0; port < channels; ++port) {
474 const ide_pci_enablebit_t *e = &(d->enablebits[port]);
475
476 if (e->reg && (pci_read_config_byte(dev, e->reg, &tmp) ||
477 (tmp & e->mask) != e->val)) {
478 printk(KERN_INFO "%s %s: IDE port disabled\n",
479 d->name, pci_name(dev));
480 continue;
481 }
482
483 if (ide_hw_configure(dev, d, port, pciirq, hw + port))
484 continue;
485
486 *(hws + port) = hw + port;
487 }
488}
489EXPORT_SYMBOL_GPL(ide_pci_setup_ports);
490
491
492
493
494
495
496
497
498
499
500
501static int do_ide_setup_pci_device(struct pci_dev *dev,
502 const struct ide_port_info *d,
503 u8 noisy)
504{
505 int pciirq, ret;
506
507
508
509
510 pciirq = dev->irq;
511
512
513
514
515
516
517
518 ret = d->init_chipset ? d->init_chipset(dev) : 0;
519 if (ret < 0)
520 goto out;
521
522
523 if ((dev->class >> 8) == PCI_CLASS_STORAGE_IDE && (dev->class & 5) != 5) {
524 if (noisy)
525 printk(KERN_INFO "%s %s: not 100%% native mode: will "
526 "probe irqs later\n", d->name, pci_name(dev));
527 pciirq = ret;
528 } else if (!pciirq && noisy) {
529 printk(KERN_WARNING "%s %s: bad irq (%d): will probe later\n",
530 d->name, pci_name(dev), pciirq);
531 } else if (noisy) {
532 printk(KERN_INFO "%s %s: 100%% native mode on irq %d\n",
533 d->name, pci_name(dev), pciirq);
534 }
535
536 ret = pciirq;
537out:
538 return ret;
539}
540
541int ide_pci_init_one(struct pci_dev *dev, const struct ide_port_info *d,
542 void *priv)
543{
544 struct ide_host *host;
545 hw_regs_t hw[4], *hws[] = { NULL, NULL, NULL, NULL };
546 int ret;
547
548 ret = ide_setup_pci_controller(dev, d, 1);
549 if (ret < 0)
550 goto out;
551
552 ide_pci_setup_ports(dev, d, 0, &hw[0], &hws[0]);
553
554 host = ide_host_alloc(d, hws);
555 if (host == NULL) {
556 ret = -ENOMEM;
557 goto out;
558 }
559
560 host->dev[0] = &dev->dev;
561
562 host->host_priv = priv;
563
564 pci_set_drvdata(dev, host);
565
566 ret = do_ide_setup_pci_device(dev, d, 1);
567 if (ret < 0)
568 goto out;
569
570
571 hw[1].irq = hw[0].irq = ret;
572
573 ret = ide_host_register(host, d, hws);
574 if (ret)
575 ide_host_free(host);
576out:
577 return ret;
578}
579EXPORT_SYMBOL_GPL(ide_pci_init_one);
580
581int ide_pci_init_two(struct pci_dev *dev1, struct pci_dev *dev2,
582 const struct ide_port_info *d, void *priv)
583{
584 struct pci_dev *pdev[] = { dev1, dev2 };
585 struct ide_host *host;
586 int ret, i;
587 hw_regs_t hw[4], *hws[] = { NULL, NULL, NULL, NULL };
588
589 for (i = 0; i < 2; i++) {
590 ret = ide_setup_pci_controller(pdev[i], d, !i);
591 if (ret < 0)
592 goto out;
593
594 ide_pci_setup_ports(pdev[i], d, 0, &hw[i*2], &hws[i*2]);
595 }
596
597 host = ide_host_alloc(d, hws);
598 if (host == NULL) {
599 ret = -ENOMEM;
600 goto out;
601 }
602
603 host->dev[0] = &dev1->dev;
604 host->dev[1] = &dev2->dev;
605
606 host->host_priv = priv;
607
608 pci_set_drvdata(pdev[0], host);
609 pci_set_drvdata(pdev[1], host);
610
611 for (i = 0; i < 2; i++) {
612 ret = do_ide_setup_pci_device(pdev[i], d, !i);
613
614
615
616
617
618 if (ret < 0)
619 goto out;
620
621
622 hw[i*2 + 1].irq = hw[i*2].irq = ret;
623 }
624
625 ret = ide_host_register(host, d, hws);
626 if (ret)
627 ide_host_free(host);
628out:
629 return ret;
630}
631EXPORT_SYMBOL_GPL(ide_pci_init_two);
632
633void ide_pci_remove(struct pci_dev *dev)
634{
635 struct ide_host *host = pci_get_drvdata(dev);
636 struct pci_dev *dev2 = host->dev[1] ? to_pci_dev(host->dev[1]) : NULL;
637 int bars;
638
639 if (host->host_flags & IDE_HFLAG_SINGLE)
640 bars = (1 << 2) - 1;
641 else
642 bars = (1 << 4) - 1;
643
644 if ((host->host_flags & IDE_HFLAG_NO_DMA) == 0) {
645 if (host->host_flags & IDE_HFLAG_CS5520)
646 bars |= (1 << 2);
647 else
648 bars |= (1 << 4);
649 }
650
651 ide_host_remove(host);
652
653 if (dev2)
654 pci_release_selected_regions(dev2, bars);
655 pci_release_selected_regions(dev, bars);
656
657 if (dev2)
658 pci_disable_device(dev2);
659 pci_disable_device(dev);
660}
661EXPORT_SYMBOL_GPL(ide_pci_remove);
662
663#ifdef CONFIG_PM
664int ide_pci_suspend(struct pci_dev *dev, pm_message_t state)
665{
666 pci_save_state(dev);
667 pci_disable_device(dev);
668 pci_set_power_state(dev, pci_choose_state(dev, state));
669
670 return 0;
671}
672EXPORT_SYMBOL_GPL(ide_pci_suspend);
673
674int ide_pci_resume(struct pci_dev *dev)
675{
676 struct ide_host *host = pci_get_drvdata(dev);
677 int rc;
678
679 pci_set_power_state(dev, PCI_D0);
680
681 rc = pci_enable_device(dev);
682 if (rc)
683 return rc;
684
685 pci_restore_state(dev);
686 pci_set_master(dev);
687
688 if (host->init_chipset)
689 host->init_chipset(dev);
690
691 return 0;
692}
693EXPORT_SYMBOL_GPL(ide_pci_resume);
694#endif
695