2 * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
3 * Due to massive hardware bugs, UltraDMA is only supported
4 * on the 646U2 and not on the 646U.
6 * Copyright (C) 1998 Eddie C. Dost (ecd@skynet.be)
7 * Copyright (C) 1998 David S. Miller (davem@redhat.com)
9 * Copyright (C) 1999-2002 Andre Hedrick <andre@linux-ide.org>
10 * Copyright (C) 2007 MontaVista Software, Inc. <source@mvista.com>
13 #include <linux/module.h>
14 #include <linux/types.h>
15 #include <linux/pci.h>
16 #include <linux/hdreg.h>
17 #include <linux/ide.h>
18 #include <linux/init.h>
25 #define cmdprintk(x...) printk(x)
27 #define cmdprintk(x...)
31 * CMD64x specific registers definition.
34 #define CFR_INTR_CH0 0x04
42 #define ARTTIM23_DIS_RA2 0x04
43 #define ARTTIM23_DIS_RA3 0x08
44 #define ARTTIM23_INTR_CH1 0x10
51 #define MRDMODE_INTR_CH0 0x04
52 #define MRDMODE_INTR_CH1 0x08
53 #define UDIDETCR0 0x73
57 #define UDIDETCR1 0x7B
60 static u8 quantize_timing(int timing, int quant)
62 return (timing + quant - 1) / quant;
66 * This routine calculates active/recovery counts and then writes them into
67 * the chipset registers.
69 static void program_cycle_times (ide_drive_t *drive, int cycle_time, int active_time)
71 struct pci_dev *dev = to_pci_dev(drive->hwif->dev);
72 int clock_time = 1000 / (ide_pci_clk ? ide_pci_clk : 33);
73 u8 cycle_count, active_count, recovery_count, drwtim;
74 static const u8 recovery_values[] =
75 {15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
76 static const u8 drwtim_regs[4] = {DRWTIM0, DRWTIM1, DRWTIM2, DRWTIM3};
78 cmdprintk("program_cycle_times parameters: total=%d, active=%d\n",
79 cycle_time, active_time);
81 cycle_count = quantize_timing( cycle_time, clock_time);
82 active_count = quantize_timing(active_time, clock_time);
83 recovery_count = cycle_count - active_count;
86 * In case we've got too long recovery phase, try to lengthen
89 if (recovery_count > 16) {
90 active_count += recovery_count - 16;
93 if (active_count > 16) /* shouldn't actually happen... */
96 cmdprintk("Final counts: total=%d, active=%d, recovery=%d\n",
97 cycle_count, active_count, recovery_count);
100 * Convert values to internal chipset representation
102 recovery_count = recovery_values[recovery_count];
103 active_count &= 0x0f;
105 /* Program the active/recovery counts into the DRWTIM register */
106 drwtim = (active_count << 4) | recovery_count;
107 (void) pci_write_config_byte(dev, drwtim_regs[drive->dn], drwtim);
108 cmdprintk("Write 0x%02x to reg 0x%x\n", drwtim, drwtim_regs[drive->dn]);
112 * This routine writes into the chipset registers
113 * PIO setup/active/recovery timings.
115 static void cmd64x_tune_pio(ide_drive_t *drive, const u8 pio)
117 ide_hwif_t *hwif = HWIF(drive);
118 struct pci_dev *dev = to_pci_dev(hwif->dev);
119 struct ide_timing *t = ide_timing_find_mode(XFER_PIO_0 + pio);
120 unsigned int cycle_time;
121 u8 setup_count, arttim = 0;
123 static const u8 setup_values[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
124 static const u8 arttim_regs[4] = {ARTTIM0, ARTTIM1, ARTTIM23, ARTTIM23};
126 cycle_time = ide_pio_cycle_time(drive, pio);
128 program_cycle_times(drive, cycle_time, t->active);
130 setup_count = quantize_timing(t->setup,
131 1000 / (ide_pci_clk ? ide_pci_clk : 33));
134 * The primary channel has individual address setup timing registers
135 * for each drive and the hardware selects the slowest timing itself.
136 * The secondary channel has one common register and we have to select
137 * the slowest address setup timing ourselves.
140 ide_drive_t *drives = hwif->drives;
142 drive->drive_data = setup_count;
143 setup_count = max(drives[0].drive_data, drives[1].drive_data);
146 if (setup_count > 5) /* shouldn't actually happen... */
148 cmdprintk("Final address setup count: %d\n", setup_count);
151 * Program the address setup clocks into the ARTTIM registers.
152 * Avoid clearing the secondary channel's interrupt bit.
154 (void) pci_read_config_byte (dev, arttim_regs[drive->dn], &arttim);
156 arttim &= ~ARTTIM23_INTR_CH1;
158 arttim |= setup_values[setup_count];
159 (void) pci_write_config_byte(dev, arttim_regs[drive->dn], arttim);
160 cmdprintk("Write 0x%02x to reg 0x%x\n", arttim, arttim_regs[drive->dn]);
164 * Attempts to set drive's PIO mode.
165 * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
168 static void cmd64x_set_pio_mode(ide_drive_t *drive, const u8 pio)
171 * Filter out the prefetch control values
172 * to prevent PIO5 from being programmed
174 if (pio == 8 || pio == 9)
177 cmd64x_tune_pio(drive, pio);
180 static void cmd64x_set_dma_mode(ide_drive_t *drive, const u8 speed)
182 ide_hwif_t *hwif = HWIF(drive);
183 struct pci_dev *dev = to_pci_dev(hwif->dev);
184 u8 unit = drive->dn & 0x01;
185 u8 regU = 0, pciU = hwif->channel ? UDIDETCR1 : UDIDETCR0;
187 if (speed >= XFER_SW_DMA_0) {
188 (void) pci_read_config_byte(dev, pciU, ®U);
189 regU &= ~(unit ? 0xCA : 0x35);
194 regU |= unit ? 0x0A : 0x05;
197 regU |= unit ? 0x4A : 0x15;
200 regU |= unit ? 0x8A : 0x25;
203 regU |= unit ? 0x42 : 0x11;
206 regU |= unit ? 0x82 : 0x21;
209 regU |= unit ? 0xC2 : 0x31;
212 program_cycle_times(drive, 120, 70);
215 program_cycle_times(drive, 150, 80);
218 program_cycle_times(drive, 480, 215);
222 if (speed >= XFER_SW_DMA_0)
223 (void) pci_write_config_byte(dev, pciU, regU);
226 static int cmd648_dma_end(ide_drive_t *drive)
228 ide_hwif_t *hwif = HWIF(drive);
229 unsigned long base = hwif->dma_base - (hwif->channel * 8);
230 int err = __ide_dma_end(drive);
231 u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
233 u8 mrdmode = inb(base + 1);
235 /* clear the interrupt bit */
236 outb((mrdmode & ~(MRDMODE_INTR_CH0 | MRDMODE_INTR_CH1)) | irq_mask,
242 static int cmd64x_dma_end(ide_drive_t *drive)
244 ide_hwif_t *hwif = HWIF(drive);
245 struct pci_dev *dev = to_pci_dev(hwif->dev);
246 int irq_reg = hwif->channel ? ARTTIM23 : CFR;
247 u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
250 int err = __ide_dma_end(drive);
252 (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
253 /* clear the interrupt bit */
254 (void) pci_write_config_byte(dev, irq_reg, irq_stat | irq_mask);
259 static int cmd648_dma_test_irq(ide_drive_t *drive)
261 ide_hwif_t *hwif = HWIF(drive);
262 unsigned long base = hwif->dma_base - (hwif->channel * 8);
263 u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
265 u8 dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
266 u8 mrdmode = inb(base + 1);
269 printk("%s: dma_stat: 0x%02x mrdmode: 0x%02x irq_mask: 0x%02x\n",
270 drive->name, dma_stat, mrdmode, irq_mask);
272 if (!(mrdmode & irq_mask))
275 /* return 1 if INTR asserted */
282 static int cmd64x_dma_test_irq(ide_drive_t *drive)
284 ide_hwif_t *hwif = HWIF(drive);
285 struct pci_dev *dev = to_pci_dev(hwif->dev);
286 int irq_reg = hwif->channel ? ARTTIM23 : CFR;
287 u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
289 u8 dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
292 (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
295 printk("%s: dma_stat: 0x%02x irq_stat: 0x%02x irq_mask: 0x%02x\n",
296 drive->name, dma_stat, irq_stat, irq_mask);
298 if (!(irq_stat & irq_mask))
301 /* return 1 if INTR asserted */
309 * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
310 * event order for DMA transfers.
313 static int cmd646_1_dma_end(ide_drive_t *drive)
315 ide_hwif_t *hwif = HWIF(drive);
316 u8 dma_stat = 0, dma_cmd = 0;
318 drive->waiting_for_dma = 0;
320 dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
321 /* read DMA command state */
322 dma_cmd = inb(hwif->dma_base + ATA_DMA_CMD);
324 outb(dma_cmd & ~1, hwif->dma_base + ATA_DMA_CMD);
325 /* clear the INTR & ERROR bits */
326 outb(dma_stat | 6, hwif->dma_base + ATA_DMA_STATUS);
327 /* and free any DMA resources */
328 ide_destroy_dmatable(drive);
329 /* verify good DMA status */
330 return (dma_stat & 7) != 4;
333 static unsigned int __devinit init_chipset_cmd64x(struct pci_dev *dev, const char *name)
337 if (dev->device == PCI_DEVICE_ID_CMD_646) {
339 switch (dev->revision) {
342 printk("%s: UltraDMA capable\n", name);
346 printk("%s: MultiWord DMA force limited\n", name);
349 printk("%s: MultiWord DMA limited, "
350 "IRQ workaround enabled\n", name);
355 /* Set a good latency timer and cache line size value. */
356 (void) pci_write_config_byte(dev, PCI_LATENCY_TIMER, 64);
357 /* FIXME: pci_set_master() to ensure a good latency timer value */
360 * Enable interrupts, select MEMORY READ LINE for reads.
362 * NOTE: although not mentioned in the PCI0646U specs,
363 * bits 0-1 are write only and won't be read back as
364 * set or not -- PCI0646U2 specs clarify this point.
366 (void) pci_read_config_byte (dev, MRDMODE, &mrdmode);
368 (void) pci_write_config_byte(dev, MRDMODE, (mrdmode | 0x02));
373 static u8 __devinit cmd64x_cable_detect(ide_hwif_t *hwif)
375 struct pci_dev *dev = to_pci_dev(hwif->dev);
376 u8 bmidecsr = 0, mask = hwif->channel ? 0x02 : 0x01;
378 switch (dev->device) {
379 case PCI_DEVICE_ID_CMD_648:
380 case PCI_DEVICE_ID_CMD_649:
381 pci_read_config_byte(dev, BMIDECSR, &bmidecsr);
382 return (bmidecsr & mask) ? ATA_CBL_PATA80 : ATA_CBL_PATA40;
384 return ATA_CBL_PATA40;
388 static const struct ide_port_ops cmd64x_port_ops = {
389 .set_pio_mode = cmd64x_set_pio_mode,
390 .set_dma_mode = cmd64x_set_dma_mode,
391 .cable_detect = cmd64x_cable_detect,
394 static const struct ide_dma_ops cmd64x_dma_ops = {
395 .dma_host_set = ide_dma_host_set,
396 .dma_setup = ide_dma_setup,
397 .dma_exec_cmd = ide_dma_exec_cmd,
398 .dma_start = ide_dma_start,
399 .dma_end = cmd64x_dma_end,
400 .dma_test_irq = cmd64x_dma_test_irq,
401 .dma_lost_irq = ide_dma_lost_irq,
402 .dma_timeout = ide_dma_timeout,
405 static const struct ide_dma_ops cmd646_rev1_dma_ops = {
406 .dma_host_set = ide_dma_host_set,
407 .dma_setup = ide_dma_setup,
408 .dma_exec_cmd = ide_dma_exec_cmd,
409 .dma_start = ide_dma_start,
410 .dma_end = cmd646_1_dma_end,
411 .dma_test_irq = ide_dma_test_irq,
412 .dma_lost_irq = ide_dma_lost_irq,
413 .dma_timeout = ide_dma_timeout,
416 static const struct ide_dma_ops cmd648_dma_ops = {
417 .dma_host_set = ide_dma_host_set,
418 .dma_setup = ide_dma_setup,
419 .dma_exec_cmd = ide_dma_exec_cmd,
420 .dma_start = ide_dma_start,
421 .dma_end = cmd648_dma_end,
422 .dma_test_irq = cmd648_dma_test_irq,
423 .dma_lost_irq = ide_dma_lost_irq,
424 .dma_timeout = ide_dma_timeout,
427 static const struct ide_port_info cmd64x_chipsets[] __devinitdata = {
430 .init_chipset = init_chipset_cmd64x,
431 .enablebits = {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
432 .port_ops = &cmd64x_port_ops,
433 .dma_ops = &cmd64x_dma_ops,
434 .host_flags = IDE_HFLAG_CLEAR_SIMPLEX |
435 IDE_HFLAG_ABUSE_PREFETCH,
436 .pio_mask = ATA_PIO5,
437 .mwdma_mask = ATA_MWDMA2,
438 .udma_mask = 0x00, /* no udma */
441 .init_chipset = init_chipset_cmd64x,
442 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
443 .chipset = ide_cmd646,
444 .port_ops = &cmd64x_port_ops,
445 .dma_ops = &cmd648_dma_ops,
446 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
447 .pio_mask = ATA_PIO5,
448 .mwdma_mask = ATA_MWDMA2,
449 .udma_mask = ATA_UDMA2,
452 .init_chipset = init_chipset_cmd64x,
453 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
454 .port_ops = &cmd64x_port_ops,
455 .dma_ops = &cmd648_dma_ops,
456 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
457 .pio_mask = ATA_PIO5,
458 .mwdma_mask = ATA_MWDMA2,
459 .udma_mask = ATA_UDMA4,
462 .init_chipset = init_chipset_cmd64x,
463 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
464 .port_ops = &cmd64x_port_ops,
465 .dma_ops = &cmd648_dma_ops,
466 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
467 .pio_mask = ATA_PIO5,
468 .mwdma_mask = ATA_MWDMA2,
469 .udma_mask = ATA_UDMA5,
473 static int __devinit cmd64x_init_one(struct pci_dev *dev, const struct pci_device_id *id)
475 struct ide_port_info d;
476 u8 idx = id->driver_data;
478 d = cmd64x_chipsets[idx];
482 * UltraDMA only supported on PCI646U and PCI646U2, which
483 * correspond to revisions 0x03, 0x05 and 0x07 respectively.
484 * Actually, although the CMD tech support people won't
485 * tell me the details, the 0x03 revision cannot support
486 * UDMA correctly without hardware modifications, and even
487 * then it only works with Quantum disks due to some
488 * hold time assumptions in the 646U part which are fixed
491 * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
493 if (dev->revision < 5) {
496 * The original PCI0646 didn't have the primary
497 * channel enable bit, it appeared starting with
498 * PCI0646U (i.e. revision ID 3).
500 if (dev->revision < 3) {
501 d.enablebits[0].reg = 0;
502 if (dev->revision == 1)
503 d.dma_ops = &cmd646_rev1_dma_ops;
505 d.dma_ops = &cmd64x_dma_ops;
510 return ide_setup_pci_device(dev, &d);
513 static const struct pci_device_id cmd64x_pci_tbl[] = {
514 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_643), 0 },
515 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_646), 1 },
516 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_648), 2 },
517 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_649), 3 },
520 MODULE_DEVICE_TABLE(pci, cmd64x_pci_tbl);
522 static struct pci_driver driver = {
523 .name = "CMD64x_IDE",
524 .id_table = cmd64x_pci_tbl,
525 .probe = cmd64x_init_one,
528 static int __init cmd64x_ide_init(void)
530 return ide_pci_register_driver(&driver);
533 module_init(cmd64x_ide_init);
535 MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick");
536 MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
537 MODULE_LICENSE("GPL");