Linux Kernel  3.7.1
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
cmd64x.c
Go to the documentation of this file.
1 /*
2  * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
3  * Due to massive hardware bugs, UltraDMA is only supported
4  * on the 646U2 and not on the 646U.
5  *
6  * Copyright (C) 1998 Eddie C. Dost ([email protected])
7  * Copyright (C) 1998 David S. Miller ([email protected])
8  *
9  * Copyright (C) 1999-2002 Andre Hedrick <[email protected]>
10  * Copyright (C) 2007-2010 Bartlomiej Zolnierkiewicz
11  * Copyright (C) 2007,2009 MontaVista Software, Inc. <[email protected]>
12  */
13 
14 #include <linux/module.h>
15 #include <linux/types.h>
16 #include <linux/pci.h>
17 #include <linux/ide.h>
18 #include <linux/init.h>
19 
20 #include <asm/io.h>
21 
22 #define DRV_NAME "cmd64x"
23 
24 /*
25  * CMD64x specific registers definition.
26  */
27 #define CFR 0x50
28 #define CFR_INTR_CH0 0x04
29 
30 #define CMDTIM 0x52
31 #define ARTTIM0 0x53
32 #define DRWTIM0 0x54
33 #define ARTTIM1 0x55
34 #define DRWTIM1 0x56
35 #define ARTTIM23 0x57
36 #define ARTTIM23_DIS_RA2 0x04
37 #define ARTTIM23_DIS_RA3 0x08
38 #define ARTTIM23_INTR_CH1 0x10
39 #define DRWTIM2 0x58
40 #define BRST 0x59
41 #define DRWTIM3 0x5b
42 
43 #define BMIDECR0 0x70
44 #define MRDMODE 0x71
45 #define MRDMODE_INTR_CH0 0x04
46 #define MRDMODE_INTR_CH1 0x08
47 #define UDIDETCR0 0x73
48 #define DTPR0 0x74
49 #define BMIDECR1 0x78
50 #define BMIDECSR 0x79
51 #define UDIDETCR1 0x7B
52 #define DTPR1 0x7C
53 
54 static void cmd64x_program_timings(ide_drive_t *drive, u8 mode)
55 {
56  ide_hwif_t *hwif = drive->hwif;
57  struct pci_dev *dev = to_pci_dev(drive->hwif->dev);
58  int bus_speed = ide_pci_clk ? ide_pci_clk : 33;
59  const unsigned long T = 1000000 / bus_speed;
60  static const u8 recovery_values[] =
61  {15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
62  static const u8 setup_values[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
63  static const u8 arttim_regs[4] = {ARTTIM0, ARTTIM1, ARTTIM23, ARTTIM23};
64  static const u8 drwtim_regs[4] = {DRWTIM0, DRWTIM1, DRWTIM2, DRWTIM3};
65  struct ide_timing t;
66  u8 arttim = 0;
67 
68  ide_timing_compute(drive, mode, &t, T, 0);
69 
70  /*
71  * In case we've got too long recovery phase, try to lengthen
72  * the active phase
73  */
74  if (t.recover > 16) {
75  t.active += t.recover - 16;
76  t.recover = 16;
77  }
78  if (t.active > 16) /* shouldn't actually happen... */
79  t.active = 16;
80 
81  /*
82  * Convert values to internal chipset representation
83  */
84  t.recover = recovery_values[t.recover];
85  t.active &= 0x0f;
86 
87  /* Program the active/recovery counts into the DRWTIM register */
88  pci_write_config_byte(dev, drwtim_regs[drive->dn],
89  (t.active << 4) | t.recover);
90 
91  /*
92  * The primary channel has individual address setup timing registers
93  * for each drive and the hardware selects the slowest timing itself.
94  * The secondary channel has one common register and we have to select
95  * the slowest address setup timing ourselves.
96  */
97  if (hwif->channel) {
98  ide_drive_t *pair = ide_get_pair_dev(drive);
99 
100  if (pair) {
101  struct ide_timing tp;
102 
103  ide_timing_compute(pair, pair->pio_mode, &tp, T, 0);
104  ide_timing_merge(&t, &tp, &t, IDE_TIMING_SETUP);
105  if (pair->dma_mode) {
106  ide_timing_compute(pair, pair->dma_mode,
107  &tp, T, 0);
108  ide_timing_merge(&tp, &t, &t, IDE_TIMING_SETUP);
109  }
110  }
111  }
112 
113  if (t.setup > 5) /* shouldn't actually happen... */
114  t.setup = 5;
115 
116  /*
117  * Program the address setup clocks into the ARTTIM registers.
118  * Avoid clearing the secondary channel's interrupt bit.
119  */
120  (void) pci_read_config_byte (dev, arttim_regs[drive->dn], &arttim);
121  if (hwif->channel)
122  arttim &= ~ARTTIM23_INTR_CH1;
123  arttim &= ~0xc0;
124  arttim |= setup_values[t.setup];
125  (void) pci_write_config_byte(dev, arttim_regs[drive->dn], arttim);
126 }
127 
128 /*
129  * Attempts to set drive's PIO mode.
130  * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
131  */
132 
133 static void cmd64x_set_pio_mode(ide_hwif_t *hwif, ide_drive_t *drive)
134 {
135  const u8 pio = drive->pio_mode - XFER_PIO_0;
136 
137  /*
138  * Filter out the prefetch control values
139  * to prevent PIO5 from being programmed
140  */
141  if (pio == 8 || pio == 9)
142  return;
143 
144  cmd64x_program_timings(drive, XFER_PIO_0 + pio);
145 }
146 
147 static void cmd64x_set_dma_mode(ide_hwif_t *hwif, ide_drive_t *drive)
148 {
149  struct pci_dev *dev = to_pci_dev(hwif->dev);
150  u8 unit = drive->dn & 0x01;
151  u8 regU = 0, pciU = hwif->channel ? UDIDETCR1 : UDIDETCR0;
152  const u8 speed = drive->dma_mode;
153 
154  pci_read_config_byte(dev, pciU, &regU);
155  regU &= ~(unit ? 0xCA : 0x35);
156 
157  switch(speed) {
158  case XFER_UDMA_5:
159  regU |= unit ? 0x0A : 0x05;
160  break;
161  case XFER_UDMA_4:
162  regU |= unit ? 0x4A : 0x15;
163  break;
164  case XFER_UDMA_3:
165  regU |= unit ? 0x8A : 0x25;
166  break;
167  case XFER_UDMA_2:
168  regU |= unit ? 0x42 : 0x11;
169  break;
170  case XFER_UDMA_1:
171  regU |= unit ? 0x82 : 0x21;
172  break;
173  case XFER_UDMA_0:
174  regU |= unit ? 0xC2 : 0x31;
175  break;
176  case XFER_MW_DMA_2:
177  case XFER_MW_DMA_1:
178  case XFER_MW_DMA_0:
179  cmd64x_program_timings(drive, speed);
180  break;
181  }
182 
183  pci_write_config_byte(dev, pciU, regU);
184 }
185 
186 static void cmd648_clear_irq(ide_drive_t *drive)
187 {
188  ide_hwif_t *hwif = drive->hwif;
189  struct pci_dev *dev = to_pci_dev(hwif->dev);
190  unsigned long base = pci_resource_start(dev, 4);
191  u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
193  u8 mrdmode = inb(base + 1);
194 
195  /* clear the interrupt bit */
196  outb((mrdmode & ~(MRDMODE_INTR_CH0 | MRDMODE_INTR_CH1)) | irq_mask,
197  base + 1);
198 }
199 
200 static void cmd64x_clear_irq(ide_drive_t *drive)
201 {
202  ide_hwif_t *hwif = drive->hwif;
203  struct pci_dev *dev = to_pci_dev(hwif->dev);
204  int irq_reg = hwif->channel ? ARTTIM23 : CFR;
205  u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
206  CFR_INTR_CH0;
207  u8 irq_stat = 0;
208 
209  (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
210  /* clear the interrupt bit */
211  (void) pci_write_config_byte(dev, irq_reg, irq_stat | irq_mask);
212 }
213 
214 static int cmd648_test_irq(ide_hwif_t *hwif)
215 {
216  struct pci_dev *dev = to_pci_dev(hwif->dev);
217  unsigned long base = pci_resource_start(dev, 4);
218  u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
220  u8 mrdmode = inb(base + 1);
221 
222  pr_debug("%s: mrdmode: 0x%02x irq_mask: 0x%02x\n",
223  hwif->name, mrdmode, irq_mask);
224 
225  return (mrdmode & irq_mask) ? 1 : 0;
226 }
227 
228 static int cmd64x_test_irq(ide_hwif_t *hwif)
229 {
230  struct pci_dev *dev = to_pci_dev(hwif->dev);
231  int irq_reg = hwif->channel ? ARTTIM23 : CFR;
232  u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
233  CFR_INTR_CH0;
234  u8 irq_stat = 0;
235 
236  (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
237 
238  pr_debug("%s: irq_stat: 0x%02x irq_mask: 0x%02x\n",
239  hwif->name, irq_stat, irq_mask);
240 
241  return (irq_stat & irq_mask) ? 1 : 0;
242 }
243 
244 /*
245  * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
246  * event order for DMA transfers.
247  */
248 
249 static int cmd646_1_dma_end(ide_drive_t *drive)
250 {
251  ide_hwif_t *hwif = drive->hwif;
252  u8 dma_stat = 0, dma_cmd = 0;
253 
254  /* get DMA status */
255  dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
256  /* read DMA command state */
257  dma_cmd = inb(hwif->dma_base + ATA_DMA_CMD);
258  /* stop DMA */
259  outb(dma_cmd & ~1, hwif->dma_base + ATA_DMA_CMD);
260  /* clear the INTR & ERROR bits */
261  outb(dma_stat | 6, hwif->dma_base + ATA_DMA_STATUS);
262  /* verify good DMA status */
263  return (dma_stat & 7) != 4;
264 }
265 
266 static int init_chipset_cmd64x(struct pci_dev *dev)
267 {
268  u8 mrdmode = 0;
269 
270  /* Set a good latency timer and cache line size value. */
271  (void) pci_write_config_byte(dev, PCI_LATENCY_TIMER, 64);
272  /* FIXME: pci_set_master() to ensure a good latency timer value */
273 
274  /*
275  * Enable interrupts, select MEMORY READ LINE for reads.
276  *
277  * NOTE: although not mentioned in the PCI0646U specs,
278  * bits 0-1 are write only and won't be read back as
279  * set or not -- PCI0646U2 specs clarify this point.
280  */
281  (void) pci_read_config_byte (dev, MRDMODE, &mrdmode);
282  mrdmode &= ~0x30;
283  (void) pci_write_config_byte(dev, MRDMODE, (mrdmode | 0x02));
284 
285  return 0;
286 }
287 
288 static u8 cmd64x_cable_detect(ide_hwif_t *hwif)
289 {
290  struct pci_dev *dev = to_pci_dev(hwif->dev);
291  u8 bmidecsr = 0, mask = hwif->channel ? 0x02 : 0x01;
292 
293  switch (dev->device) {
296  pci_read_config_byte(dev, BMIDECSR, &bmidecsr);
297  return (bmidecsr & mask) ? ATA_CBL_PATA80 : ATA_CBL_PATA40;
298  default:
299  return ATA_CBL_PATA40;
300  }
301 }
302 
303 static const struct ide_port_ops cmd64x_port_ops = {
304  .set_pio_mode = cmd64x_set_pio_mode,
305  .set_dma_mode = cmd64x_set_dma_mode,
306  .clear_irq = cmd64x_clear_irq,
307  .test_irq = cmd64x_test_irq,
308  .cable_detect = cmd64x_cable_detect,
309 };
310 
311 static const struct ide_port_ops cmd648_port_ops = {
312  .set_pio_mode = cmd64x_set_pio_mode,
313  .set_dma_mode = cmd64x_set_dma_mode,
314  .clear_irq = cmd648_clear_irq,
315  .test_irq = cmd648_test_irq,
316  .cable_detect = cmd64x_cable_detect,
317 };
318 
319 static const struct ide_dma_ops cmd646_rev1_dma_ops = {
320  .dma_host_set = ide_dma_host_set,
321  .dma_setup = ide_dma_setup,
322  .dma_start = ide_dma_start,
323  .dma_end = cmd646_1_dma_end,
324  .dma_test_irq = ide_dma_test_irq,
325  .dma_lost_irq = ide_dma_lost_irq,
326  .dma_timer_expiry = ide_dma_sff_timer_expiry,
327  .dma_sff_read_status = ide_dma_sff_read_status,
328 };
329 
330 static const struct ide_port_info cmd64x_chipsets[] __devinitconst = {
331  { /* 0: CMD643 */
332  .name = DRV_NAME,
333  .init_chipset = init_chipset_cmd64x,
334  .enablebits = {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
335  .port_ops = &cmd64x_port_ops,
336  .host_flags = IDE_HFLAG_CLEAR_SIMPLEX |
339  .pio_mask = ATA_PIO5,
340  .mwdma_mask = ATA_MWDMA2,
341  .udma_mask = 0x00, /* no udma */
342  },
343  { /* 1: CMD646 */
344  .name = DRV_NAME,
345  .init_chipset = init_chipset_cmd64x,
346  .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
347  .port_ops = &cmd648_port_ops,
348  .host_flags = IDE_HFLAG_ABUSE_PREFETCH |
350  .pio_mask = ATA_PIO5,
351  .mwdma_mask = ATA_MWDMA2,
352  .udma_mask = ATA_UDMA2,
353  },
354  { /* 2: CMD648 */
355  .name = DRV_NAME,
356  .init_chipset = init_chipset_cmd64x,
357  .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
358  .port_ops = &cmd648_port_ops,
359  .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
360  .pio_mask = ATA_PIO5,
361  .mwdma_mask = ATA_MWDMA2,
362  .udma_mask = ATA_UDMA4,
363  },
364  { /* 3: CMD649 */
365  .name = DRV_NAME,
366  .init_chipset = init_chipset_cmd64x,
367  .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
368  .port_ops = &cmd648_port_ops,
369  .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
370  .pio_mask = ATA_PIO5,
371  .mwdma_mask = ATA_MWDMA2,
372  .udma_mask = ATA_UDMA5,
373  }
374 };
375 
376 static int __devinit cmd64x_init_one(struct pci_dev *dev, const struct pci_device_id *id)
377 {
378  struct ide_port_info d;
379  u8 idx = id->driver_data;
380 
381  d = cmd64x_chipsets[idx];
382 
383  if (idx == 1) {
384  /*
385  * UltraDMA only supported on PCI646U and PCI646U2, which
386  * correspond to revisions 0x03, 0x05 and 0x07 respectively.
387  * Actually, although the CMD tech support people won't
388  * tell me the details, the 0x03 revision cannot support
389  * UDMA correctly without hardware modifications, and even
390  * then it only works with Quantum disks due to some
391  * hold time assumptions in the 646U part which are fixed
392  * in the 646U2.
393  *
394  * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
395  */
396  if (dev->revision < 5) {
397  d.udma_mask = 0x00;
398  /*
399  * The original PCI0646 didn't have the primary
400  * channel enable bit, it appeared starting with
401  * PCI0646U (i.e. revision ID 3).
402  */
403  if (dev->revision < 3) {
404  d.enablebits[0].reg = 0;
405  d.port_ops = &cmd64x_port_ops;
406  if (dev->revision == 1)
407  d.dma_ops = &cmd646_rev1_dma_ops;
408  }
409  }
410  }
411 
412  return ide_pci_init_one(dev, &d, NULL);
413 }
414 
415 static const struct pci_device_id cmd64x_pci_tbl[] = {
420  { 0, },
421 };
422 MODULE_DEVICE_TABLE(pci, cmd64x_pci_tbl);
423 
424 static struct pci_driver cmd64x_pci_driver = {
425  .name = "CMD64x_IDE",
426  .id_table = cmd64x_pci_tbl,
427  .probe = cmd64x_init_one,
428  .remove = ide_pci_remove,
429  .suspend = ide_pci_suspend,
430  .resume = ide_pci_resume,
431 };
432 
433 static int __init cmd64x_ide_init(void)
434 {
435  return ide_pci_register_driver(&cmd64x_pci_driver);
436 }
437 
438 static void __exit cmd64x_ide_exit(void)
439 {
440  pci_unregister_driver(&cmd64x_pci_driver);
441 }
442 
443 module_init(cmd64x_ide_init);
444 module_exit(cmd64x_ide_exit);
445 
446 MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick, Bartlomiej Zolnierkiewicz");
447 MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
448 MODULE_LICENSE("GPL");