Linux Kernel  3.7.1
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
pata_bf54x.c
Go to the documentation of this file.
1 /*
2  * File: drivers/ata/pata_bf54x.c
3  * Author: Sonic Zhang <[email protected]>
4  *
5  * Created:
6  * Description: PATA Driver for blackfin 54x
7  *
8  * Modified:
9  * Copyright 2007 Analog Devices Inc.
10  *
11  * Bugs: Enter bugs at http://blackfin.uclinux.org/
12  *
13  * This program is free software; you can redistribute it and/or modify
14  * it under the terms of the GNU General Public License as published by
15  * the Free Software Foundation; either version 2 of the License, or
16  * (at your option) any later version.
17  *
18  * This program is distributed in the hope that it will be useful,
19  * but WITHOUT ANY WARRANTY; without even the implied warranty of
20  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
21  * GNU General Public License for more details.
22  *
23  * You should have received a copy of the GNU General Public License
24  * along with this program; if not, see the file COPYING, or write
25  * to the Free Software Foundation, Inc.,
26  * 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
27  */
28 
29 #include <linux/kernel.h>
30 #include <linux/module.h>
31 #include <linux/pci.h>
32 #include <linux/init.h>
33 #include <linux/blkdev.h>
34 #include <linux/delay.h>
35 #include <linux/device.h>
36 #include <scsi/scsi_host.h>
37 #include <linux/libata.h>
38 #include <linux/platform_device.h>
39 #include <asm/dma.h>
40 #include <asm/gpio.h>
41 #include <asm/portmux.h>
42 
43 #define DRV_NAME "pata-bf54x"
44 #define DRV_VERSION "0.9"
45 
46 #define ATA_REG_CTRL 0x0E
47 #define ATA_REG_ALTSTATUS ATA_REG_CTRL
48 
49 /* These are the offset of the controller's registers */
50 #define ATAPI_OFFSET_CONTROL 0x00
51 #define ATAPI_OFFSET_STATUS 0x04
52 #define ATAPI_OFFSET_DEV_ADDR 0x08
53 #define ATAPI_OFFSET_DEV_TXBUF 0x0c
54 #define ATAPI_OFFSET_DEV_RXBUF 0x10
55 #define ATAPI_OFFSET_INT_MASK 0x14
56 #define ATAPI_OFFSET_INT_STATUS 0x18
57 #define ATAPI_OFFSET_XFER_LEN 0x1c
58 #define ATAPI_OFFSET_LINE_STATUS 0x20
59 #define ATAPI_OFFSET_SM_STATE 0x24
60 #define ATAPI_OFFSET_TERMINATE 0x28
61 #define ATAPI_OFFSET_PIO_TFRCNT 0x2c
62 #define ATAPI_OFFSET_DMA_TFRCNT 0x30
63 #define ATAPI_OFFSET_UMAIN_TFRCNT 0x34
64 #define ATAPI_OFFSET_UDMAOUT_TFRCNT 0x38
65 #define ATAPI_OFFSET_REG_TIM_0 0x40
66 #define ATAPI_OFFSET_PIO_TIM_0 0x44
67 #define ATAPI_OFFSET_PIO_TIM_1 0x48
68 #define ATAPI_OFFSET_MULTI_TIM_0 0x50
69 #define ATAPI_OFFSET_MULTI_TIM_1 0x54
70 #define ATAPI_OFFSET_MULTI_TIM_2 0x58
71 #define ATAPI_OFFSET_ULTRA_TIM_0 0x60
72 #define ATAPI_OFFSET_ULTRA_TIM_1 0x64
73 #define ATAPI_OFFSET_ULTRA_TIM_2 0x68
74 #define ATAPI_OFFSET_ULTRA_TIM_3 0x6c
75 
76 
77 #define ATAPI_GET_CONTROL(base)\
78  bfin_read16(base + ATAPI_OFFSET_CONTROL)
79 #define ATAPI_SET_CONTROL(base, val)\
80  bfin_write16(base + ATAPI_OFFSET_CONTROL, val)
81 #define ATAPI_GET_STATUS(base)\
82  bfin_read16(base + ATAPI_OFFSET_STATUS)
83 #define ATAPI_GET_DEV_ADDR(base)\
84  bfin_read16(base + ATAPI_OFFSET_DEV_ADDR)
85 #define ATAPI_SET_DEV_ADDR(base, val)\
86  bfin_write16(base + ATAPI_OFFSET_DEV_ADDR, val)
87 #define ATAPI_GET_DEV_TXBUF(base)\
88  bfin_read16(base + ATAPI_OFFSET_DEV_TXBUF)
89 #define ATAPI_SET_DEV_TXBUF(base, val)\
90  bfin_write16(base + ATAPI_OFFSET_DEV_TXBUF, val)
91 #define ATAPI_GET_DEV_RXBUF(base)\
92  bfin_read16(base + ATAPI_OFFSET_DEV_RXBUF)
93 #define ATAPI_SET_DEV_RXBUF(base, val)\
94  bfin_write16(base + ATAPI_OFFSET_DEV_RXBUF, val)
95 #define ATAPI_GET_INT_MASK(base)\
96  bfin_read16(base + ATAPI_OFFSET_INT_MASK)
97 #define ATAPI_SET_INT_MASK(base, val)\
98  bfin_write16(base + ATAPI_OFFSET_INT_MASK, val)
99 #define ATAPI_GET_INT_STATUS(base)\
100  bfin_read16(base + ATAPI_OFFSET_INT_STATUS)
101 #define ATAPI_SET_INT_STATUS(base, val)\
102  bfin_write16(base + ATAPI_OFFSET_INT_STATUS, val)
103 #define ATAPI_GET_XFER_LEN(base)\
104  bfin_read16(base + ATAPI_OFFSET_XFER_LEN)
105 #define ATAPI_SET_XFER_LEN(base, val)\
106  bfin_write16(base + ATAPI_OFFSET_XFER_LEN, val)
107 #define ATAPI_GET_LINE_STATUS(base)\
108  bfin_read16(base + ATAPI_OFFSET_LINE_STATUS)
109 #define ATAPI_GET_SM_STATE(base)\
110  bfin_read16(base + ATAPI_OFFSET_SM_STATE)
111 #define ATAPI_GET_TERMINATE(base)\
112  bfin_read16(base + ATAPI_OFFSET_TERMINATE)
113 #define ATAPI_SET_TERMINATE(base, val)\
114  bfin_write16(base + ATAPI_OFFSET_TERMINATE, val)
115 #define ATAPI_GET_PIO_TFRCNT(base)\
116  bfin_read16(base + ATAPI_OFFSET_PIO_TFRCNT)
117 #define ATAPI_GET_DMA_TFRCNT(base)\
118  bfin_read16(base + ATAPI_OFFSET_DMA_TFRCNT)
119 #define ATAPI_GET_UMAIN_TFRCNT(base)\
120  bfin_read16(base + ATAPI_OFFSET_UMAIN_TFRCNT)
121 #define ATAPI_GET_UDMAOUT_TFRCNT(base)\
122  bfin_read16(base + ATAPI_OFFSET_UDMAOUT_TFRCNT)
123 #define ATAPI_GET_REG_TIM_0(base)\
124  bfin_read16(base + ATAPI_OFFSET_REG_TIM_0)
125 #define ATAPI_SET_REG_TIM_0(base, val)\
126  bfin_write16(base + ATAPI_OFFSET_REG_TIM_0, val)
127 #define ATAPI_GET_PIO_TIM_0(base)\
128  bfin_read16(base + ATAPI_OFFSET_PIO_TIM_0)
129 #define ATAPI_SET_PIO_TIM_0(base, val)\
130  bfin_write16(base + ATAPI_OFFSET_PIO_TIM_0, val)
131 #define ATAPI_GET_PIO_TIM_1(base)\
132  bfin_read16(base + ATAPI_OFFSET_PIO_TIM_1)
133 #define ATAPI_SET_PIO_TIM_1(base, val)\
134  bfin_write16(base + ATAPI_OFFSET_PIO_TIM_1, val)
135 #define ATAPI_GET_MULTI_TIM_0(base)\
136  bfin_read16(base + ATAPI_OFFSET_MULTI_TIM_0)
137 #define ATAPI_SET_MULTI_TIM_0(base, val)\
138  bfin_write16(base + ATAPI_OFFSET_MULTI_TIM_0, val)
139 #define ATAPI_GET_MULTI_TIM_1(base)\
140  bfin_read16(base + ATAPI_OFFSET_MULTI_TIM_1)
141 #define ATAPI_SET_MULTI_TIM_1(base, val)\
142  bfin_write16(base + ATAPI_OFFSET_MULTI_TIM_1, val)
143 #define ATAPI_GET_MULTI_TIM_2(base)\
144  bfin_read16(base + ATAPI_OFFSET_MULTI_TIM_2)
145 #define ATAPI_SET_MULTI_TIM_2(base, val)\
146  bfin_write16(base + ATAPI_OFFSET_MULTI_TIM_2, val)
147 #define ATAPI_GET_ULTRA_TIM_0(base)\
148  bfin_read16(base + ATAPI_OFFSET_ULTRA_TIM_0)
149 #define ATAPI_SET_ULTRA_TIM_0(base, val)\
150  bfin_write16(base + ATAPI_OFFSET_ULTRA_TIM_0, val)
151 #define ATAPI_GET_ULTRA_TIM_1(base)\
152  bfin_read16(base + ATAPI_OFFSET_ULTRA_TIM_1)
153 #define ATAPI_SET_ULTRA_TIM_1(base, val)\
154  bfin_write16(base + ATAPI_OFFSET_ULTRA_TIM_1, val)
155 #define ATAPI_GET_ULTRA_TIM_2(base)\
156  bfin_read16(base + ATAPI_OFFSET_ULTRA_TIM_2)
157 #define ATAPI_SET_ULTRA_TIM_2(base, val)\
158  bfin_write16(base + ATAPI_OFFSET_ULTRA_TIM_2, val)
159 #define ATAPI_GET_ULTRA_TIM_3(base)\
160  bfin_read16(base + ATAPI_OFFSET_ULTRA_TIM_3)
161 #define ATAPI_SET_ULTRA_TIM_3(base, val)\
162  bfin_write16(base + ATAPI_OFFSET_ULTRA_TIM_3, val)
163 
167 /* mode: 0 1 2 3 4 */
168 static const u32 pio_fsclk[] =
169 { 33333333, 33333333, 33333333, 33333333, 33333333 };
170 
174 /* mode: 0 1 2 */
175 static const u32 mdma_fsclk[] = { 33333333, 33333333, 33333333 };
176 
185 /* mode: 0 1 2 3 4 5 */
186 static const u32 udma_fsclk[] =
187 { 33333333, 33333333, 40000000, 50000000, 80000000, 133333333 };
188 
192 /* mode: 0 1 2 3 4 */
193 /* Cycle Time */
194 static const u32 reg_t0min[] = { 600, 383, 330, 180, 120 };
195 /* DIOR/DIOW to end cycle */
196 static const u32 reg_t2min[] = { 290, 290, 290, 70, 25 };
197 /* DIOR/DIOW asserted pulse width */
198 static const u32 reg_teocmin[] = { 290, 290, 290, 80, 70 };
199 
203 /* mode: 0 1 2 3 4 */
204 /* Cycle Time */
205 static const u32 pio_t0min[] = { 600, 383, 240, 180, 120 };
206 /* Address valid to DIOR/DIORW */
207 static const u32 pio_t1min[] = { 70, 50, 30, 30, 25 };
208 /* DIOR/DIOW to end cycle */
209 static const u32 pio_t2min[] = { 165, 125, 100, 80, 70 };
210 /* DIOR/DIOW asserted pulse width */
211 static const u32 pio_teocmin[] = { 165, 125, 100, 70, 25 };
212 /* DIOW data hold */
213 static const u32 pio_t4min[] = { 30, 20, 15, 10, 10 };
214 
215 /* ******************************************************************
216  * Multiword DMA timing table
217  * ******************************************************************
218  */
219 /* mode: 0 1 2 */
220 /* Cycle Time */
221 static const u32 mdma_t0min[] = { 480, 150, 120 };
222 /* DIOR/DIOW asserted pulse width */
223 static const u32 mdma_tdmin[] = { 215, 80, 70 };
224 /* DMACK to read data released */
225 static const u32 mdma_thmin[] = { 20, 15, 10 };
226 /* DIOR/DIOW to DMACK hold */
227 static const u32 mdma_tjmin[] = { 20, 5, 5 };
228 /* DIOR negated pulse width */
229 static const u32 mdma_tkrmin[] = { 50, 50, 25 };
230 /* DIOR negated pulse width */
231 static const u32 mdma_tkwmin[] = { 215, 50, 25 };
232 /* CS[1:0] valid to DIOR/DIOW */
233 static const u32 mdma_tmmin[] = { 50, 30, 25 };
234 /* DMACK to read data released */
235 static const u32 mdma_tzmax[] = { 20, 25, 25 };
236 
240 /* mode: 0 1 2 3 4 5 */
241 static const u32 udma_tcycmin[] = { 112, 73, 54, 39, 25, 17 };
242 static const u32 udma_tdvsmin[] = { 70, 48, 31, 20, 7, 5 };
243 static const u32 udma_tenvmax[] = { 70, 70, 70, 55, 55, 50 };
244 static const u32 udma_trpmin[] = { 160, 125, 100, 100, 100, 85 };
245 static const u32 udma_tmin[] = { 5, 5, 5, 5, 3, 3 };
246 
247 
248 static const u32 udma_tmlimin = 20;
249 static const u32 udma_tzahmin = 20;
250 static const u32 udma_tenvmin = 20;
251 static const u32 udma_tackmin = 20;
252 static const u32 udma_tssmin = 50;
253 
254 #define BFIN_MAX_SG_SEGMENTS 4
255 
263 static unsigned short num_clocks_min(unsigned long tmin,
264  unsigned long fsclk)
265 {
266  unsigned long tmp ;
267  unsigned short result;
268 
269  tmp = tmin * (fsclk/1000/1000) / 1000;
270  result = (unsigned short)tmp;
271  if ((tmp*1000*1000) < (tmin*(fsclk/1000))) {
272  result++;
273  }
274 
275  return result;
276 }
277 
289 static void bfin_set_piomode(struct ata_port *ap, struct ata_device *adev)
290 {
291  int mode = adev->pio_mode - XFER_PIO_0;
292  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
293  unsigned int fsclk = get_sclk();
294  unsigned short teoc_reg, t2_reg, teoc_pio;
295  unsigned short t4_reg, t2_pio, t1_reg;
296  unsigned short n0, n6, t6min = 5;
297 
298  /* the most restrictive timing value is t6 and tc, the DIOW - data hold
299  * If one SCLK pulse is longer than this minimum value then register
300  * transfers cannot be supported at this frequency.
301  */
302  n6 = num_clocks_min(t6min, fsclk);
303  if (mode >= 0 && mode <= 4 && n6 >= 1) {
304  dev_dbg(adev->link->ap->dev, "set piomode: mode=%d, fsclk=%ud\n", mode, fsclk);
305  /* calculate the timing values for register transfers. */
306  while (mode > 0 && pio_fsclk[mode] > fsclk)
307  mode--;
308 
309  /* DIOR/DIOW to end cycle time */
310  t2_reg = num_clocks_min(reg_t2min[mode], fsclk);
311  /* DIOR/DIOW asserted pulse width */
312  teoc_reg = num_clocks_min(reg_teocmin[mode], fsclk);
313  /* Cycle Time */
314  n0 = num_clocks_min(reg_t0min[mode], fsclk);
315 
316  /* increase t2 until we meed the minimum cycle length */
317  if (t2_reg + teoc_reg < n0)
318  t2_reg = n0 - teoc_reg;
319 
320  /* calculate the timing values for pio transfers. */
321 
322  /* DIOR/DIOW to end cycle time */
323  t2_pio = num_clocks_min(pio_t2min[mode], fsclk);
324  /* DIOR/DIOW asserted pulse width */
325  teoc_pio = num_clocks_min(pio_teocmin[mode], fsclk);
326  /* Cycle Time */
327  n0 = num_clocks_min(pio_t0min[mode], fsclk);
328 
329  /* increase t2 until we meed the minimum cycle length */
330  if (t2_pio + teoc_pio < n0)
331  t2_pio = n0 - teoc_pio;
332 
333  /* Address valid to DIOR/DIORW */
334  t1_reg = num_clocks_min(pio_t1min[mode], fsclk);
335 
336  /* DIOW data hold */
337  t4_reg = num_clocks_min(pio_t4min[mode], fsclk);
338 
339  ATAPI_SET_REG_TIM_0(base, (teoc_reg<<8 | t2_reg));
340  ATAPI_SET_PIO_TIM_0(base, (t4_reg<<12 | t2_pio<<4 | t1_reg));
341  ATAPI_SET_PIO_TIM_1(base, teoc_pio);
342  if (mode > 2) {
343  ATAPI_SET_CONTROL(base,
344  ATAPI_GET_CONTROL(base) | IORDY_EN);
345  } else {
346  ATAPI_SET_CONTROL(base,
347  ATAPI_GET_CONTROL(base) & ~IORDY_EN);
348  }
349 
350  /* Disable host ATAPI PIO interrupts */
353  SSYNC();
354  }
355 }
356 
368 static void bfin_set_dmamode(struct ata_port *ap, struct ata_device *adev)
369 {
370  int mode;
371  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
372  unsigned long fsclk = get_sclk();
373  unsigned short tenv, tack, tcyc_tdvs, tdvs, tmli, tss, trp, tzah;
374  unsigned short tm, td, tkr, tkw, teoc, th;
375  unsigned short n0, nf, tfmin = 5;
376  unsigned short nmin, tcyc;
377 
378  mode = adev->dma_mode - XFER_UDMA_0;
379  if (mode >= 0 && mode <= 5) {
380  dev_dbg(adev->link->ap->dev, "set udmamode: mode=%d\n", mode);
381  /* the most restrictive timing value is t6 and tc,
382  * the DIOW - data hold. If one SCLK pulse is longer
383  * than this minimum value then register
384  * transfers cannot be supported at this frequency.
385  */
386  while (mode > 0 && udma_fsclk[mode] > fsclk)
387  mode--;
388 
389  nmin = num_clocks_min(udma_tmin[mode], fsclk);
390  if (nmin >= 1) {
391  /* calculate the timing values for Ultra DMA. */
392  tdvs = num_clocks_min(udma_tdvsmin[mode], fsclk);
393  tcyc = num_clocks_min(udma_tcycmin[mode], fsclk);
394  tcyc_tdvs = 2;
395 
396  /* increase tcyc - tdvs (tcyc_tdvs) until we meed
397  * the minimum cycle length
398  */
399  if (tdvs + tcyc_tdvs < tcyc)
400  tcyc_tdvs = tcyc - tdvs;
401 
402  /* Mow assign the values required for the timing
403  * registers
404  */
405  if (tcyc_tdvs < 2)
406  tcyc_tdvs = 2;
407 
408  if (tdvs < 2)
409  tdvs = 2;
410 
411  tack = num_clocks_min(udma_tackmin, fsclk);
412  tss = num_clocks_min(udma_tssmin, fsclk);
413  tmli = num_clocks_min(udma_tmlimin, fsclk);
414  tzah = num_clocks_min(udma_tzahmin, fsclk);
415  trp = num_clocks_min(udma_trpmin[mode], fsclk);
416  tenv = num_clocks_min(udma_tenvmin, fsclk);
417  if (tenv <= udma_tenvmax[mode]) {
418  ATAPI_SET_ULTRA_TIM_0(base, (tenv<<8 | tack));
420  (tcyc_tdvs<<8 | tdvs));
421  ATAPI_SET_ULTRA_TIM_2(base, (tmli<<8 | tss));
422  ATAPI_SET_ULTRA_TIM_3(base, (trp<<8 | tzah));
423  }
424  }
425  }
426 
427  mode = adev->dma_mode - XFER_MW_DMA_0;
428  if (mode >= 0 && mode <= 2) {
429  dev_dbg(adev->link->ap->dev, "set mdmamode: mode=%d\n", mode);
430  /* the most restrictive timing value is tf, the DMACK to
431  * read data released. If one SCLK pulse is longer than
432  * this maximum value then the MDMA mode
433  * cannot be supported at this frequency.
434  */
435  while (mode > 0 && mdma_fsclk[mode] > fsclk)
436  mode--;
437 
438  nf = num_clocks_min(tfmin, fsclk);
439  if (nf >= 1) {
440  /* calculate the timing values for Multi-word DMA. */
441 
442  /* DIOR/DIOW asserted pulse width */
443  td = num_clocks_min(mdma_tdmin[mode], fsclk);
444 
445  /* DIOR negated pulse width */
446  tkw = num_clocks_min(mdma_tkwmin[mode], fsclk);
447 
448  /* Cycle Time */
449  n0 = num_clocks_min(mdma_t0min[mode], fsclk);
450 
451  /* increase tk until we meed the minimum cycle length */
452  if (tkw + td < n0)
453  tkw = n0 - td;
454 
455  /* DIOR negated pulse width - read */
456  tkr = num_clocks_min(mdma_tkrmin[mode], fsclk);
457  /* CS{1:0] valid to DIOR/DIOW */
458  tm = num_clocks_min(mdma_tmmin[mode], fsclk);
459  /* DIOR/DIOW to DMACK hold */
460  teoc = num_clocks_min(mdma_tjmin[mode], fsclk);
461  /* DIOW Data hold */
462  th = num_clocks_min(mdma_thmin[mode], fsclk);
463 
464  ATAPI_SET_MULTI_TIM_0(base, (tm<<8 | td));
465  ATAPI_SET_MULTI_TIM_1(base, (tkr<<8 | tkw));
466  ATAPI_SET_MULTI_TIM_2(base, (teoc<<8 | th));
467  SSYNC();
468  }
469  }
470  return;
471 }
472 
480 static inline void wait_complete(void __iomem *base, unsigned short mask)
481 {
482  unsigned short status;
483  unsigned int i = 0;
484 
485 #define PATA_BF54X_WAIT_TIMEOUT 10000
486 
487  for (i = 0; i < PATA_BF54X_WAIT_TIMEOUT; i++) {
488  status = ATAPI_GET_INT_STATUS(base) & mask;
489  if (status)
490  break;
491  }
492 
493  ATAPI_SET_INT_STATUS(base, mask);
494 }
495 
504 static void write_atapi_register(void __iomem *base,
505  unsigned long ata_reg, unsigned short value)
506 {
507  /* Program the ATA_DEV_TXBUF register with write data (to be
508  * written into the device).
509  */
510  ATAPI_SET_DEV_TXBUF(base, value);
511 
512  /* Program the ATA_DEV_ADDR register with address of the
513  * device register (0x01 to 0x0F).
514  */
515  ATAPI_SET_DEV_ADDR(base, ata_reg);
516 
517  /* Program the ATA_CTRL register with dir set to write (1)
518  */
520 
521  /* ensure PIO DMA is not set */
523 
524  /* and start the transfer */
526 
527  /* Wait for the interrupt to indicate the end of the transfer.
528  * (We need to wait on and clear rhe ATA_DEV_INT interrupt status)
529  */
530  wait_complete(base, PIO_DONE_INT);
531 }
532 
541 static unsigned short read_atapi_register(void __iomem *base,
542  unsigned long ata_reg)
543 {
544  /* Program the ATA_DEV_ADDR register with address of the
545  * device register (0x01 to 0x0F).
546  */
547  ATAPI_SET_DEV_ADDR(base, ata_reg);
548 
549  /* Program the ATA_CTRL register with dir set to read (0) and
550  */
551  ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) & ~XFER_DIR));
552 
553  /* ensure PIO DMA is not set */
555 
556  /* and start the transfer */
558 
559  /* Wait for the interrupt to indicate the end of the transfer.
560  * (PIO_DONE interrupt is set and it doesn't seem to matter
561  * that we don't clear it)
562  */
563  wait_complete(base, PIO_DONE_INT);
564 
565  /* Read the ATA_DEV_RXBUF register with write data (to be
566  * written into the device).
567  */
568  return ATAPI_GET_DEV_RXBUF(base);
569 }
570 
579 static void write_atapi_data(void __iomem *base,
580  int len, unsigned short *buf)
581 {
582  int i;
583 
584  /* Set transfer length to 1 */
585  ATAPI_SET_XFER_LEN(base, 1);
586 
587  /* Program the ATA_DEV_ADDR register with address of the
588  * ATA_REG_DATA
589  */
591 
592  /* Program the ATA_CTRL register with dir set to write (1)
593  */
595 
596  /* ensure PIO DMA is not set */
598 
599  for (i = 0; i < len; i++) {
600  /* Program the ATA_DEV_TXBUF register with write data (to be
601  * written into the device).
602  */
603  ATAPI_SET_DEV_TXBUF(base, buf[i]);
604 
605  /* and start the transfer */
607 
608  /* Wait for the interrupt to indicate the end of the transfer.
609  * (We need to wait on and clear rhe ATA_DEV_INT
610  * interrupt status)
611  */
612  wait_complete(base, PIO_DONE_INT);
613  }
614 }
615 
624 static void read_atapi_data(void __iomem *base,
625  int len, unsigned short *buf)
626 {
627  int i;
628 
629  /* Set transfer length to 1 */
630  ATAPI_SET_XFER_LEN(base, 1);
631 
632  /* Program the ATA_DEV_ADDR register with address of the
633  * ATA_REG_DATA
634  */
636 
637  /* Program the ATA_CTRL register with dir set to read (0) and
638  */
639  ATAPI_SET_CONTROL(base, (ATAPI_GET_CONTROL(base) & ~XFER_DIR));
640 
641  /* ensure PIO DMA is not set */
643 
644  for (i = 0; i < len; i++) {
645  /* and start the transfer */
647 
648  /* Wait for the interrupt to indicate the end of the transfer.
649  * (PIO_DONE interrupt is set and it doesn't seem to matter
650  * that we don't clear it)
651  */
652  wait_complete(base, PIO_DONE_INT);
653 
654  /* Read the ATA_DEV_RXBUF register with write data (to be
655  * written into the device).
656  */
657  buf[i] = ATAPI_GET_DEV_RXBUF(base);
658  }
659 }
660 
669 static void bfin_tf_load(struct ata_port *ap, const struct ata_taskfile *tf)
670 {
671  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
672  unsigned int is_addr = tf->flags & ATA_TFLAG_ISADDR;
673 
674  if (tf->ctl != ap->last_ctl) {
675  write_atapi_register(base, ATA_REG_CTRL, tf->ctl);
676  ap->last_ctl = tf->ctl;
677  ata_wait_idle(ap);
678  }
679 
680  if (is_addr) {
681  if (tf->flags & ATA_TFLAG_LBA48) {
682  write_atapi_register(base, ATA_REG_FEATURE,
683  tf->hob_feature);
684  write_atapi_register(base, ATA_REG_NSECT,
685  tf->hob_nsect);
686  write_atapi_register(base, ATA_REG_LBAL, tf->hob_lbal);
687  write_atapi_register(base, ATA_REG_LBAM, tf->hob_lbam);
688  write_atapi_register(base, ATA_REG_LBAH, tf->hob_lbah);
689  dev_dbg(ap->dev, "hob: feat 0x%X nsect 0x%X, lba 0x%X "
690  "0x%X 0x%X\n",
691  tf->hob_feature,
692  tf->hob_nsect,
693  tf->hob_lbal,
694  tf->hob_lbam,
695  tf->hob_lbah);
696  }
697 
698  write_atapi_register(base, ATA_REG_FEATURE, tf->feature);
699  write_atapi_register(base, ATA_REG_NSECT, tf->nsect);
700  write_atapi_register(base, ATA_REG_LBAL, tf->lbal);
701  write_atapi_register(base, ATA_REG_LBAM, tf->lbam);
702  write_atapi_register(base, ATA_REG_LBAH, tf->lbah);
703  dev_dbg(ap->dev, "feat 0x%X nsect 0x%X lba 0x%X 0x%X 0x%X\n",
704  tf->feature,
705  tf->nsect,
706  tf->lbal,
707  tf->lbam,
708  tf->lbah);
709  }
710 
711  if (tf->flags & ATA_TFLAG_DEVICE) {
712  write_atapi_register(base, ATA_REG_DEVICE, tf->device);
713  dev_dbg(ap->dev, "device 0x%X\n", tf->device);
714  }
715 
716  ata_wait_idle(ap);
717 }
718 
726 static u8 bfin_check_status(struct ata_port *ap)
727 {
728  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
729  return read_atapi_register(base, ATA_REG_STATUS);
730 }
731 
740 static void bfin_tf_read(struct ata_port *ap, struct ata_taskfile *tf)
741 {
742  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
743 
744  tf->command = bfin_check_status(ap);
745  tf->feature = read_atapi_register(base, ATA_REG_ERR);
746  tf->nsect = read_atapi_register(base, ATA_REG_NSECT);
747  tf->lbal = read_atapi_register(base, ATA_REG_LBAL);
748  tf->lbam = read_atapi_register(base, ATA_REG_LBAM);
749  tf->lbah = read_atapi_register(base, ATA_REG_LBAH);
750  tf->device = read_atapi_register(base, ATA_REG_DEVICE);
751 
752  if (tf->flags & ATA_TFLAG_LBA48) {
753  write_atapi_register(base, ATA_REG_CTRL, tf->ctl | ATA_HOB);
754  tf->hob_feature = read_atapi_register(base, ATA_REG_ERR);
755  tf->hob_nsect = read_atapi_register(base, ATA_REG_NSECT);
756  tf->hob_lbal = read_atapi_register(base, ATA_REG_LBAL);
757  tf->hob_lbam = read_atapi_register(base, ATA_REG_LBAM);
758  tf->hob_lbah = read_atapi_register(base, ATA_REG_LBAH);
759  }
760 }
761 
770 static void bfin_exec_command(struct ata_port *ap,
771  const struct ata_taskfile *tf)
772 {
773  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
774  dev_dbg(ap->dev, "ata%u: cmd 0x%X\n", ap->print_id, tf->command);
775 
776  write_atapi_register(base, ATA_REG_CMD, tf->command);
777  ata_sff_pause(ap);
778 }
779 
785 static u8 bfin_check_altstatus(struct ata_port *ap)
786 {
787  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
788  return read_atapi_register(base, ATA_REG_ALTSTATUS);
789 }
790 
799 static void bfin_dev_select(struct ata_port *ap, unsigned int device)
800 {
801  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
802  u8 tmp;
803 
804  if (device == 0)
806  else
808 
809  write_atapi_register(base, ATA_REG_DEVICE, tmp);
810  ata_sff_pause(ap);
811 }
812 
819 static void bfin_set_devctl(struct ata_port *ap, u8 ctl)
820 {
821  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
822  write_atapi_register(base, ATA_REG_CTRL, ctl);
823 }
824 
832 static void bfin_bmdma_setup(struct ata_queued_cmd *qc)
833 {
834  struct ata_port *ap = qc->ap;
835  struct dma_desc_array *dma_desc_cpu = (struct dma_desc_array *)ap->bmdma_prd;
836  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
837  unsigned short config = DMAFLOW_ARRAY | NDSIZE_5 | RESTART | WDSIZE_16 | DMAEN;
838  struct scatterlist *sg;
839  unsigned int si;
840  unsigned int channel;
841  unsigned int dir;
842  unsigned int size = 0;
843 
844  dev_dbg(qc->ap->dev, "in atapi dma setup\n");
845  /* Program the ATA_CTRL register with dir */
846  if (qc->tf.flags & ATA_TFLAG_WRITE) {
847  channel = CH_ATAPI_TX;
848  dir = DMA_TO_DEVICE;
849  } else {
850  channel = CH_ATAPI_RX;
851  dir = DMA_FROM_DEVICE;
852  config |= WNR;
853  }
854 
855  dma_map_sg(ap->dev, qc->sg, qc->n_elem, dir);
856 
857  /* fill the ATAPI DMA controller */
858  for_each_sg(qc->sg, sg, qc->n_elem, si) {
859  dma_desc_cpu[si].start_addr = sg_dma_address(sg);
860  dma_desc_cpu[si].cfg = config;
861  dma_desc_cpu[si].x_count = sg_dma_len(sg) >> 1;
862  dma_desc_cpu[si].x_modify = 2;
863  size += sg_dma_len(sg);
864  }
865 
866  /* Set the last descriptor to stop mode */
867  dma_desc_cpu[qc->n_elem - 1].cfg &= ~(DMAFLOW | NDSIZE);
868 
869  flush_dcache_range((unsigned int)dma_desc_cpu,
870  (unsigned int)dma_desc_cpu +
871  qc->n_elem * sizeof(struct dma_desc_array));
872 
873  /* Enable ATA DMA operation*/
874  set_dma_curr_desc_addr(channel, (unsigned long *)ap->bmdma_prd_dma);
875  set_dma_x_count(channel, 0);
876  set_dma_x_modify(channel, 0);
877  set_dma_config(channel, config);
878 
879  SSYNC();
880 
881  /* Send ATA DMA command */
882  bfin_exec_command(ap, &qc->tf);
883 
884  if (qc->tf.flags & ATA_TFLAG_WRITE) {
885  /* set ATA DMA write direction */
887  | XFER_DIR));
888  } else {
889  /* set ATA DMA read direction */
891  & ~XFER_DIR));
892  }
893 
894  /* Reset all transfer count */
896 
897  /* Set ATAPI state machine contorl in terminate sequence */
899 
900  /* Set transfer length to the total size of sg buffers */
901  ATAPI_SET_XFER_LEN(base, size >> 1);
902 }
903 
911 static void bfin_bmdma_start(struct ata_queued_cmd *qc)
912 {
913  struct ata_port *ap = qc->ap;
914  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
915 
916  dev_dbg(qc->ap->dev, "in atapi dma start\n");
917 
918  if (!(ap->udma_mask || ap->mwdma_mask))
919  return;
920 
921  /* start ATAPI transfer*/
922  if (ap->udma_mask)
924  | ULTRA_START);
925  else
927  | MULTI_START);
928 }
929 
935 static void bfin_bmdma_stop(struct ata_queued_cmd *qc)
936 {
937  struct ata_port *ap = qc->ap;
938  unsigned int dir;
939 
940  dev_dbg(qc->ap->dev, "in atapi dma stop\n");
941 
942  if (!(ap->udma_mask || ap->mwdma_mask))
943  return;
944 
945  /* stop ATAPI DMA controller*/
946  if (qc->tf.flags & ATA_TFLAG_WRITE) {
947  dir = DMA_TO_DEVICE;
949  } else {
950  dir = DMA_FROM_DEVICE;
952  }
953 
954  dma_unmap_sg(ap->dev, qc->sg, qc->n_elem, dir);
955 }
956 
965 static unsigned int bfin_devchk(struct ata_port *ap,
966  unsigned int device)
967 {
968  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
969  u8 nsect, lbal;
970 
971  bfin_dev_select(ap, device);
972 
973  write_atapi_register(base, ATA_REG_NSECT, 0x55);
974  write_atapi_register(base, ATA_REG_LBAL, 0xaa);
975 
976  write_atapi_register(base, ATA_REG_NSECT, 0xaa);
977  write_atapi_register(base, ATA_REG_LBAL, 0x55);
978 
979  write_atapi_register(base, ATA_REG_NSECT, 0x55);
980  write_atapi_register(base, ATA_REG_LBAL, 0xaa);
981 
982  nsect = read_atapi_register(base, ATA_REG_NSECT);
983  lbal = read_atapi_register(base, ATA_REG_LBAL);
984 
985  if ((nsect == 0x55) && (lbal == 0xaa))
986  return 1; /* we found a device */
987 
988  return 0; /* nothing found */
989 }
990 
997 static void bfin_bus_post_reset(struct ata_port *ap, unsigned int devmask)
998 {
999  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1000  unsigned int dev0 = devmask & (1 << 0);
1001  unsigned int dev1 = devmask & (1 << 1);
1002  unsigned long deadline;
1003 
1004  /* if device 0 was found in ata_devchk, wait for its
1005  * BSY bit to clear
1006  */
1007  if (dev0)
1009 
1010  /* if device 1 was found in ata_devchk, wait for
1011  * register access, then wait for BSY to clear
1012  */
1013  deadline = ata_deadline(jiffies, ATA_TMOUT_BOOT);
1014  while (dev1) {
1015  u8 nsect, lbal;
1016 
1017  bfin_dev_select(ap, 1);
1018  nsect = read_atapi_register(base, ATA_REG_NSECT);
1019  lbal = read_atapi_register(base, ATA_REG_LBAL);
1020  if ((nsect == 1) && (lbal == 1))
1021  break;
1022  if (time_after(jiffies, deadline)) {
1023  dev1 = 0;
1024  break;
1025  }
1026  ata_msleep(ap, 50); /* give drive a breather */
1027  }
1028  if (dev1)
1030 
1031  /* is all this really necessary? */
1032  bfin_dev_select(ap, 0);
1033  if (dev1)
1034  bfin_dev_select(ap, 1);
1035  if (dev0)
1036  bfin_dev_select(ap, 0);
1037 }
1038 
1045 static unsigned int bfin_bus_softreset(struct ata_port *ap,
1046  unsigned int devmask)
1047 {
1048  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1049 
1050  /* software reset. causes dev0 to be selected */
1051  write_atapi_register(base, ATA_REG_CTRL, ap->ctl);
1052  udelay(20);
1053  write_atapi_register(base, ATA_REG_CTRL, ap->ctl | ATA_SRST);
1054  udelay(20);
1055  write_atapi_register(base, ATA_REG_CTRL, ap->ctl);
1056 
1057  /* spec mandates ">= 2ms" before checking status.
1058  * We wait 150ms, because that was the magic delay used for
1059  * ATAPI devices in Hale Landis's ATADRVR, for the period of time
1060  * between when the ATA command register is written, and then
1061  * status is checked. Because waiting for "a while" before
1062  * checking status is fine, post SRST, we perform this magic
1063  * delay here as well.
1064  *
1065  * Old drivers/ide uses the 2mS rule and then waits for ready
1066  */
1067  ata_msleep(ap, 150);
1068 
1069  /* Before we perform post reset processing we want to see if
1070  * the bus shows 0xFF because the odd clown forgets the D7
1071  * pulldown resistor.
1072  */
1073  if (bfin_check_status(ap) == 0xFF)
1074  return 0;
1075 
1076  bfin_bus_post_reset(ap, devmask);
1077 
1078  return 0;
1079 }
1080 
1089 static int bfin_softreset(struct ata_link *link, unsigned int *classes,
1090  unsigned long deadline)
1091 {
1092  struct ata_port *ap = link->ap;
1093  unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS;
1094  unsigned int devmask = 0, err_mask;
1095  u8 err;
1096 
1097  /* determine if device 0/1 are present */
1098  if (bfin_devchk(ap, 0))
1099  devmask |= (1 << 0);
1100  if (slave_possible && bfin_devchk(ap, 1))
1101  devmask |= (1 << 1);
1102 
1103  /* select device 0 again */
1104  bfin_dev_select(ap, 0);
1105 
1106  /* issue bus reset */
1107  err_mask = bfin_bus_softreset(ap, devmask);
1108  if (err_mask) {
1109  ata_port_err(ap, "SRST failed (err_mask=0x%x)\n",
1110  err_mask);
1111  return -EIO;
1112  }
1113 
1114  /* determine by signature whether we have ATA or ATAPI devices */
1115  classes[0] = ata_sff_dev_classify(&ap->link.device[0],
1116  devmask & (1 << 0), &err);
1117  if (slave_possible && err != 0x81)
1118  classes[1] = ata_sff_dev_classify(&ap->link.device[1],
1119  devmask & (1 << 1), &err);
1120 
1121  return 0;
1122 }
1123 
1129 static unsigned char bfin_bmdma_status(struct ata_port *ap)
1130 {
1131  unsigned char host_stat = 0;
1132  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1133 
1135  host_stat |= ATA_DMA_ACTIVE;
1136  if (ATAPI_GET_INT_STATUS(base) & ATAPI_DEV_INT)
1137  host_stat |= ATA_DMA_INTR;
1138 
1139  dev_dbg(ap->dev, "ATAPI: host_stat=0x%x\n", host_stat);
1140 
1141  return host_stat;
1142 }
1143 
1154 static unsigned int bfin_data_xfer(struct ata_device *dev, unsigned char *buf,
1155  unsigned int buflen, int rw)
1156 {
1157  struct ata_port *ap = dev->link->ap;
1158  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1159  unsigned int words = buflen >> 1;
1160  unsigned short *buf16 = (u16 *)buf;
1161 
1162  /* Transfer multiple of 2 bytes */
1163  if (rw == READ)
1164  read_atapi_data(base, words, buf16);
1165  else
1166  write_atapi_data(base, words, buf16);
1167 
1168  /* Transfer trailing 1 byte, if any. */
1169  if (unlikely(buflen & 0x01)) {
1170  unsigned short align_buf[1] = { 0 };
1171  unsigned char *trailing_buf = buf + buflen - 1;
1172 
1173  if (rw == READ) {
1174  read_atapi_data(base, 1, align_buf);
1175  memcpy(trailing_buf, align_buf, 1);
1176  } else {
1177  memcpy(align_buf, trailing_buf, 1);
1178  write_atapi_data(base, 1, align_buf);
1179  }
1180  words++;
1181  }
1182 
1183  return words << 1;
1184 }
1185 
1193 static void bfin_irq_clear(struct ata_port *ap)
1194 {
1195  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1196 
1197  dev_dbg(ap->dev, "in atapi irq clear\n");
1201 }
1202 
1210 void bfin_thaw(struct ata_port *ap)
1211 {
1212  dev_dbg(ap->dev, "in atapi dma thaw\n");
1213  bfin_check_status(ap);
1214  ata_sff_irq_on(ap);
1215 }
1216 
1225 static void bfin_postreset(struct ata_link *link, unsigned int *classes)
1226 {
1227  struct ata_port *ap = link->ap;
1228  void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr;
1229 
1230  /* re-enable interrupts */
1231  ata_sff_irq_on(ap);
1232 
1233  /* is double-select really necessary? */
1234  if (classes[0] != ATA_DEV_NONE)
1235  bfin_dev_select(ap, 1);
1236  if (classes[1] != ATA_DEV_NONE)
1237  bfin_dev_select(ap, 0);
1238 
1239  /* bail out if no device is present */
1240  if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) {
1241  return;
1242  }
1243 
1244  /* set up device control */
1245  write_atapi_register(base, ATA_REG_CTRL, ap->ctl);
1246 }
1247 
1248 static void bfin_port_stop(struct ata_port *ap)
1249 {
1250  dev_dbg(ap->dev, "in atapi port stop\n");
1251  if (ap->udma_mask != 0 || ap->mwdma_mask != 0) {
1252  dma_free_coherent(ap->dev,
1253  BFIN_MAX_SG_SEGMENTS * sizeof(struct dma_desc_array),
1254  ap->bmdma_prd,
1255  ap->bmdma_prd_dma);
1256 
1259  }
1260 }
1261 
1262 static int bfin_port_start(struct ata_port *ap)
1263 {
1264  dev_dbg(ap->dev, "in atapi port start\n");
1265  if (!(ap->udma_mask || ap->mwdma_mask))
1266  return 0;
1267 
1268  ap->bmdma_prd = dma_alloc_coherent(ap->dev,
1269  BFIN_MAX_SG_SEGMENTS * sizeof(struct dma_desc_array),
1270  &ap->bmdma_prd_dma,
1271  GFP_KERNEL);
1272 
1273  if (ap->bmdma_prd == NULL) {
1274  dev_info(ap->dev, "Unable to allocate DMA descriptor array.\n");
1275  goto out;
1276  }
1277 
1278  if (request_dma(CH_ATAPI_RX, "BFIN ATAPI RX DMA") >= 0) {
1280  "BFIN ATAPI TX DMA") >= 0)
1281  return 0;
1282 
1284  dma_free_coherent(ap->dev,
1285  BFIN_MAX_SG_SEGMENTS * sizeof(struct dma_desc_array),
1286  ap->bmdma_prd,
1287  ap->bmdma_prd_dma);
1288  }
1289 
1290 out:
1291  ap->udma_mask = 0;
1292  ap->mwdma_mask = 0;
1293  dev_err(ap->dev, "Unable to request ATAPI DMA!"
1294  " Continue in PIO mode.\n");
1295 
1296  return 0;
1297 }
1298 
1299 static unsigned int bfin_ata_host_intr(struct ata_port *ap,
1300  struct ata_queued_cmd *qc)
1301 {
1302  struct ata_eh_info *ehi = &ap->link.eh_info;
1303  u8 status, host_stat = 0;
1304 
1305  VPRINTK("ata%u: protocol %d task_state %d\n",
1306  ap->print_id, qc->tf.protocol, ap->hsm_task_state);
1307 
1308  /* Check whether we are expecting interrupt in this state */
1309  switch (ap->hsm_task_state) {
1310  case HSM_ST_FIRST:
1311  /* Some pre-ATAPI-4 devices assert INTRQ
1312  * at this state when ready to receive CDB.
1313  */
1314 
1315  /* Check the ATA_DFLAG_CDB_INTR flag is enough here.
1316  * The flag was turned on only for atapi devices.
1317  * No need to check is_atapi_taskfile(&qc->tf) again.
1318  */
1319  if (!(qc->dev->flags & ATA_DFLAG_CDB_INTR))
1320  goto idle_irq;
1321  break;
1322  case HSM_ST_LAST:
1323  if (qc->tf.protocol == ATA_PROT_DMA ||
1324  qc->tf.protocol == ATAPI_PROT_DMA) {
1325  /* check status of DMA engine */
1326  host_stat = ap->ops->bmdma_status(ap);
1327  VPRINTK("ata%u: host_stat 0x%X\n",
1328  ap->print_id, host_stat);
1329 
1330  /* if it's not our irq... */
1331  if (!(host_stat & ATA_DMA_INTR))
1332  goto idle_irq;
1333 
1334  /* before we do anything else, clear DMA-Start bit */
1335  ap->ops->bmdma_stop(qc);
1336 
1337  if (unlikely(host_stat & ATA_DMA_ERR)) {
1338  /* error when transferring data to/from memory */
1339  qc->err_mask |= AC_ERR_HOST_BUS;
1340  ap->hsm_task_state = HSM_ST_ERR;
1341  }
1342  }
1343  break;
1344  case HSM_ST:
1345  break;
1346  default:
1347  goto idle_irq;
1348  }
1349 
1350  /* check altstatus */
1351  status = ap->ops->sff_check_altstatus(ap);
1352  if (status & ATA_BUSY)
1353  goto busy_ata;
1354 
1355  /* check main status, clearing INTRQ */
1356  status = ap->ops->sff_check_status(ap);
1357  if (unlikely(status & ATA_BUSY))
1358  goto busy_ata;
1359 
1360  /* ack bmdma irq events */
1361  ap->ops->sff_irq_clear(ap);
1362 
1363  ata_sff_hsm_move(ap, qc, status, 0);
1364 
1365  if (unlikely(qc->err_mask) && (qc->tf.protocol == ATA_PROT_DMA ||
1366  qc->tf.protocol == ATAPI_PROT_DMA))
1367  ata_ehi_push_desc(ehi, "BMDMA stat 0x%x", host_stat);
1368 
1369 busy_ata:
1370  return 1; /* irq handled */
1371 
1372 idle_irq:
1373  ap->stats.idle_irq++;
1374 
1375 #ifdef ATA_IRQ_TRAP
1376  if ((ap->stats.idle_irq % 1000) == 0) {
1377  ap->ops->irq_ack(ap, 0); /* debug trap */
1378  ata_port_warn(ap, "irq trap\n");
1379  return 1;
1380  }
1381 #endif
1382  return 0; /* irq not handled */
1383 }
1384 
1385 static irqreturn_t bfin_ata_interrupt(int irq, void *dev_instance)
1386 {
1387  struct ata_host *host = dev_instance;
1388  unsigned int i;
1389  unsigned int handled = 0;
1390  unsigned long flags;
1391 
1392  /* TODO: make _irqsave conditional on x86 PCI IDE legacy mode */
1393  spin_lock_irqsave(&host->lock, flags);
1394 
1395  for (i = 0; i < host->n_ports; i++) {
1396  struct ata_port *ap = host->ports[i];
1397  struct ata_queued_cmd *qc;
1398 
1399  qc = ata_qc_from_tag(ap, ap->link.active_tag);
1400  if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING)))
1401  handled |= bfin_ata_host_intr(ap, qc);
1402  }
1403 
1404  spin_unlock_irqrestore(&host->lock, flags);
1405 
1406  return IRQ_RETVAL(handled);
1407 }
1408 
1409 
1410 static struct scsi_host_template bfin_sht = {
1412  .sg_tablesize = BFIN_MAX_SG_SEGMENTS,
1413  .dma_boundary = ATA_DMA_BOUNDARY,
1414 };
1415 
1416 static struct ata_port_operations bfin_pata_ops = {
1417  .inherits = &ata_bmdma_port_ops,
1418 
1419  .set_piomode = bfin_set_piomode,
1420  .set_dmamode = bfin_set_dmamode,
1421 
1422  .sff_tf_load = bfin_tf_load,
1423  .sff_tf_read = bfin_tf_read,
1424  .sff_exec_command = bfin_exec_command,
1425  .sff_check_status = bfin_check_status,
1426  .sff_check_altstatus = bfin_check_altstatus,
1427  .sff_dev_select = bfin_dev_select,
1428  .sff_set_devctl = bfin_set_devctl,
1429 
1430  .bmdma_setup = bfin_bmdma_setup,
1431  .bmdma_start = bfin_bmdma_start,
1432  .bmdma_stop = bfin_bmdma_stop,
1433  .bmdma_status = bfin_bmdma_status,
1434  .sff_data_xfer = bfin_data_xfer,
1435 
1436  .qc_prep = ata_noop_qc_prep,
1437 
1438  .thaw = bfin_thaw,
1439  .softreset = bfin_softreset,
1440  .postreset = bfin_postreset,
1441 
1442  .sff_irq_clear = bfin_irq_clear,
1443 
1444  .port_start = bfin_port_start,
1445  .port_stop = bfin_port_stop,
1446 };
1447 
1448 static struct ata_port_info bfin_port_info[] = {
1449  {
1450  .flags = ATA_FLAG_SLAVE_POSS,
1451  .pio_mask = ATA_PIO4,
1452  .mwdma_mask = 0,
1453  .udma_mask = 0,
1454  .port_ops = &bfin_pata_ops,
1455  },
1456 };
1457 
1462 static int bfin_reset_controller(struct ata_host *host)
1463 {
1464  void __iomem *base = (void __iomem *)host->ports[0]->ioaddr.ctl_addr;
1465  int count;
1466  unsigned short status;
1467 
1468  /* Disable all ATAPI interrupts */
1469  ATAPI_SET_INT_MASK(base, 0);
1470  SSYNC();
1471 
1472  /* Assert the RESET signal 25us*/
1474  udelay(30);
1475 
1476  /* Negate the RESET signal for 2ms*/
1478  msleep(2);
1479 
1480  /* Wait on Busy flag to clear */
1481  count = 10000000;
1482  do {
1483  status = read_atapi_register(base, ATA_REG_STATUS);
1484  } while (--count && (status & ATA_BUSY));
1485 
1486  /* Enable only ATAPI Device interrupt */
1487  ATAPI_SET_INT_MASK(base, 1);
1488  SSYNC();
1489 
1490  return (!count);
1491 }
1492 
1496 static unsigned short atapi_io_port[] = {
1497  P_ATAPI_RESET,
1498  P_ATAPI_DIOR,
1499  P_ATAPI_DIOW,
1500  P_ATAPI_CS0,
1501  P_ATAPI_CS1,
1502  P_ATAPI_DMACK,
1503  P_ATAPI_DMARQ,
1504  P_ATAPI_INTRQ,
1505  P_ATAPI_IORDY,
1506  P_ATAPI_D0A,
1507  P_ATAPI_D1A,
1508  P_ATAPI_D2A,
1509  P_ATAPI_D3A,
1510  P_ATAPI_D4A,
1511  P_ATAPI_D5A,
1512  P_ATAPI_D6A,
1513  P_ATAPI_D7A,
1514  P_ATAPI_D8A,
1515  P_ATAPI_D9A,
1516  P_ATAPI_D10A,
1517  P_ATAPI_D11A,
1518  P_ATAPI_D12A,
1519  P_ATAPI_D13A,
1520  P_ATAPI_D14A,
1521  P_ATAPI_D15A,
1522  P_ATAPI_A0A,
1523  P_ATAPI_A1A,
1524  P_ATAPI_A2A,
1525  0
1526 };
1527 
1541 static int __devinit bfin_atapi_probe(struct platform_device *pdev)
1542 {
1543  int board_idx = 0;
1544  struct resource *res;
1545  struct ata_host *host;
1546  unsigned int fsclk = get_sclk();
1547  int udma_mode = 5;
1548  const struct ata_port_info *ppi[] =
1549  { &bfin_port_info[board_idx], NULL };
1550 
1551  /*
1552  * Simple resource validation ..
1553  */
1554  if (unlikely(pdev->num_resources != 2)) {
1555  dev_err(&pdev->dev, "invalid number of resources\n");
1556  return -EINVAL;
1557  }
1558 
1559  /*
1560  * Get the register base first
1561  */
1562  res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1563  if (res == NULL)
1564  return -EINVAL;
1565 
1566  while (bfin_port_info[board_idx].udma_mask > 0 &&
1567  udma_fsclk[udma_mode] > fsclk) {
1568  udma_mode--;
1569  bfin_port_info[board_idx].udma_mask >>= 1;
1570  }
1571 
1572  /*
1573  * Now that that's out of the way, wire up the port..
1574  */
1575  host = ata_host_alloc_pinfo(&pdev->dev, ppi, 1);
1576  if (!host)
1577  return -ENOMEM;
1578 
1579  host->ports[0]->ioaddr.ctl_addr = (void *)res->start;
1580 
1581  if (peripheral_request_list(atapi_io_port, "atapi-io-port")) {
1582  dev_err(&pdev->dev, "Requesting Peripherals failed\n");
1583  return -EFAULT;
1584  }
1585 
1586  if (bfin_reset_controller(host)) {
1587  peripheral_free_list(atapi_io_port);
1588  dev_err(&pdev->dev, "Fail to reset ATAPI device\n");
1589  return -EFAULT;
1590  }
1591 
1592  if (ata_host_activate(host, platform_get_irq(pdev, 0),
1593  bfin_ata_interrupt, IRQF_SHARED, &bfin_sht) != 0) {
1594  peripheral_free_list(atapi_io_port);
1595  dev_err(&pdev->dev, "Fail to attach ATAPI device\n");
1596  return -ENODEV;
1597  }
1598 
1599  dev_set_drvdata(&pdev->dev, host);
1600 
1601  return 0;
1602 }
1603 
1611 static int __devexit bfin_atapi_remove(struct platform_device *pdev)
1612 {
1613  struct device *dev = &pdev->dev;
1614  struct ata_host *host = dev_get_drvdata(dev);
1615 
1616  ata_host_detach(host);
1617  dev_set_drvdata(&pdev->dev, NULL);
1618 
1619  peripheral_free_list(atapi_io_port);
1620 
1621  return 0;
1622 }
1623 
1624 #ifdef CONFIG_PM
1625 static int bfin_atapi_suspend(struct platform_device *pdev, pm_message_t state)
1626 {
1627  struct ata_host *host = dev_get_drvdata(&pdev->dev);
1628  if (host)
1629  return ata_host_suspend(host, state);
1630  else
1631  return 0;
1632 }
1633 
1634 static int bfin_atapi_resume(struct platform_device *pdev)
1635 {
1636  struct ata_host *host = dev_get_drvdata(&pdev->dev);
1637  int ret;
1638 
1639  if (host) {
1640  ret = bfin_reset_controller(host);
1641  if (ret) {
1642  printk(KERN_ERR DRV_NAME ": Error during HW init\n");
1643  return ret;
1644  }
1645  ata_host_resume(host);
1646  }
1647 
1648  return 0;
1649 }
1650 #else
1651 #define bfin_atapi_suspend NULL
1652 #define bfin_atapi_resume NULL
1653 #endif
1654 
1655 static struct platform_driver bfin_atapi_driver = {
1656  .probe = bfin_atapi_probe,
1657  .remove = __devexit_p(bfin_atapi_remove),
1658  .suspend = bfin_atapi_suspend,
1659  .resume = bfin_atapi_resume,
1660  .driver = {
1661  .name = DRV_NAME,
1662  .owner = THIS_MODULE,
1663  },
1664 };
1665 
1666 #define ATAPI_MODE_SIZE 10
1667 static char bfin_atapi_mode[ATAPI_MODE_SIZE];
1668 
1669 static int __init bfin_atapi_init(void)
1670 {
1671  pr_info("register bfin atapi driver\n");
1672 
1673  switch(bfin_atapi_mode[0]) {
1674  case 'p':
1675  case 'P':
1676  break;
1677  case 'm':
1678  case 'M':
1679  bfin_port_info[0].mwdma_mask = ATA_MWDMA2;
1680  break;
1681  default:
1682  bfin_port_info[0].udma_mask = ATA_UDMA5;
1683  };
1684 
1685  return platform_driver_register(&bfin_atapi_driver);
1686 }
1687 
1688 static void __exit bfin_atapi_exit(void)
1689 {
1690  platform_driver_unregister(&bfin_atapi_driver);
1691 }
1692 
1693 module_init(bfin_atapi_init);
1694 module_exit(bfin_atapi_exit);
1695 /*
1696  * ATAPI mode:
1697  * pio/PIO
1698  * udma/UDMA (default)
1699  * mwdma/MWDMA
1700  */
1701 module_param_string(bfin_atapi_mode, bfin_atapi_mode, ATAPI_MODE_SIZE, 0);
1702 
1703 MODULE_AUTHOR("Sonic Zhang <[email protected]>");
1704 MODULE_DESCRIPTION("PATA driver for blackfin 54x ATAPI controller");
1705 MODULE_LICENSE("GPL");
1707 MODULE_ALIAS("platform:" DRV_NAME);