Linux Kernel  3.7.1
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
omap_hdq.c
Go to the documentation of this file.
1 /*
2  * drivers/w1/masters/omap_hdq.c
3  *
4  * Copyright (C) 2007,2012 Texas Instruments, Inc.
5  *
6  * This file is licensed under the terms of the GNU General Public License
7  * version 2. This program is licensed "as is" without any warranty of any
8  * kind, whether express or implied.
9  *
10  */
11 #include <linux/kernel.h>
12 #include <linux/module.h>
13 #include <linux/platform_device.h>
14 #include <linux/interrupt.h>
15 #include <linux/slab.h>
16 #include <linux/err.h>
17 #include <linux/io.h>
18 #include <linux/sched.h>
19 #include <linux/pm_runtime.h>
20 
21 #include "../w1.h"
22 #include "../w1_int.h"
23 
24 #define MOD_NAME "OMAP_HDQ:"
25 
26 #define OMAP_HDQ_REVISION 0x00
27 #define OMAP_HDQ_TX_DATA 0x04
28 #define OMAP_HDQ_RX_DATA 0x08
29 #define OMAP_HDQ_CTRL_STATUS 0x0c
30 #define OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK (1<<6)
31 #define OMAP_HDQ_CTRL_STATUS_CLOCKENABLE (1<<5)
32 #define OMAP_HDQ_CTRL_STATUS_GO (1<<4)
33 #define OMAP_HDQ_CTRL_STATUS_INITIALIZATION (1<<2)
34 #define OMAP_HDQ_CTRL_STATUS_DIR (1<<1)
35 #define OMAP_HDQ_CTRL_STATUS_MODE (1<<0)
36 #define OMAP_HDQ_INT_STATUS 0x10
37 #define OMAP_HDQ_INT_STATUS_TXCOMPLETE (1<<2)
38 #define OMAP_HDQ_INT_STATUS_RXCOMPLETE (1<<1)
39 #define OMAP_HDQ_INT_STATUS_TIMEOUT (1<<0)
40 #define OMAP_HDQ_SYSCONFIG 0x14
41 #define OMAP_HDQ_SYSCONFIG_SOFTRESET (1<<1)
42 #define OMAP_HDQ_SYSCONFIG_AUTOIDLE (1<<0)
43 #define OMAP_HDQ_SYSSTATUS 0x18
44 #define OMAP_HDQ_SYSSTATUS_RESETDONE (1<<0)
45 
46 #define OMAP_HDQ_FLAG_CLEAR 0
47 #define OMAP_HDQ_FLAG_SET 1
48 #define OMAP_HDQ_TIMEOUT (HZ/5)
49 
50 #define OMAP_HDQ_MAX_USER 4
51 
52 static DECLARE_WAIT_QUEUE_HEAD(hdq_wait_queue);
53 static int w1_id;
54 
55 struct hdq_data {
56  struct device *dev;
58  /* lock status update */
59  struct mutex hdq_mutex;
62  /* device lock */
64  /*
65  * Used to control the call to omap_hdq_get and omap_hdq_put.
66  * HDQ Protocol: Write the CMD|REG_address first, followed by
67  * the data wrire or read.
68  */
70 };
71 
72 static int __devinit omap_hdq_probe(struct platform_device *pdev);
73 static int __devexit omap_hdq_remove(struct platform_device *pdev);
74 
75 static struct platform_driver omap_hdq_driver = {
76  .probe = omap_hdq_probe,
77  .remove = __devexit_p(omap_hdq_remove),
78  .driver = {
79  .name = "omap_hdq",
80  },
81 };
82 
83 static u8 omap_w1_read_byte(void *_hdq);
84 static void omap_w1_write_byte(void *_hdq, u8 byte);
85 static u8 omap_w1_reset_bus(void *_hdq);
86 static void omap_w1_search_bus(void *_hdq, struct w1_master *master_dev,
87  u8 search_type, w1_slave_found_callback slave_found);
88 
89 
90 static struct w1_bus_master omap_w1_master = {
91  .read_byte = omap_w1_read_byte,
92  .write_byte = omap_w1_write_byte,
93  .reset_bus = omap_w1_reset_bus,
94  .search = omap_w1_search_bus,
95 };
96 
97 /* HDQ register I/O routines */
98 static inline u8 hdq_reg_in(struct hdq_data *hdq_data, u32 offset)
99 {
100  return __raw_readl(hdq_data->hdq_base + offset);
101 }
102 
103 static inline void hdq_reg_out(struct hdq_data *hdq_data, u32 offset, u8 val)
104 {
105  __raw_writel(val, hdq_data->hdq_base + offset);
106 }
107 
108 static inline u8 hdq_reg_merge(struct hdq_data *hdq_data, u32 offset,
109  u8 val, u8 mask)
110 {
111  u8 new_val = (__raw_readl(hdq_data->hdq_base + offset) & ~mask)
112  | (val & mask);
113  __raw_writel(new_val, hdq_data->hdq_base + offset);
114 
115  return new_val;
116 }
117 
118 /*
119  * Wait for one or more bits in flag change.
120  * HDQ_FLAG_SET: wait until any bit in the flag is set.
121  * HDQ_FLAG_CLEAR: wait until all bits in the flag are cleared.
122  * return 0 on success and -ETIMEDOUT in the case of timeout.
123  */
124 static int hdq_wait_for_flag(struct hdq_data *hdq_data, u32 offset,
125  u8 flag, u8 flag_set, u8 *status)
126 {
127  int ret = 0;
128  unsigned long timeout = jiffies + OMAP_HDQ_TIMEOUT;
129 
130  if (flag_set == OMAP_HDQ_FLAG_CLEAR) {
131  /* wait for the flag clear */
132  while (((*status = hdq_reg_in(hdq_data, offset)) & flag)
133  && time_before(jiffies, timeout)) {
135  }
136  if (*status & flag)
137  ret = -ETIMEDOUT;
138  } else if (flag_set == OMAP_HDQ_FLAG_SET) {
139  /* wait for the flag set */
140  while (!((*status = hdq_reg_in(hdq_data, offset)) & flag)
141  && time_before(jiffies, timeout)) {
143  }
144  if (!(*status & flag))
145  ret = -ETIMEDOUT;
146  } else
147  return -EINVAL;
148 
149  return ret;
150 }
151 
152 /* write out a byte and fill *status with HDQ_INT_STATUS */
153 static int hdq_write_byte(struct hdq_data *hdq_data, u8 val, u8 *status)
154 {
155  int ret;
156  u8 tmp_status;
157  unsigned long irqflags;
158 
159  *status = 0;
160 
161  spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags);
162  /* clear interrupt flags via a dummy read */
163  hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
164  /* ISR loads it with new INT_STATUS */
165  hdq_data->hdq_irqstatus = 0;
166  spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags);
167 
168  hdq_reg_out(hdq_data, OMAP_HDQ_TX_DATA, val);
169 
170  /* set the GO bit */
171  hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, OMAP_HDQ_CTRL_STATUS_GO,
173  /* wait for the TXCOMPLETE bit */
174  ret = wait_event_timeout(hdq_wait_queue,
175  hdq_data->hdq_irqstatus, OMAP_HDQ_TIMEOUT);
176  if (ret == 0) {
177  dev_dbg(hdq_data->dev, "TX wait elapsed\n");
178  ret = -ETIMEDOUT;
179  goto out;
180  }
181 
182  *status = hdq_data->hdq_irqstatus;
183  /* check irqstatus */
184  if (!(*status & OMAP_HDQ_INT_STATUS_TXCOMPLETE)) {
185  dev_dbg(hdq_data->dev, "timeout waiting for"
186  " TXCOMPLETE/RXCOMPLETE, %x", *status);
187  ret = -ETIMEDOUT;
188  goto out;
189  }
190 
191  /* wait for the GO bit return to zero */
192  ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_CTRL_STATUS,
194  OMAP_HDQ_FLAG_CLEAR, &tmp_status);
195  if (ret) {
196  dev_dbg(hdq_data->dev, "timeout waiting GO bit"
197  " return to zero, %x", tmp_status);
198  }
199 
200 out:
201  return ret;
202 }
203 
204 /* HDQ Interrupt service routine */
205 static irqreturn_t hdq_isr(int irq, void *_hdq)
206 {
207  struct hdq_data *hdq_data = _hdq;
208  unsigned long irqflags;
209 
210  spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags);
211  hdq_data->hdq_irqstatus = hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
212  spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags);
213  dev_dbg(hdq_data->dev, "hdq_isr: %x", hdq_data->hdq_irqstatus);
214 
215  if (hdq_data->hdq_irqstatus &
216  (OMAP_HDQ_INT_STATUS_TXCOMPLETE | OMAP_HDQ_INT_STATUS_RXCOMPLETE
218  /* wake up sleeping process */
219  wake_up(&hdq_wait_queue);
220  }
221 
222  return IRQ_HANDLED;
223 }
224 
225 /* HDQ Mode: always return success */
226 static u8 omap_w1_reset_bus(void *_hdq)
227 {
228  return 0;
229 }
230 
231 /* W1 search callback function */
232 static void omap_w1_search_bus(void *_hdq, struct w1_master *master_dev,
233  u8 search_type, w1_slave_found_callback slave_found)
234 {
235  u64 module_id, rn_le, cs, id;
236 
237  if (w1_id)
238  module_id = w1_id;
239  else
240  module_id = 0x1;
241 
242  rn_le = cpu_to_le64(module_id);
243  /*
244  * HDQ might not obey truly the 1-wire spec.
245  * So calculate CRC based on module parameter.
246  */
247  cs = w1_calc_crc8((u8 *)&rn_le, 7);
248  id = (cs << 56) | module_id;
249 
250  slave_found(master_dev, id);
251 }
252 
253 static int _omap_hdq_reset(struct hdq_data *hdq_data)
254 {
255  int ret;
256  u8 tmp_status;
257 
258  hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG, OMAP_HDQ_SYSCONFIG_SOFTRESET);
259  /*
260  * Select HDQ mode & enable clocks.
261  * It is observed that INT flags can't be cleared via a read and GO/INIT
262  * won't return to zero if interrupt is disabled. So we always enable
263  * interrupt.
264  */
265  hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS,
268 
269  /* wait for reset to complete */
270  ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_SYSSTATUS,
272  if (ret)
273  dev_dbg(hdq_data->dev, "timeout waiting HDQ reset, %x",
274  tmp_status);
275  else {
276  hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS,
279  hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG,
281  }
282 
283  return ret;
284 }
285 
286 /* Issue break pulse to the device */
287 static int omap_hdq_break(struct hdq_data *hdq_data)
288 {
289  int ret = 0;
290  u8 tmp_status;
291  unsigned long irqflags;
292 
293  ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
294  if (ret < 0) {
295  dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
296  ret = -EINTR;
297  goto rtn;
298  }
299 
300  spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags);
301  /* clear interrupt flags via a dummy read */
302  hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
303  /* ISR loads it with new INT_STATUS */
304  hdq_data->hdq_irqstatus = 0;
305  spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags);
306 
307  /* set the INIT and GO bit */
308  hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS,
312 
313  /* wait for the TIMEOUT bit */
314  ret = wait_event_timeout(hdq_wait_queue,
315  hdq_data->hdq_irqstatus, OMAP_HDQ_TIMEOUT);
316  if (ret == 0) {
317  dev_dbg(hdq_data->dev, "break wait elapsed\n");
318  ret = -EINTR;
319  goto out;
320  }
321 
322  tmp_status = hdq_data->hdq_irqstatus;
323  /* check irqstatus */
324  if (!(tmp_status & OMAP_HDQ_INT_STATUS_TIMEOUT)) {
325  dev_dbg(hdq_data->dev, "timeout waiting for TIMEOUT, %x",
326  tmp_status);
327  ret = -ETIMEDOUT;
328  goto out;
329  }
330  /*
331  * wait for both INIT and GO bits rerurn to zero.
332  * zero wait time expected for interrupt mode.
333  */
334  ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_CTRL_STATUS,
337  &tmp_status);
338  if (ret)
339  dev_dbg(hdq_data->dev, "timeout waiting INIT&GO bits"
340  " return to zero, %x", tmp_status);
341 
342 out:
343  mutex_unlock(&hdq_data->hdq_mutex);
344 rtn:
345  return ret;
346 }
347 
348 static int hdq_read_byte(struct hdq_data *hdq_data, u8 *val)
349 {
350  int ret = 0;
351  u8 status;
352 
353  ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
354  if (ret < 0) {
355  ret = -EINTR;
356  goto rtn;
357  }
358 
359  if (!hdq_data->hdq_usecount) {
360  ret = -EINVAL;
361  goto out;
362  }
363 
364  if (!(hdq_data->hdq_irqstatus & OMAP_HDQ_INT_STATUS_RXCOMPLETE)) {
365  hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS,
368  /*
369  * The RX comes immediately after TX.
370  */
371  wait_event_timeout(hdq_wait_queue,
372  (hdq_data->hdq_irqstatus
374  OMAP_HDQ_TIMEOUT);
375 
376  hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, 0,
378  status = hdq_data->hdq_irqstatus;
379  /* check irqstatus */
380  if (!(status & OMAP_HDQ_INT_STATUS_RXCOMPLETE)) {
381  dev_dbg(hdq_data->dev, "timeout waiting for"
382  " RXCOMPLETE, %x", status);
383  ret = -ETIMEDOUT;
384  goto out;
385  }
386  }
387  /* the data is ready. Read it in! */
388  *val = hdq_reg_in(hdq_data, OMAP_HDQ_RX_DATA);
389 out:
390  mutex_unlock(&hdq_data->hdq_mutex);
391 rtn:
392  return ret;
393 
394 }
395 
396 /* Enable clocks and set the controller to HDQ mode */
397 static int omap_hdq_get(struct hdq_data *hdq_data)
398 {
399  int ret = 0;
400 
401  ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
402  if (ret < 0) {
403  ret = -EINTR;
404  goto rtn;
405  }
406 
407  if (OMAP_HDQ_MAX_USER == hdq_data->hdq_usecount) {
408  dev_dbg(hdq_data->dev, "attempt to exceed the max use count");
409  ret = -EINVAL;
410  goto out;
411  } else {
412  hdq_data->hdq_usecount++;
413  try_module_get(THIS_MODULE);
414  if (1 == hdq_data->hdq_usecount) {
415 
416  pm_runtime_get_sync(hdq_data->dev);
417 
418  /* make sure HDQ is out of reset */
419  if (!(hdq_reg_in(hdq_data, OMAP_HDQ_SYSSTATUS) &
421  ret = _omap_hdq_reset(hdq_data);
422  if (ret)
423  /* back up the count */
424  hdq_data->hdq_usecount--;
425  } else {
426  /* select HDQ mode & enable clocks */
427  hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS,
430  hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG,
432  hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
433  }
434  }
435  }
436 
437 out:
438  mutex_unlock(&hdq_data->hdq_mutex);
439 rtn:
440  return ret;
441 }
442 
443 /* Disable clocks to the module */
444 static int omap_hdq_put(struct hdq_data *hdq_data)
445 {
446  int ret = 0;
447 
448  ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
449  if (ret < 0)
450  return -EINTR;
451 
452  if (0 == hdq_data->hdq_usecount) {
453  dev_dbg(hdq_data->dev, "attempt to decrement use count"
454  " when it is zero");
455  ret = -EINVAL;
456  } else {
457  hdq_data->hdq_usecount--;
458  module_put(THIS_MODULE);
459  if (0 == hdq_data->hdq_usecount)
460  pm_runtime_put_sync(hdq_data->dev);
461  }
462  mutex_unlock(&hdq_data->hdq_mutex);
463 
464  return ret;
465 }
466 
467 /* Read a byte of data from the device */
468 static u8 omap_w1_read_byte(void *_hdq)
469 {
470  struct hdq_data *hdq_data = _hdq;
471  u8 val = 0;
472  int ret;
473 
474  ret = hdq_read_byte(hdq_data, &val);
475  if (ret) {
476  ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
477  if (ret < 0) {
478  dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
479  return -EINTR;
480  }
481  hdq_data->init_trans = 0;
482  mutex_unlock(&hdq_data->hdq_mutex);
483  omap_hdq_put(hdq_data);
484  return -1;
485  }
486 
487  /* Write followed by a read, release the module */
488  if (hdq_data->init_trans) {
489  ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
490  if (ret < 0) {
491  dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
492  return -EINTR;
493  }
494  hdq_data->init_trans = 0;
495  mutex_unlock(&hdq_data->hdq_mutex);
496  omap_hdq_put(hdq_data);
497  }
498 
499  return val;
500 }
501 
502 /* Write a byte of data to the device */
503 static void omap_w1_write_byte(void *_hdq, u8 byte)
504 {
505  struct hdq_data *hdq_data = _hdq;
506  int ret;
507  u8 status;
508 
509  /* First write to initialize the transfer */
510  if (hdq_data->init_trans == 0)
511  omap_hdq_get(hdq_data);
512 
513  ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
514  if (ret < 0) {
515  dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
516  return;
517  }
518  hdq_data->init_trans++;
519  mutex_unlock(&hdq_data->hdq_mutex);
520 
521  ret = hdq_write_byte(hdq_data, byte, &status);
522  if (ret < 0) {
523  dev_dbg(hdq_data->dev, "TX failure:Ctrl status %x\n", status);
524  return;
525  }
526 
527  /* Second write, data transferred. Release the module */
528  if (hdq_data->init_trans > 1) {
529  omap_hdq_put(hdq_data);
530  ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
531  if (ret < 0) {
532  dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
533  return;
534  }
535  hdq_data->init_trans = 0;
536  mutex_unlock(&hdq_data->hdq_mutex);
537  }
538 }
539 
540 static int __devinit omap_hdq_probe(struct platform_device *pdev)
541 {
542  struct device *dev = &pdev->dev;
543  struct hdq_data *hdq_data;
544  struct resource *res;
545  int ret, irq;
546  u8 rev;
547 
548  hdq_data = devm_kzalloc(dev, sizeof(*hdq_data), GFP_KERNEL);
549  if (!hdq_data) {
550  dev_dbg(&pdev->dev, "unable to allocate memory\n");
551  return -ENOMEM;
552  }
553 
554  hdq_data->dev = dev;
555  platform_set_drvdata(pdev, hdq_data);
556 
557  res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
558  if (!res) {
559  dev_dbg(&pdev->dev, "unable to get resource\n");
560  return -ENXIO;
561  }
562 
563  hdq_data->hdq_base = devm_request_and_ioremap(dev, res);
564  if (!hdq_data->hdq_base) {
565  dev_dbg(&pdev->dev, "ioremap failed\n");
566  return -ENOMEM;
567  }
568 
569  hdq_data->hdq_usecount = 0;
570  mutex_init(&hdq_data->hdq_mutex);
571 
572  pm_runtime_enable(&pdev->dev);
573  pm_runtime_get_sync(&pdev->dev);
574 
575  rev = hdq_reg_in(hdq_data, OMAP_HDQ_REVISION);
576  dev_info(&pdev->dev, "OMAP HDQ Hardware Rev %c.%c. Driver in %s mode\n",
577  (rev >> 4) + '0', (rev & 0x0f) + '0', "Interrupt");
578 
579  spin_lock_init(&hdq_data->hdq_spinlock);
580 
581  irq = platform_get_irq(pdev, 0);
582  if (irq < 0) {
583  ret = -ENXIO;
584  goto err_irq;
585  }
586 
587  ret = devm_request_irq(dev, irq, hdq_isr, IRQF_DISABLED,
588  "omap_hdq", hdq_data);
589  if (ret < 0) {
590  dev_dbg(&pdev->dev, "could not request irq\n");
591  goto err_irq;
592  }
593 
594  omap_hdq_break(hdq_data);
595 
596  pm_runtime_put_sync(&pdev->dev);
597 
598  omap_w1_master.data = hdq_data;
599 
600  ret = w1_add_master_device(&omap_w1_master);
601  if (ret) {
602  dev_dbg(&pdev->dev, "Failure in registering w1 master\n");
603  goto err_w1;
604  }
605 
606  return 0;
607 
608 err_irq:
609  pm_runtime_put_sync(&pdev->dev);
610 err_w1:
611  pm_runtime_disable(&pdev->dev);
612 
613  return ret;
614 }
615 
616 static int __devexit omap_hdq_remove(struct platform_device *pdev)
617 {
618  struct hdq_data *hdq_data = platform_get_drvdata(pdev);
619 
620  mutex_lock(&hdq_data->hdq_mutex);
621 
622  if (hdq_data->hdq_usecount) {
623  dev_dbg(&pdev->dev, "removed when use count is not zero\n");
624  mutex_unlock(&hdq_data->hdq_mutex);
625  return -EBUSY;
626  }
627 
628  mutex_unlock(&hdq_data->hdq_mutex);
629 
630  /* remove module dependency */
631  pm_runtime_disable(&pdev->dev);
632 
633  return 0;
634 }
635 
636 module_platform_driver(omap_hdq_driver);
637 
638 module_param(w1_id, int, S_IRUSR);
639 MODULE_PARM_DESC(w1_id, "1-wire id for the slave detection");
640 
641 MODULE_AUTHOR("Texas Instruments");
642 MODULE_DESCRIPTION("HDQ driver Library");
643 MODULE_LICENSE("GPL");