Commit | Line | Data |
---|---|---|
338a1281 OP |
1 | /* |
2 | * Driver for the Diolan DLN-2 USB adapter | |
3 | * | |
4 | * Copyright (c) 2014 Intel Corporation | |
5 | * | |
6 | * Derived from: | |
7 | * i2c-diolan-u2c.c | |
8 | * Copyright (c) 2010-2011 Ericsson AB | |
9 | * | |
10 | * This program is free software; you can redistribute it and/or | |
11 | * modify it under the terms of the GNU General Public License as | |
12 | * published by the Free Software Foundation, version 2. | |
13 | */ | |
14 | ||
15 | #include <linux/kernel.h> | |
16 | #include <linux/module.h> | |
17 | #include <linux/types.h> | |
18 | #include <linux/slab.h> | |
19 | #include <linux/usb.h> | |
20 | #include <linux/i2c.h> | |
21 | #include <linux/mutex.h> | |
22 | #include <linux/platform_device.h> | |
23 | #include <linux/mfd/core.h> | |
24 | #include <linux/mfd/dln2.h> | |
25 | #include <linux/rculist.h> | |
26 | ||
27 | struct dln2_header { | |
28 | __le16 size; | |
29 | __le16 id; | |
30 | __le16 echo; | |
31 | __le16 handle; | |
32 | }; | |
33 | ||
34 | struct dln2_response { | |
35 | struct dln2_header hdr; | |
36 | __le16 result; | |
37 | }; | |
38 | ||
39 | #define DLN2_GENERIC_MODULE_ID 0x00 | |
40 | #define DLN2_GENERIC_CMD(cmd) DLN2_CMD(cmd, DLN2_GENERIC_MODULE_ID) | |
41 | #define CMD_GET_DEVICE_VER DLN2_GENERIC_CMD(0x30) | |
42 | #define CMD_GET_DEVICE_SN DLN2_GENERIC_CMD(0x31) | |
43 | ||
44 | #define DLN2_HW_ID 0x200 | |
45 | #define DLN2_USB_TIMEOUT 200 /* in ms */ | |
46 | #define DLN2_MAX_RX_SLOTS 16 | |
47 | #define DLN2_MAX_URBS 16 | |
48 | #define DLN2_RX_BUF_SIZE 512 | |
49 | ||
50 | enum dln2_handle { | |
51 | DLN2_HANDLE_EVENT = 0, /* don't change, hardware defined */ | |
52 | DLN2_HANDLE_CTRL, | |
53 | DLN2_HANDLE_GPIO, | |
54 | DLN2_HANDLE_I2C, | |
55 | DLN2_HANDLES | |
56 | }; | |
57 | ||
58 | /* | |
59 | * Receive context used between the receive demultiplexer and the transfer | |
60 | * routine. While sending a request the transfer routine will look for a free | |
61 | * receive context and use it to wait for a response and to receive the URB and | |
62 | * thus the response data. | |
63 | */ | |
64 | struct dln2_rx_context { | |
65 | /* completion used to wait for a response */ | |
66 | struct completion done; | |
67 | ||
68 | /* if non-NULL the URB contains the response */ | |
69 | struct urb *urb; | |
70 | ||
71 | /* if true then this context is used to wait for a response */ | |
72 | bool in_use; | |
73 | }; | |
74 | ||
75 | /* | |
76 | * Receive contexts for a particular DLN2 module (i2c, gpio, etc.). We use the | |
77 | * handle header field to identify the module in dln2_dev.mod_rx_slots and then | |
78 | * the echo header field to index the slots field and find the receive context | |
79 | * for a particular request. | |
80 | */ | |
81 | struct dln2_mod_rx_slots { | |
82 | /* RX slots bitmap */ | |
83 | DECLARE_BITMAP(bmap, DLN2_MAX_RX_SLOTS); | |
84 | ||
85 | /* used to wait for a free RX slot */ | |
86 | wait_queue_head_t wq; | |
87 | ||
88 | /* used to wait for an RX operation to complete */ | |
89 | struct dln2_rx_context slots[DLN2_MAX_RX_SLOTS]; | |
90 | ||
91 | /* avoid races between alloc/free_rx_slot and dln2_rx_transfer */ | |
92 | spinlock_t lock; | |
93 | }; | |
94 | ||
95 | struct dln2_dev { | |
96 | struct usb_device *usb_dev; | |
97 | struct usb_interface *interface; | |
98 | u8 ep_in; | |
99 | u8 ep_out; | |
100 | ||
101 | struct urb *rx_urb[DLN2_MAX_URBS]; | |
102 | void *rx_buf[DLN2_MAX_URBS]; | |
103 | ||
104 | struct dln2_mod_rx_slots mod_rx_slots[DLN2_HANDLES]; | |
105 | ||
106 | struct list_head event_cb_list; | |
107 | spinlock_t event_cb_lock; | |
108 | ||
109 | bool disconnect; | |
110 | int active_transfers; | |
111 | wait_queue_head_t disconnect_wq; | |
112 | spinlock_t disconnect_lock; | |
113 | }; | |
114 | ||
115 | struct dln2_event_cb_entry { | |
116 | struct list_head list; | |
117 | u16 id; | |
118 | struct platform_device *pdev; | |
119 | dln2_event_cb_t callback; | |
120 | }; | |
121 | ||
122 | int dln2_register_event_cb(struct platform_device *pdev, u16 id, | |
123 | dln2_event_cb_t event_cb) | |
124 | { | |
125 | struct dln2_dev *dln2 = dev_get_drvdata(pdev->dev.parent); | |
126 | struct dln2_event_cb_entry *i, *entry; | |
127 | unsigned long flags; | |
128 | int ret = 0; | |
129 | ||
130 | entry = kzalloc(sizeof(*entry), GFP_KERNEL); | |
131 | if (!entry) | |
132 | return -ENOMEM; | |
133 | ||
134 | entry->id = id; | |
135 | entry->callback = event_cb; | |
136 | entry->pdev = pdev; | |
137 | ||
138 | spin_lock_irqsave(&dln2->event_cb_lock, flags); | |
139 | ||
140 | list_for_each_entry(i, &dln2->event_cb_list, list) { | |
141 | if (i->id == id) { | |
142 | ret = -EBUSY; | |
143 | break; | |
144 | } | |
145 | } | |
146 | ||
147 | if (!ret) | |
148 | list_add_rcu(&entry->list, &dln2->event_cb_list); | |
149 | ||
150 | spin_unlock_irqrestore(&dln2->event_cb_lock, flags); | |
151 | ||
152 | if (ret) | |
153 | kfree(entry); | |
154 | ||
155 | return ret; | |
156 | } | |
157 | EXPORT_SYMBOL(dln2_register_event_cb); | |
158 | ||
159 | void dln2_unregister_event_cb(struct platform_device *pdev, u16 id) | |
160 | { | |
161 | struct dln2_dev *dln2 = dev_get_drvdata(pdev->dev.parent); | |
162 | struct dln2_event_cb_entry *i; | |
163 | unsigned long flags; | |
164 | bool found = false; | |
165 | ||
166 | spin_lock_irqsave(&dln2->event_cb_lock, flags); | |
167 | ||
168 | list_for_each_entry(i, &dln2->event_cb_list, list) { | |
169 | if (i->id == id) { | |
170 | list_del_rcu(&i->list); | |
171 | found = true; | |
172 | break; | |
173 | } | |
174 | } | |
175 | ||
176 | spin_unlock_irqrestore(&dln2->event_cb_lock, flags); | |
177 | ||
178 | if (found) { | |
179 | synchronize_rcu(); | |
180 | kfree(i); | |
181 | } | |
182 | } | |
183 | EXPORT_SYMBOL(dln2_unregister_event_cb); | |
184 | ||
185 | /* | |
186 | * Returns true if a valid transfer slot is found. In this case the URB must not | |
187 | * be resubmitted immediately in dln2_rx as we need the data when dln2_transfer | |
188 | * is woke up. It will be resubmitted there. | |
189 | */ | |
190 | static bool dln2_transfer_complete(struct dln2_dev *dln2, struct urb *urb, | |
191 | u16 handle, u16 rx_slot) | |
192 | { | |
193 | struct device *dev = &dln2->interface->dev; | |
194 | struct dln2_mod_rx_slots *rxs = &dln2->mod_rx_slots[handle]; | |
195 | struct dln2_rx_context *rxc; | |
196 | bool valid_slot = false; | |
197 | ||
00ee7a37 OP |
198 | if (rx_slot >= DLN2_MAX_RX_SLOTS) |
199 | goto out; | |
200 | ||
338a1281 OP |
201 | rxc = &rxs->slots[rx_slot]; |
202 | ||
203 | /* | |
204 | * No need to disable interrupts as this lock is not taken in interrupt | |
205 | * context elsewhere in this driver. This function (or its callers) are | |
206 | * also not exported to other modules. | |
207 | */ | |
208 | spin_lock(&rxs->lock); | |
209 | if (rxc->in_use && !rxc->urb) { | |
210 | rxc->urb = urb; | |
211 | complete(&rxc->done); | |
212 | valid_slot = true; | |
213 | } | |
214 | spin_unlock(&rxs->lock); | |
215 | ||
00ee7a37 | 216 | out: |
338a1281 OP |
217 | if (!valid_slot) |
218 | dev_warn(dev, "bad/late response %d/%d\n", handle, rx_slot); | |
219 | ||
220 | return valid_slot; | |
221 | } | |
222 | ||
223 | static void dln2_run_event_callbacks(struct dln2_dev *dln2, u16 id, u16 echo, | |
224 | void *data, int len) | |
225 | { | |
226 | struct dln2_event_cb_entry *i; | |
227 | ||
228 | rcu_read_lock(); | |
229 | ||
230 | list_for_each_entry_rcu(i, &dln2->event_cb_list, list) { | |
231 | if (i->id == id) { | |
232 | i->callback(i->pdev, echo, data, len); | |
233 | break; | |
234 | } | |
235 | } | |
236 | ||
237 | rcu_read_unlock(); | |
238 | } | |
239 | ||
240 | static void dln2_rx(struct urb *urb) | |
241 | { | |
242 | struct dln2_dev *dln2 = urb->context; | |
243 | struct dln2_header *hdr = urb->transfer_buffer; | |
244 | struct device *dev = &dln2->interface->dev; | |
245 | u16 id, echo, handle, size; | |
246 | u8 *data; | |
247 | int len; | |
248 | int err; | |
249 | ||
250 | switch (urb->status) { | |
251 | case 0: | |
252 | /* success */ | |
253 | break; | |
254 | case -ECONNRESET: | |
255 | case -ENOENT: | |
256 | case -ESHUTDOWN: | |
257 | case -EPIPE: | |
258 | /* this urb is terminated, clean up */ | |
259 | dev_dbg(dev, "urb shutting down with status %d\n", urb->status); | |
260 | return; | |
261 | default: | |
262 | dev_dbg(dev, "nonzero urb status received %d\n", urb->status); | |
263 | goto out; | |
264 | } | |
265 | ||
266 | if (urb->actual_length < sizeof(struct dln2_header)) { | |
267 | dev_err(dev, "short response: %d\n", urb->actual_length); | |
268 | goto out; | |
269 | } | |
270 | ||
271 | handle = le16_to_cpu(hdr->handle); | |
272 | id = le16_to_cpu(hdr->id); | |
273 | echo = le16_to_cpu(hdr->echo); | |
274 | size = le16_to_cpu(hdr->size); | |
275 | ||
276 | if (size != urb->actual_length) { | |
277 | dev_err(dev, "size mismatch: handle %x cmd %x echo %x size %d actual %d\n", | |
278 | handle, id, echo, size, urb->actual_length); | |
279 | goto out; | |
280 | } | |
281 | ||
282 | if (handle >= DLN2_HANDLES) { | |
283 | dev_warn(dev, "invalid handle %d\n", handle); | |
284 | goto out; | |
285 | } | |
286 | ||
287 | data = urb->transfer_buffer + sizeof(struct dln2_header); | |
288 | len = urb->actual_length - sizeof(struct dln2_header); | |
289 | ||
290 | if (handle == DLN2_HANDLE_EVENT) { | |
291 | dln2_run_event_callbacks(dln2, id, echo, data, len); | |
292 | } else { | |
293 | /* URB will be re-submitted in _dln2_transfer (free_rx_slot) */ | |
294 | if (dln2_transfer_complete(dln2, urb, handle, echo)) | |
295 | return; | |
296 | } | |
297 | ||
298 | out: | |
299 | err = usb_submit_urb(urb, GFP_ATOMIC); | |
300 | if (err < 0) | |
301 | dev_err(dev, "failed to resubmit RX URB: %d\n", err); | |
302 | } | |
303 | ||
304 | static void *dln2_prep_buf(u16 handle, u16 cmd, u16 echo, const void *obuf, | |
305 | int *obuf_len, gfp_t gfp) | |
306 | { | |
307 | int len; | |
308 | void *buf; | |
309 | struct dln2_header *hdr; | |
310 | ||
311 | len = *obuf_len + sizeof(*hdr); | |
312 | buf = kmalloc(len, gfp); | |
313 | if (!buf) | |
314 | return NULL; | |
315 | ||
316 | hdr = (struct dln2_header *)buf; | |
317 | hdr->id = cpu_to_le16(cmd); | |
318 | hdr->size = cpu_to_le16(len); | |
319 | hdr->echo = cpu_to_le16(echo); | |
320 | hdr->handle = cpu_to_le16(handle); | |
321 | ||
322 | memcpy(buf + sizeof(*hdr), obuf, *obuf_len); | |
323 | ||
324 | *obuf_len = len; | |
325 | ||
326 | return buf; | |
327 | } | |
328 | ||
329 | static int dln2_send_wait(struct dln2_dev *dln2, u16 handle, u16 cmd, u16 echo, | |
330 | const void *obuf, int obuf_len) | |
331 | { | |
332 | int ret = 0; | |
333 | int len = obuf_len; | |
334 | void *buf; | |
335 | int actual; | |
336 | ||
337 | buf = dln2_prep_buf(handle, cmd, echo, obuf, &len, GFP_KERNEL); | |
338 | if (!buf) | |
339 | return -ENOMEM; | |
340 | ||
341 | ret = usb_bulk_msg(dln2->usb_dev, | |
342 | usb_sndbulkpipe(dln2->usb_dev, dln2->ep_out), | |
343 | buf, len, &actual, DLN2_USB_TIMEOUT); | |
344 | ||
345 | kfree(buf); | |
346 | ||
347 | return ret; | |
348 | } | |
349 | ||
350 | static bool find_free_slot(struct dln2_dev *dln2, u16 handle, int *slot) | |
351 | { | |
352 | struct dln2_mod_rx_slots *rxs; | |
353 | unsigned long flags; | |
354 | ||
355 | if (dln2->disconnect) { | |
356 | *slot = -ENODEV; | |
357 | return true; | |
358 | } | |
359 | ||
360 | rxs = &dln2->mod_rx_slots[handle]; | |
361 | ||
362 | spin_lock_irqsave(&rxs->lock, flags); | |
363 | ||
364 | *slot = find_first_zero_bit(rxs->bmap, DLN2_MAX_RX_SLOTS); | |
365 | ||
366 | if (*slot < DLN2_MAX_RX_SLOTS) { | |
367 | struct dln2_rx_context *rxc = &rxs->slots[*slot]; | |
368 | ||
369 | set_bit(*slot, rxs->bmap); | |
370 | rxc->in_use = true; | |
371 | } | |
372 | ||
373 | spin_unlock_irqrestore(&rxs->lock, flags); | |
374 | ||
375 | return *slot < DLN2_MAX_RX_SLOTS; | |
376 | } | |
377 | ||
378 | static int alloc_rx_slot(struct dln2_dev *dln2, u16 handle) | |
379 | { | |
380 | int ret; | |
381 | int slot; | |
382 | ||
383 | /* | |
384 | * No need to timeout here, the wait is bounded by the timeout in | |
385 | * _dln2_transfer. | |
386 | */ | |
387 | ret = wait_event_interruptible(dln2->mod_rx_slots[handle].wq, | |
388 | find_free_slot(dln2, handle, &slot)); | |
389 | if (ret < 0) | |
390 | return ret; | |
391 | ||
392 | return slot; | |
393 | } | |
394 | ||
395 | static void free_rx_slot(struct dln2_dev *dln2, u16 handle, int slot) | |
396 | { | |
397 | struct dln2_mod_rx_slots *rxs; | |
398 | struct urb *urb = NULL; | |
399 | unsigned long flags; | |
400 | struct dln2_rx_context *rxc; | |
401 | ||
402 | rxs = &dln2->mod_rx_slots[handle]; | |
403 | ||
404 | spin_lock_irqsave(&rxs->lock, flags); | |
405 | ||
406 | clear_bit(slot, rxs->bmap); | |
407 | ||
408 | rxc = &rxs->slots[slot]; | |
409 | rxc->in_use = false; | |
410 | urb = rxc->urb; | |
411 | rxc->urb = NULL; | |
412 | reinit_completion(&rxc->done); | |
413 | ||
414 | spin_unlock_irqrestore(&rxs->lock, flags); | |
415 | ||
416 | if (urb) { | |
417 | int err; | |
418 | struct device *dev = &dln2->interface->dev; | |
419 | ||
420 | err = usb_submit_urb(urb, GFP_KERNEL); | |
421 | if (err < 0) | |
422 | dev_err(dev, "failed to resubmit RX URB: %d\n", err); | |
423 | } | |
424 | ||
425 | wake_up_interruptible(&rxs->wq); | |
426 | } | |
427 | ||
428 | static int _dln2_transfer(struct dln2_dev *dln2, u16 handle, u16 cmd, | |
429 | const void *obuf, unsigned obuf_len, | |
430 | void *ibuf, unsigned *ibuf_len) | |
431 | { | |
432 | int ret = 0; | |
433 | int rx_slot; | |
434 | struct dln2_response *rsp; | |
435 | struct dln2_rx_context *rxc; | |
436 | struct device *dev = &dln2->interface->dev; | |
437 | const unsigned long timeout = DLN2_USB_TIMEOUT * HZ / 1000; | |
438 | struct dln2_mod_rx_slots *rxs = &dln2->mod_rx_slots[handle]; | |
2fc2b484 | 439 | int size; |
338a1281 OP |
440 | |
441 | spin_lock(&dln2->disconnect_lock); | |
442 | if (!dln2->disconnect) | |
443 | dln2->active_transfers++; | |
444 | else | |
445 | ret = -ENODEV; | |
446 | spin_unlock(&dln2->disconnect_lock); | |
447 | ||
448 | if (ret) | |
449 | return ret; | |
450 | ||
451 | rx_slot = alloc_rx_slot(dln2, handle); | |
452 | if (rx_slot < 0) { | |
453 | ret = rx_slot; | |
454 | goto out_decr; | |
455 | } | |
456 | ||
457 | ret = dln2_send_wait(dln2, handle, cmd, rx_slot, obuf, obuf_len); | |
458 | if (ret < 0) { | |
459 | dev_err(dev, "USB write failed: %d\n", ret); | |
460 | goto out_free_rx_slot; | |
461 | } | |
462 | ||
463 | rxc = &rxs->slots[rx_slot]; | |
464 | ||
465 | ret = wait_for_completion_interruptible_timeout(&rxc->done, timeout); | |
466 | if (ret <= 0) { | |
467 | if (!ret) | |
468 | ret = -ETIMEDOUT; | |
469 | goto out_free_rx_slot; | |
7ca2b1c6 OP |
470 | } else { |
471 | ret = 0; | |
338a1281 OP |
472 | } |
473 | ||
474 | if (dln2->disconnect) { | |
475 | ret = -ENODEV; | |
476 | goto out_free_rx_slot; | |
477 | } | |
478 | ||
479 | /* if we got here we know that the response header has been checked */ | |
480 | rsp = rxc->urb->transfer_buffer; | |
2fc2b484 | 481 | size = le16_to_cpu(rsp->hdr.size); |
338a1281 | 482 | |
2fc2b484 | 483 | if (size < sizeof(*rsp)) { |
338a1281 OP |
484 | ret = -EPROTO; |
485 | goto out_free_rx_slot; | |
486 | } | |
487 | ||
488 | if (le16_to_cpu(rsp->result) > 0x80) { | |
489 | dev_dbg(dev, "%d received response with error %d\n", | |
490 | handle, le16_to_cpu(rsp->result)); | |
491 | ret = -EREMOTEIO; | |
492 | goto out_free_rx_slot; | |
493 | } | |
494 | ||
7ca2b1c6 | 495 | if (!ibuf) |
338a1281 | 496 | goto out_free_rx_slot; |
338a1281 | 497 | |
2fc2b484 DC |
498 | if (*ibuf_len > size - sizeof(*rsp)) |
499 | *ibuf_len = size - sizeof(*rsp); | |
338a1281 OP |
500 | |
501 | memcpy(ibuf, rsp + 1, *ibuf_len); | |
502 | ||
503 | out_free_rx_slot: | |
504 | free_rx_slot(dln2, handle, rx_slot); | |
505 | out_decr: | |
506 | spin_lock(&dln2->disconnect_lock); | |
507 | dln2->active_transfers--; | |
508 | spin_unlock(&dln2->disconnect_lock); | |
509 | if (dln2->disconnect) | |
510 | wake_up(&dln2->disconnect_wq); | |
511 | ||
512 | return ret; | |
513 | } | |
514 | ||
515 | int dln2_transfer(struct platform_device *pdev, u16 cmd, | |
516 | const void *obuf, unsigned obuf_len, | |
517 | void *ibuf, unsigned *ibuf_len) | |
518 | { | |
519 | struct dln2_platform_data *dln2_pdata; | |
520 | struct dln2_dev *dln2; | |
521 | u16 handle; | |
522 | ||
523 | dln2 = dev_get_drvdata(pdev->dev.parent); | |
524 | dln2_pdata = dev_get_platdata(&pdev->dev); | |
525 | handle = dln2_pdata->handle; | |
526 | ||
527 | return _dln2_transfer(dln2, handle, cmd, obuf, obuf_len, ibuf, | |
528 | ibuf_len); | |
529 | } | |
530 | EXPORT_SYMBOL(dln2_transfer); | |
531 | ||
532 | static int dln2_check_hw(struct dln2_dev *dln2) | |
533 | { | |
534 | int ret; | |
535 | __le32 hw_type; | |
536 | int len = sizeof(hw_type); | |
537 | ||
538 | ret = _dln2_transfer(dln2, DLN2_HANDLE_CTRL, CMD_GET_DEVICE_VER, | |
539 | NULL, 0, &hw_type, &len); | |
540 | if (ret < 0) | |
541 | return ret; | |
542 | if (len < sizeof(hw_type)) | |
543 | return -EREMOTEIO; | |
544 | ||
545 | if (le32_to_cpu(hw_type) != DLN2_HW_ID) { | |
546 | dev_err(&dln2->interface->dev, "Device ID 0x%x not supported\n", | |
547 | le32_to_cpu(hw_type)); | |
548 | return -ENODEV; | |
549 | } | |
550 | ||
551 | return 0; | |
552 | } | |
553 | ||
554 | static int dln2_print_serialno(struct dln2_dev *dln2) | |
555 | { | |
556 | int ret; | |
557 | __le32 serial_no; | |
558 | int len = sizeof(serial_no); | |
559 | struct device *dev = &dln2->interface->dev; | |
560 | ||
561 | ret = _dln2_transfer(dln2, DLN2_HANDLE_CTRL, CMD_GET_DEVICE_SN, NULL, 0, | |
562 | &serial_no, &len); | |
563 | if (ret < 0) | |
564 | return ret; | |
565 | if (len < sizeof(serial_no)) | |
566 | return -EREMOTEIO; | |
567 | ||
568 | dev_info(dev, "Diolan DLN2 serial %u\n", le32_to_cpu(serial_no)); | |
569 | ||
570 | return 0; | |
571 | } | |
572 | ||
573 | static int dln2_hw_init(struct dln2_dev *dln2) | |
574 | { | |
575 | int ret; | |
576 | ||
577 | ret = dln2_check_hw(dln2); | |
578 | if (ret < 0) | |
579 | return ret; | |
580 | ||
581 | return dln2_print_serialno(dln2); | |
582 | } | |
583 | ||
584 | static void dln2_free_rx_urbs(struct dln2_dev *dln2) | |
585 | { | |
586 | int i; | |
587 | ||
588 | for (i = 0; i < DLN2_MAX_URBS; i++) { | |
589 | usb_kill_urb(dln2->rx_urb[i]); | |
590 | usb_free_urb(dln2->rx_urb[i]); | |
591 | kfree(dln2->rx_buf[i]); | |
592 | } | |
593 | } | |
594 | ||
595 | static void dln2_free(struct dln2_dev *dln2) | |
596 | { | |
597 | dln2_free_rx_urbs(dln2); | |
598 | usb_put_dev(dln2->usb_dev); | |
599 | kfree(dln2); | |
600 | } | |
601 | ||
602 | static int dln2_setup_rx_urbs(struct dln2_dev *dln2, | |
603 | struct usb_host_interface *hostif) | |
604 | { | |
605 | int i; | |
606 | int ret; | |
607 | const int rx_max_size = DLN2_RX_BUF_SIZE; | |
608 | struct device *dev = &dln2->interface->dev; | |
609 | ||
610 | for (i = 0; i < DLN2_MAX_URBS; i++) { | |
611 | dln2->rx_buf[i] = kmalloc(rx_max_size, GFP_KERNEL); | |
612 | if (!dln2->rx_buf[i]) | |
613 | return -ENOMEM; | |
614 | ||
615 | dln2->rx_urb[i] = usb_alloc_urb(0, GFP_KERNEL); | |
616 | if (!dln2->rx_urb[i]) | |
617 | return -ENOMEM; | |
618 | ||
619 | usb_fill_bulk_urb(dln2->rx_urb[i], dln2->usb_dev, | |
620 | usb_rcvbulkpipe(dln2->usb_dev, dln2->ep_in), | |
621 | dln2->rx_buf[i], rx_max_size, dln2_rx, dln2); | |
622 | ||
623 | ret = usb_submit_urb(dln2->rx_urb[i], GFP_KERNEL); | |
624 | if (ret < 0) { | |
625 | dev_err(dev, "failed to submit RX URB: %d\n", ret); | |
626 | return ret; | |
627 | } | |
628 | } | |
629 | ||
630 | return 0; | |
631 | } | |
632 | ||
633 | static struct dln2_platform_data dln2_pdata_gpio = { | |
634 | .handle = DLN2_HANDLE_GPIO, | |
635 | }; | |
636 | ||
637 | /* Only one I2C port seems to be supported on current hardware */ | |
638 | static struct dln2_platform_data dln2_pdata_i2c = { | |
639 | .handle = DLN2_HANDLE_I2C, | |
640 | .port = 0, | |
641 | }; | |
642 | ||
643 | static const struct mfd_cell dln2_devs[] = { | |
644 | { | |
645 | .name = "dln2-gpio", | |
646 | .platform_data = &dln2_pdata_gpio, | |
647 | .pdata_size = sizeof(struct dln2_platform_data), | |
648 | }, | |
649 | { | |
650 | .name = "dln2-i2c", | |
651 | .platform_data = &dln2_pdata_i2c, | |
652 | .pdata_size = sizeof(struct dln2_platform_data), | |
653 | }, | |
654 | }; | |
655 | ||
656 | static void dln2_disconnect(struct usb_interface *interface) | |
657 | { | |
658 | struct dln2_dev *dln2 = usb_get_intfdata(interface); | |
659 | int i, j; | |
660 | ||
661 | /* don't allow starting new transfers */ | |
662 | spin_lock(&dln2->disconnect_lock); | |
663 | dln2->disconnect = true; | |
664 | spin_unlock(&dln2->disconnect_lock); | |
665 | ||
666 | /* cancel in progress transfers */ | |
667 | for (i = 0; i < DLN2_HANDLES; i++) { | |
668 | struct dln2_mod_rx_slots *rxs = &dln2->mod_rx_slots[i]; | |
669 | unsigned long flags; | |
670 | ||
671 | spin_lock_irqsave(&rxs->lock, flags); | |
672 | ||
673 | /* cancel all response waiters */ | |
674 | for (j = 0; j < DLN2_MAX_RX_SLOTS; j++) { | |
675 | struct dln2_rx_context *rxc = &rxs->slots[j]; | |
676 | ||
677 | if (rxc->in_use) | |
678 | complete(&rxc->done); | |
679 | } | |
680 | ||
681 | spin_unlock_irqrestore(&rxs->lock, flags); | |
682 | } | |
683 | ||
684 | /* wait for transfers to end */ | |
685 | wait_event(dln2->disconnect_wq, !dln2->active_transfers); | |
686 | ||
687 | mfd_remove_devices(&interface->dev); | |
688 | ||
689 | dln2_free(dln2); | |
690 | } | |
691 | ||
692 | static int dln2_probe(struct usb_interface *interface, | |
693 | const struct usb_device_id *usb_id) | |
694 | { | |
695 | struct usb_host_interface *hostif = interface->cur_altsetting; | |
696 | struct device *dev = &interface->dev; | |
697 | struct dln2_dev *dln2; | |
698 | int ret; | |
699 | int i, j; | |
700 | ||
701 | if (hostif->desc.bInterfaceNumber != 0 || | |
702 | hostif->desc.bNumEndpoints < 2) | |
703 | return -ENODEV; | |
704 | ||
705 | dln2 = kzalloc(sizeof(*dln2), GFP_KERNEL); | |
706 | if (!dln2) | |
707 | return -ENOMEM; | |
708 | ||
709 | dln2->ep_out = hostif->endpoint[0].desc.bEndpointAddress; | |
710 | dln2->ep_in = hostif->endpoint[1].desc.bEndpointAddress; | |
711 | dln2->usb_dev = usb_get_dev(interface_to_usbdev(interface)); | |
712 | dln2->interface = interface; | |
713 | usb_set_intfdata(interface, dln2); | |
714 | init_waitqueue_head(&dln2->disconnect_wq); | |
715 | ||
716 | for (i = 0; i < DLN2_HANDLES; i++) { | |
717 | init_waitqueue_head(&dln2->mod_rx_slots[i].wq); | |
718 | spin_lock_init(&dln2->mod_rx_slots[i].lock); | |
719 | for (j = 0; j < DLN2_MAX_RX_SLOTS; j++) | |
720 | init_completion(&dln2->mod_rx_slots[i].slots[j].done); | |
721 | } | |
722 | ||
723 | spin_lock_init(&dln2->event_cb_lock); | |
724 | spin_lock_init(&dln2->disconnect_lock); | |
725 | INIT_LIST_HEAD(&dln2->event_cb_list); | |
726 | ||
727 | ret = dln2_setup_rx_urbs(dln2, hostif); | |
728 | if (ret) | |
729 | goto out_cleanup; | |
730 | ||
731 | ret = dln2_hw_init(dln2); | |
732 | if (ret < 0) { | |
733 | dev_err(dev, "failed to initialize hardware\n"); | |
734 | goto out_cleanup; | |
735 | } | |
736 | ||
737 | ret = mfd_add_hotplug_devices(dev, dln2_devs, ARRAY_SIZE(dln2_devs)); | |
738 | if (ret != 0) { | |
739 | dev_err(dev, "failed to add mfd devices to core\n"); | |
740 | goto out_cleanup; | |
741 | } | |
742 | ||
743 | return 0; | |
744 | ||
745 | out_cleanup: | |
746 | dln2_free(dln2); | |
747 | ||
748 | return ret; | |
749 | } | |
750 | ||
751 | static const struct usb_device_id dln2_table[] = { | |
752 | { USB_DEVICE(0xa257, 0x2013) }, | |
753 | { } | |
754 | }; | |
755 | ||
756 | MODULE_DEVICE_TABLE(usb, dln2_table); | |
757 | ||
758 | static struct usb_driver dln2_driver = { | |
759 | .name = "dln2", | |
760 | .probe = dln2_probe, | |
761 | .disconnect = dln2_disconnect, | |
762 | .id_table = dln2_table, | |
763 | }; | |
764 | ||
765 | module_usb_driver(dln2_driver); | |
766 | ||
767 | MODULE_AUTHOR("Octavian Purdila <octavian.purdila@intel.com>"); | |
768 | MODULE_DESCRIPTION("Core driver for the Diolan DLN2 interface adapter"); | |
769 | MODULE_LICENSE("GPL v2"); |