2 * Copyright (c) 2010-2011 Atheros Communications Inc.
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
17 #include <asm/unaligned.h>
20 /* identify firmware images */
21 #define FIRMWARE_AR7010_1_1 "htc_7010.fw"
22 #define FIRMWARE_AR9271 "htc_9271.fw"
24 MODULE_FIRMWARE(FIRMWARE_AR7010_1_1);
25 MODULE_FIRMWARE(FIRMWARE_AR9271);
27 static struct usb_device_id ath9k_hif_usb_ids[] = {
28 { USB_DEVICE(0x0cf3, 0x9271) }, /* Atheros */
29 { USB_DEVICE(0x0cf3, 0x1006) }, /* Atheros */
30 { USB_DEVICE(0x0846, 0x9030) }, /* Netgear N150 */
31 { USB_DEVICE(0x07D1, 0x3A10) }, /* Dlink Wireless 150 */
32 { USB_DEVICE(0x13D3, 0x3327) }, /* Azurewave */
33 { USB_DEVICE(0x13D3, 0x3328) }, /* Azurewave */
34 { USB_DEVICE(0x13D3, 0x3346) }, /* IMC Networks */
35 { USB_DEVICE(0x13D3, 0x3348) }, /* Azurewave */
36 { USB_DEVICE(0x13D3, 0x3349) }, /* Azurewave */
37 { USB_DEVICE(0x13D3, 0x3350) }, /* Azurewave */
38 { USB_DEVICE(0x04CA, 0x4605) }, /* Liteon */
39 { USB_DEVICE(0x040D, 0x3801) }, /* VIA */
40 { USB_DEVICE(0x0cf3, 0xb003) }, /* Ubiquiti WifiStation Ext */
41 { USB_DEVICE(0x057c, 0x8403) }, /* AVM FRITZ!WLAN 11N v2 USB */
43 { USB_DEVICE(0x0cf3, 0x7015),
44 .driver_info = AR9287_USB }, /* Atheros */
45 { USB_DEVICE(0x1668, 0x1200),
46 .driver_info = AR9287_USB }, /* Verizon */
48 { USB_DEVICE(0x0cf3, 0x7010),
49 .driver_info = AR9280_USB }, /* Atheros */
50 { USB_DEVICE(0x0846, 0x9018),
51 .driver_info = AR9280_USB }, /* Netgear WNDA3200 */
52 { USB_DEVICE(0x083A, 0xA704),
53 .driver_info = AR9280_USB }, /* SMC Networks */
54 { USB_DEVICE(0x0411, 0x017f),
55 .driver_info = AR9280_USB }, /* Sony UWA-BR100 */
56 { USB_DEVICE(0x04da, 0x3904),
57 .driver_info = AR9280_USB },
59 { USB_DEVICE(0x0cf3, 0x20ff),
60 .driver_info = STORAGE_DEVICE },
65 MODULE_DEVICE_TABLE(usb, ath9k_hif_usb_ids);
67 static int __hif_usb_tx(struct hif_device_usb *hif_dev);
69 static void hif_usb_regout_cb(struct urb *urb)
71 struct cmd_buf *cmd = (struct cmd_buf *)urb->context;
73 switch (urb->status) {
86 ath9k_htc_txcompletion_cb(cmd->hif_dev->htc_handle,
97 static int hif_usb_send_regout(struct hif_device_usb *hif_dev,
104 urb = usb_alloc_urb(0, GFP_KERNEL);
108 cmd = kzalloc(sizeof(*cmd), GFP_KERNEL);
115 cmd->hif_dev = hif_dev;
117 usb_fill_bulk_urb(urb, hif_dev->udev,
118 usb_sndbulkpipe(hif_dev->udev, USB_REG_OUT_PIPE),
120 hif_usb_regout_cb, cmd);
122 usb_anchor_urb(urb, &hif_dev->regout_submitted);
123 ret = usb_submit_urb(urb, GFP_KERNEL);
125 usb_unanchor_urb(urb);
133 static void hif_usb_mgmt_cb(struct urb *urb)
135 struct cmd_buf *cmd = (struct cmd_buf *)urb->context;
136 struct hif_device_usb *hif_dev;
139 if (!cmd || !cmd->skb || !cmd->hif_dev)
142 hif_dev = cmd->hif_dev;
144 switch (urb->status) {
154 * If the URBs are being flushed, no need to complete
157 spin_lock(&hif_dev->tx.tx_lock);
158 if (hif_dev->tx.flags & HIF_USB_TX_FLUSH) {
159 spin_unlock(&hif_dev->tx.tx_lock);
160 dev_kfree_skb_any(cmd->skb);
164 spin_unlock(&hif_dev->tx.tx_lock);
172 skb_pull(cmd->skb, 4);
173 ath9k_htc_txcompletion_cb(cmd->hif_dev->htc_handle,
178 static int hif_usb_send_mgmt(struct hif_device_usb *hif_dev,
186 urb = usb_alloc_urb(0, GFP_ATOMIC);
190 cmd = kzalloc(sizeof(*cmd), GFP_ATOMIC);
197 cmd->hif_dev = hif_dev;
199 hdr = (__le16 *) skb_push(skb, 4);
200 *hdr++ = cpu_to_le16(skb->len - 4);
201 *hdr++ = cpu_to_le16(ATH_USB_TX_STREAM_MODE_TAG);
203 usb_fill_bulk_urb(urb, hif_dev->udev,
204 usb_sndbulkpipe(hif_dev->udev, USB_WLAN_TX_PIPE),
206 hif_usb_mgmt_cb, cmd);
208 usb_anchor_urb(urb, &hif_dev->mgmt_submitted);
209 ret = usb_submit_urb(urb, GFP_ATOMIC);
211 usb_unanchor_urb(urb);
219 static inline void ath9k_skb_queue_purge(struct hif_device_usb *hif_dev,
220 struct sk_buff_head *list)
224 while ((skb = __skb_dequeue(list)) != NULL) {
225 dev_kfree_skb_any(skb);
229 static inline void ath9k_skb_queue_complete(struct hif_device_usb *hif_dev,
230 struct sk_buff_head *queue,
235 while ((skb = __skb_dequeue(queue)) != NULL) {
236 ath9k_htc_txcompletion_cb(hif_dev->htc_handle,
239 TX_STAT_INC(skb_success);
241 TX_STAT_INC(skb_failed);
245 static void hif_usb_tx_cb(struct urb *urb)
247 struct tx_buf *tx_buf = (struct tx_buf *) urb->context;
248 struct hif_device_usb *hif_dev;
251 if (!tx_buf || !tx_buf->hif_dev)
254 hif_dev = tx_buf->hif_dev;
256 switch (urb->status) {
266 * If the URBs are being flushed, no need to add this
267 * URB to the free list.
269 spin_lock(&hif_dev->tx.tx_lock);
270 if (hif_dev->tx.flags & HIF_USB_TX_FLUSH) {
271 spin_unlock(&hif_dev->tx.tx_lock);
272 ath9k_skb_queue_purge(hif_dev, &tx_buf->skb_queue);
275 spin_unlock(&hif_dev->tx.tx_lock);
283 ath9k_skb_queue_complete(hif_dev, &tx_buf->skb_queue, txok);
285 /* Re-initialize the SKB queue */
286 tx_buf->len = tx_buf->offset = 0;
287 __skb_queue_head_init(&tx_buf->skb_queue);
289 /* Add this TX buffer to the free list */
290 spin_lock(&hif_dev->tx.tx_lock);
291 list_move_tail(&tx_buf->list, &hif_dev->tx.tx_buf);
292 hif_dev->tx.tx_buf_cnt++;
293 if (!(hif_dev->tx.flags & HIF_USB_TX_STOP))
294 __hif_usb_tx(hif_dev); /* Check for pending SKBs */
295 TX_STAT_INC(buf_completed);
296 spin_unlock(&hif_dev->tx.tx_lock);
299 /* TX lock has to be taken */
300 static int __hif_usb_tx(struct hif_device_usb *hif_dev)
302 struct tx_buf *tx_buf = NULL;
303 struct sk_buff *nskb = NULL;
309 if (hif_dev->tx.tx_skb_cnt == 0)
312 /* Check if a free TX buffer is available */
313 if (list_empty(&hif_dev->tx.tx_buf))
316 tx_buf = list_first_entry(&hif_dev->tx.tx_buf, struct tx_buf, list);
317 list_move_tail(&tx_buf->list, &hif_dev->tx.tx_pending);
318 hif_dev->tx.tx_buf_cnt--;
320 tx_skb_cnt = min_t(u16, hif_dev->tx.tx_skb_cnt, MAX_TX_AGGR_NUM);
322 for (i = 0; i < tx_skb_cnt; i++) {
323 nskb = __skb_dequeue(&hif_dev->tx.tx_skb_queue);
325 /* Should never be NULL */
328 hif_dev->tx.tx_skb_cnt--;
331 buf += tx_buf->offset;
333 *hdr++ = cpu_to_le16(nskb->len);
334 *hdr++ = cpu_to_le16(ATH_USB_TX_STREAM_MODE_TAG);
336 memcpy(buf, nskb->data, nskb->len);
337 tx_buf->len = nskb->len + 4;
339 if (i < (tx_skb_cnt - 1))
340 tx_buf->offset += (((tx_buf->len - 1) / 4) + 1) * 4;
342 if (i == (tx_skb_cnt - 1))
343 tx_buf->len += tx_buf->offset;
345 __skb_queue_tail(&tx_buf->skb_queue, nskb);
346 TX_STAT_INC(skb_queued);
349 usb_fill_bulk_urb(tx_buf->urb, hif_dev->udev,
350 usb_sndbulkpipe(hif_dev->udev, USB_WLAN_TX_PIPE),
351 tx_buf->buf, tx_buf->len,
352 hif_usb_tx_cb, tx_buf);
354 ret = usb_submit_urb(tx_buf->urb, GFP_ATOMIC);
356 tx_buf->len = tx_buf->offset = 0;
357 ath9k_skb_queue_complete(hif_dev, &tx_buf->skb_queue, false);
358 __skb_queue_head_init(&tx_buf->skb_queue);
359 list_move_tail(&tx_buf->list, &hif_dev->tx.tx_buf);
360 hif_dev->tx.tx_buf_cnt++;
364 TX_STAT_INC(buf_queued);
369 static int hif_usb_send_tx(struct hif_device_usb *hif_dev, struct sk_buff *skb)
371 struct ath9k_htc_tx_ctl *tx_ctl;
375 spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
377 if (hif_dev->tx.flags & HIF_USB_TX_STOP) {
378 spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
382 /* Check if the max queue count has been reached */
383 if (hif_dev->tx.tx_skb_cnt > MAX_TX_BUF_NUM) {
384 spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
388 spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
390 tx_ctl = HTC_SKB_CB(skb);
392 /* Mgmt/Beacon frames don't use the TX buffer pool */
393 if ((tx_ctl->type == ATH9K_HTC_MGMT) ||
394 (tx_ctl->type == ATH9K_HTC_BEACON)) {
395 ret = hif_usb_send_mgmt(hif_dev, skb);
398 spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
400 if ((tx_ctl->type == ATH9K_HTC_NORMAL) ||
401 (tx_ctl->type == ATH9K_HTC_AMPDU)) {
402 __skb_queue_tail(&hif_dev->tx.tx_skb_queue, skb);
403 hif_dev->tx.tx_skb_cnt++;
406 /* Check if AMPDUs have to be sent immediately */
407 if ((hif_dev->tx.tx_buf_cnt == MAX_TX_URB_NUM) &&
408 (hif_dev->tx.tx_skb_cnt < 2)) {
409 __hif_usb_tx(hif_dev);
412 spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
417 static void hif_usb_start(void *hif_handle)
419 struct hif_device_usb *hif_dev = (struct hif_device_usb *)hif_handle;
422 hif_dev->flags |= HIF_USB_START;
424 spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
425 hif_dev->tx.flags &= ~HIF_USB_TX_STOP;
426 spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
429 static void hif_usb_stop(void *hif_handle)
431 struct hif_device_usb *hif_dev = (struct hif_device_usb *)hif_handle;
432 struct tx_buf *tx_buf = NULL, *tx_buf_tmp = NULL;
435 spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
436 ath9k_skb_queue_complete(hif_dev, &hif_dev->tx.tx_skb_queue, false);
437 hif_dev->tx.tx_skb_cnt = 0;
438 hif_dev->tx.flags |= HIF_USB_TX_STOP;
439 spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
441 /* The pending URBs have to be canceled. */
442 list_for_each_entry_safe(tx_buf, tx_buf_tmp,
443 &hif_dev->tx.tx_pending, list) {
444 usb_kill_urb(tx_buf->urb);
447 usb_kill_anchored_urbs(&hif_dev->mgmt_submitted);
450 static int hif_usb_send(void *hif_handle, u8 pipe_id, struct sk_buff *skb)
452 struct hif_device_usb *hif_dev = (struct hif_device_usb *)hif_handle;
456 case USB_WLAN_TX_PIPE:
457 ret = hif_usb_send_tx(hif_dev, skb);
459 case USB_REG_OUT_PIPE:
460 ret = hif_usb_send_regout(hif_dev, skb);
463 dev_err(&hif_dev->udev->dev,
464 "ath9k_htc: Invalid TX pipe: %d\n", pipe_id);
472 static inline bool check_index(struct sk_buff *skb, u8 idx)
474 struct ath9k_htc_tx_ctl *tx_ctl;
476 tx_ctl = HTC_SKB_CB(skb);
478 if ((tx_ctl->type == ATH9K_HTC_AMPDU) &&
479 (tx_ctl->sta_idx == idx))
485 static void hif_usb_sta_drain(void *hif_handle, u8 idx)
487 struct hif_device_usb *hif_dev = (struct hif_device_usb *)hif_handle;
488 struct sk_buff *skb, *tmp;
491 spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
493 skb_queue_walk_safe(&hif_dev->tx.tx_skb_queue, skb, tmp) {
494 if (check_index(skb, idx)) {
495 __skb_unlink(skb, &hif_dev->tx.tx_skb_queue);
496 ath9k_htc_txcompletion_cb(hif_dev->htc_handle,
498 hif_dev->tx.tx_skb_cnt--;
499 TX_STAT_INC(skb_failed);
503 spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
506 static struct ath9k_htc_hif hif_usb = {
507 .transport = ATH9K_HIF_USB,
508 .name = "ath9k_hif_usb",
510 .control_ul_pipe = USB_REG_OUT_PIPE,
511 .control_dl_pipe = USB_REG_IN_PIPE,
513 .start = hif_usb_start,
514 .stop = hif_usb_stop,
515 .sta_drain = hif_usb_sta_drain,
516 .send = hif_usb_send,
519 static void ath9k_hif_usb_rx_stream(struct hif_device_usb *hif_dev,
522 struct sk_buff *nskb, *skb_pool[MAX_PKT_NUM_IN_TRANSFER];
523 int index = 0, i = 0, len = skb->len;
524 int rx_remain_len, rx_pkt_len;
528 spin_lock(&hif_dev->rx_lock);
530 rx_remain_len = hif_dev->rx_remain_len;
531 rx_pkt_len = hif_dev->rx_transfer_len;
533 if (rx_remain_len != 0) {
534 struct sk_buff *remain_skb = hif_dev->remain_skb;
537 ptr = (u8 *) remain_skb->data;
539 index = rx_remain_len;
540 rx_remain_len -= hif_dev->rx_pad_len;
543 memcpy(ptr, skb->data, rx_remain_len);
545 rx_pkt_len += rx_remain_len;
546 hif_dev->rx_remain_len = 0;
547 skb_put(remain_skb, rx_pkt_len);
549 skb_pool[pool_index++] = remain_skb;
552 index = rx_remain_len;
556 spin_unlock(&hif_dev->rx_lock);
558 while (index < len) {
564 ptr = (u8 *) skb->data;
566 pkt_len = get_unaligned_le16(ptr + index);
567 pkt_tag = get_unaligned_le16(ptr + index + 2);
569 if (pkt_tag != ATH_USB_RX_STREAM_MODE_TAG) {
570 RX_STAT_INC(skb_dropped);
574 pad_len = 4 - (pkt_len & 0x3);
579 index = index + 4 + pkt_len + pad_len;
581 if (index > MAX_RX_BUF_SIZE) {
582 spin_lock(&hif_dev->rx_lock);
583 hif_dev->rx_remain_len = index - MAX_RX_BUF_SIZE;
584 hif_dev->rx_transfer_len =
585 MAX_RX_BUF_SIZE - chk_idx - 4;
586 hif_dev->rx_pad_len = pad_len;
588 nskb = __dev_alloc_skb(pkt_len + 32, GFP_ATOMIC);
590 dev_err(&hif_dev->udev->dev,
591 "ath9k_htc: RX memory allocation error\n");
592 spin_unlock(&hif_dev->rx_lock);
595 skb_reserve(nskb, 32);
596 RX_STAT_INC(skb_allocated);
598 memcpy(nskb->data, &(skb->data[chk_idx+4]),
599 hif_dev->rx_transfer_len);
601 /* Record the buffer pointer */
602 hif_dev->remain_skb = nskb;
603 spin_unlock(&hif_dev->rx_lock);
605 nskb = __dev_alloc_skb(pkt_len + 32, GFP_ATOMIC);
607 dev_err(&hif_dev->udev->dev,
608 "ath9k_htc: RX memory allocation error\n");
611 skb_reserve(nskb, 32);
612 RX_STAT_INC(skb_allocated);
614 memcpy(nskb->data, &(skb->data[chk_idx+4]), pkt_len);
615 skb_put(nskb, pkt_len);
616 skb_pool[pool_index++] = nskb;
621 for (i = 0; i < pool_index; i++) {
622 ath9k_htc_rx_msg(hif_dev->htc_handle, skb_pool[i],
623 skb_pool[i]->len, USB_WLAN_RX_PIPE);
624 RX_STAT_INC(skb_completed);
628 static void ath9k_hif_usb_rx_cb(struct urb *urb)
630 struct sk_buff *skb = (struct sk_buff *) urb->context;
631 struct hif_device_usb *hif_dev =
632 usb_get_intfdata(usb_ifnum_to_if(urb->dev, 0));
641 switch (urb->status) {
653 if (likely(urb->actual_length != 0)) {
654 skb_put(skb, urb->actual_length);
655 ath9k_hif_usb_rx_stream(hif_dev, skb);
659 skb_reset_tail_pointer(skb);
662 usb_anchor_urb(urb, &hif_dev->rx_submitted);
663 ret = usb_submit_urb(urb, GFP_ATOMIC);
665 usb_unanchor_urb(urb);
674 static void ath9k_hif_usb_reg_in_cb(struct urb *urb)
676 struct sk_buff *skb = (struct sk_buff *) urb->context;
677 struct sk_buff *nskb;
678 struct hif_device_usb *hif_dev =
679 usb_get_intfdata(usb_ifnum_to_if(urb->dev, 0));
688 switch (urb->status) {
697 skb_reset_tail_pointer(skb);
703 if (likely(urb->actual_length != 0)) {
704 skb_put(skb, urb->actual_length);
706 /* Process the command first */
707 ath9k_htc_rx_msg(hif_dev->htc_handle, skb,
708 skb->len, USB_REG_IN_PIPE);
711 nskb = alloc_skb(MAX_REG_IN_BUF_SIZE, GFP_ATOMIC);
713 dev_err(&hif_dev->udev->dev,
714 "ath9k_htc: REG_IN memory allocation failure\n");
719 usb_fill_bulk_urb(urb, hif_dev->udev,
720 usb_rcvbulkpipe(hif_dev->udev,
722 nskb->data, MAX_REG_IN_BUF_SIZE,
723 ath9k_hif_usb_reg_in_cb, nskb);
727 usb_anchor_urb(urb, &hif_dev->reg_in_submitted);
728 ret = usb_submit_urb(urb, GFP_ATOMIC);
730 usb_unanchor_urb(urb);
740 static void ath9k_hif_usb_dealloc_tx_urbs(struct hif_device_usb *hif_dev)
742 struct tx_buf *tx_buf = NULL, *tx_buf_tmp = NULL;
745 list_for_each_entry_safe(tx_buf, tx_buf_tmp,
746 &hif_dev->tx.tx_buf, list) {
747 usb_kill_urb(tx_buf->urb);
748 list_del(&tx_buf->list);
749 usb_free_urb(tx_buf->urb);
754 spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
755 hif_dev->tx.flags |= HIF_USB_TX_FLUSH;
756 spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
758 list_for_each_entry_safe(tx_buf, tx_buf_tmp,
759 &hif_dev->tx.tx_pending, list) {
760 usb_kill_urb(tx_buf->urb);
761 list_del(&tx_buf->list);
762 usb_free_urb(tx_buf->urb);
767 usb_kill_anchored_urbs(&hif_dev->mgmt_submitted);
770 static int ath9k_hif_usb_alloc_tx_urbs(struct hif_device_usb *hif_dev)
772 struct tx_buf *tx_buf;
775 INIT_LIST_HEAD(&hif_dev->tx.tx_buf);
776 INIT_LIST_HEAD(&hif_dev->tx.tx_pending);
777 spin_lock_init(&hif_dev->tx.tx_lock);
778 __skb_queue_head_init(&hif_dev->tx.tx_skb_queue);
779 init_usb_anchor(&hif_dev->mgmt_submitted);
781 for (i = 0; i < MAX_TX_URB_NUM; i++) {
782 tx_buf = kzalloc(sizeof(struct tx_buf), GFP_KERNEL);
786 tx_buf->buf = kzalloc(MAX_TX_BUF_SIZE, GFP_KERNEL);
790 tx_buf->urb = usb_alloc_urb(0, GFP_KERNEL);
794 tx_buf->hif_dev = hif_dev;
795 __skb_queue_head_init(&tx_buf->skb_queue);
797 list_add_tail(&tx_buf->list, &hif_dev->tx.tx_buf);
800 hif_dev->tx.tx_buf_cnt = MAX_TX_URB_NUM;
808 ath9k_hif_usb_dealloc_tx_urbs(hif_dev);
812 static void ath9k_hif_usb_dealloc_rx_urbs(struct hif_device_usb *hif_dev)
814 usb_kill_anchored_urbs(&hif_dev->rx_submitted);
817 static int ath9k_hif_usb_alloc_rx_urbs(struct hif_device_usb *hif_dev)
819 struct urb *urb = NULL;
820 struct sk_buff *skb = NULL;
823 init_usb_anchor(&hif_dev->rx_submitted);
824 spin_lock_init(&hif_dev->rx_lock);
826 for (i = 0; i < MAX_RX_URB_NUM; i++) {
829 urb = usb_alloc_urb(0, GFP_KERNEL);
835 /* Allocate buffer */
836 skb = alloc_skb(MAX_RX_BUF_SIZE, GFP_KERNEL);
842 usb_fill_bulk_urb(urb, hif_dev->udev,
843 usb_rcvbulkpipe(hif_dev->udev,
845 skb->data, MAX_RX_BUF_SIZE,
846 ath9k_hif_usb_rx_cb, skb);
849 usb_anchor_urb(urb, &hif_dev->rx_submitted);
852 ret = usb_submit_urb(urb, GFP_KERNEL);
854 usb_unanchor_urb(urb);
859 * Drop reference count.
860 * This ensures that the URB is freed when killing them.
872 ath9k_hif_usb_dealloc_rx_urbs(hif_dev);
876 static void ath9k_hif_usb_dealloc_reg_in_urbs(struct hif_device_usb *hif_dev)
878 usb_kill_anchored_urbs(&hif_dev->reg_in_submitted);
881 static int ath9k_hif_usb_alloc_reg_in_urbs(struct hif_device_usb *hif_dev)
883 struct urb *urb = NULL;
884 struct sk_buff *skb = NULL;
887 init_usb_anchor(&hif_dev->reg_in_submitted);
889 for (i = 0; i < MAX_REG_IN_URB_NUM; i++) {
892 urb = usb_alloc_urb(0, GFP_KERNEL);
898 /* Allocate buffer */
899 skb = alloc_skb(MAX_REG_IN_BUF_SIZE, GFP_KERNEL);
905 usb_fill_bulk_urb(urb, hif_dev->udev,
906 usb_rcvbulkpipe(hif_dev->udev,
908 skb->data, MAX_REG_IN_BUF_SIZE,
909 ath9k_hif_usb_reg_in_cb, skb);
912 usb_anchor_urb(urb, &hif_dev->reg_in_submitted);
915 ret = usb_submit_urb(urb, GFP_KERNEL);
917 usb_unanchor_urb(urb);
922 * Drop reference count.
923 * This ensures that the URB is freed when killing them.
935 ath9k_hif_usb_dealloc_reg_in_urbs(hif_dev);
939 static int ath9k_hif_usb_alloc_urbs(struct hif_device_usb *hif_dev)
942 init_usb_anchor(&hif_dev->regout_submitted);
945 if (ath9k_hif_usb_alloc_tx_urbs(hif_dev) < 0)
949 if (ath9k_hif_usb_alloc_rx_urbs(hif_dev) < 0)
953 if (ath9k_hif_usb_alloc_reg_in_urbs(hif_dev) < 0)
958 ath9k_hif_usb_dealloc_rx_urbs(hif_dev);
960 ath9k_hif_usb_dealloc_tx_urbs(hif_dev);
965 static void ath9k_hif_usb_dealloc_urbs(struct hif_device_usb *hif_dev)
967 usb_kill_anchored_urbs(&hif_dev->regout_submitted);
968 ath9k_hif_usb_dealloc_reg_in_urbs(hif_dev);
969 ath9k_hif_usb_dealloc_tx_urbs(hif_dev);
970 ath9k_hif_usb_dealloc_rx_urbs(hif_dev);
973 static int ath9k_hif_usb_download_fw(struct hif_device_usb *hif_dev)
976 const void *data = hif_dev->firmware->data;
977 size_t len = hif_dev->firmware->size;
978 u32 addr = AR9271_FIRMWARE;
979 u8 *buf = kzalloc(4096, GFP_KERNEL);
986 transfer = min_t(size_t, len, 4096);
987 memcpy(buf, data, transfer);
989 err = usb_control_msg(hif_dev->udev,
990 usb_sndctrlpipe(hif_dev->udev, 0),
991 FIRMWARE_DOWNLOAD, 0x40 | USB_DIR_OUT,
992 addr >> 8, 0, buf, transfer, HZ);
1004 if (IS_AR7010_DEVICE(hif_dev->usb_device_id->driver_info))
1005 firm_offset = AR7010_FIRMWARE_TEXT;
1007 firm_offset = AR9271_FIRMWARE_TEXT;
1010 * Issue FW download complete command to firmware.
1012 err = usb_control_msg(hif_dev->udev, usb_sndctrlpipe(hif_dev->udev, 0),
1013 FIRMWARE_DOWNLOAD_COMP,
1015 firm_offset >> 8, 0, NULL, 0, HZ);
1019 dev_info(&hif_dev->udev->dev, "ath9k_htc: Transferred FW: %s, size: %ld\n",
1020 hif_dev->fw_name, (unsigned long) hif_dev->firmware->size);
1025 static int ath9k_hif_usb_dev_init(struct hif_device_usb *hif_dev)
1027 struct usb_host_interface *alt = &hif_dev->interface->altsetting[0];
1028 struct usb_endpoint_descriptor *endp;
1031 ret = ath9k_hif_usb_download_fw(hif_dev);
1033 dev_err(&hif_dev->udev->dev,
1034 "ath9k_htc: Firmware - %s download failed\n",
1039 /* On downloading the firmware to the target, the USB descriptor of EP4
1040 * is 'patched' to change the type of the endpoint to Bulk. This will
1041 * bring down CPU usage during the scan period.
1043 for (idx = 0; idx < alt->desc.bNumEndpoints; idx++) {
1044 endp = &alt->endpoint[idx].desc;
1045 if ((endp->bmAttributes & USB_ENDPOINT_XFERTYPE_MASK)
1046 == USB_ENDPOINT_XFER_INT) {
1047 endp->bmAttributes &= ~USB_ENDPOINT_XFERTYPE_MASK;
1048 endp->bmAttributes |= USB_ENDPOINT_XFER_BULK;
1049 endp->bInterval = 0;
1054 ret = ath9k_hif_usb_alloc_urbs(hif_dev);
1056 dev_err(&hif_dev->udev->dev,
1057 "ath9k_htc: Unable to allocate URBs\n");
1064 static void ath9k_hif_usb_dev_deinit(struct hif_device_usb *hif_dev)
1066 ath9k_hif_usb_dealloc_urbs(hif_dev);
1070 * If initialization fails or the FW cannot be retrieved,
1071 * detach the device.
1073 static void ath9k_hif_usb_firmware_fail(struct hif_device_usb *hif_dev)
1075 struct device *parent = hif_dev->udev->dev.parent;
1077 complete(&hif_dev->fw_done);
1080 device_lock(parent);
1082 device_release_driver(&hif_dev->udev->dev);
1085 device_unlock(parent);
1088 static void ath9k_hif_usb_firmware_cb(const struct firmware *fw, void *context)
1090 struct hif_device_usb *hif_dev = context;
1094 dev_err(&hif_dev->udev->dev,
1095 "ath9k_htc: Failed to get firmware %s\n",
1100 hif_dev->htc_handle = ath9k_htc_hw_alloc(hif_dev, &hif_usb,
1101 &hif_dev->udev->dev);
1102 if (hif_dev->htc_handle == NULL) {
1106 hif_dev->firmware = fw;
1108 /* Proceed with initialization */
1110 ret = ath9k_hif_usb_dev_init(hif_dev);
1114 ret = ath9k_htc_hw_init(hif_dev->htc_handle,
1115 &hif_dev->interface->dev,
1116 hif_dev->usb_device_id->idProduct,
1117 hif_dev->udev->product,
1118 hif_dev->usb_device_id->driver_info);
1121 goto err_htc_hw_init;
1124 complete(&hif_dev->fw_done);
1129 ath9k_hif_usb_dev_deinit(hif_dev);
1131 ath9k_htc_hw_free(hif_dev->htc_handle);
1132 release_firmware(fw);
1133 hif_dev->firmware = NULL;
1135 ath9k_hif_usb_firmware_fail(hif_dev);
1139 * An exact copy of the function from zd1211rw.
1141 static int send_eject_command(struct usb_interface *interface)
1143 struct usb_device *udev = interface_to_usbdev(interface);
1144 struct usb_host_interface *iface_desc = &interface->altsetting[0];
1145 struct usb_endpoint_descriptor *endpoint;
1150 /* Find bulk out endpoint */
1151 for (r = 1; r >= 0; r--) {
1152 endpoint = &iface_desc->endpoint[r].desc;
1153 if (usb_endpoint_dir_out(endpoint) &&
1154 usb_endpoint_xfer_bulk(endpoint)) {
1155 bulk_out_ep = endpoint->bEndpointAddress;
1161 "ath9k_htc: Could not find bulk out endpoint\n");
1165 cmd = kzalloc(31, GFP_KERNEL);
1169 /* USB bulk command block */
1170 cmd[0] = 0x55; /* bulk command signature */
1171 cmd[1] = 0x53; /* bulk command signature */
1172 cmd[2] = 0x42; /* bulk command signature */
1173 cmd[3] = 0x43; /* bulk command signature */
1174 cmd[14] = 6; /* command length */
1176 cmd[15] = 0x1b; /* SCSI command: START STOP UNIT */
1177 cmd[19] = 0x2; /* eject disc */
1179 dev_info(&udev->dev, "Ejecting storage device...\n");
1180 r = usb_bulk_msg(udev, usb_sndbulkpipe(udev, bulk_out_ep),
1181 cmd, 31, NULL, 2000);
1186 /* At this point, the device disconnects and reconnects with the real
1189 usb_set_intfdata(interface, NULL);
1193 static int ath9k_hif_usb_probe(struct usb_interface *interface,
1194 const struct usb_device_id *id)
1196 struct usb_device *udev = interface_to_usbdev(interface);
1197 struct hif_device_usb *hif_dev;
1200 if (id->driver_info == STORAGE_DEVICE)
1201 return send_eject_command(interface);
1203 hif_dev = kzalloc(sizeof(struct hif_device_usb), GFP_KERNEL);
1211 hif_dev->udev = udev;
1212 hif_dev->interface = interface;
1213 hif_dev->usb_device_id = id;
1215 udev->reset_resume = 1;
1217 usb_set_intfdata(interface, hif_dev);
1219 init_completion(&hif_dev->fw_done);
1221 /* Find out which firmware to load */
1223 if (IS_AR7010_DEVICE(id->driver_info))
1224 hif_dev->fw_name = FIRMWARE_AR7010_1_1;
1226 hif_dev->fw_name = FIRMWARE_AR9271;
1228 ret = request_firmware_nowait(THIS_MODULE, true, hif_dev->fw_name,
1229 &hif_dev->udev->dev, GFP_KERNEL,
1230 hif_dev, ath9k_hif_usb_firmware_cb);
1232 dev_err(&hif_dev->udev->dev,
1233 "ath9k_htc: Async request for firmware %s failed\n",
1238 dev_info(&hif_dev->udev->dev, "ath9k_htc: Firmware %s requested\n",
1244 usb_set_intfdata(interface, NULL);
1251 static void ath9k_hif_usb_reboot(struct usb_device *udev)
1253 u32 reboot_cmd = 0xffffffff;
1257 buf = kmemdup(&reboot_cmd, 4, GFP_KERNEL);
1261 ret = usb_bulk_msg(udev, usb_sndbulkpipe(udev, USB_REG_OUT_PIPE),
1264 dev_err(&udev->dev, "ath9k_htc: USB reboot failed\n");
1269 static void ath9k_hif_usb_disconnect(struct usb_interface *interface)
1271 struct usb_device *udev = interface_to_usbdev(interface);
1272 struct hif_device_usb *hif_dev = usb_get_intfdata(interface);
1273 bool unplugged = (udev->state == USB_STATE_NOTATTACHED) ? true : false;
1278 wait_for_completion(&hif_dev->fw_done);
1280 if (hif_dev->firmware) {
1281 ath9k_htc_hw_deinit(hif_dev->htc_handle, unplugged);
1282 ath9k_htc_hw_free(hif_dev->htc_handle);
1283 ath9k_hif_usb_dev_deinit(hif_dev);
1284 release_firmware(hif_dev->firmware);
1287 usb_set_intfdata(interface, NULL);
1289 if (!unplugged && (hif_dev->flags & HIF_USB_START))
1290 ath9k_hif_usb_reboot(udev);
1293 dev_info(&udev->dev, "ath9k_htc: USB layer deinitialized\n");
1298 static int ath9k_hif_usb_suspend(struct usb_interface *interface,
1299 pm_message_t message)
1301 struct hif_device_usb *hif_dev = usb_get_intfdata(interface);
1304 * The device has to be set to FULLSLEEP mode in case no
1307 if (!(hif_dev->flags & HIF_USB_START))
1308 ath9k_htc_suspend(hif_dev->htc_handle);
1310 ath9k_hif_usb_dealloc_urbs(hif_dev);
1315 static int ath9k_hif_usb_resume(struct usb_interface *interface)
1317 struct hif_device_usb *hif_dev = usb_get_intfdata(interface);
1318 struct htc_target *htc_handle = hif_dev->htc_handle;
1321 ret = ath9k_hif_usb_alloc_urbs(hif_dev);
1325 if (hif_dev->firmware) {
1326 ret = ath9k_hif_usb_download_fw(hif_dev);
1330 ath9k_hif_usb_dealloc_urbs(hif_dev);
1336 ret = ath9k_htc_resume(htc_handle);
1344 ath9k_hif_usb_dealloc_urbs(hif_dev);
1350 static struct usb_driver ath9k_hif_usb_driver = {
1351 .name = KBUILD_MODNAME,
1352 .probe = ath9k_hif_usb_probe,
1353 .disconnect = ath9k_hif_usb_disconnect,
1355 .suspend = ath9k_hif_usb_suspend,
1356 .resume = ath9k_hif_usb_resume,
1357 .reset_resume = ath9k_hif_usb_resume,
1359 .id_table = ath9k_hif_usb_ids,
1361 .disable_hub_initiated_lpm = 1,
1364 int ath9k_hif_usb_init(void)
1366 return usb_register(&ath9k_hif_usb_driver);
1369 void ath9k_hif_usb_exit(void)
1371 usb_deregister(&ath9k_hif_usb_driver);