2 * Copyright (c) 2007-2011 Atheros Communications Inc.
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
21 #define HTC_PACKET_CONTAINER_ALLOCATION 32
22 #define HTC_CONTROL_BUFFER_SIZE (HTC_MAX_CTRL_MSG_LEN + HTC_HDR_LENGTH)
24 static int ath6kl_htc_pipe_tx(struct htc_target
*handle
,
25 struct htc_packet
*packet
);
26 static void ath6kl_htc_pipe_cleanup(struct htc_target
*handle
);
28 /* htc pipe tx path */
29 static inline void restore_tx_packet(struct htc_packet
*packet
)
31 if (packet
->info
.tx
.flags
& HTC_FLAGS_TX_FIXUP_NETBUF
) {
32 skb_pull(packet
->skb
, sizeof(struct htc_frame_hdr
));
33 packet
->info
.tx
.flags
&= ~HTC_FLAGS_TX_FIXUP_NETBUF
;
37 static void do_send_completion(struct htc_endpoint
*ep
,
38 struct list_head
*queue_to_indicate
)
40 struct htc_packet
*packet
;
42 if (list_empty(queue_to_indicate
)) {
43 /* nothing to indicate */
47 if (ep
->ep_cb
.tx_comp_multi
!= NULL
) {
48 ath6kl_dbg(ATH6KL_DBG_HTC
,
49 "%s: calling ep %d, send complete multiple callback (%d pkts)\n",
51 get_queue_depth(queue_to_indicate
));
53 * a multiple send complete handler is being used,
54 * pass the queue to the handler
56 ep
->ep_cb
.tx_comp_multi(ep
->target
, queue_to_indicate
);
58 * all packets are now owned by the callback,
59 * reset queue to be safe
61 INIT_LIST_HEAD(queue_to_indicate
);
63 /* using legacy EpTxComplete */
65 packet
= list_first_entry(queue_to_indicate
,
66 struct htc_packet
, list
);
68 list_del(&packet
->list
);
69 ath6kl_dbg(ATH6KL_DBG_HTC
,
70 "%s: calling ep %d send complete callback on packet 0x%p\n",
71 __func__
, ep
->eid
, packet
);
72 ep
->ep_cb
.tx_complete(ep
->target
, packet
);
73 } while (!list_empty(queue_to_indicate
));
77 static void send_packet_completion(struct htc_target
*target
,
78 struct htc_packet
*packet
)
80 struct htc_endpoint
*ep
= &target
->endpoint
[packet
->endpoint
];
81 struct list_head container
;
83 restore_tx_packet(packet
);
84 INIT_LIST_HEAD(&container
);
85 list_add_tail(&packet
->list
, &container
);
88 do_send_completion(ep
, &container
);
91 static void get_htc_packet_credit_based(struct htc_target
*target
,
92 struct htc_endpoint
*ep
,
93 struct list_head
*queue
)
98 struct htc_packet
*packet
;
99 unsigned int transfer_len
;
101 /* NOTE : the TX lock is held when this function is called */
103 /* loop until we can grab as many packets out of the queue as we can */
106 if (list_empty(&ep
->txq
))
109 /* get packet at head, but don't remove it */
110 packet
= list_first_entry(&ep
->txq
, struct htc_packet
, list
);
112 ath6kl_dbg(ATH6KL_DBG_HTC
,
113 "%s: got head packet:0x%p , queue depth: %d\n",
114 __func__
, packet
, get_queue_depth(&ep
->txq
));
116 transfer_len
= packet
->act_len
+ HTC_HDR_LENGTH
;
118 if (transfer_len
<= target
->tgt_cred_sz
) {
119 credits_required
= 1;
121 /* figure out how many credits this message requires */
122 credits_required
= transfer_len
/ target
->tgt_cred_sz
;
123 remainder
= transfer_len
% target
->tgt_cred_sz
;
129 ath6kl_dbg(ATH6KL_DBG_HTC
, "%s: creds required:%d got:%d\n",
130 __func__
, credits_required
, ep
->cred_dist
.credits
);
132 if (ep
->eid
== ENDPOINT_0
) {
134 * endpoint 0 is special, it always has a credit and
135 * does not require credit based flow control
137 credits_required
= 0;
140 if (ep
->cred_dist
.credits
< credits_required
)
143 ep
->cred_dist
.credits
-= credits_required
;
144 ep
->ep_st
.cred_cosumd
+= credits_required
;
146 /* check if we need credits back from the target */
147 if (ep
->cred_dist
.credits
<
148 ep
->cred_dist
.cred_per_msg
) {
149 /* tell the target we need credits ASAP! */
150 send_flags
|= HTC_FLAGS_NEED_CREDIT_UPDATE
;
151 ep
->ep_st
.cred_low_indicate
+= 1;
152 ath6kl_dbg(ATH6KL_DBG_HTC
,
153 "%s: host needs credits\n",
158 /* now we can fully dequeue */
159 packet
= list_first_entry(&ep
->txq
, struct htc_packet
, list
);
161 list_del(&packet
->list
);
162 /* save the number of credits this packet consumed */
163 packet
->info
.tx
.cred_used
= credits_required
;
164 /* save send flags */
165 packet
->info
.tx
.flags
= send_flags
;
166 packet
->info
.tx
.seqno
= ep
->seqno
;
168 /* queue this packet into the caller's queue */
169 list_add_tail(&packet
->list
, queue
);
173 static void get_htc_packet(struct htc_target
*target
,
174 struct htc_endpoint
*ep
,
175 struct list_head
*queue
, int resources
)
177 struct htc_packet
*packet
;
179 /* NOTE : the TX lock is held when this function is called */
181 /* loop until we can grab as many packets out of the queue as we can */
183 if (list_empty(&ep
->txq
))
186 packet
= list_first_entry(&ep
->txq
, struct htc_packet
, list
);
187 list_del(&packet
->list
);
189 ath6kl_dbg(ATH6KL_DBG_HTC
,
190 "%s: got packet:0x%p , new queue depth: %d\n",
191 __func__
, packet
, get_queue_depth(&ep
->txq
));
192 packet
->info
.tx
.seqno
= ep
->seqno
;
193 packet
->info
.tx
.flags
= 0;
194 packet
->info
.tx
.cred_used
= 0;
197 /* queue this packet into the caller's queue */
198 list_add_tail(&packet
->list
, queue
);
203 static int htc_issue_packets(struct htc_target
*target
,
204 struct htc_endpoint
*ep
,
205 struct list_head
*pkt_queue
)
210 struct htc_frame_hdr
*htc_hdr
;
211 struct htc_packet
*packet
;
213 ath6kl_dbg(ATH6KL_DBG_HTC
,
214 "%s: queue: 0x%p, pkts %d\n", __func__
,
215 pkt_queue
, get_queue_depth(pkt_queue
));
217 while (!list_empty(pkt_queue
)) {
218 packet
= list_first_entry(pkt_queue
, struct htc_packet
, list
);
219 list_del(&packet
->list
);
228 payload_len
= packet
->act_len
;
230 /* setup HTC frame header */
231 htc_hdr
= skb_push(skb
, sizeof(*htc_hdr
));
238 packet
->info
.tx
.flags
|= HTC_FLAGS_TX_FIXUP_NETBUF
;
241 put_unaligned((u16
) payload_len
, &htc_hdr
->payld_len
);
242 htc_hdr
->flags
= packet
->info
.tx
.flags
;
243 htc_hdr
->eid
= (u8
) packet
->endpoint
;
244 htc_hdr
->ctrl
[0] = 0;
245 htc_hdr
->ctrl
[1] = (u8
) packet
->info
.tx
.seqno
;
247 spin_lock_bh(&target
->tx_lock
);
249 /* store in look up queue to match completions */
250 list_add_tail(&packet
->list
, &ep
->pipe
.tx_lookup_queue
);
251 ep
->ep_st
.tx_issued
+= 1;
252 spin_unlock_bh(&target
->tx_lock
);
254 status
= ath6kl_hif_pipe_send(target
->dev
->ar
,
255 ep
->pipe
.pipeid_ul
, NULL
, skb
);
258 if (status
!= -ENOMEM
) {
259 /* TODO: if more than 1 endpoint maps to the
260 * same PipeID, it is possible to run out of
261 * resources in the HIF layer.
262 * Don't emit the error
264 ath6kl_dbg(ATH6KL_DBG_HTC
,
265 "%s: failed status:%d\n",
268 spin_lock_bh(&target
->tx_lock
);
269 list_del(&packet
->list
);
271 /* reclaim credits */
272 ep
->cred_dist
.credits
+= packet
->info
.tx
.cred_used
;
273 spin_unlock_bh(&target
->tx_lock
);
275 /* put it back into the callers queue */
276 list_add(&packet
->list
, pkt_queue
);
282 while (!list_empty(pkt_queue
)) {
283 if (status
!= -ENOMEM
) {
284 ath6kl_dbg(ATH6KL_DBG_HTC
,
285 "%s: failed pkt:0x%p status:%d\n",
286 __func__
, packet
, status
);
289 packet
= list_first_entry(pkt_queue
,
290 struct htc_packet
, list
);
291 list_del(&packet
->list
);
292 packet
->status
= status
;
293 send_packet_completion(target
, packet
);
300 static enum htc_send_queue_result
htc_try_send(struct htc_target
*target
,
301 struct htc_endpoint
*ep
,
302 struct list_head
*txq
)
304 struct list_head send_queue
; /* temp queue to hold packets */
305 struct htc_packet
*packet
, *tmp_pkt
;
306 struct ath6kl
*ar
= target
->dev
->ar
;
307 enum htc_send_full_action action
;
308 int tx_resources
, overflow
, txqueue_depth
, i
, good_pkts
;
311 ath6kl_dbg(ATH6KL_DBG_HTC
, "%s: (queue:0x%p depth:%d)\n",
313 (txq
== NULL
) ? 0 : get_queue_depth(txq
));
315 /* init the local send queue */
316 INIT_LIST_HEAD(&send_queue
);
319 * txq equals to NULL means
320 * caller didn't provide a queue, just wants us to
321 * check queues and send
324 if (list_empty(txq
)) {
326 return HTC_SEND_QUEUE_DROP
;
329 spin_lock_bh(&target
->tx_lock
);
330 txqueue_depth
= get_queue_depth(&ep
->txq
);
331 spin_unlock_bh(&target
->tx_lock
);
333 if (txqueue_depth
>= ep
->max_txq_depth
) {
334 /* we've already overflowed */
335 overflow
= get_queue_depth(txq
);
337 /* get how much we will overflow by */
338 overflow
= txqueue_depth
;
339 overflow
+= get_queue_depth(txq
);
340 /* get how much we will overflow the TX queue by */
341 overflow
-= ep
->max_txq_depth
;
344 /* if overflow is negative or zero, we are okay */
346 ath6kl_dbg(ATH6KL_DBG_HTC
,
347 "%s: Endpoint %d, TX queue will overflow :%d, Tx Depth:%d, Max:%d\n",
348 __func__
, ep
->eid
, overflow
, txqueue_depth
,
351 if ((overflow
<= 0) ||
352 (ep
->ep_cb
.tx_full
== NULL
)) {
354 * all packets will fit or caller did not provide send
355 * full indication handler -- just move all of them
356 * to the local send_queue object
358 list_splice_tail_init(txq
, &send_queue
);
360 good_pkts
= get_queue_depth(txq
) - overflow
;
363 return HTC_SEND_QUEUE_DROP
;
366 /* we have overflowed, and a callback is provided */
367 /* dequeue all non-overflow packets to the sendqueue */
368 for (i
= 0; i
< good_pkts
; i
++) {
369 /* pop off caller's queue */
370 packet
= list_first_entry(txq
,
373 /* move to local queue */
374 list_move_tail(&packet
->list
, &send_queue
);
378 * the caller's queue has all the packets that won't fit
379 * walk through the caller's queue and indicate each to
380 * the send full handler
382 list_for_each_entry_safe(packet
, tmp_pkt
,
384 ath6kl_dbg(ATH6KL_DBG_HTC
,
385 "%s: Indicate overflowed TX pkts: %p\n",
387 action
= ep
->ep_cb
.tx_full(ep
->target
, packet
);
388 if (action
== HTC_SEND_FULL_DROP
) {
389 /* callback wants the packet dropped */
390 ep
->ep_st
.tx_dropped
+= 1;
392 /* leave this one in the caller's queue
395 /* callback wants to keep this packet,
396 * move from caller's queue to the send
398 list_move_tail(&packet
->list
,
403 if (list_empty(&send_queue
)) {
404 /* no packets made it in, caller will cleanup */
405 return HTC_SEND_QUEUE_DROP
;
410 if (!ep
->pipe
.tx_credit_flow_enabled
) {
412 ath6kl_hif_pipe_get_free_queue_number(ar
,
418 spin_lock_bh(&target
->tx_lock
);
419 if (!list_empty(&send_queue
)) {
420 /* transfer packets to tail */
421 list_splice_tail_init(&send_queue
, &ep
->txq
);
422 if (!list_empty(&send_queue
)) {
424 spin_unlock_bh(&target
->tx_lock
);
425 return HTC_SEND_QUEUE_DROP
;
427 INIT_LIST_HEAD(&send_queue
);
430 /* increment tx processing count on entry */
433 if (ep
->tx_proc_cnt
> 1) {
435 * Another thread or task is draining the TX queues on this
436 * endpoint that thread will reset the tx processing count
437 * when the queue is drained.
440 spin_unlock_bh(&target
->tx_lock
);
441 return HTC_SEND_QUEUE_OK
;
444 /***** beyond this point only 1 thread may enter ******/
447 * Now drain the endpoint TX queue for transmission as long as we have
448 * enough transmit resources.
451 if (get_queue_depth(&ep
->txq
) == 0)
454 if (ep
->pipe
.tx_credit_flow_enabled
) {
456 * Credit based mechanism provides flow control
457 * based on target transmit resource availability,
458 * we assume that the HIF layer will always have
459 * bus resources greater than target transmit
462 get_htc_packet_credit_based(target
, ep
, &send_queue
);
465 * Get all packets for this endpoint that we can
468 get_htc_packet(target
, ep
, &send_queue
, tx_resources
);
471 if (get_queue_depth(&send_queue
) == 0) {
473 * Didn't get packets due to out of resources or TX
479 spin_unlock_bh(&target
->tx_lock
);
481 /* send what we can */
482 htc_issue_packets(target
, ep
, &send_queue
);
484 if (!ep
->pipe
.tx_credit_flow_enabled
) {
485 pipeid
= ep
->pipe
.pipeid_ul
;
487 ath6kl_hif_pipe_get_free_queue_number(ar
, pipeid
);
490 spin_lock_bh(&target
->tx_lock
);
493 /* done with this endpoint, we can clear the count */
495 spin_unlock_bh(&target
->tx_lock
);
497 return HTC_SEND_QUEUE_OK
;
500 /* htc control packet manipulation */
501 static void destroy_htc_txctrl_packet(struct htc_packet
*packet
)
509 static struct htc_packet
*build_htc_txctrl_packet(void)
511 struct htc_packet
*packet
= NULL
;
514 packet
= kzalloc(sizeof(struct htc_packet
), GFP_KERNEL
);
518 skb
= __dev_alloc_skb(HTC_CONTROL_BUFFER_SIZE
, GFP_KERNEL
);
529 static void htc_free_txctrl_packet(struct htc_target
*target
,
530 struct htc_packet
*packet
)
532 destroy_htc_txctrl_packet(packet
);
535 static struct htc_packet
*htc_alloc_txctrl_packet(struct htc_target
*target
)
537 return build_htc_txctrl_packet();
540 static void htc_txctrl_complete(struct htc_target
*target
,
541 struct htc_packet
*packet
)
543 htc_free_txctrl_packet(target
, packet
);
546 #define MAX_MESSAGE_SIZE 1536
548 static int htc_setup_target_buffer_assignments(struct htc_target
*target
)
550 int status
, credits
, credit_per_maxmsg
, i
;
551 struct htc_pipe_txcredit_alloc
*entry
;
552 unsigned int hif_usbaudioclass
= 0;
554 credit_per_maxmsg
= MAX_MESSAGE_SIZE
/ target
->tgt_cred_sz
;
555 if (MAX_MESSAGE_SIZE
% target
->tgt_cred_sz
)
558 /* TODO, this should be configured by the caller! */
560 credits
= target
->tgt_creds
;
561 entry
= &target
->pipe
.txcredit_alloc
[0];
565 /* FIXME: hif_usbaudioclass is always zero */
566 if (hif_usbaudioclass
) {
567 ath6kl_dbg(ATH6KL_DBG_HTC
,
568 "%s: For USB Audio Class- Total:%d\n",
572 /* Setup VO Service To have Max Credits */
573 entry
->service_id
= WMI_DATA_VO_SVC
;
574 entry
->credit_alloc
= (credits
- 6);
575 if (entry
->credit_alloc
== 0)
576 entry
->credit_alloc
++;
578 credits
-= (int) entry
->credit_alloc
;
583 entry
->service_id
= WMI_CONTROL_SVC
;
584 entry
->credit_alloc
= credit_per_maxmsg
;
585 credits
-= (int) entry
->credit_alloc
;
589 /* leftovers go to best effort */
592 entry
->service_id
= WMI_DATA_BE_SVC
;
593 entry
->credit_alloc
= (u8
) credits
;
597 entry
->service_id
= WMI_DATA_VI_SVC
;
598 entry
->credit_alloc
= credits
/ 4;
599 if (entry
->credit_alloc
== 0)
600 entry
->credit_alloc
++;
602 credits
-= (int) entry
->credit_alloc
;
607 entry
->service_id
= WMI_DATA_VO_SVC
;
608 entry
->credit_alloc
= credits
/ 4;
609 if (entry
->credit_alloc
== 0)
610 entry
->credit_alloc
++;
612 credits
-= (int) entry
->credit_alloc
;
617 entry
->service_id
= WMI_CONTROL_SVC
;
618 entry
->credit_alloc
= credit_per_maxmsg
;
619 credits
-= (int) entry
->credit_alloc
;
624 entry
->service_id
= WMI_DATA_BK_SVC
;
625 entry
->credit_alloc
= credit_per_maxmsg
;
626 credits
-= (int) entry
->credit_alloc
;
630 /* leftovers go to best effort */
632 entry
->service_id
= WMI_DATA_BE_SVC
;
633 entry
->credit_alloc
= (u8
) credits
;
638 for (i
= 0; i
< ENDPOINT_MAX
; i
++) {
639 if (target
->pipe
.txcredit_alloc
[i
].service_id
!= 0) {
640 ath6kl_dbg(ATH6KL_DBG_HTC
,
641 "HTC Service Index : %d TX : 0x%2.2X : alloc:%d\n",
643 target
->pipe
.txcredit_alloc
[i
].
645 target
->pipe
.txcredit_alloc
[i
].
653 /* process credit reports and call distribution function */
654 static void htc_process_credit_report(struct htc_target
*target
,
655 struct htc_credit_report
*rpt
,
657 enum htc_endpoint_id from_ep
)
659 int total_credits
= 0, i
;
660 struct htc_endpoint
*ep
;
662 /* lock out TX while we update credits */
663 spin_lock_bh(&target
->tx_lock
);
665 for (i
= 0; i
< num_entries
; i
++, rpt
++) {
666 if (rpt
->eid
>= ENDPOINT_MAX
) {
668 spin_unlock_bh(&target
->tx_lock
);
672 ep
= &target
->endpoint
[rpt
->eid
];
673 ep
->cred_dist
.credits
+= rpt
->credits
;
675 if (ep
->cred_dist
.credits
&& get_queue_depth(&ep
->txq
)) {
676 spin_unlock_bh(&target
->tx_lock
);
677 htc_try_send(target
, ep
, NULL
);
678 spin_lock_bh(&target
->tx_lock
);
681 total_credits
+= rpt
->credits
;
683 ath6kl_dbg(ATH6KL_DBG_HTC
,
684 "Report indicated %d credits to distribute\n",
687 spin_unlock_bh(&target
->tx_lock
);
690 /* flush endpoint TX queue */
691 static void htc_flush_tx_endpoint(struct htc_target
*target
,
692 struct htc_endpoint
*ep
, u16 tag
)
694 struct htc_packet
*packet
;
696 spin_lock_bh(&target
->tx_lock
);
697 while (get_queue_depth(&ep
->txq
)) {
698 packet
= list_first_entry(&ep
->txq
, struct htc_packet
, list
);
699 list_del(&packet
->list
);
701 send_packet_completion(target
, packet
);
703 spin_unlock_bh(&target
->tx_lock
);
707 * In the adapted HIF layer, struct sk_buff * are passed between HIF and HTC,
708 * since upper layers expects struct htc_packet containers we use the completed
709 * skb and lookup it's corresponding HTC packet buffer from a lookup list.
710 * This is extra overhead that can be fixed by re-aligning HIF interfaces with
713 static struct htc_packet
*htc_lookup_tx_packet(struct htc_target
*target
,
714 struct htc_endpoint
*ep
,
717 struct htc_packet
*packet
, *tmp_pkt
, *found_packet
= NULL
;
719 spin_lock_bh(&target
->tx_lock
);
722 * interate from the front of tx lookup queue
723 * this lookup should be fast since lower layers completes in-order and
724 * so the completed packet should be at the head of the list generally
726 list_for_each_entry_safe(packet
, tmp_pkt
, &ep
->pipe
.tx_lookup_queue
,
728 /* check for removal */
729 if (skb
== packet
->skb
) {
731 list_del(&packet
->list
);
732 found_packet
= packet
;
737 spin_unlock_bh(&target
->tx_lock
);
742 static int ath6kl_htc_pipe_tx_complete(struct ath6kl
*ar
, struct sk_buff
*skb
)
744 struct htc_target
*target
= ar
->htc_target
;
745 struct htc_frame_hdr
*htc_hdr
;
746 struct htc_endpoint
*ep
;
747 struct htc_packet
*packet
;
754 htc_hdr
= (struct htc_frame_hdr
*) netdata
;
756 ep_id
= htc_hdr
->eid
;
757 ep
= &target
->endpoint
[ep_id
];
759 packet
= htc_lookup_tx_packet(target
, ep
, skb
);
760 if (packet
== NULL
) {
761 /* may have already been flushed and freed */
762 ath6kl_err("HTC TX lookup failed!\n");
764 /* will be giving this buffer back to upper layers */
766 send_packet_completion(target
, packet
);
770 if (!ep
->pipe
.tx_credit_flow_enabled
) {
772 * note: when using TX credit flow, the re-checking of queues
773 * happens when credits flow back from the target. in the
774 * non-TX credit case, we recheck after the packet completes
776 htc_try_send(target
, ep
, NULL
);
782 static int htc_send_packets_multiple(struct htc_target
*target
,
783 struct list_head
*pkt_queue
)
785 struct htc_endpoint
*ep
;
786 struct htc_packet
*packet
, *tmp_pkt
;
788 if (list_empty(pkt_queue
))
791 /* get first packet to find out which ep the packets will go into */
792 packet
= list_first_entry(pkt_queue
, struct htc_packet
, list
);
794 if (packet
->endpoint
>= ENDPOINT_MAX
) {
798 ep
= &target
->endpoint
[packet
->endpoint
];
800 htc_try_send(target
, ep
, pkt_queue
);
802 /* do completion on any packets that couldn't get in */
803 if (!list_empty(pkt_queue
)) {
804 list_for_each_entry_safe(packet
, tmp_pkt
, pkt_queue
, list
) {
805 packet
->status
= -ENOMEM
;
808 do_send_completion(ep
, pkt_queue
);
814 /* htc pipe rx path */
815 static struct htc_packet
*alloc_htc_packet_container(struct htc_target
*target
)
817 struct htc_packet
*packet
;
818 spin_lock_bh(&target
->rx_lock
);
820 if (target
->pipe
.htc_packet_pool
== NULL
) {
821 spin_unlock_bh(&target
->rx_lock
);
825 packet
= target
->pipe
.htc_packet_pool
;
826 target
->pipe
.htc_packet_pool
= (struct htc_packet
*) packet
->list
.next
;
828 spin_unlock_bh(&target
->rx_lock
);
830 packet
->list
.next
= NULL
;
834 static void free_htc_packet_container(struct htc_target
*target
,
835 struct htc_packet
*packet
)
837 struct list_head
*lh
;
839 spin_lock_bh(&target
->rx_lock
);
841 if (target
->pipe
.htc_packet_pool
== NULL
) {
842 target
->pipe
.htc_packet_pool
= packet
;
843 packet
->list
.next
= NULL
;
845 lh
= (struct list_head
*) target
->pipe
.htc_packet_pool
;
846 packet
->list
.next
= lh
;
847 target
->pipe
.htc_packet_pool
= packet
;
850 spin_unlock_bh(&target
->rx_lock
);
853 static int htc_process_trailer(struct htc_target
*target
, u8
*buffer
,
854 int len
, enum htc_endpoint_id from_ep
)
856 struct htc_credit_report
*report
;
857 struct htc_record_hdr
*record
;
858 u8
*record_buf
, *orig_buf
;
859 int orig_len
, status
;
866 if (len
< sizeof(struct htc_record_hdr
)) {
871 /* these are byte aligned structs */
872 record
= (struct htc_record_hdr
*) buffer
;
873 len
-= sizeof(struct htc_record_hdr
);
874 buffer
+= sizeof(struct htc_record_hdr
);
876 if (record
->len
> len
) {
877 /* no room left in buffer for record */
878 ath6kl_dbg(ATH6KL_DBG_HTC
,
879 "invalid length: %d (id:%d) buffer has: %d bytes left\n",
880 record
->len
, record
->rec_id
, len
);
885 /* start of record follows the header */
888 switch (record
->rec_id
) {
889 case HTC_RECORD_CREDITS
:
890 if (record
->len
< sizeof(struct htc_credit_report
)) {
895 report
= (struct htc_credit_report
*) record_buf
;
896 htc_process_credit_report(target
, report
,
897 record
->len
/ sizeof(*report
),
901 ath6kl_dbg(ATH6KL_DBG_HTC
,
902 "unhandled record: id:%d length:%d\n",
903 record
->rec_id
, record
->len
);
910 /* advance buffer past this record for next time around */
911 buffer
+= record
->len
;
918 static void do_recv_completion(struct htc_endpoint
*ep
,
919 struct list_head
*queue_to_indicate
)
921 struct htc_packet
*packet
;
923 if (list_empty(queue_to_indicate
)) {
924 /* nothing to indicate */
928 /* using legacy EpRecv */
929 while (!list_empty(queue_to_indicate
)) {
930 packet
= list_first_entry(queue_to_indicate
,
931 struct htc_packet
, list
);
932 list_del(&packet
->list
);
933 ep
->ep_cb
.rx(ep
->target
, packet
);
939 static void recv_packet_completion(struct htc_target
*target
,
940 struct htc_endpoint
*ep
,
941 struct htc_packet
*packet
)
943 struct list_head container
;
944 INIT_LIST_HEAD(&container
);
945 list_add_tail(&packet
->list
, &container
);
948 do_recv_completion(ep
, &container
);
951 static int ath6kl_htc_pipe_rx_complete(struct ath6kl
*ar
, struct sk_buff
*skb
,
954 struct htc_target
*target
= ar
->htc_target
;
955 u8
*netdata
, *trailer
, hdr_info
;
956 struct htc_frame_hdr
*htc_hdr
;
957 u32 netlen
, trailerlen
= 0;
958 struct htc_packet
*packet
;
959 struct htc_endpoint
*ep
;
964 * ar->htc_target can be NULL due to a race condition that can occur
965 * during driver initialization(we do 'ath6kl_hif_power_on' before
966 * initializing 'ar->htc_target' via 'ath6kl_htc_create').
967 * 'ath6kl_hif_power_on' assigns 'ath6kl_recv_complete' as
968 * usb_complete_t/callback function for 'usb_fill_bulk_urb'.
969 * Thus the possibility of ar->htc_target being NULL
970 * via ath6kl_recv_complete -> ath6kl_usb_io_comp_work.
972 if (WARN_ON_ONCE(!target
)) {
973 ath6kl_err("Target not yet initialized\n");
982 htc_hdr
= (struct htc_frame_hdr
*) netdata
;
984 if (htc_hdr
->eid
>= ENDPOINT_MAX
) {
985 ath6kl_dbg(ATH6KL_DBG_HTC
,
986 "HTC Rx: invalid EndpointID=%d\n",
991 ep
= &target
->endpoint
[htc_hdr
->eid
];
993 payload_len
= le16_to_cpu(get_unaligned(&htc_hdr
->payld_len
));
995 if (netlen
< (payload_len
+ HTC_HDR_LENGTH
)) {
996 ath6kl_dbg(ATH6KL_DBG_HTC
,
997 "HTC Rx: insufficient length, got:%d expected =%zu\n",
998 netlen
, payload_len
+ HTC_HDR_LENGTH
);
1003 /* get flags to check for trailer */
1004 hdr_info
= htc_hdr
->flags
;
1005 if (hdr_info
& HTC_FLG_RX_TRAILER
) {
1006 /* extract the trailer length */
1007 hdr_info
= htc_hdr
->ctrl
[0];
1008 if ((hdr_info
< sizeof(struct htc_record_hdr
)) ||
1009 (hdr_info
> payload_len
)) {
1010 ath6kl_dbg(ATH6KL_DBG_HTC
,
1011 "invalid header: payloadlen should be %d, CB[0]: %d\n",
1012 payload_len
, hdr_info
);
1017 trailerlen
= hdr_info
;
1018 /* process trailer after hdr/apps payload */
1019 trailer
= (u8
*) htc_hdr
+ HTC_HDR_LENGTH
+
1020 payload_len
- hdr_info
;
1021 status
= htc_process_trailer(target
, trailer
, hdr_info
,
1027 if (((int) payload_len
- (int) trailerlen
) <= 0) {
1028 /* zero length packet with trailer, just drop these */
1032 if (htc_hdr
->eid
== ENDPOINT_0
) {
1033 /* handle HTC control message */
1034 if (target
->htc_flags
& HTC_OP_STATE_SETUP_COMPLETE
) {
1036 * fatal: target should not send unsolicited
1037 * messageson the endpoint 0
1039 ath6kl_dbg(ATH6KL_DBG_HTC
,
1040 "HTC ignores Rx Ctrl after setup complete\n");
1045 /* remove HTC header */
1046 skb_pull(skb
, HTC_HDR_LENGTH
);
1048 netdata
= skb
->data
;
1051 spin_lock_bh(&target
->rx_lock
);
1053 target
->pipe
.ctrl_response_valid
= true;
1054 target
->pipe
.ctrl_response_len
= min_t(int, netlen
,
1055 HTC_MAX_CTRL_MSG_LEN
);
1056 memcpy(target
->pipe
.ctrl_response_buf
, netdata
,
1057 target
->pipe
.ctrl_response_len
);
1059 spin_unlock_bh(&target
->rx_lock
);
1068 * TODO: the message based HIF architecture allocates net bufs
1069 * for recv packets since it bridges that HIF to upper layers,
1070 * which expects HTC packets, we form the packets here
1072 packet
= alloc_htc_packet_container(target
);
1073 if (packet
== NULL
) {
1079 packet
->endpoint
= htc_hdr
->eid
;
1080 packet
->pkt_cntxt
= skb
;
1082 /* TODO: for backwards compatibility */
1083 packet
->buf
= skb_push(skb
, 0) + HTC_HDR_LENGTH
;
1084 packet
->act_len
= netlen
- HTC_HDR_LENGTH
- trailerlen
;
1087 * TODO: this is a hack because the driver layer will set the
1088 * actual len of the skb again which will just double the len
1092 recv_packet_completion(target
, ep
, packet
);
1094 /* recover the packet container */
1095 free_htc_packet_container(target
, packet
);
1104 static void htc_flush_rx_queue(struct htc_target
*target
,
1105 struct htc_endpoint
*ep
)
1107 struct list_head container
;
1108 struct htc_packet
*packet
;
1110 spin_lock_bh(&target
->rx_lock
);
1113 if (list_empty(&ep
->rx_bufq
))
1116 packet
= list_first_entry(&ep
->rx_bufq
,
1117 struct htc_packet
, list
);
1118 list_del(&packet
->list
);
1120 spin_unlock_bh(&target
->rx_lock
);
1121 packet
->status
= -ECANCELED
;
1122 packet
->act_len
= 0;
1124 ath6kl_dbg(ATH6KL_DBG_HTC
,
1125 "Flushing RX packet:0x%p, length:%d, ep:%d\n",
1126 packet
, packet
->buf_len
,
1129 INIT_LIST_HEAD(&container
);
1130 list_add_tail(&packet
->list
, &container
);
1132 /* give the packet back */
1133 do_recv_completion(ep
, &container
);
1134 spin_lock_bh(&target
->rx_lock
);
1137 spin_unlock_bh(&target
->rx_lock
);
1140 /* polling routine to wait for a control packet to be received */
1141 static int htc_wait_recv_ctrl_message(struct htc_target
*target
)
1143 int count
= HTC_TARGET_RESPONSE_POLL_COUNT
;
1146 spin_lock_bh(&target
->rx_lock
);
1148 if (target
->pipe
.ctrl_response_valid
) {
1149 target
->pipe
.ctrl_response_valid
= false;
1150 spin_unlock_bh(&target
->rx_lock
);
1154 spin_unlock_bh(&target
->rx_lock
);
1158 msleep_interruptible(HTC_TARGET_RESPONSE_POLL_WAIT
);
1162 ath6kl_warn("htc pipe control receive timeout!\n");
1169 static void htc_rxctrl_complete(struct htc_target
*context
,
1170 struct htc_packet
*packet
)
1172 struct sk_buff
*skb
= packet
->skb
;
1174 if (packet
->endpoint
== ENDPOINT_0
&&
1175 packet
->status
== -ECANCELED
&&
1180 /* htc pipe initialization */
1181 static void reset_endpoint_states(struct htc_target
*target
)
1183 struct htc_endpoint
*ep
;
1186 for (i
= ENDPOINT_0
; i
< ENDPOINT_MAX
; i
++) {
1187 ep
= &target
->endpoint
[i
];
1190 ep
->max_txq_depth
= 0;
1192 INIT_LIST_HEAD(&ep
->txq
);
1193 INIT_LIST_HEAD(&ep
->pipe
.tx_lookup_queue
);
1194 INIT_LIST_HEAD(&ep
->rx_bufq
);
1195 ep
->target
= target
;
1196 ep
->pipe
.tx_credit_flow_enabled
= true;
1200 /* start HTC, this is called after all services are connected */
1201 static int htc_config_target_hif_pipe(struct htc_target
*target
)
1206 /* htc service functions */
1207 static u8
htc_get_credit_alloc(struct htc_target
*target
, u16 service_id
)
1212 for (i
= 0; i
< ENDPOINT_MAX
; i
++) {
1213 if (target
->pipe
.txcredit_alloc
[i
].service_id
== service_id
)
1215 target
->pipe
.txcredit_alloc
[i
].credit_alloc
;
1218 if (allocation
== 0) {
1219 ath6kl_dbg(ATH6KL_DBG_HTC
,
1220 "HTC Service TX : 0x%2.2X : allocation is zero!\n",
1227 static int ath6kl_htc_pipe_conn_service(struct htc_target
*target
,
1228 struct htc_service_connect_req
*conn_req
,
1229 struct htc_service_connect_resp
*conn_resp
)
1231 struct ath6kl
*ar
= target
->dev
->ar
;
1232 struct htc_packet
*packet
= NULL
;
1233 struct htc_conn_service_resp
*resp_msg
;
1234 struct htc_conn_service_msg
*conn_msg
;
1235 enum htc_endpoint_id assigned_epid
= ENDPOINT_MAX
;
1236 bool disable_credit_flowctrl
= false;
1237 unsigned int max_msg_size
= 0;
1238 struct htc_endpoint
*ep
;
1239 int length
, status
= 0;
1240 struct sk_buff
*skb
;
1244 if (conn_req
->svc_id
== 0) {
1250 if (conn_req
->svc_id
== HTC_CTRL_RSVD_SVC
) {
1251 /* special case for pseudo control service */
1252 assigned_epid
= ENDPOINT_0
;
1253 max_msg_size
= HTC_MAX_CTRL_MSG_LEN
;
1257 tx_alloc
= htc_get_credit_alloc(target
, conn_req
->svc_id
);
1258 if (tx_alloc
== 0) {
1263 /* allocate a packet to send to the target */
1264 packet
= htc_alloc_txctrl_packet(target
);
1266 if (packet
== NULL
) {
1273 length
= sizeof(struct htc_conn_service_msg
);
1275 /* assemble connect service message */
1276 conn_msg
= skb_put(skb
, length
);
1277 if (conn_msg
== NULL
) {
1284 sizeof(struct htc_conn_service_msg
));
1285 conn_msg
->msg_id
= cpu_to_le16(HTC_MSG_CONN_SVC_ID
);
1286 conn_msg
->svc_id
= cpu_to_le16(conn_req
->svc_id
);
1287 conn_msg
->conn_flags
= cpu_to_le16(conn_req
->conn_flags
&
1288 ~HTC_CONN_FLGS_SET_RECV_ALLOC_MASK
);
1290 /* tell target desired recv alloc for this ep */
1291 flags
= tx_alloc
<< HTC_CONN_FLGS_SET_RECV_ALLOC_SHIFT
;
1292 conn_msg
->conn_flags
|= cpu_to_le16(flags
);
1294 if (conn_req
->conn_flags
&
1295 HTC_CONN_FLGS_DISABLE_CRED_FLOW_CTRL
) {
1296 disable_credit_flowctrl
= true;
1299 set_htc_pkt_info(packet
, NULL
, (u8
*) conn_msg
,
1301 ENDPOINT_0
, HTC_SERVICE_TX_PACKET_TAG
);
1303 status
= ath6kl_htc_pipe_tx(target
, packet
);
1305 /* we don't own it anymore */
1310 /* wait for response */
1311 status
= htc_wait_recv_ctrl_message(target
);
1315 /* we controlled the buffer creation so it has to be
1318 resp_msg
= (struct htc_conn_service_resp
*)
1319 target
->pipe
.ctrl_response_buf
;
1321 if (resp_msg
->msg_id
!= cpu_to_le16(HTC_MSG_CONN_SVC_RESP_ID
) ||
1322 (target
->pipe
.ctrl_response_len
< sizeof(*resp_msg
))) {
1323 /* this message is not valid */
1329 ath6kl_dbg(ATH6KL_DBG_TRC
,
1330 "%s: service 0x%X conn resp: status: %d ep: %d\n",
1331 __func__
, resp_msg
->svc_id
, resp_msg
->status
,
1334 conn_resp
->resp_code
= resp_msg
->status
;
1335 /* check response status */
1336 if (resp_msg
->status
!= HTC_SERVICE_SUCCESS
) {
1337 ath6kl_dbg(ATH6KL_DBG_HTC
,
1338 "Target failed service 0x%X connect request (status:%d)\n",
1339 resp_msg
->svc_id
, resp_msg
->status
);
1344 assigned_epid
= (enum htc_endpoint_id
) resp_msg
->eid
;
1345 max_msg_size
= le16_to_cpu(resp_msg
->max_msg_sz
);
1348 /* the rest are parameter checks so set the error status */
1351 if (assigned_epid
>= ENDPOINT_MAX
) {
1356 if (max_msg_size
== 0) {
1361 ep
= &target
->endpoint
[assigned_epid
];
1362 ep
->eid
= assigned_epid
;
1363 if (ep
->svc_id
!= 0) {
1364 /* endpoint already in use! */
1369 /* return assigned endpoint to caller */
1370 conn_resp
->endpoint
= assigned_epid
;
1371 conn_resp
->len_max
= max_msg_size
;
1373 /* setup the endpoint */
1374 ep
->svc_id
= conn_req
->svc_id
; /* this marks ep in use */
1375 ep
->max_txq_depth
= conn_req
->max_txq_depth
;
1376 ep
->len_max
= max_msg_size
;
1377 ep
->cred_dist
.credits
= tx_alloc
;
1378 ep
->cred_dist
.cred_sz
= target
->tgt_cred_sz
;
1379 ep
->cred_dist
.cred_per_msg
= max_msg_size
/ target
->tgt_cred_sz
;
1380 if (max_msg_size
% target
->tgt_cred_sz
)
1381 ep
->cred_dist
.cred_per_msg
++;
1383 /* copy all the callbacks */
1384 ep
->ep_cb
= conn_req
->ep_cb
;
1386 /* initialize tx_drop_packet_threshold */
1387 ep
->tx_drop_packet_threshold
= MAX_HI_COOKIE_NUM
;
1389 status
= ath6kl_hif_pipe_map_service(ar
, ep
->svc_id
,
1390 &ep
->pipe
.pipeid_ul
,
1391 &ep
->pipe
.pipeid_dl
);
1395 ath6kl_dbg(ATH6KL_DBG_HTC
,
1396 "SVC Ready: 0x%4.4X: ULpipe:%d DLpipe:%d id:%d\n",
1397 ep
->svc_id
, ep
->pipe
.pipeid_ul
,
1398 ep
->pipe
.pipeid_dl
, ep
->eid
);
1400 if (disable_credit_flowctrl
&& ep
->pipe
.tx_credit_flow_enabled
) {
1401 ep
->pipe
.tx_credit_flow_enabled
= false;
1402 ath6kl_dbg(ATH6KL_DBG_HTC
,
1403 "SVC: 0x%4.4X ep:%d TX flow control off\n",
1404 ep
->svc_id
, assigned_epid
);
1409 htc_free_txctrl_packet(target
, packet
);
1413 /* htc export functions */
1414 static void *ath6kl_htc_pipe_create(struct ath6kl
*ar
)
1417 struct htc_endpoint
*ep
= NULL
;
1418 struct htc_target
*target
= NULL
;
1419 struct htc_packet
*packet
;
1422 target
= kzalloc(sizeof(struct htc_target
), GFP_KERNEL
);
1423 if (target
== NULL
) {
1424 ath6kl_err("htc create unable to allocate memory\n");
1426 goto fail_htc_create
;
1429 spin_lock_init(&target
->htc_lock
);
1430 spin_lock_init(&target
->rx_lock
);
1431 spin_lock_init(&target
->tx_lock
);
1433 reset_endpoint_states(target
);
1435 for (i
= 0; i
< HTC_PACKET_CONTAINER_ALLOCATION
; i
++) {
1436 packet
= kzalloc(sizeof(struct htc_packet
), GFP_KERNEL
);
1439 free_htc_packet_container(target
, packet
);
1442 target
->dev
= kzalloc(sizeof(*target
->dev
), GFP_KERNEL
);
1444 ath6kl_err("unable to allocate memory\n");
1446 goto fail_htc_create
;
1448 target
->dev
->ar
= ar
;
1449 target
->dev
->htc_cnxt
= target
;
1451 /* Get HIF default pipe for HTC message exchange */
1452 ep
= &target
->endpoint
[ENDPOINT_0
];
1454 ath6kl_hif_pipe_get_default(ar
, &ep
->pipe
.pipeid_ul
,
1455 &ep
->pipe
.pipeid_dl
);
1462 ath6kl_htc_pipe_cleanup(target
);
1469 /* cleanup the HTC instance */
1470 static void ath6kl_htc_pipe_cleanup(struct htc_target
*target
)
1472 struct htc_packet
*packet
;
1475 packet
= alloc_htc_packet_container(target
);
1483 /* kfree our instance */
1487 static int ath6kl_htc_pipe_start(struct htc_target
*target
)
1489 struct sk_buff
*skb
;
1490 struct htc_setup_comp_ext_msg
*setup
;
1491 struct htc_packet
*packet
;
1493 htc_config_target_hif_pipe(target
);
1495 /* allocate a buffer to send */
1496 packet
= htc_alloc_txctrl_packet(target
);
1497 if (packet
== NULL
) {
1504 /* assemble setup complete message */
1505 setup
= skb_put(skb
, sizeof(*setup
));
1506 memset(setup
, 0, sizeof(struct htc_setup_comp_ext_msg
));
1507 setup
->msg_id
= cpu_to_le16(HTC_MSG_SETUP_COMPLETE_EX_ID
);
1509 ath6kl_dbg(ATH6KL_DBG_HTC
, "HTC using TX credit flow control\n");
1511 set_htc_pkt_info(packet
, NULL
, (u8
*) setup
,
1512 sizeof(struct htc_setup_comp_ext_msg
),
1513 ENDPOINT_0
, HTC_SERVICE_TX_PACKET_TAG
);
1515 target
->htc_flags
|= HTC_OP_STATE_SETUP_COMPLETE
;
1517 return ath6kl_htc_pipe_tx(target
, packet
);
1520 static void ath6kl_htc_pipe_stop(struct htc_target
*target
)
1523 struct htc_endpoint
*ep
;
1525 /* cleanup endpoints */
1526 for (i
= 0; i
< ENDPOINT_MAX
; i
++) {
1527 ep
= &target
->endpoint
[i
];
1528 htc_flush_rx_queue(target
, ep
);
1529 htc_flush_tx_endpoint(target
, ep
, HTC_TX_PACKET_TAG_ALL
);
1532 reset_endpoint_states(target
);
1533 target
->htc_flags
&= ~HTC_OP_STATE_SETUP_COMPLETE
;
1536 static int ath6kl_htc_pipe_get_rxbuf_num(struct htc_target
*target
,
1537 enum htc_endpoint_id endpoint
)
1541 spin_lock_bh(&target
->rx_lock
);
1542 num
= get_queue_depth(&(target
->endpoint
[endpoint
].rx_bufq
));
1543 spin_unlock_bh(&target
->rx_lock
);
1548 static int ath6kl_htc_pipe_tx(struct htc_target
*target
,
1549 struct htc_packet
*packet
)
1551 struct list_head queue
;
1553 ath6kl_dbg(ATH6KL_DBG_HTC
,
1554 "%s: endPointId: %d, buffer: 0x%p, length: %d\n",
1555 __func__
, packet
->endpoint
, packet
->buf
,
1558 INIT_LIST_HEAD(&queue
);
1559 list_add_tail(&packet
->list
, &queue
);
1561 return htc_send_packets_multiple(target
, &queue
);
1564 static int ath6kl_htc_pipe_wait_target(struct htc_target
*target
)
1566 struct htc_ready_ext_msg
*ready_msg
;
1567 struct htc_service_connect_req connect
;
1568 struct htc_service_connect_resp resp
;
1571 status
= htc_wait_recv_ctrl_message(target
);
1576 if (target
->pipe
.ctrl_response_len
< sizeof(*ready_msg
)) {
1577 ath6kl_warn("invalid htc pipe ready msg len: %d\n",
1578 target
->pipe
.ctrl_response_len
);
1582 ready_msg
= (struct htc_ready_ext_msg
*) target
->pipe
.ctrl_response_buf
;
1584 if (ready_msg
->ver2_0_info
.msg_id
!= cpu_to_le16(HTC_MSG_READY_ID
)) {
1585 ath6kl_warn("invalid htc pipe ready msg: 0x%x\n",
1586 ready_msg
->ver2_0_info
.msg_id
);
1590 ath6kl_dbg(ATH6KL_DBG_HTC
,
1591 "Target Ready! : transmit resources : %d size:%d\n",
1592 ready_msg
->ver2_0_info
.cred_cnt
,
1593 ready_msg
->ver2_0_info
.cred_sz
);
1595 target
->tgt_creds
= le16_to_cpu(ready_msg
->ver2_0_info
.cred_cnt
);
1596 target
->tgt_cred_sz
= le16_to_cpu(ready_msg
->ver2_0_info
.cred_sz
);
1598 if ((target
->tgt_creds
== 0) || (target
->tgt_cred_sz
== 0))
1601 htc_setup_target_buffer_assignments(target
);
1603 /* setup our pseudo HTC control endpoint connection */
1604 memset(&connect
, 0, sizeof(connect
));
1605 memset(&resp
, 0, sizeof(resp
));
1606 connect
.ep_cb
.tx_complete
= htc_txctrl_complete
;
1607 connect
.ep_cb
.rx
= htc_rxctrl_complete
;
1608 connect
.max_txq_depth
= NUM_CONTROL_TX_BUFFERS
;
1609 connect
.svc_id
= HTC_CTRL_RSVD_SVC
;
1611 /* connect fake service */
1612 status
= ath6kl_htc_pipe_conn_service(target
, &connect
, &resp
);
1617 static void ath6kl_htc_pipe_flush_txep(struct htc_target
*target
,
1618 enum htc_endpoint_id endpoint
, u16 tag
)
1620 struct htc_endpoint
*ep
= &target
->endpoint
[endpoint
];
1622 if (ep
->svc_id
== 0) {
1628 htc_flush_tx_endpoint(target
, ep
, tag
);
1631 static int ath6kl_htc_pipe_add_rxbuf_multiple(struct htc_target
*target
,
1632 struct list_head
*pkt_queue
)
1634 struct htc_packet
*packet
, *tmp_pkt
, *first
;
1635 struct htc_endpoint
*ep
;
1638 if (list_empty(pkt_queue
))
1641 first
= list_first_entry(pkt_queue
, struct htc_packet
, list
);
1643 if (first
->endpoint
>= ENDPOINT_MAX
) {
1648 ath6kl_dbg(ATH6KL_DBG_HTC
, "%s: epid: %d, cnt:%d, len: %d\n",
1649 __func__
, first
->endpoint
, get_queue_depth(pkt_queue
),
1652 ep
= &target
->endpoint
[first
->endpoint
];
1654 spin_lock_bh(&target
->rx_lock
);
1656 /* store receive packets */
1657 list_splice_tail_init(pkt_queue
, &ep
->rx_bufq
);
1659 spin_unlock_bh(&target
->rx_lock
);
1662 /* walk through queue and mark each one canceled */
1663 list_for_each_entry_safe(packet
, tmp_pkt
, pkt_queue
, list
) {
1664 packet
->status
= -ECANCELED
;
1667 do_recv_completion(ep
, pkt_queue
);
1673 static void ath6kl_htc_pipe_activity_changed(struct htc_target
*target
,
1674 enum htc_endpoint_id ep
,
1680 static void ath6kl_htc_pipe_flush_rx_buf(struct htc_target
*target
)
1682 struct htc_endpoint
*endpoint
;
1683 struct htc_packet
*packet
, *tmp_pkt
;
1686 for (i
= ENDPOINT_0
; i
< ENDPOINT_MAX
; i
++) {
1687 endpoint
= &target
->endpoint
[i
];
1689 spin_lock_bh(&target
->rx_lock
);
1691 list_for_each_entry_safe(packet
, tmp_pkt
,
1692 &endpoint
->rx_bufq
, list
) {
1693 list_del(&packet
->list
);
1694 spin_unlock_bh(&target
->rx_lock
);
1695 ath6kl_dbg(ATH6KL_DBG_HTC
,
1696 "htc rx flush pkt 0x%p len %d ep %d\n",
1697 packet
, packet
->buf_len
,
1699 dev_kfree_skb(packet
->pkt_cntxt
);
1700 spin_lock_bh(&target
->rx_lock
);
1703 spin_unlock_bh(&target
->rx_lock
);
1707 static int ath6kl_htc_pipe_credit_setup(struct htc_target
*target
,
1708 struct ath6kl_htc_credit_info
*info
)
1713 static const struct ath6kl_htc_ops ath6kl_htc_pipe_ops
= {
1714 .create
= ath6kl_htc_pipe_create
,
1715 .wait_target
= ath6kl_htc_pipe_wait_target
,
1716 .start
= ath6kl_htc_pipe_start
,
1717 .conn_service
= ath6kl_htc_pipe_conn_service
,
1718 .tx
= ath6kl_htc_pipe_tx
,
1719 .stop
= ath6kl_htc_pipe_stop
,
1720 .cleanup
= ath6kl_htc_pipe_cleanup
,
1721 .flush_txep
= ath6kl_htc_pipe_flush_txep
,
1722 .flush_rx_buf
= ath6kl_htc_pipe_flush_rx_buf
,
1723 .activity_changed
= ath6kl_htc_pipe_activity_changed
,
1724 .get_rxbuf_num
= ath6kl_htc_pipe_get_rxbuf_num
,
1725 .add_rxbuf_multiple
= ath6kl_htc_pipe_add_rxbuf_multiple
,
1726 .credit_setup
= ath6kl_htc_pipe_credit_setup
,
1727 .tx_complete
= ath6kl_htc_pipe_tx_complete
,
1728 .rx_complete
= ath6kl_htc_pipe_rx_complete
,
1731 void ath6kl_htc_pipe_attach(struct ath6kl
*ar
)
1733 ar
->htc_ops
= &ath6kl_htc_pipe_ops
;