2 * Copyright (c) 2007-2011 Atheros Communications Inc.
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
21 #define HTC_PACKET_CONTAINER_ALLOCATION 32
22 #define HTC_CONTROL_BUFFER_SIZE (HTC_MAX_CTRL_MSG_LEN + HTC_HDR_LENGTH)
24 static int ath6kl_htc_pipe_tx(struct htc_target
*handle
,
25 struct htc_packet
*packet
);
26 static void ath6kl_htc_pipe_cleanup(struct htc_target
*handle
);
28 /* htc pipe tx path */
29 static inline void restore_tx_packet(struct htc_packet
*packet
)
31 if (packet
->info
.tx
.flags
& HTC_FLAGS_TX_FIXUP_NETBUF
) {
32 skb_pull(packet
->skb
, sizeof(struct htc_frame_hdr
));
33 packet
->info
.tx
.flags
&= ~HTC_FLAGS_TX_FIXUP_NETBUF
;
37 static void do_send_completion(struct htc_endpoint
*ep
,
38 struct list_head
*queue_to_indicate
)
40 struct htc_packet
*packet
;
42 if (list_empty(queue_to_indicate
)) {
43 /* nothing to indicate */
47 if (ep
->ep_cb
.tx_comp_multi
!= NULL
) {
48 ath6kl_dbg(ATH6KL_DBG_HTC
,
49 "%s: calling ep %d, send complete multiple callback (%d pkts)\n",
51 get_queue_depth(queue_to_indicate
));
53 * a multiple send complete handler is being used,
54 * pass the queue to the handler
56 ep
->ep_cb
.tx_comp_multi(ep
->target
, queue_to_indicate
);
58 * all packets are now owned by the callback,
59 * reset queue to be safe
61 INIT_LIST_HEAD(queue_to_indicate
);
63 /* using legacy EpTxComplete */
65 packet
= list_first_entry(queue_to_indicate
,
66 struct htc_packet
, list
);
68 list_del(&packet
->list
);
69 ath6kl_dbg(ATH6KL_DBG_HTC
,
70 "%s: calling ep %d send complete callback on packet 0x%p\n",
71 __func__
, ep
->eid
, packet
);
72 ep
->ep_cb
.tx_complete(ep
->target
, packet
);
73 } while (!list_empty(queue_to_indicate
));
77 static void send_packet_completion(struct htc_target
*target
,
78 struct htc_packet
*packet
)
80 struct htc_endpoint
*ep
= &target
->endpoint
[packet
->endpoint
];
81 struct list_head container
;
83 restore_tx_packet(packet
);
84 INIT_LIST_HEAD(&container
);
85 list_add_tail(&packet
->list
, &container
);
88 do_send_completion(ep
, &container
);
91 static void get_htc_packet_credit_based(struct htc_target
*target
,
92 struct htc_endpoint
*ep
,
93 struct list_head
*queue
)
98 struct htc_packet
*packet
;
99 unsigned int transfer_len
;
101 /* NOTE : the TX lock is held when this function is called */
103 /* loop until we can grab as many packets out of the queue as we can */
106 if (list_empty(&ep
->txq
))
109 /* get packet at head, but don't remove it */
110 packet
= list_first_entry(&ep
->txq
, struct htc_packet
, list
);
112 ath6kl_dbg(ATH6KL_DBG_HTC
,
113 "%s: got head packet:0x%p , queue depth: %d\n",
114 __func__
, packet
, get_queue_depth(&ep
->txq
));
116 transfer_len
= packet
->act_len
+ HTC_HDR_LENGTH
;
118 if (transfer_len
<= target
->tgt_cred_sz
) {
119 credits_required
= 1;
121 /* figure out how many credits this message requires */
122 credits_required
= transfer_len
/ target
->tgt_cred_sz
;
123 remainder
= transfer_len
% target
->tgt_cred_sz
;
129 ath6kl_dbg(ATH6KL_DBG_HTC
, "%s: creds required:%d got:%d\n",
130 __func__
, credits_required
, ep
->cred_dist
.credits
);
132 if (ep
->eid
== ENDPOINT_0
) {
134 * endpoint 0 is special, it always has a credit and
135 * does not require credit based flow control
137 credits_required
= 0;
141 if (ep
->cred_dist
.credits
< credits_required
)
144 ep
->cred_dist
.credits
-= credits_required
;
145 ep
->ep_st
.cred_cosumd
+= credits_required
;
147 /* check if we need credits back from the target */
148 if (ep
->cred_dist
.credits
<
149 ep
->cred_dist
.cred_per_msg
) {
150 /* tell the target we need credits ASAP! */
151 send_flags
|= HTC_FLAGS_NEED_CREDIT_UPDATE
;
152 ep
->ep_st
.cred_low_indicate
+= 1;
153 ath6kl_dbg(ATH6KL_DBG_HTC
,
154 "%s: host needs credits\n",
159 /* now we can fully dequeue */
160 packet
= list_first_entry(&ep
->txq
, struct htc_packet
, list
);
162 list_del(&packet
->list
);
163 /* save the number of credits this packet consumed */
164 packet
->info
.tx
.cred_used
= credits_required
;
165 /* save send flags */
166 packet
->info
.tx
.flags
= send_flags
;
167 packet
->info
.tx
.seqno
= ep
->seqno
;
169 /* queue this packet into the caller's queue */
170 list_add_tail(&packet
->list
, queue
);
175 static void get_htc_packet(struct htc_target
*target
,
176 struct htc_endpoint
*ep
,
177 struct list_head
*queue
, int resources
)
179 struct htc_packet
*packet
;
181 /* NOTE : the TX lock is held when this function is called */
183 /* loop until we can grab as many packets out of the queue as we can */
185 if (list_empty(&ep
->txq
))
188 packet
= list_first_entry(&ep
->txq
, struct htc_packet
, list
);
189 list_del(&packet
->list
);
191 ath6kl_dbg(ATH6KL_DBG_HTC
,
192 "%s: got packet:0x%p , new queue depth: %d\n",
193 __func__
, packet
, get_queue_depth(&ep
->txq
));
194 packet
->info
.tx
.seqno
= ep
->seqno
;
195 packet
->info
.tx
.flags
= 0;
196 packet
->info
.tx
.cred_used
= 0;
199 /* queue this packet into the caller's queue */
200 list_add_tail(&packet
->list
, queue
);
205 static int htc_issue_packets(struct htc_target
*target
,
206 struct htc_endpoint
*ep
,
207 struct list_head
*pkt_queue
)
212 struct htc_frame_hdr
*htc_hdr
;
213 struct htc_packet
*packet
;
215 ath6kl_dbg(ATH6KL_DBG_HTC
,
216 "%s: queue: 0x%p, pkts %d\n", __func__
,
217 pkt_queue
, get_queue_depth(pkt_queue
));
219 while (!list_empty(pkt_queue
)) {
220 packet
= list_first_entry(pkt_queue
, struct htc_packet
, list
);
221 list_del(&packet
->list
);
230 payload_len
= packet
->act_len
;
232 /* setup HTC frame header */
233 htc_hdr
= (struct htc_frame_hdr
*) skb_push(skb
,
241 packet
->info
.tx
.flags
|= HTC_FLAGS_TX_FIXUP_NETBUF
;
244 put_unaligned((u16
) payload_len
, &htc_hdr
->payld_len
);
245 htc_hdr
->flags
= packet
->info
.tx
.flags
;
246 htc_hdr
->eid
= (u8
) packet
->endpoint
;
247 htc_hdr
->ctrl
[0] = 0;
248 htc_hdr
->ctrl
[1] = (u8
) packet
->info
.tx
.seqno
;
250 spin_lock_bh(&target
->tx_lock
);
252 /* store in look up queue to match completions */
253 list_add_tail(&packet
->list
, &ep
->pipe
.tx_lookup_queue
);
254 ep
->ep_st
.tx_issued
+= 1;
255 spin_unlock_bh(&target
->tx_lock
);
257 status
= ath6kl_hif_pipe_send(target
->dev
->ar
,
258 ep
->pipe
.pipeid_ul
, NULL
, skb
);
261 if (status
!= -ENOMEM
) {
262 /* TODO: if more than 1 endpoint maps to the
263 * same PipeID, it is possible to run out of
264 * resources in the HIF layer.
265 * Don't emit the error
267 ath6kl_dbg(ATH6KL_DBG_HTC
,
268 "%s: failed status:%d\n",
271 spin_lock_bh(&target
->tx_lock
);
272 list_del(&packet
->list
);
274 /* reclaim credits */
275 ep
->cred_dist
.credits
+= packet
->info
.tx
.cred_used
;
276 spin_unlock_bh(&target
->tx_lock
);
278 /* put it back into the callers queue */
279 list_add(&packet
->list
, pkt_queue
);
286 while (!list_empty(pkt_queue
)) {
287 if (status
!= -ENOMEM
) {
288 ath6kl_dbg(ATH6KL_DBG_HTC
,
289 "%s: failed pkt:0x%p status:%d\n",
290 __func__
, packet
, status
);
293 packet
= list_first_entry(pkt_queue
,
294 struct htc_packet
, list
);
295 list_del(&packet
->list
);
296 packet
->status
= status
;
297 send_packet_completion(target
, packet
);
304 static enum htc_send_queue_result
htc_try_send(struct htc_target
*target
,
305 struct htc_endpoint
*ep
,
306 struct list_head
*txq
)
308 struct list_head send_queue
; /* temp queue to hold packets */
309 struct htc_packet
*packet
, *tmp_pkt
;
310 struct ath6kl
*ar
= target
->dev
->ar
;
311 enum htc_send_full_action action
;
312 int tx_resources
, overflow
, txqueue_depth
, i
, good_pkts
;
315 ath6kl_dbg(ATH6KL_DBG_HTC
, "%s: (queue:0x%p depth:%d)\n",
317 (txq
== NULL
) ? 0 : get_queue_depth(txq
));
319 /* init the local send queue */
320 INIT_LIST_HEAD(&send_queue
);
323 * txq equals to NULL means
324 * caller didn't provide a queue, just wants us to
325 * check queues and send
328 if (list_empty(txq
)) {
330 return HTC_SEND_QUEUE_DROP
;
333 spin_lock_bh(&target
->tx_lock
);
334 txqueue_depth
= get_queue_depth(&ep
->txq
);
335 spin_unlock_bh(&target
->tx_lock
);
337 if (txqueue_depth
>= ep
->max_txq_depth
) {
338 /* we've already overflowed */
339 overflow
= get_queue_depth(txq
);
341 /* get how much we will overflow by */
342 overflow
= txqueue_depth
;
343 overflow
+= get_queue_depth(txq
);
344 /* get how much we will overflow the TX queue by */
345 overflow
-= ep
->max_txq_depth
;
348 /* if overflow is negative or zero, we are okay */
350 ath6kl_dbg(ATH6KL_DBG_HTC
,
351 "%s: Endpoint %d, TX queue will overflow :%d, Tx Depth:%d, Max:%d\n",
352 __func__
, ep
->eid
, overflow
, txqueue_depth
,
355 if ((overflow
<= 0) ||
356 (ep
->ep_cb
.tx_full
== NULL
)) {
358 * all packets will fit or caller did not provide send
359 * full indication handler -- just move all of them
360 * to the local send_queue object
362 list_splice_tail_init(txq
, &send_queue
);
364 good_pkts
= get_queue_depth(txq
) - overflow
;
367 return HTC_SEND_QUEUE_DROP
;
370 /* we have overflowed, and a callback is provided */
371 /* dequeue all non-overflow packets to the sendqueue */
372 for (i
= 0; i
< good_pkts
; i
++) {
373 /* pop off caller's queue */
374 packet
= list_first_entry(txq
,
377 /* move to local queue */
378 list_move_tail(&packet
->list
, &send_queue
);
382 * the caller's queue has all the packets that won't fit
383 * walk through the caller's queue and indicate each to
384 * the send full handler
386 list_for_each_entry_safe(packet
, tmp_pkt
,
389 ath6kl_dbg(ATH6KL_DBG_HTC
,
390 "%s: Indicat overflowed TX pkts: %p\n",
392 action
= ep
->ep_cb
.tx_full(ep
->target
, packet
);
393 if (action
== HTC_SEND_FULL_DROP
) {
394 /* callback wants the packet dropped */
395 ep
->ep_st
.tx_dropped
+= 1;
397 /* leave this one in the caller's queue
400 /* callback wants to keep this packet,
401 * move from caller's queue to the send
403 list_move_tail(&packet
->list
,
409 if (list_empty(&send_queue
)) {
410 /* no packets made it in, caller will cleanup */
411 return HTC_SEND_QUEUE_DROP
;
416 if (!ep
->pipe
.tx_credit_flow_enabled
) {
418 ath6kl_hif_pipe_get_free_queue_number(ar
,
424 spin_lock_bh(&target
->tx_lock
);
425 if (!list_empty(&send_queue
)) {
426 /* transfer packets to tail */
427 list_splice_tail_init(&send_queue
, &ep
->txq
);
428 if (!list_empty(&send_queue
)) {
430 spin_unlock_bh(&target
->tx_lock
);
431 return HTC_SEND_QUEUE_DROP
;
433 INIT_LIST_HEAD(&send_queue
);
436 /* increment tx processing count on entry */
439 if (ep
->tx_proc_cnt
> 1) {
441 * Another thread or task is draining the TX queues on this
442 * endpoint that thread will reset the tx processing count
443 * when the queue is drained.
446 spin_unlock_bh(&target
->tx_lock
);
447 return HTC_SEND_QUEUE_OK
;
450 /***** beyond this point only 1 thread may enter ******/
453 * Now drain the endpoint TX queue for transmission as long as we have
454 * enough transmit resources.
458 if (get_queue_depth(&ep
->txq
) == 0)
461 if (ep
->pipe
.tx_credit_flow_enabled
) {
463 * Credit based mechanism provides flow control
464 * based on target transmit resource availability,
465 * we assume that the HIF layer will always have
466 * bus resources greater than target transmit
469 get_htc_packet_credit_based(target
, ep
, &send_queue
);
472 * Get all packets for this endpoint that we can
475 get_htc_packet(target
, ep
, &send_queue
, tx_resources
);
478 if (get_queue_depth(&send_queue
) == 0) {
480 * Didn't get packets due to out of resources or TX
486 spin_unlock_bh(&target
->tx_lock
);
488 /* send what we can */
489 htc_issue_packets(target
, ep
, &send_queue
);
491 if (!ep
->pipe
.tx_credit_flow_enabled
) {
492 pipeid
= ep
->pipe
.pipeid_ul
;
494 ath6kl_hif_pipe_get_free_queue_number(ar
, pipeid
);
497 spin_lock_bh(&target
->tx_lock
);
500 /* done with this endpoint, we can clear the count */
502 spin_unlock_bh(&target
->tx_lock
);
504 return HTC_SEND_QUEUE_OK
;
507 /* htc control packet manipulation */
508 static void destroy_htc_txctrl_packet(struct htc_packet
*packet
)
516 static struct htc_packet
*build_htc_txctrl_packet(void)
518 struct htc_packet
*packet
= NULL
;
521 packet
= kzalloc(sizeof(struct htc_packet
), GFP_KERNEL
);
525 skb
= __dev_alloc_skb(HTC_CONTROL_BUFFER_SIZE
, GFP_KERNEL
);
536 static void htc_free_txctrl_packet(struct htc_target
*target
,
537 struct htc_packet
*packet
)
539 destroy_htc_txctrl_packet(packet
);
542 static struct htc_packet
*htc_alloc_txctrl_packet(struct htc_target
*target
)
544 return build_htc_txctrl_packet();
547 static void htc_txctrl_complete(struct htc_target
*target
,
548 struct htc_packet
*packet
)
550 htc_free_txctrl_packet(target
, packet
);
553 #define MAX_MESSAGE_SIZE 1536
555 static int htc_setup_target_buffer_assignments(struct htc_target
*target
)
557 int status
, credits
, credit_per_maxmsg
, i
;
558 struct htc_pipe_txcredit_alloc
*entry
;
559 unsigned int hif_usbaudioclass
= 0;
561 credit_per_maxmsg
= MAX_MESSAGE_SIZE
/ target
->tgt_cred_sz
;
562 if (MAX_MESSAGE_SIZE
% target
->tgt_cred_sz
)
565 /* TODO, this should be configured by the caller! */
567 credits
= target
->tgt_creds
;
568 entry
= &target
->pipe
.txcredit_alloc
[0];
572 /* FIXME: hif_usbaudioclass is always zero */
573 if (hif_usbaudioclass
) {
574 ath6kl_dbg(ATH6KL_DBG_HTC
,
575 "%s: For USB Audio Class- Total:%d\n",
579 /* Setup VO Service To have Max Credits */
580 entry
->service_id
= WMI_DATA_VO_SVC
;
581 entry
->credit_alloc
= (credits
- 6);
582 if (entry
->credit_alloc
== 0)
583 entry
->credit_alloc
++;
585 credits
-= (int) entry
->credit_alloc
;
590 entry
->service_id
= WMI_CONTROL_SVC
;
591 entry
->credit_alloc
= credit_per_maxmsg
;
592 credits
-= (int) entry
->credit_alloc
;
596 /* leftovers go to best effort */
599 entry
->service_id
= WMI_DATA_BE_SVC
;
600 entry
->credit_alloc
= (u8
) credits
;
604 entry
->service_id
= WMI_DATA_VI_SVC
;
605 entry
->credit_alloc
= credits
/ 4;
606 if (entry
->credit_alloc
== 0)
607 entry
->credit_alloc
++;
609 credits
-= (int) entry
->credit_alloc
;
614 entry
->service_id
= WMI_DATA_VO_SVC
;
615 entry
->credit_alloc
= credits
/ 4;
616 if (entry
->credit_alloc
== 0)
617 entry
->credit_alloc
++;
619 credits
-= (int) entry
->credit_alloc
;
624 entry
->service_id
= WMI_CONTROL_SVC
;
625 entry
->credit_alloc
= credit_per_maxmsg
;
626 credits
-= (int) entry
->credit_alloc
;
631 entry
->service_id
= WMI_DATA_BK_SVC
;
632 entry
->credit_alloc
= credit_per_maxmsg
;
633 credits
-= (int) entry
->credit_alloc
;
637 /* leftovers go to best effort */
639 entry
->service_id
= WMI_DATA_BE_SVC
;
640 entry
->credit_alloc
= (u8
) credits
;
645 for (i
= 0; i
< ENDPOINT_MAX
; i
++) {
646 if (target
->pipe
.txcredit_alloc
[i
].service_id
!= 0) {
647 ath6kl_dbg(ATH6KL_DBG_HTC
,
648 "HTC Service Index : %d TX : 0x%2.2X : alloc:%d\n",
650 target
->pipe
.txcredit_alloc
[i
].
652 target
->pipe
.txcredit_alloc
[i
].
660 /* process credit reports and call distribution function */
661 static void htc_process_credit_report(struct htc_target
*target
,
662 struct htc_credit_report
*rpt
,
664 enum htc_endpoint_id from_ep
)
666 int total_credits
= 0, i
;
667 struct htc_endpoint
*ep
;
669 /* lock out TX while we update credits */
670 spin_lock_bh(&target
->tx_lock
);
672 for (i
= 0; i
< num_entries
; i
++, rpt
++) {
673 if (rpt
->eid
>= ENDPOINT_MAX
) {
675 spin_unlock_bh(&target
->tx_lock
);
679 ep
= &target
->endpoint
[rpt
->eid
];
680 ep
->cred_dist
.credits
+= rpt
->credits
;
682 if (ep
->cred_dist
.credits
&& get_queue_depth(&ep
->txq
)) {
683 spin_unlock_bh(&target
->tx_lock
);
684 htc_try_send(target
, ep
, NULL
);
685 spin_lock_bh(&target
->tx_lock
);
688 total_credits
+= rpt
->credits
;
690 ath6kl_dbg(ATH6KL_DBG_HTC
,
691 "Report indicated %d credits to distribute\n",
694 spin_unlock_bh(&target
->tx_lock
);
697 /* flush endpoint TX queue */
698 static void htc_flush_tx_endpoint(struct htc_target
*target
,
699 struct htc_endpoint
*ep
, u16 tag
)
701 struct htc_packet
*packet
;
703 spin_lock_bh(&target
->tx_lock
);
704 while (get_queue_depth(&ep
->txq
)) {
705 packet
= list_first_entry(&ep
->txq
, struct htc_packet
, list
);
706 list_del(&packet
->list
);
708 send_packet_completion(target
, packet
);
710 spin_unlock_bh(&target
->tx_lock
);
714 * In the adapted HIF layer, struct sk_buff * are passed between HIF and HTC,
715 * since upper layers expects struct htc_packet containers we use the completed
716 * skb and lookup it's corresponding HTC packet buffer from a lookup list.
717 * This is extra overhead that can be fixed by re-aligning HIF interfaces with
720 static struct htc_packet
*htc_lookup_tx_packet(struct htc_target
*target
,
721 struct htc_endpoint
*ep
,
724 struct htc_packet
*packet
, *tmp_pkt
, *found_packet
= NULL
;
726 spin_lock_bh(&target
->tx_lock
);
729 * interate from the front of tx lookup queue
730 * this lookup should be fast since lower layers completes in-order and
731 * so the completed packet should be at the head of the list generally
733 list_for_each_entry_safe(packet
, tmp_pkt
, &ep
->pipe
.tx_lookup_queue
,
735 /* check for removal */
736 if (skb
== packet
->skb
) {
738 list_del(&packet
->list
);
739 found_packet
= packet
;
744 spin_unlock_bh(&target
->tx_lock
);
749 static int ath6kl_htc_pipe_tx_complete(struct ath6kl
*ar
, struct sk_buff
*skb
)
751 struct htc_target
*target
= ar
->htc_target
;
752 struct htc_frame_hdr
*htc_hdr
;
753 struct htc_endpoint
*ep
;
754 struct htc_packet
*packet
;
761 htc_hdr
= (struct htc_frame_hdr
*) netdata
;
763 ep_id
= htc_hdr
->eid
;
764 ep
= &target
->endpoint
[ep_id
];
766 packet
= htc_lookup_tx_packet(target
, ep
, skb
);
767 if (packet
== NULL
) {
768 /* may have already been flushed and freed */
769 ath6kl_err("HTC TX lookup failed!\n");
771 /* will be giving this buffer back to upper layers */
773 send_packet_completion(target
, packet
);
777 if (!ep
->pipe
.tx_credit_flow_enabled
) {
779 * note: when using TX credit flow, the re-checking of queues
780 * happens when credits flow back from the target. in the
781 * non-TX credit case, we recheck after the packet completes
783 htc_try_send(target
, ep
, NULL
);
789 static int htc_send_packets_multiple(struct htc_target
*target
,
790 struct list_head
*pkt_queue
)
792 struct htc_endpoint
*ep
;
793 struct htc_packet
*packet
, *tmp_pkt
;
795 if (list_empty(pkt_queue
))
798 /* get first packet to find out which ep the packets will go into */
799 packet
= list_first_entry(pkt_queue
, struct htc_packet
, list
);
801 if (packet
->endpoint
>= ENDPOINT_MAX
) {
805 ep
= &target
->endpoint
[packet
->endpoint
];
807 htc_try_send(target
, ep
, pkt_queue
);
809 /* do completion on any packets that couldn't get in */
810 if (!list_empty(pkt_queue
)) {
811 list_for_each_entry_safe(packet
, tmp_pkt
, pkt_queue
, list
) {
812 packet
->status
= -ENOMEM
;
815 do_send_completion(ep
, pkt_queue
);
821 /* htc pipe rx path */
822 static struct htc_packet
*alloc_htc_packet_container(struct htc_target
*target
)
824 struct htc_packet
*packet
;
825 spin_lock_bh(&target
->rx_lock
);
827 if (target
->pipe
.htc_packet_pool
== NULL
) {
828 spin_unlock_bh(&target
->rx_lock
);
832 packet
= target
->pipe
.htc_packet_pool
;
833 target
->pipe
.htc_packet_pool
= (struct htc_packet
*) packet
->list
.next
;
835 spin_unlock_bh(&target
->rx_lock
);
837 packet
->list
.next
= NULL
;
841 static void free_htc_packet_container(struct htc_target
*target
,
842 struct htc_packet
*packet
)
844 struct list_head
*lh
;
846 spin_lock_bh(&target
->rx_lock
);
848 if (target
->pipe
.htc_packet_pool
== NULL
) {
849 target
->pipe
.htc_packet_pool
= packet
;
850 packet
->list
.next
= NULL
;
852 lh
= (struct list_head
*) target
->pipe
.htc_packet_pool
;
853 packet
->list
.next
= lh
;
854 target
->pipe
.htc_packet_pool
= packet
;
857 spin_unlock_bh(&target
->rx_lock
);
860 static int htc_process_trailer(struct htc_target
*target
, u8
*buffer
,
861 int len
, enum htc_endpoint_id from_ep
)
863 struct htc_credit_report
*report
;
864 struct htc_record_hdr
*record
;
865 u8
*record_buf
, *orig_buf
;
866 int orig_len
, status
;
873 if (len
< sizeof(struct htc_record_hdr
)) {
878 /* these are byte aligned structs */
879 record
= (struct htc_record_hdr
*) buffer
;
880 len
-= sizeof(struct htc_record_hdr
);
881 buffer
+= sizeof(struct htc_record_hdr
);
883 if (record
->len
> len
) {
884 /* no room left in buffer for record */
885 ath6kl_dbg(ATH6KL_DBG_HTC
,
886 "invalid length: %d (id:%d) buffer has: %d bytes left\n",
887 record
->len
, record
->rec_id
, len
);
892 /* start of record follows the header */
895 switch (record
->rec_id
) {
896 case HTC_RECORD_CREDITS
:
897 if (record
->len
< sizeof(struct htc_credit_report
)) {
902 report
= (struct htc_credit_report
*) record_buf
;
903 htc_process_credit_report(target
, report
,
904 record
->len
/ sizeof(*report
),
908 ath6kl_dbg(ATH6KL_DBG_HTC
,
909 "unhandled record: id:%d length:%d\n",
910 record
->rec_id
, record
->len
);
917 /* advance buffer past this record for next time around */
918 buffer
+= record
->len
;
925 static void do_recv_completion(struct htc_endpoint
*ep
,
926 struct list_head
*queue_to_indicate
)
928 struct htc_packet
*packet
;
930 if (list_empty(queue_to_indicate
)) {
931 /* nothing to indicate */
935 /* using legacy EpRecv */
936 while (!list_empty(queue_to_indicate
)) {
937 packet
= list_first_entry(queue_to_indicate
,
938 struct htc_packet
, list
);
939 list_del(&packet
->list
);
940 ep
->ep_cb
.rx(ep
->target
, packet
);
946 static void recv_packet_completion(struct htc_target
*target
,
947 struct htc_endpoint
*ep
,
948 struct htc_packet
*packet
)
950 struct list_head container
;
951 INIT_LIST_HEAD(&container
);
952 list_add_tail(&packet
->list
, &container
);
955 do_recv_completion(ep
, &container
);
958 static int ath6kl_htc_pipe_rx_complete(struct ath6kl
*ar
, struct sk_buff
*skb
,
961 struct htc_target
*target
= ar
->htc_target
;
962 u8
*netdata
, *trailer
, hdr_info
;
963 struct htc_frame_hdr
*htc_hdr
;
964 u32 netlen
, trailerlen
= 0;
965 struct htc_packet
*packet
;
966 struct htc_endpoint
*ep
;
971 * ar->htc_target can be NULL due to a race condition that can occur
972 * during driver initialization(we do 'ath6kl_hif_power_on' before
973 * initializing 'ar->htc_target' via 'ath6kl_htc_create').
974 * 'ath6kl_hif_power_on' assigns 'ath6kl_recv_complete' as
975 * usb_complete_t/callback function for 'usb_fill_bulk_urb'.
976 * Thus the possibility of ar->htc_target being NULL
977 * via ath6kl_recv_complete -> ath6kl_usb_io_comp_work.
979 if (WARN_ON_ONCE(!target
)) {
980 ath6kl_err("Target not yet initialized\n");
989 htc_hdr
= (struct htc_frame_hdr
*) netdata
;
991 if (htc_hdr
->eid
>= ENDPOINT_MAX
) {
992 ath6kl_dbg(ATH6KL_DBG_HTC
,
993 "HTC Rx: invalid EndpointID=%d\n",
998 ep
= &target
->endpoint
[htc_hdr
->eid
];
1000 payload_len
= le16_to_cpu(get_unaligned(&htc_hdr
->payld_len
));
1002 if (netlen
< (payload_len
+ HTC_HDR_LENGTH
)) {
1003 ath6kl_dbg(ATH6KL_DBG_HTC
,
1004 "HTC Rx: insufficient length, got:%d expected =%u\n",
1005 netlen
, payload_len
+ HTC_HDR_LENGTH
);
1010 /* get flags to check for trailer */
1011 hdr_info
= htc_hdr
->flags
;
1012 if (hdr_info
& HTC_FLG_RX_TRAILER
) {
1013 /* extract the trailer length */
1014 hdr_info
= htc_hdr
->ctrl
[0];
1015 if ((hdr_info
< sizeof(struct htc_record_hdr
)) ||
1016 (hdr_info
> payload_len
)) {
1017 ath6kl_dbg(ATH6KL_DBG_HTC
,
1018 "invalid header: payloadlen should be %d, CB[0]: %d\n",
1019 payload_len
, hdr_info
);
1024 trailerlen
= hdr_info
;
1025 /* process trailer after hdr/apps payload */
1026 trailer
= (u8
*) htc_hdr
+ HTC_HDR_LENGTH
+
1027 payload_len
- hdr_info
;
1028 status
= htc_process_trailer(target
, trailer
, hdr_info
,
1034 if (((int) payload_len
- (int) trailerlen
) <= 0) {
1035 /* zero length packet with trailer, just drop these */
1039 if (htc_hdr
->eid
== ENDPOINT_0
) {
1040 /* handle HTC control message */
1041 if (target
->htc_flags
& HTC_OP_STATE_SETUP_COMPLETE
) {
1043 * fatal: target should not send unsolicited
1044 * messageson the endpoint 0
1046 ath6kl_dbg(ATH6KL_DBG_HTC
,
1047 "HTC ignores Rx Ctrl after setup complete\n");
1052 /* remove HTC header */
1053 skb_pull(skb
, HTC_HDR_LENGTH
);
1055 netdata
= skb
->data
;
1058 spin_lock_bh(&target
->rx_lock
);
1060 target
->pipe
.ctrl_response_valid
= true;
1061 target
->pipe
.ctrl_response_len
= min_t(int, netlen
,
1062 HTC_MAX_CTRL_MSG_LEN
);
1063 memcpy(target
->pipe
.ctrl_response_buf
, netdata
,
1064 target
->pipe
.ctrl_response_len
);
1066 spin_unlock_bh(&target
->rx_lock
);
1075 * TODO: the message based HIF architecture allocates net bufs
1076 * for recv packets since it bridges that HIF to upper layers,
1077 * which expects HTC packets, we form the packets here
1079 packet
= alloc_htc_packet_container(target
);
1080 if (packet
== NULL
) {
1086 packet
->endpoint
= htc_hdr
->eid
;
1087 packet
->pkt_cntxt
= skb
;
1089 /* TODO: for backwards compatibility */
1090 packet
->buf
= skb_push(skb
, 0) + HTC_HDR_LENGTH
;
1091 packet
->act_len
= netlen
- HTC_HDR_LENGTH
- trailerlen
;
1094 * TODO: this is a hack because the driver layer will set the
1095 * actual len of the skb again which will just double the len
1099 recv_packet_completion(target
, ep
, packet
);
1101 /* recover the packet container */
1102 free_htc_packet_container(target
, packet
);
1112 static void htc_flush_rx_queue(struct htc_target
*target
,
1113 struct htc_endpoint
*ep
)
1115 struct list_head container
;
1116 struct htc_packet
*packet
;
1118 spin_lock_bh(&target
->rx_lock
);
1121 if (list_empty(&ep
->rx_bufq
))
1124 packet
= list_first_entry(&ep
->rx_bufq
,
1125 struct htc_packet
, list
);
1126 list_del(&packet
->list
);
1128 spin_unlock_bh(&target
->rx_lock
);
1129 packet
->status
= -ECANCELED
;
1130 packet
->act_len
= 0;
1132 ath6kl_dbg(ATH6KL_DBG_HTC
,
1133 "Flushing RX packet:0x%p, length:%d, ep:%d\n",
1134 packet
, packet
->buf_len
,
1137 INIT_LIST_HEAD(&container
);
1138 list_add_tail(&packet
->list
, &container
);
1140 /* give the packet back */
1141 do_recv_completion(ep
, &container
);
1142 spin_lock_bh(&target
->rx_lock
);
1145 spin_unlock_bh(&target
->rx_lock
);
1148 /* polling routine to wait for a control packet to be received */
1149 static int htc_wait_recv_ctrl_message(struct htc_target
*target
)
1151 int count
= HTC_TARGET_RESPONSE_POLL_COUNT
;
1154 spin_lock_bh(&target
->rx_lock
);
1156 if (target
->pipe
.ctrl_response_valid
) {
1157 target
->pipe
.ctrl_response_valid
= false;
1158 spin_unlock_bh(&target
->rx_lock
);
1162 spin_unlock_bh(&target
->rx_lock
);
1166 msleep_interruptible(HTC_TARGET_RESPONSE_POLL_WAIT
);
1170 ath6kl_warn("htc pipe control receive timeout!\n");
1177 static void htc_rxctrl_complete(struct htc_target
*context
,
1178 struct htc_packet
*packet
)
1180 /* TODO, can't really receive HTC control messages yet.... */
1181 ath6kl_dbg(ATH6KL_DBG_HTC
, "%s: invalid call function\n", __func__
);
1184 /* htc pipe initialization */
1185 static void reset_endpoint_states(struct htc_target
*target
)
1187 struct htc_endpoint
*ep
;
1190 for (i
= ENDPOINT_0
; i
< ENDPOINT_MAX
; i
++) {
1191 ep
= &target
->endpoint
[i
];
1194 ep
->max_txq_depth
= 0;
1196 INIT_LIST_HEAD(&ep
->txq
);
1197 INIT_LIST_HEAD(&ep
->pipe
.tx_lookup_queue
);
1198 INIT_LIST_HEAD(&ep
->rx_bufq
);
1199 ep
->target
= target
;
1200 ep
->pipe
.tx_credit_flow_enabled
= true;
1204 /* start HTC, this is called after all services are connected */
1205 static int htc_config_target_hif_pipe(struct htc_target
*target
)
1210 /* htc service functions */
1211 static u8
htc_get_credit_alloc(struct htc_target
*target
, u16 service_id
)
1216 for (i
= 0; i
< ENDPOINT_MAX
; i
++) {
1217 if (target
->pipe
.txcredit_alloc
[i
].service_id
== service_id
)
1219 target
->pipe
.txcredit_alloc
[i
].credit_alloc
;
1222 if (allocation
== 0) {
1223 ath6kl_dbg(ATH6KL_DBG_HTC
,
1224 "HTC Service TX : 0x%2.2X : allocation is zero!\n",
1231 static int ath6kl_htc_pipe_conn_service(struct htc_target
*target
,
1232 struct htc_service_connect_req
*conn_req
,
1233 struct htc_service_connect_resp
*conn_resp
)
1235 struct ath6kl
*ar
= target
->dev
->ar
;
1236 struct htc_packet
*packet
= NULL
;
1237 struct htc_conn_service_resp
*resp_msg
;
1238 struct htc_conn_service_msg
*conn_msg
;
1239 enum htc_endpoint_id assigned_epid
= ENDPOINT_MAX
;
1240 bool disable_credit_flowctrl
= false;
1241 unsigned int max_msg_size
= 0;
1242 struct htc_endpoint
*ep
;
1243 int length
, status
= 0;
1244 struct sk_buff
*skb
;
1248 if (conn_req
->svc_id
== 0) {
1254 if (conn_req
->svc_id
== HTC_CTRL_RSVD_SVC
) {
1255 /* special case for pseudo control service */
1256 assigned_epid
= ENDPOINT_0
;
1257 max_msg_size
= HTC_MAX_CTRL_MSG_LEN
;
1262 tx_alloc
= htc_get_credit_alloc(target
, conn_req
->svc_id
);
1263 if (tx_alloc
== 0) {
1268 /* allocate a packet to send to the target */
1269 packet
= htc_alloc_txctrl_packet(target
);
1271 if (packet
== NULL
) {
1278 length
= sizeof(struct htc_conn_service_msg
);
1280 /* assemble connect service message */
1281 conn_msg
= (struct htc_conn_service_msg
*) skb_put(skb
,
1283 if (conn_msg
== NULL
) {
1290 sizeof(struct htc_conn_service_msg
));
1291 conn_msg
->msg_id
= cpu_to_le16(HTC_MSG_CONN_SVC_ID
);
1292 conn_msg
->svc_id
= cpu_to_le16(conn_req
->svc_id
);
1293 conn_msg
->conn_flags
= cpu_to_le16(conn_req
->conn_flags
&
1294 ~HTC_CONN_FLGS_SET_RECV_ALLOC_MASK
);
1296 /* tell target desired recv alloc for this ep */
1297 flags
= tx_alloc
<< HTC_CONN_FLGS_SET_RECV_ALLOC_SHIFT
;
1298 conn_msg
->conn_flags
|= cpu_to_le16(flags
);
1300 if (conn_req
->conn_flags
&
1301 HTC_CONN_FLGS_DISABLE_CRED_FLOW_CTRL
) {
1302 disable_credit_flowctrl
= true;
1305 set_htc_pkt_info(packet
, NULL
, (u8
*) conn_msg
,
1307 ENDPOINT_0
, HTC_SERVICE_TX_PACKET_TAG
);
1309 status
= ath6kl_htc_pipe_tx(target
, packet
);
1311 /* we don't own it anymore */
1316 /* wait for response */
1317 status
= htc_wait_recv_ctrl_message(target
);
1321 /* we controlled the buffer creation so it has to be
1324 resp_msg
= (struct htc_conn_service_resp
*)
1325 target
->pipe
.ctrl_response_buf
;
1327 if (resp_msg
->msg_id
!= cpu_to_le16(HTC_MSG_CONN_SVC_RESP_ID
) ||
1328 (target
->pipe
.ctrl_response_len
< sizeof(*resp_msg
))) {
1329 /* this message is not valid */
1335 ath6kl_dbg(ATH6KL_DBG_TRC
,
1336 "%s: service 0x%X conn resp: status: %d ep: %d\n",
1337 __func__
, resp_msg
->svc_id
, resp_msg
->status
,
1340 conn_resp
->resp_code
= resp_msg
->status
;
1341 /* check response status */
1342 if (resp_msg
->status
!= HTC_SERVICE_SUCCESS
) {
1343 ath6kl_dbg(ATH6KL_DBG_HTC
,
1344 "Target failed service 0x%X connect request (status:%d)\n",
1345 resp_msg
->svc_id
, resp_msg
->status
);
1350 assigned_epid
= (enum htc_endpoint_id
) resp_msg
->eid
;
1351 max_msg_size
= le16_to_cpu(resp_msg
->max_msg_sz
);
1354 /* the rest are parameter checks so set the error status */
1357 if (assigned_epid
>= ENDPOINT_MAX
) {
1362 if (max_msg_size
== 0) {
1367 ep
= &target
->endpoint
[assigned_epid
];
1368 ep
->eid
= assigned_epid
;
1369 if (ep
->svc_id
!= 0) {
1370 /* endpoint already in use! */
1375 /* return assigned endpoint to caller */
1376 conn_resp
->endpoint
= assigned_epid
;
1377 conn_resp
->len_max
= max_msg_size
;
1379 /* setup the endpoint */
1380 ep
->svc_id
= conn_req
->svc_id
; /* this marks ep in use */
1381 ep
->max_txq_depth
= conn_req
->max_txq_depth
;
1382 ep
->len_max
= max_msg_size
;
1383 ep
->cred_dist
.credits
= tx_alloc
;
1384 ep
->cred_dist
.cred_sz
= target
->tgt_cred_sz
;
1385 ep
->cred_dist
.cred_per_msg
= max_msg_size
/ target
->tgt_cred_sz
;
1386 if (max_msg_size
% target
->tgt_cred_sz
)
1387 ep
->cred_dist
.cred_per_msg
++;
1389 /* copy all the callbacks */
1390 ep
->ep_cb
= conn_req
->ep_cb
;
1392 /* initialize tx_drop_packet_threshold */
1393 ep
->tx_drop_packet_threshold
= MAX_HI_COOKIE_NUM
;
1395 status
= ath6kl_hif_pipe_map_service(ar
, ep
->svc_id
,
1396 &ep
->pipe
.pipeid_ul
,
1397 &ep
->pipe
.pipeid_dl
);
1401 ath6kl_dbg(ATH6KL_DBG_HTC
,
1402 "SVC Ready: 0x%4.4X: ULpipe:%d DLpipe:%d id:%d\n",
1403 ep
->svc_id
, ep
->pipe
.pipeid_ul
,
1404 ep
->pipe
.pipeid_dl
, ep
->eid
);
1406 if (disable_credit_flowctrl
&& ep
->pipe
.tx_credit_flow_enabled
) {
1407 ep
->pipe
.tx_credit_flow_enabled
= false;
1408 ath6kl_dbg(ATH6KL_DBG_HTC
,
1409 "SVC: 0x%4.4X ep:%d TX flow control off\n",
1410 ep
->svc_id
, assigned_epid
);
1415 htc_free_txctrl_packet(target
, packet
);
1419 /* htc export functions */
1420 static void *ath6kl_htc_pipe_create(struct ath6kl
*ar
)
1423 struct htc_endpoint
*ep
= NULL
;
1424 struct htc_target
*target
= NULL
;
1425 struct htc_packet
*packet
;
1428 target
= kzalloc(sizeof(struct htc_target
), GFP_KERNEL
);
1429 if (target
== NULL
) {
1430 ath6kl_err("htc create unable to allocate memory\n");
1432 goto fail_htc_create
;
1435 spin_lock_init(&target
->htc_lock
);
1436 spin_lock_init(&target
->rx_lock
);
1437 spin_lock_init(&target
->tx_lock
);
1439 reset_endpoint_states(target
);
1441 for (i
= 0; i
< HTC_PACKET_CONTAINER_ALLOCATION
; i
++) {
1442 packet
= kzalloc(sizeof(struct htc_packet
), GFP_KERNEL
);
1445 free_htc_packet_container(target
, packet
);
1448 target
->dev
= kzalloc(sizeof(*target
->dev
), GFP_KERNEL
);
1450 ath6kl_err("unable to allocate memory\n");
1452 goto fail_htc_create
;
1454 target
->dev
->ar
= ar
;
1455 target
->dev
->htc_cnxt
= target
;
1457 /* Get HIF default pipe for HTC message exchange */
1458 ep
= &target
->endpoint
[ENDPOINT_0
];
1460 ath6kl_hif_pipe_get_default(ar
, &ep
->pipe
.pipeid_ul
,
1461 &ep
->pipe
.pipeid_dl
);
1468 ath6kl_htc_pipe_cleanup(target
);
1475 /* cleanup the HTC instance */
1476 static void ath6kl_htc_pipe_cleanup(struct htc_target
*target
)
1478 struct htc_packet
*packet
;
1481 packet
= alloc_htc_packet_container(target
);
1489 /* kfree our instance */
1493 static int ath6kl_htc_pipe_start(struct htc_target
*target
)
1495 struct sk_buff
*skb
;
1496 struct htc_setup_comp_ext_msg
*setup
;
1497 struct htc_packet
*packet
;
1499 htc_config_target_hif_pipe(target
);
1501 /* allocate a buffer to send */
1502 packet
= htc_alloc_txctrl_packet(target
);
1503 if (packet
== NULL
) {
1510 /* assemble setup complete message */
1511 setup
= (struct htc_setup_comp_ext_msg
*) skb_put(skb
,
1513 memset(setup
, 0, sizeof(struct htc_setup_comp_ext_msg
));
1514 setup
->msg_id
= cpu_to_le16(HTC_MSG_SETUP_COMPLETE_EX_ID
);
1516 ath6kl_dbg(ATH6KL_DBG_HTC
, "HTC using TX credit flow control\n");
1518 set_htc_pkt_info(packet
, NULL
, (u8
*) setup
,
1519 sizeof(struct htc_setup_comp_ext_msg
),
1520 ENDPOINT_0
, HTC_SERVICE_TX_PACKET_TAG
);
1522 target
->htc_flags
|= HTC_OP_STATE_SETUP_COMPLETE
;
1524 return ath6kl_htc_pipe_tx(target
, packet
);
1527 static void ath6kl_htc_pipe_stop(struct htc_target
*target
)
1530 struct htc_endpoint
*ep
;
1532 /* cleanup endpoints */
1533 for (i
= 0; i
< ENDPOINT_MAX
; i
++) {
1534 ep
= &target
->endpoint
[i
];
1535 htc_flush_rx_queue(target
, ep
);
1536 htc_flush_tx_endpoint(target
, ep
, HTC_TX_PACKET_TAG_ALL
);
1539 reset_endpoint_states(target
);
1540 target
->htc_flags
&= ~HTC_OP_STATE_SETUP_COMPLETE
;
1543 static int ath6kl_htc_pipe_get_rxbuf_num(struct htc_target
*target
,
1544 enum htc_endpoint_id endpoint
)
1548 spin_lock_bh(&target
->rx_lock
);
1549 num
= get_queue_depth(&(target
->endpoint
[endpoint
].rx_bufq
));
1550 spin_unlock_bh(&target
->rx_lock
);
1555 static int ath6kl_htc_pipe_tx(struct htc_target
*target
,
1556 struct htc_packet
*packet
)
1558 struct list_head queue
;
1560 ath6kl_dbg(ATH6KL_DBG_HTC
,
1561 "%s: endPointId: %d, buffer: 0x%p, length: %d\n",
1562 __func__
, packet
->endpoint
, packet
->buf
,
1565 INIT_LIST_HEAD(&queue
);
1566 list_add_tail(&packet
->list
, &queue
);
1568 return htc_send_packets_multiple(target
, &queue
);
1571 static int ath6kl_htc_pipe_wait_target(struct htc_target
*target
)
1573 struct htc_ready_ext_msg
*ready_msg
;
1574 struct htc_service_connect_req connect
;
1575 struct htc_service_connect_resp resp
;
1578 status
= htc_wait_recv_ctrl_message(target
);
1583 if (target
->pipe
.ctrl_response_len
< sizeof(*ready_msg
)) {
1584 ath6kl_warn("invalid htc pipe ready msg len: %d\n",
1585 target
->pipe
.ctrl_response_len
);
1589 ready_msg
= (struct htc_ready_ext_msg
*) target
->pipe
.ctrl_response_buf
;
1591 if (ready_msg
->ver2_0_info
.msg_id
!= cpu_to_le16(HTC_MSG_READY_ID
)) {
1592 ath6kl_warn("invalid htc pipe ready msg: 0x%x\n",
1593 ready_msg
->ver2_0_info
.msg_id
);
1597 ath6kl_dbg(ATH6KL_DBG_HTC
,
1598 "Target Ready! : transmit resources : %d size:%d\n",
1599 ready_msg
->ver2_0_info
.cred_cnt
,
1600 ready_msg
->ver2_0_info
.cred_sz
);
1602 target
->tgt_creds
= le16_to_cpu(ready_msg
->ver2_0_info
.cred_cnt
);
1603 target
->tgt_cred_sz
= le16_to_cpu(ready_msg
->ver2_0_info
.cred_sz
);
1605 if ((target
->tgt_creds
== 0) || (target
->tgt_cred_sz
== 0))
1608 htc_setup_target_buffer_assignments(target
);
1610 /* setup our pseudo HTC control endpoint connection */
1611 memset(&connect
, 0, sizeof(connect
));
1612 memset(&resp
, 0, sizeof(resp
));
1613 connect
.ep_cb
.tx_complete
= htc_txctrl_complete
;
1614 connect
.ep_cb
.rx
= htc_rxctrl_complete
;
1615 connect
.max_txq_depth
= NUM_CONTROL_TX_BUFFERS
;
1616 connect
.svc_id
= HTC_CTRL_RSVD_SVC
;
1618 /* connect fake service */
1619 status
= ath6kl_htc_pipe_conn_service(target
, &connect
, &resp
);
1624 static void ath6kl_htc_pipe_flush_txep(struct htc_target
*target
,
1625 enum htc_endpoint_id endpoint
, u16 tag
)
1627 struct htc_endpoint
*ep
= &target
->endpoint
[endpoint
];
1629 if (ep
->svc_id
== 0) {
1635 htc_flush_tx_endpoint(target
, ep
, tag
);
1638 static int ath6kl_htc_pipe_add_rxbuf_multiple(struct htc_target
*target
,
1639 struct list_head
*pkt_queue
)
1641 struct htc_packet
*packet
, *tmp_pkt
, *first
;
1642 struct htc_endpoint
*ep
;
1645 if (list_empty(pkt_queue
))
1648 first
= list_first_entry(pkt_queue
, struct htc_packet
, list
);
1650 if (first
->endpoint
>= ENDPOINT_MAX
) {
1655 ath6kl_dbg(ATH6KL_DBG_HTC
, "%s: epid: %d, cnt:%d, len: %d\n",
1656 __func__
, first
->endpoint
, get_queue_depth(pkt_queue
),
1659 ep
= &target
->endpoint
[first
->endpoint
];
1661 spin_lock_bh(&target
->rx_lock
);
1663 /* store receive packets */
1664 list_splice_tail_init(pkt_queue
, &ep
->rx_bufq
);
1666 spin_unlock_bh(&target
->rx_lock
);
1669 /* walk through queue and mark each one canceled */
1670 list_for_each_entry_safe(packet
, tmp_pkt
, pkt_queue
, list
) {
1671 packet
->status
= -ECANCELED
;
1674 do_recv_completion(ep
, pkt_queue
);
1680 static void ath6kl_htc_pipe_activity_changed(struct htc_target
*target
,
1681 enum htc_endpoint_id ep
,
1687 static void ath6kl_htc_pipe_flush_rx_buf(struct htc_target
*target
)
1692 static int ath6kl_htc_pipe_credit_setup(struct htc_target
*target
,
1693 struct ath6kl_htc_credit_info
*info
)
1698 static const struct ath6kl_htc_ops ath6kl_htc_pipe_ops
= {
1699 .create
= ath6kl_htc_pipe_create
,
1700 .wait_target
= ath6kl_htc_pipe_wait_target
,
1701 .start
= ath6kl_htc_pipe_start
,
1702 .conn_service
= ath6kl_htc_pipe_conn_service
,
1703 .tx
= ath6kl_htc_pipe_tx
,
1704 .stop
= ath6kl_htc_pipe_stop
,
1705 .cleanup
= ath6kl_htc_pipe_cleanup
,
1706 .flush_txep
= ath6kl_htc_pipe_flush_txep
,
1707 .flush_rx_buf
= ath6kl_htc_pipe_flush_rx_buf
,
1708 .activity_changed
= ath6kl_htc_pipe_activity_changed
,
1709 .get_rxbuf_num
= ath6kl_htc_pipe_get_rxbuf_num
,
1710 .add_rxbuf_multiple
= ath6kl_htc_pipe_add_rxbuf_multiple
,
1711 .credit_setup
= ath6kl_htc_pipe_credit_setup
,
1712 .tx_complete
= ath6kl_htc_pipe_tx_complete
,
1713 .rx_complete
= ath6kl_htc_pipe_rx_complete
,
1716 void ath6kl_htc_pipe_attach(struct ath6kl
*ar
)
1718 ar
->htc_ops
= &ath6kl_htc_pipe_ops
;