2 * Copyright (c) 2005-2011 Atheros Communications Inc.
3 * Copyright (c) 2011-2013 Qualcomm Atheros, Inc.
5 * Permission to use, copy, modify, and/or distribute this software for any
6 * purpose with or without fee is hereby granted, provided that the above
7 * copyright notice and this permission notice appear in all copies.
9 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
10 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
11 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
12 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
13 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
14 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
15 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
24 static void ath10k_report_offchan_tx(struct ath10k
*ar
, struct sk_buff
*skb
)
26 if (!ATH10K_SKB_CB(skb
)->htt
.is_offchan
)
29 /* If the original wait_for_completion() timed out before
30 * {data,mgmt}_tx_completed() was called then we could complete
31 * offchan_tx_completed for a different skb. Prevent this by using
33 spin_lock_bh(&ar
->data_lock
);
34 if (ar
->offchan_tx_skb
!= skb
) {
35 ath10k_warn(ar
, "completed old offchannel frame\n");
39 complete(&ar
->offchan_tx_completed
);
40 ar
->offchan_tx_skb
= NULL
; /* just for sanity */
42 ath10k_dbg(ar
, ATH10K_DBG_HTT
, "completed offchannel skb %p\n", skb
);
44 spin_unlock_bh(&ar
->data_lock
);
47 void ath10k_txrx_tx_unref(struct ath10k_htt
*htt
,
48 const struct htt_tx_done
*tx_done
)
50 struct ath10k
*ar
= htt
->ar
;
51 struct device
*dev
= ar
->dev
;
52 struct ieee80211_tx_info
*info
;
53 struct ath10k_skb_cb
*skb_cb
;
55 struct ieee80211_hdr
*hdr
;
57 bool limit_mgmt_desc
= false;
59 ath10k_dbg(ar
, ATH10K_DBG_HTT
,
60 "htt tx completion msdu_id %u discard %d no_ack %d success %d\n",
61 tx_done
->msdu_id
, !!tx_done
->discard
,
62 !!tx_done
->no_ack
, !!tx_done
->success
);
64 if (tx_done
->msdu_id
>= htt
->max_num_pending_tx
) {
65 ath10k_warn(ar
, "warning: msdu_id %d too big, ignoring\n",
70 spin_lock_bh(&htt
->tx_lock
);
71 msdu
= idr_find(&htt
->pending_tx
, tx_done
->msdu_id
);
73 ath10k_warn(ar
, "received tx completion for invalid msdu_id: %d\n",
75 spin_unlock_bh(&htt
->tx_lock
);
79 hdr
= (struct ieee80211_hdr
*)msdu
->data
;
80 fc
= hdr
->frame_control
;
82 if (unlikely(ieee80211_is_mgmt(fc
)) &&
83 ar
->hw_params
.max_probe_resp_desc_thres
)
84 limit_mgmt_desc
= true;
86 ath10k_htt_tx_free_msdu_id(htt
, tx_done
->msdu_id
);
87 __ath10k_htt_tx_dec_pending(htt
, limit_mgmt_desc
);
88 if (htt
->num_pending_tx
== 0)
89 wake_up(&htt
->empty_tx_wq
);
90 spin_unlock_bh(&htt
->tx_lock
);
92 skb_cb
= ATH10K_SKB_CB(msdu
);
93 dma_unmap_single(dev
, skb_cb
->paddr
, msdu
->len
, DMA_TO_DEVICE
);
95 ath10k_report_offchan_tx(htt
->ar
, msdu
);
97 info
= IEEE80211_SKB_CB(msdu
);
98 memset(&info
->status
, 0, sizeof(info
->status
));
99 trace_ath10k_txrx_tx_unref(ar
, tx_done
->msdu_id
);
101 if (tx_done
->discard
) {
102 ieee80211_free_txskb(htt
->ar
->hw
, msdu
);
106 if (!(info
->flags
& IEEE80211_TX_CTL_NO_ACK
))
107 info
->flags
|= IEEE80211_TX_STAT_ACK
;
110 info
->flags
&= ~IEEE80211_TX_STAT_ACK
;
112 if (tx_done
->success
&& (info
->flags
& IEEE80211_TX_CTL_NO_ACK
))
113 info
->flags
|= IEEE80211_TX_STAT_NOACK_TRANSMITTED
;
115 ieee80211_tx_status(htt
->ar
->hw
, msdu
);
116 /* we do not own the msdu anymore */
119 struct ath10k_peer
*ath10k_peer_find(struct ath10k
*ar
, int vdev_id
,
122 struct ath10k_peer
*peer
;
124 lockdep_assert_held(&ar
->data_lock
);
126 list_for_each_entry(peer
, &ar
->peers
, list
) {
127 if (peer
->vdev_id
!= vdev_id
)
129 if (memcmp(peer
->addr
, addr
, ETH_ALEN
))
138 struct ath10k_peer
*ath10k_peer_find_by_id(struct ath10k
*ar
, int peer_id
)
140 struct ath10k_peer
*peer
;
142 lockdep_assert_held(&ar
->data_lock
);
144 list_for_each_entry(peer
, &ar
->peers
, list
)
145 if (test_bit(peer_id
, peer
->peer_ids
))
151 static int ath10k_wait_for_peer_common(struct ath10k
*ar
, int vdev_id
,
152 const u8
*addr
, bool expect_mapped
)
156 time_left
= wait_event_timeout(ar
->peer_mapping_wq
, ({
159 spin_lock_bh(&ar
->data_lock
);
160 mapped
= !!ath10k_peer_find(ar
, vdev_id
, addr
);
161 spin_unlock_bh(&ar
->data_lock
);
163 (mapped
== expect_mapped
||
164 test_bit(ATH10K_FLAG_CRASH_FLUSH
, &ar
->dev_flags
));
173 int ath10k_wait_for_peer_created(struct ath10k
*ar
, int vdev_id
, const u8
*addr
)
175 return ath10k_wait_for_peer_common(ar
, vdev_id
, addr
, true);
178 int ath10k_wait_for_peer_deleted(struct ath10k
*ar
, int vdev_id
, const u8
*addr
)
180 return ath10k_wait_for_peer_common(ar
, vdev_id
, addr
, false);
183 void ath10k_peer_map_event(struct ath10k_htt
*htt
,
184 struct htt_peer_map_event
*ev
)
186 struct ath10k
*ar
= htt
->ar
;
187 struct ath10k_peer
*peer
;
189 spin_lock_bh(&ar
->data_lock
);
190 peer
= ath10k_peer_find(ar
, ev
->vdev_id
, ev
->addr
);
192 peer
= kzalloc(sizeof(*peer
), GFP_ATOMIC
);
196 peer
->vdev_id
= ev
->vdev_id
;
197 ether_addr_copy(peer
->addr
, ev
->addr
);
198 list_add(&peer
->list
, &ar
->peers
);
199 wake_up(&ar
->peer_mapping_wq
);
202 ath10k_dbg(ar
, ATH10K_DBG_HTT
, "htt peer map vdev %d peer %pM id %d\n",
203 ev
->vdev_id
, ev
->addr
, ev
->peer_id
);
205 set_bit(ev
->peer_id
, peer
->peer_ids
);
207 spin_unlock_bh(&ar
->data_lock
);
210 void ath10k_peer_unmap_event(struct ath10k_htt
*htt
,
211 struct htt_peer_unmap_event
*ev
)
213 struct ath10k
*ar
= htt
->ar
;
214 struct ath10k_peer
*peer
;
216 spin_lock_bh(&ar
->data_lock
);
217 peer
= ath10k_peer_find_by_id(ar
, ev
->peer_id
);
219 ath10k_warn(ar
, "peer-unmap-event: unknown peer id %d\n",
224 ath10k_dbg(ar
, ATH10K_DBG_HTT
, "htt peer unmap vdev %d peer %pM id %d\n",
225 peer
->vdev_id
, peer
->addr
, ev
->peer_id
);
227 clear_bit(ev
->peer_id
, peer
->peer_ids
);
229 if (bitmap_empty(peer
->peer_ids
, ATH10K_MAX_NUM_PEER_IDS
)) {
230 list_del(&peer
->list
);
232 wake_up(&ar
->peer_mapping_wq
);
236 spin_unlock_bh(&ar
->data_lock
);