ARM: mm: Recreate kernel mappings in early_paging_init()
[linux/fpc-iii.git] / drivers / net / wireless / iwlwifi / dvm / lib.c
blob3d5bdc4217a8901a3c4fb46ad6ffc091809bf9b4
1 /******************************************************************************
3 * GPL LICENSE SUMMARY
5 * Copyright(c) 2008 - 2013 Intel Corporation. All rights reserved.
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of version 2 of the GNU General Public License as
9 * published by the Free Software Foundation.
11 * This program is distributed in the hope that it will be useful, but
12 * WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
19 * USA
21 * The full GNU General Public License is included in this distribution
22 * in the file called COPYING.
24 * Contact Information:
25 * Intel Linux Wireless <ilw@linux.intel.com>
26 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
28 *****************************************************************************/
29 #include <linux/etherdevice.h>
30 #include <linux/kernel.h>
31 #include <linux/module.h>
32 #include <linux/init.h>
33 #include <linux/sched.h>
34 #include <net/mac80211.h>
36 #include "iwl-io.h"
37 #include "iwl-agn-hw.h"
38 #include "iwl-trans.h"
39 #include "iwl-modparams.h"
41 #include "dev.h"
42 #include "agn.h"
44 int iwlagn_hw_valid_rtc_data_addr(u32 addr)
46 return (addr >= IWLAGN_RTC_DATA_LOWER_BOUND) &&
47 (addr < IWLAGN_RTC_DATA_UPPER_BOUND);
50 int iwlagn_send_tx_power(struct iwl_priv *priv)
52 struct iwlagn_tx_power_dbm_cmd tx_power_cmd;
53 u8 tx_ant_cfg_cmd;
55 if (WARN_ONCE(test_bit(STATUS_SCAN_HW, &priv->status),
56 "TX Power requested while scanning!\n"))
57 return -EAGAIN;
59 /* half dBm need to multiply */
60 tx_power_cmd.global_lmt = (s8)(2 * priv->tx_power_user_lmt);
62 if (tx_power_cmd.global_lmt > priv->nvm_data->max_tx_pwr_half_dbm) {
64 * For the newer devices which using enhanced/extend tx power
65 * table in EEPROM, the format is in half dBm. driver need to
66 * convert to dBm format before report to mac80211.
67 * By doing so, there is a possibility of 1/2 dBm resolution
68 * lost. driver will perform "round-up" operation before
69 * reporting, but it will cause 1/2 dBm tx power over the
70 * regulatory limit. Perform the checking here, if the
71 * "tx_power_user_lmt" is higher than EEPROM value (in
72 * half-dBm format), lower the tx power based on EEPROM
74 tx_power_cmd.global_lmt =
75 priv->nvm_data->max_tx_pwr_half_dbm;
77 tx_power_cmd.flags = IWLAGN_TX_POWER_NO_CLOSED;
78 tx_power_cmd.srv_chan_lmt = IWLAGN_TX_POWER_AUTO;
80 if (IWL_UCODE_API(priv->fw->ucode_ver) == 1)
81 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD_V1;
82 else
83 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD;
85 return iwl_dvm_send_cmd_pdu(priv, tx_ant_cfg_cmd, CMD_SYNC,
86 sizeof(tx_power_cmd), &tx_power_cmd);
89 void iwlagn_temperature(struct iwl_priv *priv)
91 lockdep_assert_held(&priv->statistics.lock);
93 /* store temperature from correct statistics (in Celsius) */
94 priv->temperature = le32_to_cpu(priv->statistics.common.temperature);
95 iwl_tt_handler(priv);
98 int iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags, enum ieee80211_band band)
100 int idx = 0;
101 int band_offset = 0;
103 /* HT rate format: mac80211 wants an MCS number, which is just LSB */
104 if (rate_n_flags & RATE_MCS_HT_MSK) {
105 idx = (rate_n_flags & 0xff);
106 return idx;
107 /* Legacy rate format, search for match in table */
108 } else {
109 if (band == IEEE80211_BAND_5GHZ)
110 band_offset = IWL_FIRST_OFDM_RATE;
111 for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++)
112 if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF))
113 return idx - band_offset;
116 return -1;
119 int iwlagn_manage_ibss_station(struct iwl_priv *priv,
120 struct ieee80211_vif *vif, bool add)
122 struct iwl_vif_priv *vif_priv = (void *)vif->drv_priv;
124 if (add)
125 return iwlagn_add_bssid_station(priv, vif_priv->ctx,
126 vif->bss_conf.bssid,
127 &vif_priv->ibss_bssid_sta_id);
128 return iwl_remove_station(priv, vif_priv->ibss_bssid_sta_id,
129 vif->bss_conf.bssid);
133 * iwlagn_txfifo_flush: send REPLY_TXFIFO_FLUSH command to uCode
135 * pre-requirements:
136 * 1. acquire mutex before calling
137 * 2. make sure rf is on and not in exit state
139 int iwlagn_txfifo_flush(struct iwl_priv *priv, u32 scd_q_msk)
141 struct iwl_txfifo_flush_cmd flush_cmd;
142 struct iwl_host_cmd cmd = {
143 .id = REPLY_TXFIFO_FLUSH,
144 .len = { sizeof(struct iwl_txfifo_flush_cmd), },
145 .flags = CMD_SYNC,
146 .data = { &flush_cmd, },
149 memset(&flush_cmd, 0, sizeof(flush_cmd));
151 flush_cmd.queue_control = IWL_SCD_VO_MSK | IWL_SCD_VI_MSK |
152 IWL_SCD_BE_MSK | IWL_SCD_BK_MSK |
153 IWL_SCD_MGMT_MSK;
154 if ((priv->valid_contexts != BIT(IWL_RXON_CTX_BSS)))
155 flush_cmd.queue_control |= IWL_PAN_SCD_VO_MSK |
156 IWL_PAN_SCD_VI_MSK |
157 IWL_PAN_SCD_BE_MSK |
158 IWL_PAN_SCD_BK_MSK |
159 IWL_PAN_SCD_MGMT_MSK |
160 IWL_PAN_SCD_MULTICAST_MSK;
162 if (priv->nvm_data->sku_cap_11n_enable)
163 flush_cmd.queue_control |= IWL_AGG_TX_QUEUE_MSK;
165 if (scd_q_msk)
166 flush_cmd.queue_control = cpu_to_le32(scd_q_msk);
168 IWL_DEBUG_INFO(priv, "queue control: 0x%x\n",
169 flush_cmd.queue_control);
170 flush_cmd.flush_control = cpu_to_le16(IWL_DROP_ALL);
172 return iwl_dvm_send_cmd(priv, &cmd);
175 void iwlagn_dev_txfifo_flush(struct iwl_priv *priv)
177 mutex_lock(&priv->mutex);
178 ieee80211_stop_queues(priv->hw);
179 if (iwlagn_txfifo_flush(priv, 0)) {
180 IWL_ERR(priv, "flush request fail\n");
181 goto done;
183 IWL_DEBUG_INFO(priv, "wait transmit/flush all frames\n");
184 iwl_trans_wait_tx_queue_empty(priv->trans);
185 done:
186 ieee80211_wake_queues(priv->hw);
187 mutex_unlock(&priv->mutex);
191 * BT coex
193 /* Notmal TDM */
194 static const __le32 iwlagn_def_3w_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
195 cpu_to_le32(0xaaaaaaaa),
196 cpu_to_le32(0xaaaaaaaa),
197 cpu_to_le32(0xaeaaaaaa),
198 cpu_to_le32(0xaaaaaaaa),
199 cpu_to_le32(0xcc00ff28),
200 cpu_to_le32(0x0000aaaa),
201 cpu_to_le32(0xcc00aaaa),
202 cpu_to_le32(0x0000aaaa),
203 cpu_to_le32(0xc0004000),
204 cpu_to_le32(0x00004000),
205 cpu_to_le32(0xf0005000),
206 cpu_to_le32(0xf0005000),
210 /* Loose Coex */
211 static const __le32 iwlagn_loose_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
212 cpu_to_le32(0xaaaaaaaa),
213 cpu_to_le32(0xaaaaaaaa),
214 cpu_to_le32(0xaeaaaaaa),
215 cpu_to_le32(0xaaaaaaaa),
216 cpu_to_le32(0xcc00ff28),
217 cpu_to_le32(0x0000aaaa),
218 cpu_to_le32(0xcc00aaaa),
219 cpu_to_le32(0x0000aaaa),
220 cpu_to_le32(0x00000000),
221 cpu_to_le32(0x00000000),
222 cpu_to_le32(0xf0005000),
223 cpu_to_le32(0xf0005000),
226 /* Full concurrency */
227 static const __le32 iwlagn_concurrent_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
228 cpu_to_le32(0xaaaaaaaa),
229 cpu_to_le32(0xaaaaaaaa),
230 cpu_to_le32(0xaaaaaaaa),
231 cpu_to_le32(0xaaaaaaaa),
232 cpu_to_le32(0xaaaaaaaa),
233 cpu_to_le32(0xaaaaaaaa),
234 cpu_to_le32(0xaaaaaaaa),
235 cpu_to_le32(0xaaaaaaaa),
236 cpu_to_le32(0x00000000),
237 cpu_to_le32(0x00000000),
238 cpu_to_le32(0x00000000),
239 cpu_to_le32(0x00000000),
242 void iwlagn_send_advance_bt_config(struct iwl_priv *priv)
244 struct iwl_basic_bt_cmd basic = {
245 .max_kill = IWLAGN_BT_MAX_KILL_DEFAULT,
246 .bt3_timer_t7_value = IWLAGN_BT3_T7_DEFAULT,
247 .bt3_prio_sample_time = IWLAGN_BT3_PRIO_SAMPLE_DEFAULT,
248 .bt3_timer_t2_value = IWLAGN_BT3_T2_DEFAULT,
250 struct iwl_bt_cmd_v1 bt_cmd_v1;
251 struct iwl_bt_cmd_v2 bt_cmd_v2;
252 int ret;
254 BUILD_BUG_ON(sizeof(iwlagn_def_3w_lookup) !=
255 sizeof(basic.bt3_lookup_table));
257 if (priv->lib->bt_params) {
259 * newer generation of devices (2000 series and newer)
260 * use the version 2 of the bt command
261 * we need to make sure sending the host command
262 * with correct data structure to avoid uCode assert
264 if (priv->lib->bt_params->bt_session_2) {
265 bt_cmd_v2.prio_boost = cpu_to_le32(
266 priv->lib->bt_params->bt_prio_boost);
267 bt_cmd_v2.tx_prio_boost = 0;
268 bt_cmd_v2.rx_prio_boost = 0;
269 } else {
270 /* older version only has 8 bits */
271 WARN_ON(priv->lib->bt_params->bt_prio_boost & ~0xFF);
272 bt_cmd_v1.prio_boost =
273 priv->lib->bt_params->bt_prio_boost;
274 bt_cmd_v1.tx_prio_boost = 0;
275 bt_cmd_v1.rx_prio_boost = 0;
277 } else {
278 IWL_ERR(priv, "failed to construct BT Coex Config\n");
279 return;
283 * Possible situations when BT needs to take over for receive,
284 * at the same time where STA needs to response to AP's frame(s),
285 * reduce the tx power of the required response frames, by that,
286 * allow the concurrent BT receive & WiFi transmit
287 * (BT - ANT A, WiFi -ANT B), without interference to one another
289 * Reduced tx power apply to control frames only (ACK/Back/CTS)
290 * when indicated by the BT config command
292 basic.kill_ack_mask = priv->kill_ack_mask;
293 basic.kill_cts_mask = priv->kill_cts_mask;
294 if (priv->reduced_txpower)
295 basic.reduce_txpower = IWLAGN_BT_REDUCED_TX_PWR;
296 basic.valid = priv->bt_valid;
299 * Configure BT coex mode to "no coexistence" when the
300 * user disabled BT coexistence, we have no interface
301 * (might be in monitor mode), or the interface is in
302 * IBSS mode (no proper uCode support for coex then).
304 if (!iwlwifi_mod_params.bt_coex_active ||
305 priv->iw_mode == NL80211_IFTYPE_ADHOC) {
306 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_DISABLED;
307 } else {
308 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_3W <<
309 IWLAGN_BT_FLAG_COEX_MODE_SHIFT;
311 if (!priv->bt_enable_pspoll)
312 basic.flags |= IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
313 else
314 basic.flags &= ~IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
316 if (priv->bt_ch_announce)
317 basic.flags |= IWLAGN_BT_FLAG_CHANNEL_INHIBITION;
318 IWL_DEBUG_COEX(priv, "BT coex flag: 0X%x\n", basic.flags);
320 priv->bt_enable_flag = basic.flags;
321 if (priv->bt_full_concurrent)
322 memcpy(basic.bt3_lookup_table, iwlagn_concurrent_lookup,
323 sizeof(iwlagn_concurrent_lookup));
324 else
325 memcpy(basic.bt3_lookup_table, iwlagn_def_3w_lookup,
326 sizeof(iwlagn_def_3w_lookup));
328 IWL_DEBUG_COEX(priv, "BT coex %s in %s mode\n",
329 basic.flags ? "active" : "disabled",
330 priv->bt_full_concurrent ?
331 "full concurrency" : "3-wire");
333 if (priv->lib->bt_params->bt_session_2) {
334 memcpy(&bt_cmd_v2.basic, &basic,
335 sizeof(basic));
336 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
337 CMD_SYNC, sizeof(bt_cmd_v2), &bt_cmd_v2);
338 } else {
339 memcpy(&bt_cmd_v1.basic, &basic,
340 sizeof(basic));
341 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
342 CMD_SYNC, sizeof(bt_cmd_v1), &bt_cmd_v1);
344 if (ret)
345 IWL_ERR(priv, "failed to send BT Coex Config\n");
349 void iwlagn_bt_adjust_rssi_monitor(struct iwl_priv *priv, bool rssi_ena)
351 struct iwl_rxon_context *ctx, *found_ctx = NULL;
352 bool found_ap = false;
354 lockdep_assert_held(&priv->mutex);
356 /* Check whether AP or GO mode is active. */
357 if (rssi_ena) {
358 for_each_context(priv, ctx) {
359 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_AP &&
360 iwl_is_associated_ctx(ctx)) {
361 found_ap = true;
362 break;
368 * If disable was received or If GO/AP mode, disable RSSI
369 * measurements.
371 if (!rssi_ena || found_ap) {
372 if (priv->cur_rssi_ctx) {
373 ctx = priv->cur_rssi_ctx;
374 ieee80211_disable_rssi_reports(ctx->vif);
375 priv->cur_rssi_ctx = NULL;
377 return;
381 * If rssi measurements need to be enabled, consider all cases now.
382 * Figure out how many contexts are active.
384 for_each_context(priv, ctx) {
385 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION &&
386 iwl_is_associated_ctx(ctx)) {
387 found_ctx = ctx;
388 break;
393 * rssi monitor already enabled for the correct interface...nothing
394 * to do.
396 if (found_ctx == priv->cur_rssi_ctx)
397 return;
400 * Figure out if rssi monitor is currently enabled, and needs
401 * to be changed. If rssi monitor is already enabled, disable
402 * it first else just enable rssi measurements on the
403 * interface found above.
405 if (priv->cur_rssi_ctx) {
406 ctx = priv->cur_rssi_ctx;
407 if (ctx->vif)
408 ieee80211_disable_rssi_reports(ctx->vif);
411 priv->cur_rssi_ctx = found_ctx;
413 if (!found_ctx)
414 return;
416 ieee80211_enable_rssi_reports(found_ctx->vif,
417 IWLAGN_BT_PSP_MIN_RSSI_THRESHOLD,
418 IWLAGN_BT_PSP_MAX_RSSI_THRESHOLD);
421 static bool iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg *uart_msg)
423 return BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3 >>
424 BT_UART_MSG_FRAME3SCOESCO_POS;
427 static void iwlagn_bt_traffic_change_work(struct work_struct *work)
429 struct iwl_priv *priv =
430 container_of(work, struct iwl_priv, bt_traffic_change_work);
431 struct iwl_rxon_context *ctx;
432 int smps_request = -1;
434 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
435 /* bt coex disabled */
436 return;
440 * Note: bt_traffic_load can be overridden by scan complete and
441 * coex profile notifications. Ignore that since only bad consequence
442 * can be not matching debug print with actual state.
444 IWL_DEBUG_COEX(priv, "BT traffic load changes: %d\n",
445 priv->bt_traffic_load);
447 switch (priv->bt_traffic_load) {
448 case IWL_BT_COEX_TRAFFIC_LOAD_NONE:
449 if (priv->bt_status)
450 smps_request = IEEE80211_SMPS_DYNAMIC;
451 else
452 smps_request = IEEE80211_SMPS_AUTOMATIC;
453 break;
454 case IWL_BT_COEX_TRAFFIC_LOAD_LOW:
455 smps_request = IEEE80211_SMPS_DYNAMIC;
456 break;
457 case IWL_BT_COEX_TRAFFIC_LOAD_HIGH:
458 case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS:
459 smps_request = IEEE80211_SMPS_STATIC;
460 break;
461 default:
462 IWL_ERR(priv, "Invalid BT traffic load: %d\n",
463 priv->bt_traffic_load);
464 break;
467 mutex_lock(&priv->mutex);
470 * We can not send command to firmware while scanning. When the scan
471 * complete we will schedule this work again. We do check with mutex
472 * locked to prevent new scan request to arrive. We do not check
473 * STATUS_SCANNING to avoid race when queue_work two times from
474 * different notifications, but quit and not perform any work at all.
476 if (test_bit(STATUS_SCAN_HW, &priv->status))
477 goto out;
479 iwl_update_chain_flags(priv);
481 if (smps_request != -1) {
482 priv->current_ht_config.smps = smps_request;
483 for_each_context(priv, ctx) {
484 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION)
485 ieee80211_request_smps(ctx->vif, smps_request);
490 * Dynamic PS poll related functionality. Adjust RSSI measurements if
491 * necessary.
493 iwlagn_bt_coex_rssi_monitor(priv);
494 out:
495 mutex_unlock(&priv->mutex);
499 * If BT sco traffic, and RSSI monitor is enabled, move measurements to the
500 * correct interface or disable it if this is the last interface to be
501 * removed.
503 void iwlagn_bt_coex_rssi_monitor(struct iwl_priv *priv)
505 if (priv->bt_is_sco &&
506 priv->bt_traffic_load == IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS)
507 iwlagn_bt_adjust_rssi_monitor(priv, true);
508 else
509 iwlagn_bt_adjust_rssi_monitor(priv, false);
512 static void iwlagn_print_uartmsg(struct iwl_priv *priv,
513 struct iwl_bt_uart_msg *uart_msg)
515 IWL_DEBUG_COEX(priv, "Message Type = 0x%X, SSN = 0x%X, "
516 "Update Req = 0x%X\n",
517 (BT_UART_MSG_FRAME1MSGTYPE_MSK & uart_msg->frame1) >>
518 BT_UART_MSG_FRAME1MSGTYPE_POS,
519 (BT_UART_MSG_FRAME1SSN_MSK & uart_msg->frame1) >>
520 BT_UART_MSG_FRAME1SSN_POS,
521 (BT_UART_MSG_FRAME1UPDATEREQ_MSK & uart_msg->frame1) >>
522 BT_UART_MSG_FRAME1UPDATEREQ_POS);
524 IWL_DEBUG_COEX(priv, "Open connections = 0x%X, Traffic load = 0x%X, "
525 "Chl_SeqN = 0x%X, In band = 0x%X\n",
526 (BT_UART_MSG_FRAME2OPENCONNECTIONS_MSK & uart_msg->frame2) >>
527 BT_UART_MSG_FRAME2OPENCONNECTIONS_POS,
528 (BT_UART_MSG_FRAME2TRAFFICLOAD_MSK & uart_msg->frame2) >>
529 BT_UART_MSG_FRAME2TRAFFICLOAD_POS,
530 (BT_UART_MSG_FRAME2CHLSEQN_MSK & uart_msg->frame2) >>
531 BT_UART_MSG_FRAME2CHLSEQN_POS,
532 (BT_UART_MSG_FRAME2INBAND_MSK & uart_msg->frame2) >>
533 BT_UART_MSG_FRAME2INBAND_POS);
535 IWL_DEBUG_COEX(priv, "SCO/eSCO = 0x%X, Sniff = 0x%X, A2DP = 0x%X, "
536 "ACL = 0x%X, Master = 0x%X, OBEX = 0x%X\n",
537 (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
538 BT_UART_MSG_FRAME3SCOESCO_POS,
539 (BT_UART_MSG_FRAME3SNIFF_MSK & uart_msg->frame3) >>
540 BT_UART_MSG_FRAME3SNIFF_POS,
541 (BT_UART_MSG_FRAME3A2DP_MSK & uart_msg->frame3) >>
542 BT_UART_MSG_FRAME3A2DP_POS,
543 (BT_UART_MSG_FRAME3ACL_MSK & uart_msg->frame3) >>
544 BT_UART_MSG_FRAME3ACL_POS,
545 (BT_UART_MSG_FRAME3MASTER_MSK & uart_msg->frame3) >>
546 BT_UART_MSG_FRAME3MASTER_POS,
547 (BT_UART_MSG_FRAME3OBEX_MSK & uart_msg->frame3) >>
548 BT_UART_MSG_FRAME3OBEX_POS);
550 IWL_DEBUG_COEX(priv, "Idle duration = 0x%X\n",
551 (BT_UART_MSG_FRAME4IDLEDURATION_MSK & uart_msg->frame4) >>
552 BT_UART_MSG_FRAME4IDLEDURATION_POS);
554 IWL_DEBUG_COEX(priv, "Tx Activity = 0x%X, Rx Activity = 0x%X, "
555 "eSCO Retransmissions = 0x%X\n",
556 (BT_UART_MSG_FRAME5TXACTIVITY_MSK & uart_msg->frame5) >>
557 BT_UART_MSG_FRAME5TXACTIVITY_POS,
558 (BT_UART_MSG_FRAME5RXACTIVITY_MSK & uart_msg->frame5) >>
559 BT_UART_MSG_FRAME5RXACTIVITY_POS,
560 (BT_UART_MSG_FRAME5ESCORETRANSMIT_MSK & uart_msg->frame5) >>
561 BT_UART_MSG_FRAME5ESCORETRANSMIT_POS);
563 IWL_DEBUG_COEX(priv, "Sniff Interval = 0x%X, Discoverable = 0x%X\n",
564 (BT_UART_MSG_FRAME6SNIFFINTERVAL_MSK & uart_msg->frame6) >>
565 BT_UART_MSG_FRAME6SNIFFINTERVAL_POS,
566 (BT_UART_MSG_FRAME6DISCOVERABLE_MSK & uart_msg->frame6) >>
567 BT_UART_MSG_FRAME6DISCOVERABLE_POS);
569 IWL_DEBUG_COEX(priv, "Sniff Activity = 0x%X, Page = "
570 "0x%X, Inquiry = 0x%X, Connectable = 0x%X\n",
571 (BT_UART_MSG_FRAME7SNIFFACTIVITY_MSK & uart_msg->frame7) >>
572 BT_UART_MSG_FRAME7SNIFFACTIVITY_POS,
573 (BT_UART_MSG_FRAME7PAGE_MSK & uart_msg->frame7) >>
574 BT_UART_MSG_FRAME7PAGE_POS,
575 (BT_UART_MSG_FRAME7INQUIRY_MSK & uart_msg->frame7) >>
576 BT_UART_MSG_FRAME7INQUIRY_POS,
577 (BT_UART_MSG_FRAME7CONNECTABLE_MSK & uart_msg->frame7) >>
578 BT_UART_MSG_FRAME7CONNECTABLE_POS);
581 static bool iwlagn_set_kill_msk(struct iwl_priv *priv,
582 struct iwl_bt_uart_msg *uart_msg)
584 bool need_update = false;
585 u8 kill_msk = IWL_BT_KILL_REDUCE;
586 static const __le32 bt_kill_ack_msg[3] = {
587 IWLAGN_BT_KILL_ACK_MASK_DEFAULT,
588 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
589 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
590 static const __le32 bt_kill_cts_msg[3] = {
591 IWLAGN_BT_KILL_CTS_MASK_DEFAULT,
592 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
593 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
595 if (!priv->reduced_txpower)
596 kill_msk = (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3)
597 ? IWL_BT_KILL_OVERRIDE : IWL_BT_KILL_DEFAULT;
598 if (priv->kill_ack_mask != bt_kill_ack_msg[kill_msk] ||
599 priv->kill_cts_mask != bt_kill_cts_msg[kill_msk]) {
600 priv->bt_valid |= IWLAGN_BT_VALID_KILL_ACK_MASK;
601 priv->kill_ack_mask = bt_kill_ack_msg[kill_msk];
602 priv->bt_valid |= IWLAGN_BT_VALID_KILL_CTS_MASK;
603 priv->kill_cts_mask = bt_kill_cts_msg[kill_msk];
604 need_update = true;
606 return need_update;
610 * Upon RSSI changes, sends a bt config command with following changes
611 * 1. enable/disable "reduced control frames tx power
612 * 2. update the "kill)ack_mask" and "kill_cts_mask"
614 * If "reduced tx power" is enabled, uCode shall
615 * 1. ACK/Back/CTS rate shall reduced to 6Mbps
616 * 2. not use duplciate 20/40MHz mode
618 static bool iwlagn_fill_txpower_mode(struct iwl_priv *priv,
619 struct iwl_bt_uart_msg *uart_msg)
621 bool need_update = false;
622 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
623 int ave_rssi;
625 if (!ctx->vif || (ctx->vif->type != NL80211_IFTYPE_STATION)) {
626 IWL_DEBUG_INFO(priv, "BSS ctx not active or not in sta mode\n");
627 return false;
630 ave_rssi = ieee80211_ave_rssi(ctx->vif);
631 if (!ave_rssi) {
632 /* no rssi data, no changes to reduce tx power */
633 IWL_DEBUG_COEX(priv, "no rssi data available\n");
634 return need_update;
636 if (!priv->reduced_txpower &&
637 !iwl_is_associated(priv, IWL_RXON_CTX_PAN) &&
638 (ave_rssi > BT_ENABLE_REDUCED_TXPOWER_THRESHOLD) &&
639 (uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
640 BT_UART_MSG_FRAME3OBEX_MSK)) &&
641 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
642 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK))) {
643 /* enabling reduced tx power */
644 priv->reduced_txpower = true;
645 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
646 need_update = true;
647 } else if (priv->reduced_txpower &&
648 (iwl_is_associated(priv, IWL_RXON_CTX_PAN) ||
649 (ave_rssi < BT_DISABLE_REDUCED_TXPOWER_THRESHOLD) ||
650 (uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
651 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK)) ||
652 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
653 BT_UART_MSG_FRAME3OBEX_MSK)))) {
654 /* disable reduced tx power */
655 priv->reduced_txpower = false;
656 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
657 need_update = true;
660 return need_update;
663 int iwlagn_bt_coex_profile_notif(struct iwl_priv *priv,
664 struct iwl_rx_cmd_buffer *rxb,
665 struct iwl_device_cmd *cmd)
667 struct iwl_rx_packet *pkt = rxb_addr(rxb);
668 struct iwl_bt_coex_profile_notif *coex = (void *)pkt->data;
669 struct iwl_bt_uart_msg *uart_msg = &coex->last_bt_uart_msg;
671 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
672 /* bt coex disabled */
673 return 0;
676 IWL_DEBUG_COEX(priv, "BT Coex notification:\n");
677 IWL_DEBUG_COEX(priv, " status: %d\n", coex->bt_status);
678 IWL_DEBUG_COEX(priv, " traffic load: %d\n", coex->bt_traffic_load);
679 IWL_DEBUG_COEX(priv, " CI compliance: %d\n",
680 coex->bt_ci_compliance);
681 iwlagn_print_uartmsg(priv, uart_msg);
683 priv->last_bt_traffic_load = priv->bt_traffic_load;
684 priv->bt_is_sco = iwlagn_bt_traffic_is_sco(uart_msg);
686 if (priv->iw_mode != NL80211_IFTYPE_ADHOC) {
687 if (priv->bt_status != coex->bt_status ||
688 priv->last_bt_traffic_load != coex->bt_traffic_load) {
689 if (coex->bt_status) {
690 /* BT on */
691 if (!priv->bt_ch_announce)
692 priv->bt_traffic_load =
693 IWL_BT_COEX_TRAFFIC_LOAD_HIGH;
694 else
695 priv->bt_traffic_load =
696 coex->bt_traffic_load;
697 } else {
698 /* BT off */
699 priv->bt_traffic_load =
700 IWL_BT_COEX_TRAFFIC_LOAD_NONE;
702 priv->bt_status = coex->bt_status;
703 queue_work(priv->workqueue,
704 &priv->bt_traffic_change_work);
708 /* schedule to send runtime bt_config */
709 /* check reduce power before change ack/cts kill mask */
710 if (iwlagn_fill_txpower_mode(priv, uart_msg) ||
711 iwlagn_set_kill_msk(priv, uart_msg))
712 queue_work(priv->workqueue, &priv->bt_runtime_config);
715 /* FIXME: based on notification, adjust the prio_boost */
717 priv->bt_ci_compliance = coex->bt_ci_compliance;
718 return 0;
721 void iwlagn_bt_rx_handler_setup(struct iwl_priv *priv)
723 priv->rx_handlers[REPLY_BT_COEX_PROFILE_NOTIF] =
724 iwlagn_bt_coex_profile_notif;
727 void iwlagn_bt_setup_deferred_work(struct iwl_priv *priv)
729 INIT_WORK(&priv->bt_traffic_change_work,
730 iwlagn_bt_traffic_change_work);
733 void iwlagn_bt_cancel_deferred_work(struct iwl_priv *priv)
735 cancel_work_sync(&priv->bt_traffic_change_work);
738 static bool is_single_rx_stream(struct iwl_priv *priv)
740 return priv->current_ht_config.smps == IEEE80211_SMPS_STATIC ||
741 priv->current_ht_config.single_chain_sufficient;
744 #define IWL_NUM_RX_CHAINS_MULTIPLE 3
745 #define IWL_NUM_RX_CHAINS_SINGLE 2
746 #define IWL_NUM_IDLE_CHAINS_DUAL 2
747 #define IWL_NUM_IDLE_CHAINS_SINGLE 1
750 * Determine how many receiver/antenna chains to use.
752 * More provides better reception via diversity. Fewer saves power
753 * at the expense of throughput, but only when not in powersave to
754 * start with.
756 * MIMO (dual stream) requires at least 2, but works better with 3.
757 * This does not determine *which* chains to use, just how many.
759 static int iwl_get_active_rx_chain_count(struct iwl_priv *priv)
761 if (priv->lib->bt_params &&
762 priv->lib->bt_params->advanced_bt_coexist &&
763 (priv->bt_full_concurrent ||
764 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
766 * only use chain 'A' in bt high traffic load or
767 * full concurrency mode
769 return IWL_NUM_RX_CHAINS_SINGLE;
771 /* # of Rx chains to use when expecting MIMO. */
772 if (is_single_rx_stream(priv))
773 return IWL_NUM_RX_CHAINS_SINGLE;
774 else
775 return IWL_NUM_RX_CHAINS_MULTIPLE;
779 * When we are in power saving mode, unless device support spatial
780 * multiplexing power save, use the active count for rx chain count.
782 static int iwl_get_idle_rx_chain_count(struct iwl_priv *priv, int active_cnt)
784 /* # Rx chains when idling, depending on SMPS mode */
785 switch (priv->current_ht_config.smps) {
786 case IEEE80211_SMPS_STATIC:
787 case IEEE80211_SMPS_DYNAMIC:
788 return IWL_NUM_IDLE_CHAINS_SINGLE;
789 case IEEE80211_SMPS_AUTOMATIC:
790 case IEEE80211_SMPS_OFF:
791 return active_cnt;
792 default:
793 WARN(1, "invalid SMPS mode %d",
794 priv->current_ht_config.smps);
795 return active_cnt;
799 /* up to 4 chains */
800 static u8 iwl_count_chain_bitmap(u32 chain_bitmap)
802 u8 res;
803 res = (chain_bitmap & BIT(0)) >> 0;
804 res += (chain_bitmap & BIT(1)) >> 1;
805 res += (chain_bitmap & BIT(2)) >> 2;
806 res += (chain_bitmap & BIT(3)) >> 3;
807 return res;
811 * iwlagn_set_rxon_chain - Set up Rx chain usage in "staging" RXON image
813 * Selects how many and which Rx receivers/antennas/chains to use.
814 * This should not be used for scan command ... it puts data in wrong place.
816 void iwlagn_set_rxon_chain(struct iwl_priv *priv, struct iwl_rxon_context *ctx)
818 bool is_single = is_single_rx_stream(priv);
819 bool is_cam = !test_bit(STATUS_POWER_PMI, &priv->status);
820 u8 idle_rx_cnt, active_rx_cnt, valid_rx_cnt;
821 u32 active_chains;
822 u16 rx_chain;
824 /* Tell uCode which antennas are actually connected.
825 * Before first association, we assume all antennas are connected.
826 * Just after first association, iwl_chain_noise_calibration()
827 * checks which antennas actually *are* connected. */
828 if (priv->chain_noise_data.active_chains)
829 active_chains = priv->chain_noise_data.active_chains;
830 else
831 active_chains = priv->nvm_data->valid_rx_ant;
833 if (priv->lib->bt_params &&
834 priv->lib->bt_params->advanced_bt_coexist &&
835 (priv->bt_full_concurrent ||
836 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
838 * only use chain 'A' in bt high traffic load or
839 * full concurrency mode
841 active_chains = first_antenna(active_chains);
844 rx_chain = active_chains << RXON_RX_CHAIN_VALID_POS;
846 /* How many receivers should we use? */
847 active_rx_cnt = iwl_get_active_rx_chain_count(priv);
848 idle_rx_cnt = iwl_get_idle_rx_chain_count(priv, active_rx_cnt);
851 /* correct rx chain count according hw settings
852 * and chain noise calibration
854 valid_rx_cnt = iwl_count_chain_bitmap(active_chains);
855 if (valid_rx_cnt < active_rx_cnt)
856 active_rx_cnt = valid_rx_cnt;
858 if (valid_rx_cnt < idle_rx_cnt)
859 idle_rx_cnt = valid_rx_cnt;
861 rx_chain |= active_rx_cnt << RXON_RX_CHAIN_MIMO_CNT_POS;
862 rx_chain |= idle_rx_cnt << RXON_RX_CHAIN_CNT_POS;
864 ctx->staging.rx_chain = cpu_to_le16(rx_chain);
866 if (!is_single && (active_rx_cnt >= IWL_NUM_RX_CHAINS_SINGLE) && is_cam)
867 ctx->staging.rx_chain |= RXON_RX_CHAIN_MIMO_FORCE_MSK;
868 else
869 ctx->staging.rx_chain &= ~RXON_RX_CHAIN_MIMO_FORCE_MSK;
871 IWL_DEBUG_ASSOC(priv, "rx_chain=0x%X active=%d idle=%d\n",
872 ctx->staging.rx_chain,
873 active_rx_cnt, idle_rx_cnt);
875 WARN_ON(active_rx_cnt == 0 || idle_rx_cnt == 0 ||
876 active_rx_cnt < idle_rx_cnt);
879 u8 iwl_toggle_tx_ant(struct iwl_priv *priv, u8 ant, u8 valid)
881 int i;
882 u8 ind = ant;
884 if (priv->band == IEEE80211_BAND_2GHZ &&
885 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)
886 return 0;
888 for (i = 0; i < RATE_ANT_NUM - 1; i++) {
889 ind = (ind + 1) < RATE_ANT_NUM ? ind + 1 : 0;
890 if (valid & BIT(ind))
891 return ind;
893 return ant;
896 #ifdef CONFIG_PM_SLEEP
897 static void iwlagn_convert_p1k(u16 *p1k, __le16 *out)
899 int i;
901 for (i = 0; i < IWLAGN_P1K_SIZE; i++)
902 out[i] = cpu_to_le16(p1k[i]);
905 struct wowlan_key_data {
906 struct iwl_rxon_context *ctx;
907 struct iwlagn_wowlan_rsc_tsc_params_cmd *rsc_tsc;
908 struct iwlagn_wowlan_tkip_params_cmd *tkip;
909 const u8 *bssid;
910 bool error, use_rsc_tsc, use_tkip;
914 static void iwlagn_wowlan_program_keys(struct ieee80211_hw *hw,
915 struct ieee80211_vif *vif,
916 struct ieee80211_sta *sta,
917 struct ieee80211_key_conf *key,
918 void *_data)
920 struct iwl_priv *priv = IWL_MAC80211_GET_DVM(hw);
921 struct wowlan_key_data *data = _data;
922 struct iwl_rxon_context *ctx = data->ctx;
923 struct aes_sc *aes_sc, *aes_tx_sc = NULL;
924 struct tkip_sc *tkip_sc, *tkip_tx_sc = NULL;
925 struct iwlagn_p1k_cache *rx_p1ks;
926 u8 *rx_mic_key;
927 struct ieee80211_key_seq seq;
928 u32 cur_rx_iv32 = 0;
929 u16 p1k[IWLAGN_P1K_SIZE];
930 int ret, i;
932 mutex_lock(&priv->mutex);
934 if ((key->cipher == WLAN_CIPHER_SUITE_WEP40 ||
935 key->cipher == WLAN_CIPHER_SUITE_WEP104) &&
936 !sta && !ctx->key_mapping_keys)
937 ret = iwl_set_default_wep_key(priv, ctx, key);
938 else
939 ret = iwl_set_dynamic_key(priv, ctx, key, sta);
941 if (ret) {
942 IWL_ERR(priv, "Error setting key during suspend!\n");
943 data->error = true;
946 switch (key->cipher) {
947 case WLAN_CIPHER_SUITE_TKIP:
948 if (sta) {
949 tkip_sc = data->rsc_tsc->all_tsc_rsc.tkip.unicast_rsc;
950 tkip_tx_sc = &data->rsc_tsc->all_tsc_rsc.tkip.tsc;
952 rx_p1ks = data->tkip->rx_uni;
954 ieee80211_get_key_tx_seq(key, &seq);
955 tkip_tx_sc->iv16 = cpu_to_le16(seq.tkip.iv16);
956 tkip_tx_sc->iv32 = cpu_to_le32(seq.tkip.iv32);
958 ieee80211_get_tkip_p1k_iv(key, seq.tkip.iv32, p1k);
959 iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
961 memcpy(data->tkip->mic_keys.tx,
962 &key->key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY],
963 IWLAGN_MIC_KEY_SIZE);
965 rx_mic_key = data->tkip->mic_keys.rx_unicast;
966 } else {
967 tkip_sc =
968 data->rsc_tsc->all_tsc_rsc.tkip.multicast_rsc;
969 rx_p1ks = data->tkip->rx_multi;
970 rx_mic_key = data->tkip->mic_keys.rx_mcast;
974 * For non-QoS this relies on the fact that both the uCode and
975 * mac80211 use TID 0 (as they need to to avoid replay attacks)
976 * for checking the IV in the frames.
978 for (i = 0; i < IWLAGN_NUM_RSC; i++) {
979 ieee80211_get_key_rx_seq(key, i, &seq);
980 tkip_sc[i].iv16 = cpu_to_le16(seq.tkip.iv16);
981 tkip_sc[i].iv32 = cpu_to_le32(seq.tkip.iv32);
982 /* wrapping isn't allowed, AP must rekey */
983 if (seq.tkip.iv32 > cur_rx_iv32)
984 cur_rx_iv32 = seq.tkip.iv32;
987 ieee80211_get_tkip_rx_p1k(key, data->bssid, cur_rx_iv32, p1k);
988 iwlagn_convert_p1k(p1k, rx_p1ks[0].p1k);
989 ieee80211_get_tkip_rx_p1k(key, data->bssid,
990 cur_rx_iv32 + 1, p1k);
991 iwlagn_convert_p1k(p1k, rx_p1ks[1].p1k);
993 memcpy(rx_mic_key,
994 &key->key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY],
995 IWLAGN_MIC_KEY_SIZE);
997 data->use_tkip = true;
998 data->use_rsc_tsc = true;
999 break;
1000 case WLAN_CIPHER_SUITE_CCMP:
1001 if (sta) {
1002 u8 *pn = seq.ccmp.pn;
1004 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.unicast_rsc;
1005 aes_tx_sc = &data->rsc_tsc->all_tsc_rsc.aes.tsc;
1007 ieee80211_get_key_tx_seq(key, &seq);
1008 aes_tx_sc->pn = cpu_to_le64(
1009 (u64)pn[5] |
1010 ((u64)pn[4] << 8) |
1011 ((u64)pn[3] << 16) |
1012 ((u64)pn[2] << 24) |
1013 ((u64)pn[1] << 32) |
1014 ((u64)pn[0] << 40));
1015 } else
1016 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.multicast_rsc;
1019 * For non-QoS this relies on the fact that both the uCode and
1020 * mac80211 use TID 0 for checking the IV in the frames.
1022 for (i = 0; i < IWLAGN_NUM_RSC; i++) {
1023 u8 *pn = seq.ccmp.pn;
1025 ieee80211_get_key_rx_seq(key, i, &seq);
1026 aes_sc->pn = cpu_to_le64(
1027 (u64)pn[5] |
1028 ((u64)pn[4] << 8) |
1029 ((u64)pn[3] << 16) |
1030 ((u64)pn[2] << 24) |
1031 ((u64)pn[1] << 32) |
1032 ((u64)pn[0] << 40));
1034 data->use_rsc_tsc = true;
1035 break;
1038 mutex_unlock(&priv->mutex);
1041 int iwlagn_send_patterns(struct iwl_priv *priv,
1042 struct cfg80211_wowlan *wowlan)
1044 struct iwlagn_wowlan_patterns_cmd *pattern_cmd;
1045 struct iwl_host_cmd cmd = {
1046 .id = REPLY_WOWLAN_PATTERNS,
1047 .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1048 .flags = CMD_SYNC,
1050 int i, err;
1052 if (!wowlan->n_patterns)
1053 return 0;
1055 cmd.len[0] = sizeof(*pattern_cmd) +
1056 wowlan->n_patterns * sizeof(struct iwlagn_wowlan_pattern);
1058 pattern_cmd = kmalloc(cmd.len[0], GFP_KERNEL);
1059 if (!pattern_cmd)
1060 return -ENOMEM;
1062 pattern_cmd->n_patterns = cpu_to_le32(wowlan->n_patterns);
1064 for (i = 0; i < wowlan->n_patterns; i++) {
1065 int mask_len = DIV_ROUND_UP(wowlan->patterns[i].pattern_len, 8);
1067 memcpy(&pattern_cmd->patterns[i].mask,
1068 wowlan->patterns[i].mask, mask_len);
1069 memcpy(&pattern_cmd->patterns[i].pattern,
1070 wowlan->patterns[i].pattern,
1071 wowlan->patterns[i].pattern_len);
1072 pattern_cmd->patterns[i].mask_size = mask_len;
1073 pattern_cmd->patterns[i].pattern_size =
1074 wowlan->patterns[i].pattern_len;
1077 cmd.data[0] = pattern_cmd;
1078 err = iwl_dvm_send_cmd(priv, &cmd);
1079 kfree(pattern_cmd);
1080 return err;
1083 int iwlagn_suspend(struct iwl_priv *priv, struct cfg80211_wowlan *wowlan)
1085 struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd;
1086 struct iwl_rxon_cmd rxon;
1087 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
1088 struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd;
1089 struct iwlagn_wowlan_tkip_params_cmd tkip_cmd = {};
1090 struct iwlagn_d3_config_cmd d3_cfg_cmd = {
1092 * Program the minimum sleep time to 10 seconds, as many
1093 * platforms have issues processing a wakeup signal while
1094 * still being in the process of suspending.
1096 .min_sleep_time = cpu_to_le32(10 * 1000 * 1000),
1098 struct wowlan_key_data key_data = {
1099 .ctx = ctx,
1100 .bssid = ctx->active.bssid_addr,
1101 .use_rsc_tsc = false,
1102 .tkip = &tkip_cmd,
1103 .use_tkip = false,
1105 int ret, i;
1106 u16 seq;
1108 key_data.rsc_tsc = kzalloc(sizeof(*key_data.rsc_tsc), GFP_KERNEL);
1109 if (!key_data.rsc_tsc)
1110 return -ENOMEM;
1112 memset(&wakeup_filter_cmd, 0, sizeof(wakeup_filter_cmd));
1115 * We know the last used seqno, and the uCode expects to know that
1116 * one, it will increment before TX.
1118 seq = le16_to_cpu(priv->last_seq_ctl) & IEEE80211_SCTL_SEQ;
1119 wakeup_filter_cmd.non_qos_seq = cpu_to_le16(seq);
1122 * For QoS counters, we store the one to use next, so subtract 0x10
1123 * since the uCode will add 0x10 before using the value.
1125 for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
1126 seq = priv->tid_data[IWL_AP_ID][i].seq_number;
1127 seq -= 0x10;
1128 wakeup_filter_cmd.qos_seq[i] = cpu_to_le16(seq);
1131 if (wowlan->disconnect)
1132 wakeup_filter_cmd.enabled |=
1133 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_BEACON_MISS |
1134 IWLAGN_WOWLAN_WAKEUP_LINK_CHANGE);
1135 if (wowlan->magic_pkt)
1136 wakeup_filter_cmd.enabled |=
1137 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_MAGIC_PACKET);
1138 if (wowlan->gtk_rekey_failure)
1139 wakeup_filter_cmd.enabled |=
1140 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_GTK_REKEY_FAIL);
1141 if (wowlan->eap_identity_req)
1142 wakeup_filter_cmd.enabled |=
1143 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_EAP_IDENT_REQ);
1144 if (wowlan->four_way_handshake)
1145 wakeup_filter_cmd.enabled |=
1146 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_4WAY_HANDSHAKE);
1147 if (wowlan->n_patterns)
1148 wakeup_filter_cmd.enabled |=
1149 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_PATTERN_MATCH);
1151 if (wowlan->rfkill_release)
1152 d3_cfg_cmd.wakeup_flags |=
1153 cpu_to_le32(IWLAGN_D3_WAKEUP_RFKILL);
1155 iwl_scan_cancel_timeout(priv, 200);
1157 memcpy(&rxon, &ctx->active, sizeof(rxon));
1159 priv->ucode_loaded = false;
1160 iwl_trans_stop_device(priv->trans);
1162 priv->wowlan = true;
1164 ret = iwl_load_ucode_wait_alive(priv, IWL_UCODE_WOWLAN);
1165 if (ret)
1166 goto out;
1168 /* now configure WoWLAN ucode */
1169 ret = iwl_alive_start(priv);
1170 if (ret)
1171 goto out;
1173 memcpy(&ctx->staging, &rxon, sizeof(rxon));
1174 ret = iwlagn_commit_rxon(priv, ctx);
1175 if (ret)
1176 goto out;
1178 ret = iwl_power_update_mode(priv, true);
1179 if (ret)
1180 goto out;
1182 if (!iwlwifi_mod_params.sw_crypto) {
1183 /* mark all keys clear */
1184 priv->ucode_key_table = 0;
1185 ctx->key_mapping_keys = 0;
1188 * This needs to be unlocked due to lock ordering
1189 * constraints. Since we're in the suspend path
1190 * that isn't really a problem though.
1192 mutex_unlock(&priv->mutex);
1193 ieee80211_iter_keys(priv->hw, ctx->vif,
1194 iwlagn_wowlan_program_keys,
1195 &key_data);
1196 mutex_lock(&priv->mutex);
1197 if (key_data.error) {
1198 ret = -EIO;
1199 goto out;
1202 if (key_data.use_rsc_tsc) {
1203 struct iwl_host_cmd rsc_tsc_cmd = {
1204 .id = REPLY_WOWLAN_TSC_RSC_PARAMS,
1205 .flags = CMD_SYNC,
1206 .data[0] = key_data.rsc_tsc,
1207 .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1208 .len[0] = sizeof(*key_data.rsc_tsc),
1211 ret = iwl_dvm_send_cmd(priv, &rsc_tsc_cmd);
1212 if (ret)
1213 goto out;
1216 if (key_data.use_tkip) {
1217 ret = iwl_dvm_send_cmd_pdu(priv,
1218 REPLY_WOWLAN_TKIP_PARAMS,
1219 CMD_SYNC, sizeof(tkip_cmd),
1220 &tkip_cmd);
1221 if (ret)
1222 goto out;
1225 if (priv->have_rekey_data) {
1226 memset(&kek_kck_cmd, 0, sizeof(kek_kck_cmd));
1227 memcpy(kek_kck_cmd.kck, priv->kck, NL80211_KCK_LEN);
1228 kek_kck_cmd.kck_len = cpu_to_le16(NL80211_KCK_LEN);
1229 memcpy(kek_kck_cmd.kek, priv->kek, NL80211_KEK_LEN);
1230 kek_kck_cmd.kek_len = cpu_to_le16(NL80211_KEK_LEN);
1231 kek_kck_cmd.replay_ctr = priv->replay_ctr;
1233 ret = iwl_dvm_send_cmd_pdu(priv,
1234 REPLY_WOWLAN_KEK_KCK_MATERIAL,
1235 CMD_SYNC, sizeof(kek_kck_cmd),
1236 &kek_kck_cmd);
1237 if (ret)
1238 goto out;
1242 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_D3_CONFIG, CMD_SYNC,
1243 sizeof(d3_cfg_cmd), &d3_cfg_cmd);
1244 if (ret)
1245 goto out;
1247 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_WOWLAN_WAKEUP_FILTER,
1248 CMD_SYNC, sizeof(wakeup_filter_cmd),
1249 &wakeup_filter_cmd);
1250 if (ret)
1251 goto out;
1253 ret = iwlagn_send_patterns(priv, wowlan);
1254 out:
1255 kfree(key_data.rsc_tsc);
1256 return ret;
1258 #endif
1260 int iwl_dvm_send_cmd(struct iwl_priv *priv, struct iwl_host_cmd *cmd)
1262 if (iwl_is_rfkill(priv) || iwl_is_ctkill(priv)) {
1263 IWL_WARN(priv, "Not sending command - %s KILL\n",
1264 iwl_is_rfkill(priv) ? "RF" : "CT");
1265 return -EIO;
1268 if (test_bit(STATUS_FW_ERROR, &priv->status)) {
1269 IWL_ERR(priv, "Command %s failed: FW Error\n",
1270 iwl_dvm_get_cmd_string(cmd->id));
1271 return -EIO;
1275 * This can happen upon FW ASSERT: we clear the STATUS_FW_ERROR flag
1276 * in iwl_down but cancel the workers only later.
1278 if (!priv->ucode_loaded) {
1279 IWL_ERR(priv, "Fw not loaded - dropping CMD: %x\n", cmd->id);
1280 return -EIO;
1284 * Synchronous commands from this op-mode must hold
1285 * the mutex, this ensures we don't try to send two
1286 * (or more) synchronous commands at a time.
1288 if (!(cmd->flags & CMD_ASYNC))
1289 lockdep_assert_held(&priv->mutex);
1291 return iwl_trans_send_cmd(priv->trans, cmd);
1294 int iwl_dvm_send_cmd_pdu(struct iwl_priv *priv, u8 id,
1295 u32 flags, u16 len, const void *data)
1297 struct iwl_host_cmd cmd = {
1298 .id = id,
1299 .len = { len, },
1300 .data = { data, },
1301 .flags = flags,
1304 return iwl_dvm_send_cmd(priv, &cmd);