gpio: rcar: Fix runtime PM imbalance on error
[linux/fpc-iii.git] / drivers / net / wireless / intel / iwlwifi / mvm / mvm.h
blobafcf2b98a9cb913c18fdeb39c44a2949d6c024f0
1 /******************************************************************************
3 * This file is provided under a dual BSD/GPLv2 license. When using or
4 * redistributing this file, you may do so under either license.
6 * GPL LICENSE SUMMARY
8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
10 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
11 * Copyright(c) 2018 - 2019 Intel Corporation
13 * This program is free software; you can redistribute it and/or modify
14 * it under the terms of version 2 of the GNU General Public License as
15 * published by the Free Software Foundation.
17 * This program is distributed in the hope that it will be useful, but
18 * WITHOUT ANY WARRANTY; without even the implied warranty of
19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
20 * General Public License for more details.
22 * The full GNU General Public License is included in this distribution
23 * in the file called COPYING.
25 * Contact Information:
26 * Intel Linux Wireless <linuxwifi@intel.com>
27 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
29 * BSD LICENSE
31 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
32 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
33 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
34 * Copyright(c) 2018 - 2019 Intel Corporation
35 * All rights reserved.
37 * Redistribution and use in source and binary forms, with or without
38 * modification, are permitted provided that the following conditions
39 * are met:
41 * * Redistributions of source code must retain the above copyright
42 * notice, this list of conditions and the following disclaimer.
43 * * Redistributions in binary form must reproduce the above copyright
44 * notice, this list of conditions and the following disclaimer in
45 * the documentation and/or other materials provided with the
46 * distribution.
47 * * Neither the name Intel Corporation nor the names of its
48 * contributors may be used to endorse or promote products derived
49 * from this software without specific prior written permission.
51 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
52 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
53 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
54 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
55 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
56 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
57 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
58 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
59 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
60 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
61 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
63 *****************************************************************************/
65 #ifndef __IWL_MVM_H__
66 #define __IWL_MVM_H__
68 #include <linux/list.h>
69 #include <linux/spinlock.h>
70 #include <linux/leds.h>
71 #include <linux/in6.h>
73 #ifdef CONFIG_THERMAL
74 #include <linux/thermal.h>
75 #endif
77 #include "iwl-op-mode.h"
78 #include "iwl-trans.h"
79 #include "fw/notif-wait.h"
80 #include "iwl-eeprom-parse.h"
81 #include "fw/file.h"
82 #include "iwl-config.h"
83 #include "sta.h"
84 #include "fw-api.h"
85 #include "constants.h"
86 #include "fw/runtime.h"
87 #include "fw/dbg.h"
88 #include "fw/acpi.h"
89 #include "iwl-nvm-parse.h"
91 #include <linux/average.h>
93 #define IWL_MVM_MAX_ADDRESSES 5
94 /* RSSI offset for WkP */
95 #define IWL_RSSI_OFFSET 50
96 #define IWL_MVM_MISSED_BEACONS_THRESHOLD 8
97 #define IWL_MVM_MISSED_BEACONS_THRESHOLD_LONG 16
99 /* A TimeUnit is 1024 microsecond */
100 #define MSEC_TO_TU(_msec) (_msec*1000/1024)
102 /* For GO, this value represents the number of TUs before CSA "beacon
103 * 0" TBTT when the CSA time-event needs to be scheduled to start. It
104 * must be big enough to ensure that we switch in time.
106 #define IWL_MVM_CHANNEL_SWITCH_TIME_GO 40
108 /* For client, this value represents the number of TUs before CSA
109 * "beacon 1" TBTT, instead. This is because we don't know when the
110 * GO/AP will be in the new channel, so we switch early enough.
112 #define IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT 10
115 * This value (in TUs) is used to fine tune the CSA NoA end time which should
116 * be just before "beacon 0" TBTT.
118 #define IWL_MVM_CHANNEL_SWITCH_MARGIN 4
121 * Number of beacons to transmit on a new channel until we unblock tx to
122 * the stations, even if we didn't identify them on a new channel
124 #define IWL_MVM_CS_UNBLOCK_TX_TIMEOUT 3
126 /* offchannel queue towards mac80211 */
127 #define IWL_MVM_OFFCHANNEL_QUEUE 0
129 extern const struct ieee80211_ops iwl_mvm_hw_ops;
132 * struct iwl_mvm_mod_params - module parameters for iwlmvm
133 * @init_dbg: if true, then the NIC won't be stopped if the INIT fw asserted.
134 * We will register to mac80211 to have testmode working. The NIC must not
135 * be up'ed after the INIT fw asserted. This is useful to be able to use
136 * proprietary tools over testmode to debug the INIT fw.
137 * @tfd_q_hang_detect: enabled the detection of hung transmit queues
138 * @power_scheme: one of enum iwl_power_scheme
140 struct iwl_mvm_mod_params {
141 bool init_dbg;
142 bool tfd_q_hang_detect;
143 int power_scheme;
145 extern struct iwl_mvm_mod_params iwlmvm_mod_params;
147 struct iwl_mvm_phy_ctxt {
148 u16 id;
149 u16 color;
150 u32 ref;
152 enum nl80211_chan_width width;
155 * TODO: This should probably be removed. Currently here only for rate
156 * scaling algorithm
158 struct ieee80211_channel *channel;
161 struct iwl_mvm_time_event_data {
162 struct ieee80211_vif *vif;
163 struct list_head list;
164 unsigned long end_jiffies;
165 u32 duration;
166 bool running;
167 u32 uid;
170 * The access to the 'id' field must be done when the
171 * mvm->time_event_lock is held, as it value is used to indicate
172 * if the te is in the time event list or not (when id == TE_MAX)
174 u32 id;
177 /* Power management */
180 * enum iwl_power_scheme
181 * @IWL_POWER_LEVEL_CAM - Continuously Active Mode
182 * @IWL_POWER_LEVEL_BPS - Balanced Power Save (default)
183 * @IWL_POWER_LEVEL_LP - Low Power
185 enum iwl_power_scheme {
186 IWL_POWER_SCHEME_CAM = 1,
187 IWL_POWER_SCHEME_BPS,
188 IWL_POWER_SCHEME_LP
191 union geo_tx_power_profiles_cmd {
192 struct iwl_geo_tx_power_profiles_cmd geo_cmd;
193 struct iwl_geo_tx_power_profiles_cmd_v1 geo_cmd_v1;
196 #define IWL_CONN_MAX_LISTEN_INTERVAL 10
197 #define IWL_UAPSD_MAX_SP IEEE80211_WMM_IE_STA_QOSINFO_SP_ALL
199 #ifdef CONFIG_IWLWIFI_DEBUGFS
200 enum iwl_dbgfs_pm_mask {
201 MVM_DEBUGFS_PM_KEEP_ALIVE = BIT(0),
202 MVM_DEBUGFS_PM_SKIP_OVER_DTIM = BIT(1),
203 MVM_DEBUGFS_PM_SKIP_DTIM_PERIODS = BIT(2),
204 MVM_DEBUGFS_PM_RX_DATA_TIMEOUT = BIT(3),
205 MVM_DEBUGFS_PM_TX_DATA_TIMEOUT = BIT(4),
206 MVM_DEBUGFS_PM_LPRX_ENA = BIT(6),
207 MVM_DEBUGFS_PM_LPRX_RSSI_THRESHOLD = BIT(7),
208 MVM_DEBUGFS_PM_SNOOZE_ENABLE = BIT(8),
209 MVM_DEBUGFS_PM_UAPSD_MISBEHAVING = BIT(9),
210 MVM_DEBUGFS_PM_USE_PS_POLL = BIT(10),
213 struct iwl_dbgfs_pm {
214 u16 keep_alive_seconds;
215 u32 rx_data_timeout;
216 u32 tx_data_timeout;
217 bool skip_over_dtim;
218 u8 skip_dtim_periods;
219 bool lprx_ena;
220 u32 lprx_rssi_threshold;
221 bool snooze_ena;
222 bool uapsd_misbehaving;
223 bool use_ps_poll;
224 int mask;
227 /* beacon filtering */
229 enum iwl_dbgfs_bf_mask {
230 MVM_DEBUGFS_BF_ENERGY_DELTA = BIT(0),
231 MVM_DEBUGFS_BF_ROAMING_ENERGY_DELTA = BIT(1),
232 MVM_DEBUGFS_BF_ROAMING_STATE = BIT(2),
233 MVM_DEBUGFS_BF_TEMP_THRESHOLD = BIT(3),
234 MVM_DEBUGFS_BF_TEMP_FAST_FILTER = BIT(4),
235 MVM_DEBUGFS_BF_TEMP_SLOW_FILTER = BIT(5),
236 MVM_DEBUGFS_BF_ENABLE_BEACON_FILTER = BIT(6),
237 MVM_DEBUGFS_BF_DEBUG_FLAG = BIT(7),
238 MVM_DEBUGFS_BF_ESCAPE_TIMER = BIT(8),
239 MVM_DEBUGFS_BA_ESCAPE_TIMER = BIT(9),
240 MVM_DEBUGFS_BA_ENABLE_BEACON_ABORT = BIT(10),
243 struct iwl_dbgfs_bf {
244 u32 bf_energy_delta;
245 u32 bf_roaming_energy_delta;
246 u32 bf_roaming_state;
247 u32 bf_temp_threshold;
248 u32 bf_temp_fast_filter;
249 u32 bf_temp_slow_filter;
250 u32 bf_enable_beacon_filter;
251 u32 bf_debug_flag;
252 u32 bf_escape_timer;
253 u32 ba_escape_timer;
254 u32 ba_enable_beacon_abort;
255 int mask;
257 #endif
259 enum iwl_mvm_smps_type_request {
260 IWL_MVM_SMPS_REQ_BT_COEX,
261 IWL_MVM_SMPS_REQ_TT,
262 IWL_MVM_SMPS_REQ_PROT,
263 NUM_IWL_MVM_SMPS_REQ,
266 enum iwl_bt_force_ant_mode {
267 BT_FORCE_ANT_DIS = 0,
268 BT_FORCE_ANT_AUTO,
269 BT_FORCE_ANT_BT,
270 BT_FORCE_ANT_WIFI,
272 BT_FORCE_ANT_MAX,
276 * struct iwl_mvm_low_latency_force - low latency force mode set by debugfs
277 * @LOW_LATENCY_FORCE_UNSET: unset force mode
278 * @LOW_LATENCY_FORCE_ON: for low latency on
279 * @LOW_LATENCY_FORCE_OFF: for low latency off
280 * @NUM_LOW_LATENCY_FORCE: max num of modes
282 enum iwl_mvm_low_latency_force {
283 LOW_LATENCY_FORCE_UNSET,
284 LOW_LATENCY_FORCE_ON,
285 LOW_LATENCY_FORCE_OFF,
286 NUM_LOW_LATENCY_FORCE
290 * struct iwl_mvm_low_latency_cause - low latency set causes
291 * @LOW_LATENCY_TRAFFIC: indicates low latency traffic was detected
292 * @LOW_LATENCY_DEBUGFS: low latency mode set from debugfs
293 * @LOW_LATENCY_VCMD: low latency mode set from vendor command
294 * @LOW_LATENCY_VIF_TYPE: low latency mode set because of vif type (ap)
295 * @LOW_LATENCY_DEBUGFS_FORCE_ENABLE: indicate that force mode is enabled
296 * the actual set/unset is done with LOW_LATENCY_DEBUGFS_FORCE
297 * @LOW_LATENCY_DEBUGFS_FORCE: low latency force mode from debugfs
298 * set this with LOW_LATENCY_DEBUGFS_FORCE_ENABLE flag
299 * in low_latency.
301 enum iwl_mvm_low_latency_cause {
302 LOW_LATENCY_TRAFFIC = BIT(0),
303 LOW_LATENCY_DEBUGFS = BIT(1),
304 LOW_LATENCY_VCMD = BIT(2),
305 LOW_LATENCY_VIF_TYPE = BIT(3),
306 LOW_LATENCY_DEBUGFS_FORCE_ENABLE = BIT(4),
307 LOW_LATENCY_DEBUGFS_FORCE = BIT(5),
311 * struct iwl_mvm_vif_bf_data - beacon filtering related data
312 * @bf_enabled: indicates if beacon filtering is enabled
313 * @ba_enabled: indicated if beacon abort is enabled
314 * @ave_beacon_signal: average beacon signal
315 * @last_cqm_event: rssi of the last cqm event
316 * @bt_coex_min_thold: minimum threshold for BT coex
317 * @bt_coex_max_thold: maximum threshold for BT coex
318 * @last_bt_coex_event: rssi of the last BT coex event
320 struct iwl_mvm_vif_bf_data {
321 bool bf_enabled;
322 bool ba_enabled;
323 int ave_beacon_signal;
324 int last_cqm_event;
325 int bt_coex_min_thold;
326 int bt_coex_max_thold;
327 int last_bt_coex_event;
331 * struct iwl_probe_resp_data - data for NoA/CSA updates
332 * @rcu_head: used for freeing the data on update
333 * @notif: notification data
334 * @noa_len: length of NoA attribute, calculated from the notification
336 struct iwl_probe_resp_data {
337 struct rcu_head rcu_head;
338 struct iwl_probe_resp_data_notif notif;
339 int noa_len;
343 * struct iwl_mvm_vif - data per Virtual Interface, it is a MAC context
344 * @id: between 0 and 3
345 * @color: to solve races upon MAC addition and removal
346 * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA
347 * @bssid: BSSID for this (client) interface
348 * @associated: indicates that we're currently associated, used only for
349 * managing the firmware state in iwl_mvm_bss_info_changed_station()
350 * @ap_assoc_sta_count: count of stations associated to us - valid only
351 * if VIF type is AP
352 * @uploaded: indicates the MAC context has been added to the device
353 * @ap_ibss_active: indicates that AP/IBSS is configured and that the interface
354 * should get quota etc.
355 * @pm_enabled - Indicate if MAC power management is allowed
356 * @monitor_active: indicates that monitor context is configured, and that the
357 * interface should get quota etc.
358 * @low_latency: bit flags for low latency
359 * see enum &iwl_mvm_low_latency_cause for causes.
360 * @low_latency_actual: boolean, indicates low latency is set,
361 * as a result from low_latency bit flags and takes force into account.
362 * @ps_disabled: indicates that this interface requires PS to be disabled
363 * @queue_params: QoS params for this MAC
364 * @bcast_sta: station used for broadcast packets. Used by the following
365 * vifs: P2P_DEVICE, GO and AP.
366 * @beacon_skb: the skb used to hold the AP/GO beacon template
367 * @smps_requests: the SMPS requests of different parts of the driver,
368 * combined on update to yield the overall request to mac80211.
369 * @beacon_stats: beacon statistics, containing the # of received beacons,
370 * # of received beacons accumulated over FW restart, and the current
371 * average signal of beacons retrieved from the firmware
372 * @csa_failed: CSA failed to schedule time event, report an error later
373 * @features: hw features active for this vif
374 * @probe_resp_data: data from FW notification to store NOA and CSA related
375 * data to be inserted into probe response.
377 struct iwl_mvm_vif {
378 struct iwl_mvm *mvm;
379 u16 id;
380 u16 color;
381 u8 ap_sta_id;
383 u8 bssid[ETH_ALEN];
384 bool associated;
385 u8 ap_assoc_sta_count;
387 u16 cab_queue;
389 bool uploaded;
390 bool ap_ibss_active;
391 bool pm_enabled;
392 bool monitor_active;
393 u8 low_latency: 6;
394 u8 low_latency_actual: 1;
395 bool ps_disabled;
396 struct iwl_mvm_vif_bf_data bf_data;
398 struct {
399 u32 num_beacons, accu_num_beacons;
400 u8 avg_signal;
401 } beacon_stats;
403 u32 ap_beacon_time;
405 enum iwl_tsf_id tsf_id;
408 * QoS data from mac80211, need to store this here
409 * as mac80211 has a separate callback but we need
410 * to have the data for the MAC context
412 struct ieee80211_tx_queue_params queue_params[IEEE80211_NUM_ACS];
413 struct iwl_mvm_time_event_data time_event_data;
414 struct iwl_mvm_time_event_data hs_time_event_data;
416 struct iwl_mvm_int_sta bcast_sta;
417 struct iwl_mvm_int_sta mcast_sta;
420 * Assigned while mac80211 has the interface in a channel context,
421 * or, for P2P Device, while it exists.
423 struct iwl_mvm_phy_ctxt *phy_ctxt;
425 #ifdef CONFIG_PM
426 /* WoWLAN GTK rekey data */
427 struct {
428 u8 kck[NL80211_KCK_LEN], kek[NL80211_KEK_LEN];
429 __le64 replay_ctr;
430 bool valid;
431 } rekey_data;
433 int tx_key_idx;
435 bool seqno_valid;
436 u16 seqno;
437 #endif
439 #if IS_ENABLED(CONFIG_IPV6)
440 /* IPv6 addresses for WoWLAN */
441 struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX];
442 unsigned long tentative_addrs[BITS_TO_LONGS(IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX)];
443 int num_target_ipv6_addrs;
444 #endif
446 #ifdef CONFIG_IWLWIFI_DEBUGFS
447 struct dentry *dbgfs_dir;
448 struct dentry *dbgfs_slink;
449 struct iwl_dbgfs_pm dbgfs_pm;
450 struct iwl_dbgfs_bf dbgfs_bf;
451 struct iwl_mac_power_cmd mac_pwr_cmd;
452 int dbgfs_quota_min;
453 #endif
455 enum ieee80211_smps_mode smps_requests[NUM_IWL_MVM_SMPS_REQ];
457 /* FW identified misbehaving AP */
458 u8 uapsd_misbehaving_bssid[ETH_ALEN];
460 struct delayed_work uapsd_nonagg_detected_wk;
462 /* Indicates that CSA countdown may be started */
463 bool csa_countdown;
464 bool csa_failed;
465 u16 csa_target_freq;
466 u16 csa_count;
467 u16 csa_misbehave;
468 struct delayed_work csa_work;
470 /* Indicates that we are waiting for a beacon on a new channel */
471 bool csa_bcn_pending;
473 /* TCP Checksum Offload */
474 netdev_features_t features;
476 struct iwl_probe_resp_data __rcu *probe_resp_data;
478 /* we can only have 2 GTK + 2 IGTK active at a time */
479 struct ieee80211_key_conf *ap_early_keys[4];
481 /* 26-tone RU OFDMA transmissions should be blocked */
482 bool he_ru_2mhz_block;
485 static inline struct iwl_mvm_vif *
486 iwl_mvm_vif_from_mac80211(struct ieee80211_vif *vif)
488 if (!vif)
489 return NULL;
490 return (void *)vif->drv_priv;
493 extern const u8 tid_to_mac80211_ac[];
495 #define IWL_MVM_SCAN_STOPPING_SHIFT 8
497 enum iwl_scan_status {
498 IWL_MVM_SCAN_REGULAR = BIT(0),
499 IWL_MVM_SCAN_SCHED = BIT(1),
500 IWL_MVM_SCAN_NETDETECT = BIT(2),
502 IWL_MVM_SCAN_STOPPING_REGULAR = BIT(8),
503 IWL_MVM_SCAN_STOPPING_SCHED = BIT(9),
504 IWL_MVM_SCAN_STOPPING_NETDETECT = BIT(10),
506 IWL_MVM_SCAN_REGULAR_MASK = IWL_MVM_SCAN_REGULAR |
507 IWL_MVM_SCAN_STOPPING_REGULAR,
508 IWL_MVM_SCAN_SCHED_MASK = IWL_MVM_SCAN_SCHED |
509 IWL_MVM_SCAN_STOPPING_SCHED,
510 IWL_MVM_SCAN_NETDETECT_MASK = IWL_MVM_SCAN_NETDETECT |
511 IWL_MVM_SCAN_STOPPING_NETDETECT,
513 IWL_MVM_SCAN_STOPPING_MASK = 0xff << IWL_MVM_SCAN_STOPPING_SHIFT,
514 IWL_MVM_SCAN_MASK = 0xff,
517 enum iwl_mvm_scan_type {
518 IWL_SCAN_TYPE_NOT_SET,
519 IWL_SCAN_TYPE_UNASSOC,
520 IWL_SCAN_TYPE_WILD,
521 IWL_SCAN_TYPE_MILD,
522 IWL_SCAN_TYPE_FRAGMENTED,
523 IWL_SCAN_TYPE_FAST_BALANCE,
526 enum iwl_mvm_sched_scan_pass_all_states {
527 SCHED_SCAN_PASS_ALL_DISABLED,
528 SCHED_SCAN_PASS_ALL_ENABLED,
529 SCHED_SCAN_PASS_ALL_FOUND,
533 * struct iwl_mvm_tt_mgnt - Thermal Throttling Management structure
534 * @ct_kill_exit: worker to exit thermal kill
535 * @dynamic_smps: Is thermal throttling enabled dynamic_smps?
536 * @tx_backoff: The current thremal throttling tx backoff in uSec.
537 * @min_backoff: The minimal tx backoff due to power restrictions
538 * @params: Parameters to configure the thermal throttling algorithm.
539 * @throttle: Is thermal throttling is active?
541 struct iwl_mvm_tt_mgmt {
542 struct delayed_work ct_kill_exit;
543 bool dynamic_smps;
544 u32 tx_backoff;
545 u32 min_backoff;
546 struct iwl_tt_params params;
547 bool throttle;
550 #ifdef CONFIG_THERMAL
552 *struct iwl_mvm_thermal_device - thermal zone related data
553 * @temp_trips: temperature thresholds for report
554 * @fw_trips_index: keep indexes to original array - temp_trips
555 * @tzone: thermal zone device data
557 struct iwl_mvm_thermal_device {
558 s16 temp_trips[IWL_MAX_DTS_TRIPS];
559 u8 fw_trips_index[IWL_MAX_DTS_TRIPS];
560 struct thermal_zone_device *tzone;
564 * struct iwl_mvm_cooling_device
565 * @cur_state: current state
566 * @cdev: struct thermal cooling device
568 struct iwl_mvm_cooling_device {
569 u32 cur_state;
570 struct thermal_cooling_device *cdev;
572 #endif
574 #define IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES 8
576 struct iwl_mvm_frame_stats {
577 u32 legacy_frames;
578 u32 ht_frames;
579 u32 vht_frames;
580 u32 bw_20_frames;
581 u32 bw_40_frames;
582 u32 bw_80_frames;
583 u32 bw_160_frames;
584 u32 sgi_frames;
585 u32 ngi_frames;
586 u32 siso_frames;
587 u32 mimo2_frames;
588 u32 agg_frames;
589 u32 ampdu_count;
590 u32 success_frames;
591 u32 fail_frames;
592 u32 last_rates[IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES];
593 int last_frame_idx;
596 #define IWL_MVM_DEBUG_SET_TEMPERATURE_DISABLE 0xff
597 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MIN -100
598 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MAX 200
600 enum iwl_mvm_tdls_cs_state {
601 IWL_MVM_TDLS_SW_IDLE = 0,
602 IWL_MVM_TDLS_SW_REQ_SENT,
603 IWL_MVM_TDLS_SW_RESP_RCVD,
604 IWL_MVM_TDLS_SW_REQ_RCVD,
605 IWL_MVM_TDLS_SW_ACTIVE,
608 enum iwl_mvm_traffic_load {
609 IWL_MVM_TRAFFIC_LOW,
610 IWL_MVM_TRAFFIC_MEDIUM,
611 IWL_MVM_TRAFFIC_HIGH,
614 DECLARE_EWMA(rate, 16, 16)
616 struct iwl_mvm_tcm_mac {
617 struct {
618 u32 pkts[IEEE80211_NUM_ACS];
619 u32 airtime;
620 } tx;
621 struct {
622 u32 pkts[IEEE80211_NUM_ACS];
623 u32 airtime;
624 u32 last_ampdu_ref;
625 } rx;
626 struct {
627 /* track AP's transfer in client mode */
628 u64 rx_bytes;
629 struct ewma_rate rate;
630 bool detected;
631 } uapsd_nonagg_detect;
632 bool opened_rx_ba_sessions;
635 struct iwl_mvm_tcm {
636 struct delayed_work work;
637 spinlock_t lock; /* used when time elapsed */
638 unsigned long ts; /* timestamp when period ends */
639 unsigned long ll_ts;
640 unsigned long uapsd_nonagg_ts;
641 bool paused;
642 struct iwl_mvm_tcm_mac data[NUM_MAC_INDEX_DRIVER];
643 struct {
644 u32 elapsed; /* milliseconds for this TCM period */
645 u32 airtime[NUM_MAC_INDEX_DRIVER];
646 enum iwl_mvm_traffic_load load[NUM_MAC_INDEX_DRIVER];
647 enum iwl_mvm_traffic_load band_load[NUM_NL80211_BANDS];
648 enum iwl_mvm_traffic_load global_load;
649 bool low_latency[NUM_MAC_INDEX_DRIVER];
650 bool change[NUM_MAC_INDEX_DRIVER];
651 bool global_change;
652 } result;
656 * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer
657 * @head_sn: reorder window head sn
658 * @num_stored: number of mpdus stored in the buffer
659 * @buf_size: the reorder buffer size as set by the last addba request
660 * @queue: queue of this reorder buffer
661 * @last_amsdu: track last ASMDU SN for duplication detection
662 * @last_sub_index: track ASMDU sub frame index for duplication detection
663 * @reorder_timer: timer for frames are in the reorder buffer. For AMSDU
664 * it is the time of last received sub-frame
665 * @removed: prevent timer re-arming
666 * @valid: reordering is valid for this queue
667 * @lock: protect reorder buffer internal state
668 * @mvm: mvm pointer, needed for frame timer context
669 * @consec_oldsn_drops: consecutive drops due to old SN
670 * @consec_oldsn_ampdu_gp2: A-MPDU GP2 timestamp to track
671 * when to apply old SN consecutive drop workaround
672 * @consec_oldsn_prev_drop: track whether or not an MPDU
673 * that was single/part of the previous A-MPDU was
674 * dropped due to old SN
676 struct iwl_mvm_reorder_buffer {
677 u16 head_sn;
678 u16 num_stored;
679 u16 buf_size;
680 int queue;
681 u16 last_amsdu;
682 u8 last_sub_index;
683 struct timer_list reorder_timer;
684 bool removed;
685 bool valid;
686 spinlock_t lock;
687 struct iwl_mvm *mvm;
688 unsigned int consec_oldsn_drops;
689 u32 consec_oldsn_ampdu_gp2;
690 unsigned int consec_oldsn_prev_drop:1;
691 } ____cacheline_aligned_in_smp;
694 * struct _iwl_mvm_reorder_buf_entry - reorder buffer entry per-queue/per-seqno
695 * @frames: list of skbs stored
696 * @reorder_time: time the packet was stored in the reorder buffer
698 struct _iwl_mvm_reorder_buf_entry {
699 struct sk_buff_head frames;
700 unsigned long reorder_time;
703 /* make this indirection to get the aligned thing */
704 struct iwl_mvm_reorder_buf_entry {
705 struct _iwl_mvm_reorder_buf_entry e;
707 #ifndef __CHECKER__
708 /* sparse doesn't like this construct: "bad integer constant expression" */
709 __aligned(roundup_pow_of_two(sizeof(struct _iwl_mvm_reorder_buf_entry)))
710 #endif
714 * struct iwl_mvm_baid_data - BA session data
715 * @sta_id: station id
716 * @tid: tid of the session
717 * @baid baid of the session
718 * @timeout: the timeout set in the addba request
719 * @entries_per_queue: # of buffers per queue, this actually gets
720 * aligned up to avoid cache line sharing between queues
721 * @last_rx: last rx jiffies, updated only if timeout passed from last update
722 * @session_timer: timer to check if BA session expired, runs at 2 * timeout
723 * @mvm: mvm pointer, needed for timer context
724 * @reorder_buf: reorder buffer, allocated per queue
725 * @reorder_buf_data: data
727 struct iwl_mvm_baid_data {
728 struct rcu_head rcu_head;
729 u8 sta_id;
730 u8 tid;
731 u8 baid;
732 u16 timeout;
733 u16 entries_per_queue;
734 unsigned long last_rx;
735 struct timer_list session_timer;
736 struct iwl_mvm_baid_data __rcu **rcu_ptr;
737 struct iwl_mvm *mvm;
738 struct iwl_mvm_reorder_buffer reorder_buf[IWL_MAX_RX_HW_QUEUES];
739 struct iwl_mvm_reorder_buf_entry entries[];
742 static inline struct iwl_mvm_baid_data *
743 iwl_mvm_baid_data_from_reorder_buf(struct iwl_mvm_reorder_buffer *buf)
745 return (void *)((u8 *)buf -
746 offsetof(struct iwl_mvm_baid_data, reorder_buf) -
747 sizeof(*buf) * buf->queue);
751 * enum iwl_mvm_queue_status - queue status
752 * @IWL_MVM_QUEUE_FREE: the queue is not allocated nor reserved
753 * Basically, this means that this queue can be used for any purpose
754 * @IWL_MVM_QUEUE_RESERVED: queue is reserved but not yet in use
755 * This is the state of a queue that has been dedicated for some RATID
756 * (agg'd or not), but that hasn't yet gone through the actual enablement
757 * of iwl_mvm_enable_txq(), and therefore no traffic can go through it yet.
758 * Note that in this state there is no requirement to already know what TID
759 * should be used with this queue, it is just marked as a queue that will
760 * be used, and shouldn't be allocated to anyone else.
761 * @IWL_MVM_QUEUE_READY: queue is ready to be used
762 * This is the state of a queue that has been fully configured (including
763 * SCD pointers, etc), has a specific RA/TID assigned to it, and can be
764 * used to send traffic.
765 * @IWL_MVM_QUEUE_SHARED: queue is shared, or in a process of becoming shared
766 * This is a state in which a single queue serves more than one TID, all of
767 * which are not aggregated. Note that the queue is only associated to one
768 * RA.
770 enum iwl_mvm_queue_status {
771 IWL_MVM_QUEUE_FREE,
772 IWL_MVM_QUEUE_RESERVED,
773 IWL_MVM_QUEUE_READY,
774 IWL_MVM_QUEUE_SHARED,
777 #define IWL_MVM_DQA_QUEUE_TIMEOUT (5 * HZ)
778 #define IWL_MVM_INVALID_QUEUE 0xFFFF
780 #define IWL_MVM_NUM_CIPHERS 10
783 struct iwl_mvm_txq {
784 struct list_head list;
785 u16 txq_id;
786 atomic_t tx_request;
787 bool stopped;
790 static inline struct iwl_mvm_txq *
791 iwl_mvm_txq_from_mac80211(struct ieee80211_txq *txq)
793 return (void *)txq->drv_priv;
796 static inline struct iwl_mvm_txq *
797 iwl_mvm_txq_from_tid(struct ieee80211_sta *sta, u8 tid)
799 if (tid == IWL_MAX_TID_COUNT)
800 tid = IEEE80211_NUM_TIDS;
802 return (void *)sta->txq[tid]->drv_priv;
806 * struct iwl_mvm_tvqm_txq_info - maps TVQM hw queue to tid
808 * @sta_id: sta id
809 * @txq_tid: txq tid
811 struct iwl_mvm_tvqm_txq_info {
812 u8 sta_id;
813 u8 txq_tid;
816 struct iwl_mvm_dqa_txq_info {
817 u8 ra_sta_id; /* The RA this queue is mapped to, if exists */
818 bool reserved; /* Is this the TXQ reserved for a STA */
819 u8 mac80211_ac; /* The mac80211 AC this queue is mapped to */
820 u8 txq_tid; /* The TID "owner" of this queue*/
821 u16 tid_bitmap; /* Bitmap of the TIDs mapped to this queue */
822 /* Timestamp for inactivation per TID of this queue */
823 unsigned long last_frame_time[IWL_MAX_TID_COUNT + 1];
824 enum iwl_mvm_queue_status status;
827 struct iwl_mvm {
828 /* for logger access */
829 struct device *dev;
831 struct iwl_trans *trans;
832 const struct iwl_fw *fw;
833 const struct iwl_cfg *cfg;
834 struct iwl_phy_db *phy_db;
835 struct ieee80211_hw *hw;
837 /* for protecting access to iwl_mvm */
838 struct mutex mutex;
839 struct list_head async_handlers_list;
840 spinlock_t async_handlers_lock;
841 struct work_struct async_handlers_wk;
843 struct work_struct roc_done_wk;
845 unsigned long init_status;
847 unsigned long status;
849 u32 queue_sync_cookie;
850 atomic_t queue_sync_counter;
852 * for beacon filtering -
853 * currently only one interface can be supported
855 struct iwl_mvm_vif *bf_allowed_vif;
857 bool hw_registered;
858 bool rfkill_safe_init_done;
859 bool support_umac_log;
861 u32 ampdu_ref;
862 bool ampdu_toggle;
864 struct iwl_notif_wait_data notif_wait;
866 union {
867 struct mvm_statistics_rx_v3 rx_stats_v3;
868 struct mvm_statistics_rx rx_stats;
871 struct {
872 u64 rx_time;
873 u64 tx_time;
874 u64 on_time_rf;
875 u64 on_time_scan;
876 } radio_stats, accu_radio_stats;
878 struct list_head add_stream_txqs;
879 union {
880 struct iwl_mvm_dqa_txq_info queue_info[IWL_MAX_HW_QUEUES];
881 struct iwl_mvm_tvqm_txq_info tvqm_info[IWL_MAX_TVQM_QUEUES];
883 struct work_struct add_stream_wk; /* To add streams to queues */
885 const char *nvm_file_name;
886 struct iwl_nvm_data *nvm_data;
887 /* NVM sections */
888 struct iwl_nvm_section nvm_sections[NVM_MAX_NUM_SECTIONS];
890 struct iwl_fw_runtime fwrt;
892 /* EEPROM MAC addresses */
893 struct mac_address addresses[IWL_MVM_MAX_ADDRESSES];
895 /* data related to data path */
896 struct iwl_rx_phy_info last_phy_info;
897 struct ieee80211_sta __rcu *fw_id_to_mac_id[IWL_MVM_STATION_COUNT];
898 u8 rx_ba_sessions;
900 /* configured by mac80211 */
901 u32 rts_threshold;
903 /* Scan status, cmd (pre-allocated) and auxiliary station */
904 unsigned int scan_status;
905 void *scan_cmd;
906 struct iwl_mcast_filter_cmd *mcast_filter_cmd;
907 /* For CDB this is low band scan type, for non-CDB - type. */
908 enum iwl_mvm_scan_type scan_type;
909 enum iwl_mvm_scan_type hb_scan_type;
911 enum iwl_mvm_sched_scan_pass_all_states sched_scan_pass_all;
912 struct delayed_work scan_timeout_dwork;
914 /* max number of simultaneous scans the FW supports */
915 unsigned int max_scans;
917 /* UMAC scan tracking */
918 u32 scan_uid_status[IWL_MVM_MAX_UMAC_SCANS];
920 /* start time of last scan in TSF of the mac that requested the scan */
921 u64 scan_start;
923 /* the vif that requested the current scan */
924 struct iwl_mvm_vif *scan_vif;
926 /* rx chain antennas set through debugfs for the scan command */
927 u8 scan_rx_ant;
929 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING
930 /* broadcast filters to configure for each associated station */
931 const struct iwl_fw_bcast_filter *bcast_filters;
932 #ifdef CONFIG_IWLWIFI_DEBUGFS
933 struct {
934 bool override;
935 struct iwl_bcast_filter_cmd cmd;
936 } dbgfs_bcast_filtering;
937 #endif
938 #endif
940 /* Internal station */
941 struct iwl_mvm_int_sta aux_sta;
942 struct iwl_mvm_int_sta snif_sta;
944 bool last_ebs_successful;
946 u8 scan_last_antenna_idx; /* to toggle TX between antennas */
947 u8 mgmt_last_antenna_idx;
949 /* last smart fifo state that was successfully sent to firmware */
950 enum iwl_sf_state sf_state;
952 #ifdef CONFIG_IWLWIFI_DEBUGFS
953 struct dentry *debugfs_dir;
954 u32 dbgfs_sram_offset, dbgfs_sram_len;
955 u32 dbgfs_prph_reg_addr;
956 bool disable_power_off;
957 bool disable_power_off_d3;
958 bool beacon_inject_active;
960 bool scan_iter_notif_enabled;
962 struct debugfs_blob_wrapper nvm_hw_blob;
963 struct debugfs_blob_wrapper nvm_sw_blob;
964 struct debugfs_blob_wrapper nvm_calib_blob;
965 struct debugfs_blob_wrapper nvm_prod_blob;
966 struct debugfs_blob_wrapper nvm_phy_sku_blob;
967 struct debugfs_blob_wrapper nvm_reg_blob;
969 struct iwl_mvm_frame_stats drv_rx_stats;
970 spinlock_t drv_stats_lock;
971 u16 dbgfs_rx_phyinfo;
972 #endif
974 struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX];
976 struct list_head time_event_list;
977 spinlock_t time_event_lock;
980 * A bitmap indicating the index of the key in use. The firmware
981 * can hold 16 keys at most. Reflect this fact.
983 unsigned long fw_key_table[BITS_TO_LONGS(STA_KEY_MAX_NUM)];
984 u8 fw_key_deleted[STA_KEY_MAX_NUM];
986 u8 vif_count;
987 struct ieee80211_vif __rcu *vif_id_to_mac[NUM_MAC_INDEX_DRIVER];
989 /* -1 for always, 0 for never, >0 for that many times */
990 s8 fw_restart;
991 u8 *error_recovery_buf;
993 #ifdef CONFIG_IWLWIFI_LEDS
994 struct led_classdev led;
995 #endif
997 struct ieee80211_vif *p2p_device_vif;
999 #ifdef CONFIG_PM
1000 struct wiphy_wowlan_support wowlan;
1001 int gtk_ivlen, gtk_icvlen, ptk_ivlen, ptk_icvlen;
1003 /* sched scan settings for net detect */
1004 struct ieee80211_scan_ies nd_ies;
1005 struct cfg80211_match_set *nd_match_sets;
1006 int n_nd_match_sets;
1007 struct ieee80211_channel **nd_channels;
1008 int n_nd_channels;
1009 bool net_detect;
1010 u8 offload_tid;
1011 #ifdef CONFIG_IWLWIFI_DEBUGFS
1012 bool d3_wake_sysassert;
1013 bool d3_test_active;
1014 u32 d3_test_pme_ptr;
1015 struct ieee80211_vif *keep_vif;
1016 u32 last_netdetect_scans; /* no. of scans in the last net-detect wake */
1017 #endif
1018 #endif
1020 wait_queue_head_t rx_sync_waitq;
1022 /* BT-Coex */
1023 struct iwl_bt_coex_profile_notif last_bt_notif;
1024 struct iwl_bt_coex_ci_cmd last_bt_ci_cmd;
1026 u8 bt_tx_prio;
1027 enum iwl_bt_force_ant_mode bt_force_ant_mode;
1029 /* Aux ROC */
1030 struct list_head aux_roc_te_list;
1032 /* Thermal Throttling and CTkill */
1033 struct iwl_mvm_tt_mgmt thermal_throttle;
1034 #ifdef CONFIG_THERMAL
1035 struct iwl_mvm_thermal_device tz_device;
1036 struct iwl_mvm_cooling_device cooling_dev;
1037 #endif
1039 s32 temperature; /* Celsius */
1041 * Debug option to set the NIC temperature. This option makes the
1042 * driver think this is the actual NIC temperature, and ignore the
1043 * real temperature that is received from the fw
1045 bool temperature_test; /* Debug test temperature is enabled */
1047 unsigned long bt_coex_last_tcm_ts;
1048 struct iwl_mvm_tcm tcm;
1050 u8 uapsd_noagg_bssid_write_idx;
1051 struct mac_address uapsd_noagg_bssids[IWL_MVM_UAPSD_NOAGG_BSSIDS_NUM]
1052 __aligned(2);
1054 struct iwl_time_quota_cmd last_quota_cmd;
1056 #ifdef CONFIG_NL80211_TESTMODE
1057 u32 noa_duration;
1058 struct ieee80211_vif *noa_vif;
1059 #endif
1061 /* Tx queues */
1062 u16 aux_queue;
1063 u16 snif_queue;
1064 u16 probe_queue;
1065 u16 p2p_dev_queue;
1067 /* Indicate if device power save is allowed */
1068 u8 ps_disabled; /* u8 instead of bool to ease debugfs_create_* usage */
1069 /* Indicate if 32Khz external clock is valid */
1070 u32 ext_clock_valid;
1072 struct ieee80211_vif __rcu *csa_vif;
1073 struct ieee80211_vif __rcu *csa_tx_blocked_vif;
1074 u8 csa_tx_block_bcn_timeout;
1076 /* system time of last beacon (for AP/GO interface) */
1077 u32 ap_last_beacon_gp2;
1079 /* indicates that we transmitted the last beacon */
1080 bool ibss_manager;
1082 bool lar_regdom_set;
1083 enum iwl_mcc_source mcc_src;
1085 /* TDLS channel switch data */
1086 struct {
1087 struct delayed_work dwork;
1088 enum iwl_mvm_tdls_cs_state state;
1091 * Current cs sta - might be different from periodic cs peer
1092 * station. Value is meaningless when the cs-state is idle.
1094 u8 cur_sta_id;
1096 /* TDLS periodic channel-switch peer */
1097 struct {
1098 u8 sta_id;
1099 u8 op_class;
1100 bool initiator; /* are we the link initiator */
1101 struct cfg80211_chan_def chandef;
1102 struct sk_buff *skb; /* ch sw template */
1103 u32 ch_sw_tm_ie;
1105 /* timestamp of last ch-sw request sent (GP2 time) */
1106 u32 sent_timestamp;
1107 } peer;
1108 } tdls_cs;
1111 u32 ciphers[IWL_MVM_NUM_CIPHERS];
1112 struct ieee80211_cipher_scheme cs[IWL_UCODE_MAX_CS];
1114 struct cfg80211_ftm_responder_stats ftm_resp_stats;
1115 struct {
1116 struct cfg80211_pmsr_request *req;
1117 struct wireless_dev *req_wdev;
1118 struct list_head loc_list;
1119 int responses[IWL_MVM_TOF_MAX_APS];
1120 } ftm_initiator;
1122 struct {
1123 u8 d0i3_resp;
1124 } cmd_ver;
1126 struct ieee80211_vif *nan_vif;
1127 #define IWL_MAX_BAID 32
1128 struct iwl_mvm_baid_data __rcu *baid_map[IWL_MAX_BAID];
1131 * Drop beacons from other APs in AP mode when there are no connected
1132 * clients.
1134 bool drop_bcn_ap_mode;
1136 struct delayed_work cs_tx_unblock_dwork;
1138 /* does a monitor vif exist (only one can exist hence bool) */
1139 bool monitor_on;
1141 /* sniffer data to include in radiotap */
1142 __le16 cur_aid;
1143 u8 cur_bssid[ETH_ALEN];
1146 /* Extract MVM priv from op_mode and _hw */
1147 #define IWL_OP_MODE_GET_MVM(_iwl_op_mode) \
1148 ((struct iwl_mvm *)(_iwl_op_mode)->op_mode_specific)
1150 #define IWL_MAC80211_GET_MVM(_hw) \
1151 IWL_OP_MODE_GET_MVM((struct iwl_op_mode *)((_hw)->priv))
1154 * enum iwl_mvm_status - MVM status bits
1155 * @IWL_MVM_STATUS_HW_RFKILL: HW RF-kill is asserted
1156 * @IWL_MVM_STATUS_HW_CTKILL: CT-kill is active
1157 * @IWL_MVM_STATUS_ROC_RUNNING: remain-on-channel is running
1158 * @IWL_MVM_STATUS_HW_RESTART_REQUESTED: HW restart was requested
1159 * @IWL_MVM_STATUS_IN_HW_RESTART: HW restart is active
1160 * @IWL_MVM_STATUS_ROC_AUX_RUNNING: AUX remain-on-channel is running
1161 * @IWL_MVM_STATUS_FIRMWARE_RUNNING: firmware is running
1162 * @IWL_MVM_STATUS_NEED_FLUSH_P2P: need to flush P2P bcast STA
1163 * @IWL_MVM_STATUS_IN_D3: in D3 (or at least about to go into it)
1165 enum iwl_mvm_status {
1166 IWL_MVM_STATUS_HW_RFKILL,
1167 IWL_MVM_STATUS_HW_CTKILL,
1168 IWL_MVM_STATUS_ROC_RUNNING,
1169 IWL_MVM_STATUS_HW_RESTART_REQUESTED,
1170 IWL_MVM_STATUS_IN_HW_RESTART,
1171 IWL_MVM_STATUS_ROC_AUX_RUNNING,
1172 IWL_MVM_STATUS_FIRMWARE_RUNNING,
1173 IWL_MVM_STATUS_NEED_FLUSH_P2P,
1174 IWL_MVM_STATUS_IN_D3,
1177 /* Keep track of completed init configuration */
1178 enum iwl_mvm_init_status {
1179 IWL_MVM_INIT_STATUS_THERMAL_INIT_COMPLETE = BIT(0),
1180 IWL_MVM_INIT_STATUS_LEDS_INIT_COMPLETE = BIT(1),
1183 static inline bool iwl_mvm_is_radio_killed(struct iwl_mvm *mvm)
1185 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status) ||
1186 test_bit(IWL_MVM_STATUS_HW_CTKILL, &mvm->status);
1189 static inline bool iwl_mvm_is_radio_hw_killed(struct iwl_mvm *mvm)
1191 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status);
1194 static inline bool iwl_mvm_firmware_running(struct iwl_mvm *mvm)
1196 return test_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1199 /* Must be called with rcu_read_lock() held and it can only be
1200 * released when mvmsta is not needed anymore.
1202 static inline struct iwl_mvm_sta *
1203 iwl_mvm_sta_from_staid_rcu(struct iwl_mvm *mvm, u8 sta_id)
1205 struct ieee80211_sta *sta;
1207 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id))
1208 return NULL;
1210 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]);
1212 /* This can happen if the station has been removed right now */
1213 if (IS_ERR_OR_NULL(sta))
1214 return NULL;
1216 return iwl_mvm_sta_from_mac80211(sta);
1219 static inline struct iwl_mvm_sta *
1220 iwl_mvm_sta_from_staid_protected(struct iwl_mvm *mvm, u8 sta_id)
1222 struct ieee80211_sta *sta;
1224 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id))
1225 return NULL;
1227 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id],
1228 lockdep_is_held(&mvm->mutex));
1230 /* This can happen if the station has been removed right now */
1231 if (IS_ERR_OR_NULL(sta))
1232 return NULL;
1234 return iwl_mvm_sta_from_mac80211(sta);
1237 static inline struct ieee80211_vif *
1238 iwl_mvm_rcu_dereference_vif_id(struct iwl_mvm *mvm, u8 vif_id, bool rcu)
1240 if (WARN_ON(vif_id >= ARRAY_SIZE(mvm->vif_id_to_mac)))
1241 return NULL;
1243 if (rcu)
1244 return rcu_dereference(mvm->vif_id_to_mac[vif_id]);
1246 return rcu_dereference_protected(mvm->vif_id_to_mac[vif_id],
1247 lockdep_is_held(&mvm->mutex));
1250 static inline bool iwl_mvm_is_adaptive_dwell_supported(struct iwl_mvm *mvm)
1252 return fw_has_api(&mvm->fw->ucode_capa,
1253 IWL_UCODE_TLV_API_ADAPTIVE_DWELL);
1256 static inline bool iwl_mvm_is_adaptive_dwell_v2_supported(struct iwl_mvm *mvm)
1258 return fw_has_api(&mvm->fw->ucode_capa,
1259 IWL_UCODE_TLV_API_ADAPTIVE_DWELL_V2);
1262 static inline bool iwl_mvm_is_adwell_hb_ap_num_supported(struct iwl_mvm *mvm)
1264 return fw_has_api(&mvm->fw->ucode_capa,
1265 IWL_UCODE_TLV_API_ADWELL_HB_DEF_N_AP);
1268 static inline bool iwl_mvm_is_oce_supported(struct iwl_mvm *mvm)
1270 /* OCE should never be enabled for LMAC scan FWs */
1271 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_OCE);
1274 static inline bool iwl_mvm_is_frag_ebs_supported(struct iwl_mvm *mvm)
1276 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_FRAG_EBS);
1279 static inline bool iwl_mvm_is_short_beacon_notif_supported(struct iwl_mvm *mvm)
1281 return fw_has_api(&mvm->fw->ucode_capa,
1282 IWL_UCODE_TLV_API_SHORT_BEACON_NOTIF);
1285 static inline bool iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue)
1287 return (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE) &&
1288 (queue <= IWL_MVM_DQA_MAX_DATA_QUEUE);
1291 static inline bool iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue)
1293 return (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE) &&
1294 (queue <= IWL_MVM_DQA_MAX_MGMT_QUEUE);
1297 static inline bool iwl_mvm_is_lar_supported(struct iwl_mvm *mvm)
1299 bool nvm_lar = mvm->nvm_data->lar_enabled;
1300 bool tlv_lar = fw_has_capa(&mvm->fw->ucode_capa,
1301 IWL_UCODE_TLV_CAPA_LAR_SUPPORT);
1304 * Enable LAR only if it is supported by the FW (TLV) &&
1305 * enabled in the NVM
1307 if (mvm->cfg->nvm_type == IWL_NVM_EXT)
1308 return nvm_lar && tlv_lar;
1309 else
1310 return tlv_lar;
1313 static inline bool iwl_mvm_is_wifi_mcc_supported(struct iwl_mvm *mvm)
1315 return fw_has_api(&mvm->fw->ucode_capa,
1316 IWL_UCODE_TLV_API_WIFI_MCC_UPDATE) ||
1317 fw_has_capa(&mvm->fw->ucode_capa,
1318 IWL_UCODE_TLV_CAPA_LAR_MULTI_MCC);
1321 static inline bool iwl_mvm_bt_is_rrc_supported(struct iwl_mvm *mvm)
1323 return fw_has_capa(&mvm->fw->ucode_capa,
1324 IWL_UCODE_TLV_CAPA_BT_COEX_RRC) &&
1325 IWL_MVM_BT_COEX_RRC;
1328 static inline bool iwl_mvm_is_csum_supported(struct iwl_mvm *mvm)
1330 return fw_has_capa(&mvm->fw->ucode_capa,
1331 IWL_UCODE_TLV_CAPA_CSUM_SUPPORT) &&
1332 !IWL_MVM_HW_CSUM_DISABLE;
1335 static inline bool iwl_mvm_is_mplut_supported(struct iwl_mvm *mvm)
1337 return fw_has_capa(&mvm->fw->ucode_capa,
1338 IWL_UCODE_TLV_CAPA_BT_MPLUT_SUPPORT) &&
1339 IWL_MVM_BT_COEX_MPLUT;
1342 static inline
1343 bool iwl_mvm_is_p2p_scm_uapsd_supported(struct iwl_mvm *mvm)
1345 return fw_has_capa(&mvm->fw->ucode_capa,
1346 IWL_UCODE_TLV_CAPA_P2P_SCM_UAPSD) &&
1347 !(iwlwifi_mod_params.uapsd_disable &
1348 IWL_DISABLE_UAPSD_P2P_CLIENT);
1351 static inline bool iwl_mvm_has_new_rx_api(struct iwl_mvm *mvm)
1353 return fw_has_capa(&mvm->fw->ucode_capa,
1354 IWL_UCODE_TLV_CAPA_MULTI_QUEUE_RX_SUPPORT);
1357 static inline bool iwl_mvm_has_new_tx_api(struct iwl_mvm *mvm)
1359 /* TODO - replace with TLV once defined */
1360 return mvm->trans->trans_cfg->use_tfh;
1363 static inline bool iwl_mvm_has_unified_ucode(struct iwl_mvm *mvm)
1365 /* TODO - better define this */
1366 return mvm->trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_22000;
1369 static inline bool iwl_mvm_is_cdb_supported(struct iwl_mvm *mvm)
1372 * TODO:
1373 * The issue of how to determine CDB APIs and usage is still not fully
1374 * defined.
1375 * There is a compilation for CDB and non-CDB FW, but there may
1376 * be also runtime check.
1377 * For now there is a TLV for checking compilation mode, but a
1378 * runtime check will also have to be here - once defined.
1380 return fw_has_capa(&mvm->fw->ucode_capa,
1381 IWL_UCODE_TLV_CAPA_CDB_SUPPORT);
1384 static inline bool iwl_mvm_cdb_scan_api(struct iwl_mvm *mvm)
1387 * TODO: should this be the same as iwl_mvm_is_cdb_supported()?
1388 * but then there's a little bit of code in scan that won't make
1389 * any sense...
1391 return mvm->trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_22000;
1394 static inline bool iwl_mvm_is_scan_ext_chan_supported(struct iwl_mvm *mvm)
1396 return fw_has_api(&mvm->fw->ucode_capa,
1397 IWL_UCODE_TLV_API_SCAN_EXT_CHAN_VER);
1401 static inline bool iwl_mvm_is_reduced_config_scan_supported(struct iwl_mvm *mvm)
1403 return fw_has_api(&mvm->fw->ucode_capa,
1404 IWL_UCODE_TLV_API_REDUCED_SCAN_CONFIG);
1407 static inline bool iwl_mvm_is_band_in_rx_supported(struct iwl_mvm *mvm)
1409 return fw_has_api(&mvm->fw->ucode_capa,
1410 IWL_UCODE_TLV_API_BAND_IN_RX_DATA);
1413 static inline bool iwl_mvm_has_new_rx_stats_api(struct iwl_mvm *mvm)
1415 return fw_has_api(&mvm->fw->ucode_capa,
1416 IWL_UCODE_TLV_API_NEW_RX_STATS);
1419 static inline bool iwl_mvm_has_quota_low_latency(struct iwl_mvm *mvm)
1421 return fw_has_api(&mvm->fw->ucode_capa,
1422 IWL_UCODE_TLV_API_QUOTA_LOW_LATENCY);
1425 static inline bool iwl_mvm_has_tlc_offload(const struct iwl_mvm *mvm)
1427 return fw_has_capa(&mvm->fw->ucode_capa,
1428 IWL_UCODE_TLV_CAPA_TLC_OFFLOAD);
1431 static inline struct agg_tx_status *
1432 iwl_mvm_get_agg_status(struct iwl_mvm *mvm, void *tx_resp)
1434 if (iwl_mvm_has_new_tx_api(mvm))
1435 return &((struct iwl_mvm_tx_resp *)tx_resp)->status;
1436 else
1437 return ((struct iwl_mvm_tx_resp_v3 *)tx_resp)->status;
1440 static inline bool iwl_mvm_is_tt_in_fw(struct iwl_mvm *mvm)
1442 /* these two TLV are redundant since the responsibility to CT-kill by
1443 * FW happens only after we send at least one command of
1444 * temperature THs report.
1446 return fw_has_capa(&mvm->fw->ucode_capa,
1447 IWL_UCODE_TLV_CAPA_CT_KILL_BY_FW) &&
1448 fw_has_capa(&mvm->fw->ucode_capa,
1449 IWL_UCODE_TLV_CAPA_TEMP_THS_REPORT_SUPPORT);
1452 static inline bool iwl_mvm_is_ctdp_supported(struct iwl_mvm *mvm)
1454 return fw_has_capa(&mvm->fw->ucode_capa,
1455 IWL_UCODE_TLV_CAPA_CTDP_SUPPORT);
1458 extern const u8 iwl_mvm_ac_to_tx_fifo[];
1459 extern const u8 iwl_mvm_ac_to_gen2_tx_fifo[];
1461 static inline u8 iwl_mvm_mac_ac_to_tx_fifo(struct iwl_mvm *mvm,
1462 enum ieee80211_ac_numbers ac)
1464 return iwl_mvm_has_new_tx_api(mvm) ?
1465 iwl_mvm_ac_to_gen2_tx_fifo[ac] : iwl_mvm_ac_to_tx_fifo[ac];
1468 struct iwl_rate_info {
1469 u8 plcp; /* uCode API: IWL_RATE_6M_PLCP, etc. */
1470 u8 plcp_siso; /* uCode API: IWL_RATE_SISO_6M_PLCP, etc. */
1471 u8 plcp_mimo2; /* uCode API: IWL_RATE_MIMO2_6M_PLCP, etc. */
1472 u8 plcp_mimo3; /* uCode API: IWL_RATE_MIMO3_6M_PLCP, etc. */
1473 u8 ieee; /* MAC header: IWL_RATE_6M_IEEE, etc. */
1476 void __iwl_mvm_mac_stop(struct iwl_mvm *mvm);
1477 int __iwl_mvm_mac_start(struct iwl_mvm *mvm);
1479 /******************
1480 * MVM Methods
1481 ******************/
1482 /* uCode */
1483 int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm);
1485 /* Utils */
1486 int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags,
1487 enum nl80211_band band);
1488 void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags,
1489 enum nl80211_band band,
1490 struct ieee80211_tx_rate *r);
1491 u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx);
1492 u8 iwl_mvm_mac80211_ac_to_ucode_ac(enum ieee80211_ac_numbers ac);
1493 void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm);
1494 u8 first_antenna(u8 mask);
1495 u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx);
1496 void iwl_mvm_get_sync_time(struct iwl_mvm *mvm, u32 *gp2, u64 *boottime);
1497 u32 iwl_mvm_get_systime(struct iwl_mvm *mvm);
1499 /* Tx / Host Commands */
1500 int __must_check iwl_mvm_send_cmd(struct iwl_mvm *mvm,
1501 struct iwl_host_cmd *cmd);
1502 int __must_check iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u32 id,
1503 u32 flags, u16 len, const void *data);
1504 int __must_check iwl_mvm_send_cmd_status(struct iwl_mvm *mvm,
1505 struct iwl_host_cmd *cmd,
1506 u32 *status);
1507 int __must_check iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u32 id,
1508 u16 len, const void *data,
1509 u32 *status);
1510 int iwl_mvm_tx_skb_sta(struct iwl_mvm *mvm, struct sk_buff *skb,
1511 struct ieee80211_sta *sta);
1512 int iwl_mvm_tx_skb_non_sta(struct iwl_mvm *mvm, struct sk_buff *skb);
1513 void iwl_mvm_set_tx_cmd(struct iwl_mvm *mvm, struct sk_buff *skb,
1514 struct iwl_tx_cmd *tx_cmd,
1515 struct ieee80211_tx_info *info, u8 sta_id);
1516 void iwl_mvm_set_tx_cmd_rate(struct iwl_mvm *mvm, struct iwl_tx_cmd *tx_cmd,
1517 struct ieee80211_tx_info *info,
1518 struct ieee80211_sta *sta, __le16 fc);
1519 void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq);
1520 unsigned int iwl_mvm_max_amsdu_size(struct iwl_mvm *mvm,
1521 struct ieee80211_sta *sta,
1522 unsigned int tid);
1524 #ifdef CONFIG_IWLWIFI_DEBUG
1525 const char *iwl_mvm_get_tx_fail_reason(u32 status);
1526 #else
1527 static inline const char *iwl_mvm_get_tx_fail_reason(u32 status) { return ""; }
1528 #endif
1529 int iwl_mvm_flush_tx_path(struct iwl_mvm *mvm, u32 tfd_msk, u32 flags);
1530 int iwl_mvm_flush_sta(struct iwl_mvm *mvm, void *sta, bool internal, u32 flags);
1531 int iwl_mvm_flush_sta_tids(struct iwl_mvm *mvm, u32 sta_id,
1532 u16 tids, u32 flags);
1534 void iwl_mvm_async_handlers_purge(struct iwl_mvm *mvm);
1536 static inline void iwl_mvm_set_tx_cmd_ccmp(struct ieee80211_tx_info *info,
1537 struct iwl_tx_cmd *tx_cmd)
1539 struct ieee80211_key_conf *keyconf = info->control.hw_key;
1541 tx_cmd->sec_ctl = TX_CMD_SEC_CCM;
1542 memcpy(tx_cmd->key, keyconf->key, keyconf->keylen);
1545 static inline void iwl_mvm_wait_for_async_handlers(struct iwl_mvm *mvm)
1547 flush_work(&mvm->async_handlers_wk);
1550 /* Statistics */
1551 void iwl_mvm_handle_rx_statistics(struct iwl_mvm *mvm,
1552 struct iwl_rx_packet *pkt);
1553 void iwl_mvm_rx_statistics(struct iwl_mvm *mvm,
1554 struct iwl_rx_cmd_buffer *rxb);
1555 int iwl_mvm_request_statistics(struct iwl_mvm *mvm, bool clear);
1556 void iwl_mvm_accu_radio_stats(struct iwl_mvm *mvm);
1558 /* NVM */
1559 int iwl_nvm_init(struct iwl_mvm *mvm);
1560 int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm);
1562 static inline u8 iwl_mvm_get_valid_tx_ant(struct iwl_mvm *mvm)
1564 return mvm->nvm_data && mvm->nvm_data->valid_tx_ant ?
1565 mvm->fw->valid_tx_ant & mvm->nvm_data->valid_tx_ant :
1566 mvm->fw->valid_tx_ant;
1569 static inline u8 iwl_mvm_get_valid_rx_ant(struct iwl_mvm *mvm)
1571 return mvm->nvm_data && mvm->nvm_data->valid_rx_ant ?
1572 mvm->fw->valid_rx_ant & mvm->nvm_data->valid_rx_ant :
1573 mvm->fw->valid_rx_ant;
1576 static inline void iwl_mvm_toggle_tx_ant(struct iwl_mvm *mvm, u8 *ant)
1578 *ant = iwl_mvm_next_antenna(mvm, iwl_mvm_get_valid_tx_ant(mvm), *ant);
1581 static inline u32 iwl_mvm_get_phy_config(struct iwl_mvm *mvm)
1583 u32 phy_config = ~(FW_PHY_CFG_TX_CHAIN |
1584 FW_PHY_CFG_RX_CHAIN);
1585 u32 valid_rx_ant = iwl_mvm_get_valid_rx_ant(mvm);
1586 u32 valid_tx_ant = iwl_mvm_get_valid_tx_ant(mvm);
1588 phy_config |= valid_tx_ant << FW_PHY_CFG_TX_CHAIN_POS |
1589 valid_rx_ant << FW_PHY_CFG_RX_CHAIN_POS;
1591 return mvm->fw->phy_config & phy_config;
1594 int iwl_mvm_up(struct iwl_mvm *mvm);
1595 int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm);
1597 int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm);
1598 bool iwl_mvm_bcast_filter_build_cmd(struct iwl_mvm *mvm,
1599 struct iwl_bcast_filter_cmd *cmd);
1602 * FW notifications / CMD responses handlers
1603 * Convention: iwl_mvm_rx_<NAME OF THE CMD>
1605 void iwl_mvm_rx_rx_phy_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1606 void iwl_mvm_rx_rx_mpdu(struct iwl_mvm *mvm, struct napi_struct *napi,
1607 struct iwl_rx_cmd_buffer *rxb);
1608 void iwl_mvm_rx_mpdu_mq(struct iwl_mvm *mvm, struct napi_struct *napi,
1609 struct iwl_rx_cmd_buffer *rxb, int queue);
1610 void iwl_mvm_rx_monitor_no_data(struct iwl_mvm *mvm, struct napi_struct *napi,
1611 struct iwl_rx_cmd_buffer *rxb, int queue);
1612 void iwl_mvm_rx_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi,
1613 struct iwl_rx_cmd_buffer *rxb, int queue);
1614 void iwl_mvm_rx_bar_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi,
1615 struct iwl_rx_cmd_buffer *rxb, int queue);
1616 int iwl_mvm_notify_rx_queue(struct iwl_mvm *mvm, u32 rxq_mask,
1617 const u8 *data, u32 count, bool async);
1618 void iwl_mvm_rx_queue_notif(struct iwl_mvm *mvm, struct napi_struct *napi,
1619 struct iwl_rx_cmd_buffer *rxb, int queue);
1620 void iwl_mvm_rx_tx_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1621 void iwl_mvm_mfu_assert_dump_notif(struct iwl_mvm *mvm,
1622 struct iwl_rx_cmd_buffer *rxb);
1623 void iwl_mvm_send_recovery_cmd(struct iwl_mvm *mvm, u32 flags);
1624 void iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1625 void iwl_mvm_rx_ant_coupling_notif(struct iwl_mvm *mvm,
1626 struct iwl_rx_cmd_buffer *rxb);
1627 void iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1628 void iwl_mvm_rx_card_state_notif(struct iwl_mvm *mvm,
1629 struct iwl_rx_cmd_buffer *rxb);
1630 void iwl_mvm_rx_mfuart_notif(struct iwl_mvm *mvm,
1631 struct iwl_rx_cmd_buffer *rxb);
1632 void iwl_mvm_rx_shared_mem_cfg_notif(struct iwl_mvm *mvm,
1633 struct iwl_rx_cmd_buffer *rxb);
1635 /* MVM PHY */
1636 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1637 struct cfg80211_chan_def *chandef,
1638 u8 chains_static, u8 chains_dynamic);
1639 int iwl_mvm_phy_ctxt_changed(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1640 struct cfg80211_chan_def *chandef,
1641 u8 chains_static, u8 chains_dynamic);
1642 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm,
1643 struct iwl_mvm_phy_ctxt *ctxt);
1644 void iwl_mvm_phy_ctxt_unref(struct iwl_mvm *mvm,
1645 struct iwl_mvm_phy_ctxt *ctxt);
1646 int iwl_mvm_phy_ctx_count(struct iwl_mvm *mvm);
1647 u8 iwl_mvm_get_channel_width(struct cfg80211_chan_def *chandef);
1648 u8 iwl_mvm_get_ctrl_pos(struct cfg80211_chan_def *chandef);
1650 /* MAC (virtual interface) programming */
1651 int iwl_mvm_mac_ctxt_init(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1652 int iwl_mvm_mac_ctxt_add(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1653 int iwl_mvm_mac_ctxt_changed(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1654 bool force_assoc_off, const u8 *bssid_override);
1655 int iwl_mvm_mac_ctxt_remove(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1656 int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm,
1657 struct ieee80211_vif *vif);
1658 int iwl_mvm_mac_ctxt_send_beacon(struct iwl_mvm *mvm,
1659 struct ieee80211_vif *vif,
1660 struct sk_buff *beacon);
1661 int iwl_mvm_mac_ctxt_send_beacon_cmd(struct iwl_mvm *mvm,
1662 struct sk_buff *beacon,
1663 void *data, int len);
1664 u8 iwl_mvm_mac_ctxt_get_lowest_rate(struct ieee80211_tx_info *info,
1665 struct ieee80211_vif *vif);
1666 void iwl_mvm_mac_ctxt_set_tim(struct iwl_mvm *mvm,
1667 __le32 *tim_index, __le32 *tim_size,
1668 u8 *beacon, u32 frame_size);
1669 void iwl_mvm_rx_beacon_notif(struct iwl_mvm *mvm,
1670 struct iwl_rx_cmd_buffer *rxb);
1671 void iwl_mvm_rx_missed_beacons_notif(struct iwl_mvm *mvm,
1672 struct iwl_rx_cmd_buffer *rxb);
1673 void iwl_mvm_rx_stored_beacon_notif(struct iwl_mvm *mvm,
1674 struct iwl_rx_cmd_buffer *rxb);
1675 void iwl_mvm_mu_mimo_grp_notif(struct iwl_mvm *mvm,
1676 struct iwl_rx_cmd_buffer *rxb);
1677 void iwl_mvm_sta_pm_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1678 void iwl_mvm_window_status_notif(struct iwl_mvm *mvm,
1679 struct iwl_rx_cmd_buffer *rxb);
1680 void iwl_mvm_mac_ctxt_recalc_tsf_id(struct iwl_mvm *mvm,
1681 struct ieee80211_vif *vif);
1682 void iwl_mvm_probe_resp_data_notif(struct iwl_mvm *mvm,
1683 struct iwl_rx_cmd_buffer *rxb);
1684 void iwl_mvm_rx_missed_vap_notif(struct iwl_mvm *mvm,
1685 struct iwl_rx_cmd_buffer *rxb);
1686 void iwl_mvm_channel_switch_noa_notif(struct iwl_mvm *mvm,
1687 struct iwl_rx_cmd_buffer *rxb);
1688 /* Bindings */
1689 int iwl_mvm_binding_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1690 int iwl_mvm_binding_remove_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1692 /* Quota management */
1693 static inline size_t iwl_mvm_quota_cmd_size(struct iwl_mvm *mvm)
1695 return iwl_mvm_has_quota_low_latency(mvm) ?
1696 sizeof(struct iwl_time_quota_cmd) :
1697 sizeof(struct iwl_time_quota_cmd_v1);
1700 static inline struct iwl_time_quota_data
1701 *iwl_mvm_quota_cmd_get_quota(struct iwl_mvm *mvm,
1702 struct iwl_time_quota_cmd *cmd,
1703 int i)
1705 struct iwl_time_quota_data_v1 *quotas;
1707 if (iwl_mvm_has_quota_low_latency(mvm))
1708 return &cmd->quotas[i];
1710 quotas = (struct iwl_time_quota_data_v1 *)cmd->quotas;
1711 return (struct iwl_time_quota_data *)&quotas[i];
1714 int iwl_mvm_update_quotas(struct iwl_mvm *mvm, bool force_upload,
1715 struct ieee80211_vif *disabled_vif);
1717 /* Scanning */
1718 int iwl_mvm_reg_scan_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1719 struct cfg80211_scan_request *req,
1720 struct ieee80211_scan_ies *ies);
1721 int iwl_mvm_scan_size(struct iwl_mvm *mvm);
1722 int iwl_mvm_scan_stop(struct iwl_mvm *mvm, int type, bool notify);
1723 int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm);
1724 void iwl_mvm_report_scan_aborted(struct iwl_mvm *mvm);
1725 void iwl_mvm_scan_timeout_wk(struct work_struct *work);
1727 /* Scheduled scan */
1728 void iwl_mvm_rx_lmac_scan_complete_notif(struct iwl_mvm *mvm,
1729 struct iwl_rx_cmd_buffer *rxb);
1730 void iwl_mvm_rx_lmac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1731 struct iwl_rx_cmd_buffer *rxb);
1732 int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm,
1733 struct ieee80211_vif *vif,
1734 struct cfg80211_sched_scan_request *req,
1735 struct ieee80211_scan_ies *ies,
1736 int type);
1737 void iwl_mvm_rx_scan_match_found(struct iwl_mvm *mvm,
1738 struct iwl_rx_cmd_buffer *rxb);
1740 /* UMAC scan */
1741 int iwl_mvm_config_scan(struct iwl_mvm *mvm);
1742 void iwl_mvm_rx_umac_scan_complete_notif(struct iwl_mvm *mvm,
1743 struct iwl_rx_cmd_buffer *rxb);
1744 void iwl_mvm_rx_umac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1745 struct iwl_rx_cmd_buffer *rxb);
1747 /* MVM debugfs */
1748 #ifdef CONFIG_IWLWIFI_DEBUGFS
1749 void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, struct dentry *dbgfs_dir);
1750 void iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1751 void iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1752 #else
1753 static inline void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm,
1754 struct dentry *dbgfs_dir)
1757 static inline void
1758 iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1761 static inline void
1762 iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1765 #endif /* CONFIG_IWLWIFI_DEBUGFS */
1767 /* rate scaling */
1768 int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq);
1769 void iwl_mvm_update_frame_stats(struct iwl_mvm *mvm, u32 rate, bool agg);
1770 int rs_pretty_print_rate(char *buf, int bufsz, const u32 rate);
1771 void rs_update_last_rssi(struct iwl_mvm *mvm,
1772 struct iwl_mvm_sta *mvmsta,
1773 struct ieee80211_rx_status *rx_status);
1775 /* power management */
1776 int iwl_mvm_power_update_device(struct iwl_mvm *mvm);
1777 int iwl_mvm_power_update_mac(struct iwl_mvm *mvm);
1778 int iwl_mvm_power_update_ps(struct iwl_mvm *mvm);
1779 int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1780 char *buf, int bufsz);
1782 void iwl_mvm_power_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1783 void iwl_mvm_power_uapsd_misbehaving_ap_notif(struct iwl_mvm *mvm,
1784 struct iwl_rx_cmd_buffer *rxb);
1786 #ifdef CONFIG_IWLWIFI_LEDS
1787 int iwl_mvm_leds_init(struct iwl_mvm *mvm);
1788 void iwl_mvm_leds_exit(struct iwl_mvm *mvm);
1789 void iwl_mvm_leds_sync(struct iwl_mvm *mvm);
1790 #else
1791 static inline int iwl_mvm_leds_init(struct iwl_mvm *mvm)
1793 return 0;
1795 static inline void iwl_mvm_leds_exit(struct iwl_mvm *mvm)
1798 static inline void iwl_mvm_leds_sync(struct iwl_mvm *mvm)
1801 #endif
1803 /* D3 (WoWLAN, NetDetect) */
1804 int iwl_mvm_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wowlan);
1805 int iwl_mvm_resume(struct ieee80211_hw *hw);
1806 void iwl_mvm_set_wakeup(struct ieee80211_hw *hw, bool enabled);
1807 void iwl_mvm_set_rekey_data(struct ieee80211_hw *hw,
1808 struct ieee80211_vif *vif,
1809 struct cfg80211_gtk_rekey_data *data);
1810 void iwl_mvm_ipv6_addr_change(struct ieee80211_hw *hw,
1811 struct ieee80211_vif *vif,
1812 struct inet6_dev *idev);
1813 void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw,
1814 struct ieee80211_vif *vif, int idx);
1815 extern const struct file_operations iwl_dbgfs_d3_test_ops;
1816 struct iwl_wowlan_status *iwl_mvm_send_wowlan_get_status(struct iwl_mvm *mvm);
1817 #ifdef CONFIG_PM
1818 void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm,
1819 struct ieee80211_vif *vif);
1820 #else
1821 static inline void
1822 iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1825 #endif
1826 void iwl_mvm_set_wowlan_qos_seq(struct iwl_mvm_sta *mvm_ap_sta,
1827 struct iwl_wowlan_config_cmd *cmd);
1828 int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm,
1829 struct ieee80211_vif *vif,
1830 bool disable_offloading,
1831 bool offload_ns,
1832 u32 cmd_flags);
1834 /* BT Coex */
1835 int iwl_mvm_send_bt_init_conf(struct iwl_mvm *mvm);
1836 void iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm,
1837 struct iwl_rx_cmd_buffer *rxb);
1838 void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1839 enum ieee80211_rssi_event_data);
1840 void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm);
1841 u16 iwl_mvm_coex_agg_time_limit(struct iwl_mvm *mvm,
1842 struct ieee80211_sta *sta);
1843 bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm,
1844 struct ieee80211_sta *sta);
1845 bool iwl_mvm_bt_coex_is_ant_avail(struct iwl_mvm *mvm, u8 ant);
1846 bool iwl_mvm_bt_coex_is_shared_ant_avail(struct iwl_mvm *mvm);
1847 bool iwl_mvm_bt_coex_is_tpc_allowed(struct iwl_mvm *mvm,
1848 enum nl80211_band band);
1849 u8 iwl_mvm_bt_coex_get_single_ant_msk(struct iwl_mvm *mvm, u8 enabled_ants);
1850 u8 iwl_mvm_bt_coex_tx_prio(struct iwl_mvm *mvm, struct ieee80211_hdr *hdr,
1851 struct ieee80211_tx_info *info, u8 ac);
1853 /* beacon filtering */
1854 #ifdef CONFIG_IWLWIFI_DEBUGFS
1855 void
1856 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1857 struct iwl_beacon_filter_cmd *cmd);
1858 #else
1859 static inline void
1860 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1861 struct iwl_beacon_filter_cmd *cmd)
1863 #endif
1864 int iwl_mvm_enable_beacon_filter(struct iwl_mvm *mvm,
1865 struct ieee80211_vif *vif,
1866 u32 flags);
1867 int iwl_mvm_disable_beacon_filter(struct iwl_mvm *mvm,
1868 struct ieee80211_vif *vif,
1869 u32 flags);
1870 /* SMPS */
1871 void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1872 enum iwl_mvm_smps_type_request req_type,
1873 enum ieee80211_smps_mode smps_request);
1874 bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm);
1876 /* Low latency */
1877 int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1878 bool low_latency,
1879 enum iwl_mvm_low_latency_cause cause);
1880 /* get SystemLowLatencyMode - only needed for beacon threshold? */
1881 bool iwl_mvm_low_latency(struct iwl_mvm *mvm);
1882 bool iwl_mvm_low_latency_band(struct iwl_mvm *mvm, enum nl80211_band band);
1883 void iwl_mvm_send_low_latency_cmd(struct iwl_mvm *mvm, bool low_latency,
1884 u16 mac_id);
1886 /* get VMACLowLatencyMode */
1887 static inline bool iwl_mvm_vif_low_latency(struct iwl_mvm_vif *mvmvif)
1890 * should this consider associated/active/... state?
1892 * Normally low-latency should only be active on interfaces
1893 * that are active, but at least with debugfs it can also be
1894 * enabled on interfaces that aren't active. However, when
1895 * interface aren't active then they aren't added into the
1896 * binding, so this has no real impact. For now, just return
1897 * the current desired low-latency state.
1899 return mvmvif->low_latency_actual;
1902 static inline
1903 void iwl_mvm_vif_set_low_latency(struct iwl_mvm_vif *mvmvif, bool set,
1904 enum iwl_mvm_low_latency_cause cause)
1906 u8 new_state;
1908 if (set)
1909 mvmvif->low_latency |= cause;
1910 else
1911 mvmvif->low_latency &= ~cause;
1914 * if LOW_LATENCY_DEBUGFS_FORCE_ENABLE is enabled no changes are
1915 * allowed to actual mode.
1917 if (mvmvif->low_latency & LOW_LATENCY_DEBUGFS_FORCE_ENABLE &&
1918 cause != LOW_LATENCY_DEBUGFS_FORCE_ENABLE)
1919 return;
1921 if (cause == LOW_LATENCY_DEBUGFS_FORCE_ENABLE && set)
1923 * We enter force state
1925 new_state = !!(mvmvif->low_latency &
1926 LOW_LATENCY_DEBUGFS_FORCE);
1927 else
1929 * Check if any other one set low latency
1931 new_state = !!(mvmvif->low_latency &
1932 ~(LOW_LATENCY_DEBUGFS_FORCE_ENABLE |
1933 LOW_LATENCY_DEBUGFS_FORCE));
1935 mvmvif->low_latency_actual = new_state;
1938 /* Return a bitmask with all the hw supported queues, except for the
1939 * command queue, which can't be flushed.
1941 static inline u32 iwl_mvm_flushable_queues(struct iwl_mvm *mvm)
1943 return ((BIT(mvm->trans->trans_cfg->base_params->num_of_queues) - 1) &
1944 ~BIT(IWL_MVM_DQA_CMD_QUEUE));
1947 static inline void iwl_mvm_stop_device(struct iwl_mvm *mvm)
1949 lockdep_assert_held(&mvm->mutex);
1950 iwl_fw_cancel_timestamp(&mvm->fwrt);
1951 clear_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1952 iwl_fw_dbg_stop_sync(&mvm->fwrt);
1953 iwl_trans_stop_device(mvm->trans);
1954 iwl_free_fw_paging(&mvm->fwrt);
1955 iwl_fw_dump_conf_clear(&mvm->fwrt);
1958 /* Re-configure the SCD for a queue that has already been configured */
1959 int iwl_mvm_reconfig_scd(struct iwl_mvm *mvm, int queue, int fifo, int sta_id,
1960 int tid, int frame_limit, u16 ssn);
1962 /* Thermal management and CT-kill */
1963 void iwl_mvm_tt_tx_backoff(struct iwl_mvm *mvm, u32 backoff);
1964 void iwl_mvm_tt_temp_changed(struct iwl_mvm *mvm, u32 temp);
1965 void iwl_mvm_temp_notif(struct iwl_mvm *mvm,
1966 struct iwl_rx_cmd_buffer *rxb);
1967 void iwl_mvm_tt_handler(struct iwl_mvm *mvm);
1968 void iwl_mvm_thermal_initialize(struct iwl_mvm *mvm, u32 min_backoff);
1969 void iwl_mvm_thermal_exit(struct iwl_mvm *mvm);
1970 void iwl_mvm_set_hw_ctkill_state(struct iwl_mvm *mvm, bool state);
1971 int iwl_mvm_get_temp(struct iwl_mvm *mvm, s32 *temp);
1972 void iwl_mvm_ct_kill_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1973 void iwl_mvm_enter_ctkill(struct iwl_mvm *mvm);
1974 int iwl_mvm_send_temp_report_ths_cmd(struct iwl_mvm *mvm);
1975 int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 budget);
1977 /* Location Aware Regulatory */
1978 struct iwl_mcc_update_resp *
1979 iwl_mvm_update_mcc(struct iwl_mvm *mvm, const char *alpha2,
1980 enum iwl_mcc_source src_id);
1981 int iwl_mvm_init_mcc(struct iwl_mvm *mvm);
1982 void iwl_mvm_rx_chub_update_mcc(struct iwl_mvm *mvm,
1983 struct iwl_rx_cmd_buffer *rxb);
1984 struct ieee80211_regdomain *iwl_mvm_get_regdomain(struct wiphy *wiphy,
1985 const char *alpha2,
1986 enum iwl_mcc_source src_id,
1987 bool *changed);
1988 struct ieee80211_regdomain *iwl_mvm_get_current_regdomain(struct iwl_mvm *mvm,
1989 bool *changed);
1990 int iwl_mvm_init_fw_regd(struct iwl_mvm *mvm);
1991 void iwl_mvm_update_changed_regdom(struct iwl_mvm *mvm);
1993 /* smart fifo */
1994 int iwl_mvm_sf_update(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1995 bool added_vif);
1997 /* FTM responder */
1998 int iwl_mvm_ftm_start_responder(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1999 void iwl_mvm_ftm_restart_responder(struct iwl_mvm *mvm,
2000 struct ieee80211_vif *vif);
2001 void iwl_mvm_ftm_responder_stats(struct iwl_mvm *mvm,
2002 struct iwl_rx_cmd_buffer *rxb);
2004 /* FTM initiator */
2005 void iwl_mvm_ftm_restart(struct iwl_mvm *mvm);
2006 void iwl_mvm_ftm_range_resp(struct iwl_mvm *mvm,
2007 struct iwl_rx_cmd_buffer *rxb);
2008 void iwl_mvm_ftm_lc_notif(struct iwl_mvm *mvm,
2009 struct iwl_rx_cmd_buffer *rxb);
2010 int iwl_mvm_ftm_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2011 struct cfg80211_pmsr_request *request);
2012 void iwl_mvm_ftm_abort(struct iwl_mvm *mvm, struct cfg80211_pmsr_request *req);
2014 /* TDLS */
2017 * We use TID 4 (VI) as a FW-used-only TID when TDLS connections are present.
2018 * This TID is marked as used vs the AP and all connected TDLS peers.
2020 #define IWL_MVM_TDLS_FW_TID 4
2022 int iwl_mvm_tdls_sta_count(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
2023 void iwl_mvm_teardown_tdls_peers(struct iwl_mvm *mvm);
2024 void iwl_mvm_recalc_tdls_state(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2025 bool sta_added);
2026 void iwl_mvm_mac_mgd_protect_tdls_discover(struct ieee80211_hw *hw,
2027 struct ieee80211_vif *vif);
2028 int iwl_mvm_tdls_channel_switch(struct ieee80211_hw *hw,
2029 struct ieee80211_vif *vif,
2030 struct ieee80211_sta *sta, u8 oper_class,
2031 struct cfg80211_chan_def *chandef,
2032 struct sk_buff *tmpl_skb, u32 ch_sw_tm_ie);
2033 void iwl_mvm_tdls_recv_channel_switch(struct ieee80211_hw *hw,
2034 struct ieee80211_vif *vif,
2035 struct ieee80211_tdls_ch_sw_params *params);
2036 void iwl_mvm_tdls_cancel_channel_switch(struct ieee80211_hw *hw,
2037 struct ieee80211_vif *vif,
2038 struct ieee80211_sta *sta);
2039 void iwl_mvm_rx_tdls_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
2040 void iwl_mvm_tdls_ch_switch_work(struct work_struct *work);
2042 void iwl_mvm_sync_rx_queues_internal(struct iwl_mvm *mvm,
2043 struct iwl_mvm_internal_rxq_notif *notif,
2044 u32 size);
2045 void iwl_mvm_reorder_timer_expired(struct timer_list *t);
2046 struct ieee80211_vif *iwl_mvm_get_bss_vif(struct iwl_mvm *mvm);
2047 bool iwl_mvm_is_vif_assoc(struct iwl_mvm *mvm);
2049 #define MVM_TCM_PERIOD_MSEC 500
2050 #define MVM_TCM_PERIOD (HZ * MVM_TCM_PERIOD_MSEC / 1000)
2051 #define MVM_LL_PERIOD (10 * HZ)
2052 void iwl_mvm_tcm_work(struct work_struct *work);
2053 void iwl_mvm_recalc_tcm(struct iwl_mvm *mvm);
2054 void iwl_mvm_pause_tcm(struct iwl_mvm *mvm, bool with_cancel);
2055 void iwl_mvm_resume_tcm(struct iwl_mvm *mvm);
2056 void iwl_mvm_tcm_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
2057 void iwl_mvm_tcm_rm_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
2058 u8 iwl_mvm_tcm_load_percentage(u32 airtime, u32 elapsed);
2060 void iwl_mvm_nic_restart(struct iwl_mvm *mvm, bool fw_error);
2061 unsigned int iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm,
2062 struct ieee80211_vif *vif,
2063 bool tdls, bool cmd_q);
2064 void iwl_mvm_connection_loss(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2065 const char *errmsg);
2066 void iwl_mvm_event_frame_timeout_callback(struct iwl_mvm *mvm,
2067 struct ieee80211_vif *vif,
2068 const struct ieee80211_sta *sta,
2069 u16 tid);
2071 int iwl_mvm_sar_select_profile(struct iwl_mvm *mvm, int prof_a, int prof_b);
2072 int iwl_mvm_get_sar_geo_profile(struct iwl_mvm *mvm);
2073 int iwl_mvm_ppag_send_cmd(struct iwl_mvm *mvm);
2074 #ifdef CONFIG_IWLWIFI_DEBUGFS
2075 void iwl_mvm_sta_add_debugfs(struct ieee80211_hw *hw,
2076 struct ieee80211_vif *vif,
2077 struct ieee80211_sta *sta,
2078 struct dentry *dir);
2079 #endif
2081 static inline u8 iwl_mvm_phy_band_from_nl80211(enum nl80211_band band)
2083 switch (band) {
2084 case NL80211_BAND_2GHZ:
2085 return PHY_BAND_24;
2086 case NL80211_BAND_5GHZ:
2087 return PHY_BAND_5;
2088 default:
2089 WARN_ONCE(1, "Unsupported band (%u)\n", band);
2090 return PHY_BAND_5;
2094 /* Channel info utils */
2095 static inline bool iwl_mvm_has_ultra_hb_channel(struct iwl_mvm *mvm)
2097 return fw_has_capa(&mvm->fw->ucode_capa,
2098 IWL_UCODE_TLV_CAPA_ULTRA_HB_CHANNELS);
2101 static inline void *iwl_mvm_chan_info_cmd_tail(struct iwl_mvm *mvm,
2102 struct iwl_fw_channel_info *ci)
2104 return (u8 *)ci + (iwl_mvm_has_ultra_hb_channel(mvm) ?
2105 sizeof(struct iwl_fw_channel_info) :
2106 sizeof(struct iwl_fw_channel_info_v1));
2109 static inline size_t iwl_mvm_chan_info_padding(struct iwl_mvm *mvm)
2111 return iwl_mvm_has_ultra_hb_channel(mvm) ? 0 :
2112 sizeof(struct iwl_fw_channel_info) -
2113 sizeof(struct iwl_fw_channel_info_v1);
2116 static inline void iwl_mvm_set_chan_info(struct iwl_mvm *mvm,
2117 struct iwl_fw_channel_info *ci,
2118 u32 chan, u8 band, u8 width,
2119 u8 ctrl_pos)
2121 if (iwl_mvm_has_ultra_hb_channel(mvm)) {
2122 ci->channel = cpu_to_le32(chan);
2123 ci->band = band;
2124 ci->width = width;
2125 ci->ctrl_pos = ctrl_pos;
2126 } else {
2127 struct iwl_fw_channel_info_v1 *ci_v1 =
2128 (struct iwl_fw_channel_info_v1 *)ci;
2130 ci_v1->channel = chan;
2131 ci_v1->band = band;
2132 ci_v1->width = width;
2133 ci_v1->ctrl_pos = ctrl_pos;
2137 static inline void
2138 iwl_mvm_set_chan_info_chandef(struct iwl_mvm *mvm,
2139 struct iwl_fw_channel_info *ci,
2140 struct cfg80211_chan_def *chandef)
2142 enum nl80211_band band = chandef->chan->band;
2144 iwl_mvm_set_chan_info(mvm, ci, chandef->chan->hw_value,
2145 iwl_mvm_phy_band_from_nl80211(band),
2146 iwl_mvm_get_channel_width(chandef),
2147 iwl_mvm_get_ctrl_pos(chandef));
2150 static inline int iwl_umac_scan_get_max_profiles(const struct iwl_fw *fw)
2152 u8 ver = iwl_mvm_lookup_cmd_ver(fw, IWL_ALWAYS_LONG_GROUP,
2153 SCAN_OFFLOAD_UPDATE_PROFILES_CMD);
2154 return (ver == IWL_FW_CMD_VER_UNKNOWN || ver < 3) ?
2155 IWL_SCAN_MAX_PROFILES : IWL_SCAN_MAX_PROFILES_V2;
2157 #endif /* __IWL_MVM_H__ */