1 // SPDX-License-Identifier: ISC
2 /* Copyright (C) 2019 MediaTek Inc.
4 * Author: Roy Luo <royluo@google.com>
5 * Ryder Lee <ryder.lee@mediatek.com>
8 #include <linux/firmware.h>
14 static bool prefer_offload_fw
= true;
15 module_param(prefer_offload_fw
, bool, 0644);
16 MODULE_PARM_DESC(prefer_offload_fw
,
17 "Prefer client mode offload firmware (MT7663)");
19 struct mt7615_patch_hdr
{
27 struct mt7615_fw_trailer
{
37 #define FW_V3_COMMON_TAILER_SIZE 36
38 #define FW_V3_REGION_TAILER_SIZE 40
39 #define FW_START_OVERRIDE BIT(0)
40 #define FW_START_DLYCAL BIT(1)
41 #define FW_START_WORKING_PDA_CR4 BIT(2)
43 struct mt7663_fw_buf
{
53 #define MT7615_PATCH_ADDRESS 0x80000
54 #define MT7622_PATCH_ADDRESS 0x9c000
55 #define MT7663_PATCH_ADDRESS 0xdc000
57 #define N9_REGION_NUM 2
58 #define CR4_REGION_NUM 1
62 void mt7615_mcu_fill_msg(struct mt7615_dev
*dev
, struct sk_buff
*skb
,
63 int cmd
, int *wait_seq
)
65 int txd_len
, mcu_cmd
= FIELD_GET(__MCU_CMD_FIELD_ID
, cmd
);
66 struct mt7615_uni_txd
*uni_txd
;
67 struct mt7615_mcu_txd
*mcu_txd
;
68 u8 seq
, q_idx
, pkt_fmt
;
72 /* TODO: make dynamic based on msg type */
73 dev
->mt76
.mcu
.timeout
= 20 * HZ
;
75 seq
= ++dev
->mt76
.mcu
.msg_seq
& 0xf;
77 seq
= ++dev
->mt76
.mcu
.msg_seq
& 0xf;
81 txd_len
= cmd
& __MCU_CMD_FIELD_UNI
? sizeof(*uni_txd
) : sizeof(*mcu_txd
);
82 txd
= (__le32
*)skb_push(skb
, txd_len
);
84 if (cmd
!= MCU_CMD(FW_SCATTER
)) {
85 q_idx
= MT_TX_MCU_PORT_RX_Q0
;
86 pkt_fmt
= MT_TX_TYPE_CMD
;
88 q_idx
= MT_TX_MCU_PORT_RX_FWDL
;
89 pkt_fmt
= MT_TX_TYPE_FW
;
92 val
= FIELD_PREP(MT_TXD0_TX_BYTES
, skb
->len
) |
93 FIELD_PREP(MT_TXD0_P_IDX
, MT_TX_PORT_IDX_MCU
) |
94 FIELD_PREP(MT_TXD0_Q_IDX
, q_idx
);
95 txd
[0] = cpu_to_le32(val
);
97 val
= MT_TXD1_LONG_FORMAT
|
98 FIELD_PREP(MT_TXD1_HDR_FORMAT
, MT_HDR_FORMAT_CMD
) |
99 FIELD_PREP(MT_TXD1_PKT_FMT
, pkt_fmt
);
100 txd
[1] = cpu_to_le32(val
);
102 if (cmd
& __MCU_CMD_FIELD_UNI
) {
103 uni_txd
= (struct mt7615_uni_txd
*)txd
;
104 uni_txd
->len
= cpu_to_le16(skb
->len
- sizeof(uni_txd
->txd
));
105 uni_txd
->option
= MCU_CMD_UNI_EXT_ACK
;
106 uni_txd
->cid
= cpu_to_le16(mcu_cmd
);
107 uni_txd
->s2d_index
= MCU_S2D_H2N
;
108 uni_txd
->pkt_type
= MCU_PKT_ID
;
114 mcu_txd
= (struct mt7615_mcu_txd
*)txd
;
115 mcu_txd
->len
= cpu_to_le16(skb
->len
- sizeof(mcu_txd
->txd
));
116 mcu_txd
->pq_id
= cpu_to_le16(MCU_PQ_ID(MT_TX_PORT_IDX_MCU
, q_idx
));
117 mcu_txd
->s2d_index
= MCU_S2D_H2N
;
118 mcu_txd
->pkt_type
= MCU_PKT_ID
;
120 mcu_txd
->cid
= mcu_cmd
;
121 mcu_txd
->ext_cid
= FIELD_GET(__MCU_CMD_FIELD_EXT_ID
, cmd
);
123 if (mcu_txd
->ext_cid
|| (cmd
& __MCU_CMD_FIELD_CE
)) {
124 if (cmd
& __MCU_CMD_FIELD_QUERY
)
125 mcu_txd
->set_query
= MCU_Q_QUERY
;
127 mcu_txd
->set_query
= MCU_Q_SET
;
128 mcu_txd
->ext_cid_ack
= !!mcu_txd
->ext_cid
;
130 mcu_txd
->set_query
= MCU_Q_NA
;
133 EXPORT_SYMBOL_GPL(mt7615_mcu_fill_msg
);
135 int mt7615_mcu_parse_response(struct mt76_dev
*mdev
, int cmd
,
136 struct sk_buff
*skb
, int seq
)
138 struct mt7615_mcu_rxd
*rxd
;
142 dev_err(mdev
->dev
, "Message %08x (seq %d) timeout\n",
147 rxd
= (struct mt7615_mcu_rxd
*)skb
->data
;
151 if (cmd
== MCU_CMD(PATCH_SEM_CONTROL
)) {
152 skb_pull(skb
, sizeof(*rxd
) - 4);
154 } else if (cmd
== MCU_EXT_CMD(THERMAL_CTRL
)) {
155 skb_pull(skb
, sizeof(*rxd
));
156 ret
= le32_to_cpu(*(__le32
*)skb
->data
);
157 } else if (cmd
== MCU_EXT_QUERY(RF_REG_ACCESS
)) {
158 skb_pull(skb
, sizeof(*rxd
));
159 ret
= le32_to_cpu(*(__le32
*)&skb
->data
[8]);
160 } else if (cmd
== MCU_UNI_CMD(DEV_INFO_UPDATE
) ||
161 cmd
== MCU_UNI_CMD(BSS_INFO_UPDATE
) ||
162 cmd
== MCU_UNI_CMD(STA_REC_UPDATE
) ||
163 cmd
== MCU_UNI_CMD(HIF_CTRL
) ||
164 cmd
== MCU_UNI_CMD(OFFLOAD
) ||
165 cmd
== MCU_UNI_CMD(SUSPEND
)) {
166 struct mt76_connac_mcu_uni_event
*event
;
168 skb_pull(skb
, sizeof(*rxd
));
169 event
= (struct mt76_connac_mcu_uni_event
*)skb
->data
;
170 ret
= le32_to_cpu(event
->status
);
171 } else if (cmd
== MCU_CE_QUERY(REG_READ
)) {
172 struct mt76_connac_mcu_reg_event
*event
;
174 skb_pull(skb
, sizeof(*rxd
));
175 event
= (struct mt76_connac_mcu_reg_event
*)skb
->data
;
176 ret
= (int)le32_to_cpu(event
->val
);
181 EXPORT_SYMBOL_GPL(mt7615_mcu_parse_response
);
184 mt7615_mcu_send_message(struct mt76_dev
*mdev
, struct sk_buff
*skb
,
187 struct mt7615_dev
*dev
= container_of(mdev
, struct mt7615_dev
, mt76
);
188 enum mt76_mcuq_id qid
;
190 mt7615_mcu_fill_msg(dev
, skb
, cmd
, seq
);
191 if (test_bit(MT76_STATE_MCU_RUNNING
, &dev
->mphy
.state
))
196 return mt76_tx_queue_skb_raw(dev
, dev
->mt76
.q_mcu
[qid
], skb
, 0);
199 u32
mt7615_rf_rr(struct mt7615_dev
*dev
, u32 wf
, u32 reg
)
206 .wifi_stream
= cpu_to_le32(wf
),
207 .address
= cpu_to_le32(reg
),
210 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_QUERY(RF_REG_ACCESS
),
211 &req
, sizeof(req
), true);
214 int mt7615_rf_wr(struct mt7615_dev
*dev
, u32 wf
, u32 reg
, u32 val
)
221 .wifi_stream
= cpu_to_le32(wf
),
222 .address
= cpu_to_le32(reg
),
223 .data
= cpu_to_le32(val
),
226 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(RF_REG_ACCESS
),
227 &req
, sizeof(req
), false);
230 void mt7622_trigger_hif_int(struct mt7615_dev
*dev
, bool en
)
232 if (!is_mt7622(&dev
->mt76
))
235 regmap_update_bits(dev
->infracfg
, MT_INFRACFG_MISC
,
236 MT_INFRACFG_MISC_AP2CONN_WAKE
,
237 !en
* MT_INFRACFG_MISC_AP2CONN_WAKE
);
239 EXPORT_SYMBOL_GPL(mt7622_trigger_hif_int
);
241 static int mt7615_mcu_drv_pmctrl(struct mt7615_dev
*dev
)
243 struct mt76_phy
*mphy
= &dev
->mt76
.phy
;
244 struct mt76_connac_pm
*pm
= &dev
->pm
;
245 struct mt76_dev
*mdev
= &dev
->mt76
;
249 if (is_mt7663(mdev
)) {
250 /* Clear firmware own via N9 eint */
251 mt76_wr(dev
, MT_PCIE_DOORBELL_PUSH
, MT_CFG_LPCR_HOST_DRV_OWN
);
252 mt76_poll(dev
, MT_CONN_ON_MISC
, MT_CFG_LPCR_HOST_FW_OWN
, 0, 3000);
254 addr
= MT_CONN_HIF_ON_LPCTL
;
256 addr
= MT_CFG_LPCR_HOST
;
259 mt76_wr(dev
, addr
, MT_CFG_LPCR_HOST_DRV_OWN
);
261 mt7622_trigger_hif_int(dev
, true);
263 err
= !mt76_poll_msec(dev
, addr
, MT_CFG_LPCR_HOST_FW_OWN
, 0, 3000);
265 mt7622_trigger_hif_int(dev
, false);
268 dev_err(mdev
->dev
, "driver own failed\n");
272 clear_bit(MT76_STATE_PM
, &mphy
->state
);
274 pm
->stats
.last_wake_event
= jiffies
;
275 pm
->stats
.doze_time
+= pm
->stats
.last_wake_event
-
276 pm
->stats
.last_doze_event
;
281 static int mt7615_mcu_lp_drv_pmctrl(struct mt7615_dev
*dev
)
283 struct mt76_phy
*mphy
= &dev
->mt76
.phy
;
284 struct mt76_connac_pm
*pm
= &dev
->pm
;
287 mutex_lock(&pm
->mutex
);
289 if (!test_bit(MT76_STATE_PM
, &mphy
->state
))
292 for (i
= 0; i
< MT7615_DRV_OWN_RETRY_COUNT
; i
++) {
293 mt76_wr(dev
, MT_PCIE_DOORBELL_PUSH
, MT_CFG_LPCR_HOST_DRV_OWN
);
294 if (mt76_poll_msec(dev
, MT_CONN_HIF_ON_LPCTL
,
295 MT_CFG_LPCR_HOST_FW_OWN
, 0, 50))
299 if (i
== MT7615_DRV_OWN_RETRY_COUNT
) {
300 dev_err(dev
->mt76
.dev
, "driver own failed\n");
304 clear_bit(MT76_STATE_PM
, &mphy
->state
);
306 pm
->stats
.last_wake_event
= jiffies
;
307 pm
->stats
.doze_time
+= pm
->stats
.last_wake_event
-
308 pm
->stats
.last_doze_event
;
310 mutex_unlock(&pm
->mutex
);
315 static int mt7615_mcu_fw_pmctrl(struct mt7615_dev
*dev
)
317 struct mt76_phy
*mphy
= &dev
->mt76
.phy
;
318 struct mt76_connac_pm
*pm
= &dev
->pm
;
322 mutex_lock(&pm
->mutex
);
324 if (mt76_connac_skip_fw_pmctrl(mphy
, pm
))
327 mt7622_trigger_hif_int(dev
, true);
329 addr
= is_mt7663(&dev
->mt76
) ? MT_CONN_HIF_ON_LPCTL
: MT_CFG_LPCR_HOST
;
330 mt76_wr(dev
, addr
, MT_CFG_LPCR_HOST_FW_OWN
);
332 if (is_mt7622(&dev
->mt76
) &&
333 !mt76_poll_msec(dev
, addr
, MT_CFG_LPCR_HOST_FW_OWN
,
334 MT_CFG_LPCR_HOST_FW_OWN
, 3000)) {
335 dev_err(dev
->mt76
.dev
, "Timeout for firmware own\n");
336 clear_bit(MT76_STATE_PM
, &mphy
->state
);
340 mt7622_trigger_hif_int(dev
, false);
342 pm
->stats
.last_doze_event
= jiffies
;
343 pm
->stats
.awake_time
+= pm
->stats
.last_doze_event
-
344 pm
->stats
.last_wake_event
;
347 mutex_unlock(&pm
->mutex
);
353 mt7615_mcu_csa_finish(void *priv
, u8
*mac
, struct ieee80211_vif
*vif
)
355 if (vif
->bss_conf
.csa_active
)
356 ieee80211_csa_finish(vif
, 0);
360 mt7615_mcu_rx_csa_notify(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
362 struct mt7615_phy
*ext_phy
= mt7615_ext_phy(dev
);
363 struct mt76_phy
*mphy
= &dev
->mt76
.phy
;
364 struct mt7615_mcu_csa_notify
*c
;
366 c
= (struct mt7615_mcu_csa_notify
*)skb
->data
;
368 if (c
->omac_idx
> EXT_BSSID_MAX
)
371 if (ext_phy
&& ext_phy
->omac_mask
& BIT_ULL(c
->omac_idx
))
372 mphy
= dev
->mt76
.phys
[MT_BAND1
];
374 ieee80211_iterate_active_interfaces_atomic(mphy
->hw
,
375 IEEE80211_IFACE_ITER_RESUME_ALL
,
376 mt7615_mcu_csa_finish
, mphy
->hw
);
380 mt7615_mcu_rx_radar_detected(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
382 struct mt76_phy
*mphy
= &dev
->mt76
.phy
;
383 struct mt7615_mcu_rdd_report
*r
;
385 r
= (struct mt7615_mcu_rdd_report
*)skb
->data
;
387 if (!dev
->radar_pattern
.n_pulses
&& !r
->long_detected
&&
388 !r
->constant_prf_detected
&& !r
->staggered_prf_detected
)
391 if (r
->band_idx
&& dev
->mt76
.phys
[MT_BAND1
])
392 mphy
= dev
->mt76
.phys
[MT_BAND1
];
394 if (mt76_phy_dfs_state(mphy
) < MT_DFS_STATE_CAC
)
397 ieee80211_radar_detected(mphy
->hw
, NULL
);
402 mt7615_mcu_rx_log_message(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
404 struct mt7615_mcu_rxd
*rxd
= (struct mt7615_mcu_rxd
*)skb
->data
;
405 const char *data
= (char *)&rxd
[1];
408 switch (rxd
->s2d_index
) {
420 wiphy_info(mt76_hw(dev
)->wiphy
, "%s: %.*s", type
,
421 (int)(skb
->len
- sizeof(*rxd
)), data
);
425 mt7615_mcu_rx_ext_event(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
427 struct mt7615_mcu_rxd
*rxd
= (struct mt7615_mcu_rxd
*)skb
->data
;
429 switch (rxd
->ext_eid
) {
430 case MCU_EXT_EVENT_RDD_REPORT
:
431 mt7615_mcu_rx_radar_detected(dev
, skb
);
433 case MCU_EXT_EVENT_CSA_NOTIFY
:
434 mt7615_mcu_rx_csa_notify(dev
, skb
);
436 case MCU_EXT_EVENT_FW_LOG_2_HOST
:
437 mt7615_mcu_rx_log_message(dev
, skb
);
445 mt7615_mcu_scan_event(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
447 u8
*seq_num
= skb
->data
+ sizeof(struct mt7615_mcu_rxd
);
448 struct mt7615_phy
*phy
;
449 struct mt76_phy
*mphy
;
451 if (*seq_num
& BIT(7) && dev
->mt76
.phys
[MT_BAND1
])
452 mphy
= dev
->mt76
.phys
[MT_BAND1
];
454 mphy
= &dev
->mt76
.phy
;
458 spin_lock_bh(&dev
->mt76
.lock
);
459 __skb_queue_tail(&phy
->scan_event_list
, skb
);
460 spin_unlock_bh(&dev
->mt76
.lock
);
462 ieee80211_queue_delayed_work(mphy
->hw
, &phy
->scan_work
,
463 MT7615_HW_SCAN_TIMEOUT
);
467 mt7615_mcu_roc_event(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
469 struct mt7615_roc_tlv
*event
;
470 struct mt7615_phy
*phy
;
471 struct mt76_phy
*mphy
;
474 skb_pull(skb
, sizeof(struct mt7615_mcu_rxd
));
475 event
= (struct mt7615_roc_tlv
*)skb
->data
;
477 if (event
->dbdc_band
&& dev
->mt76
.phys
[MT_BAND1
])
478 mphy
= dev
->mt76
.phys
[MT_BAND1
];
480 mphy
= &dev
->mt76
.phy
;
482 ieee80211_ready_on_channel(mphy
->hw
);
485 phy
->roc_grant
= true;
486 wake_up(&phy
->roc_wait
);
488 duration
= le32_to_cpu(event
->max_interval
);
489 mod_timer(&phy
->roc_timer
,
490 round_jiffies_up(jiffies
+ msecs_to_jiffies(duration
)));
494 mt7615_mcu_beacon_loss_event(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
496 struct mt76_connac_beacon_loss_event
*event
;
497 struct mt76_phy
*mphy
;
498 u8 band_idx
= 0; /* DBDC support */
500 skb_pull(skb
, sizeof(struct mt7615_mcu_rxd
));
501 event
= (struct mt76_connac_beacon_loss_event
*)skb
->data
;
502 if (band_idx
&& dev
->mt76
.phys
[MT_BAND1
])
503 mphy
= dev
->mt76
.phys
[MT_BAND1
];
505 mphy
= &dev
->mt76
.phy
;
507 ieee80211_iterate_active_interfaces_atomic(mphy
->hw
,
508 IEEE80211_IFACE_ITER_RESUME_ALL
,
509 mt76_connac_mcu_beacon_loss_iter
,
514 mt7615_mcu_bss_event(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
516 struct mt76_connac_mcu_bss_event
*event
;
517 struct mt76_phy
*mphy
;
518 u8 band_idx
= 0; /* DBDC support */
520 skb_pull(skb
, sizeof(struct mt7615_mcu_rxd
));
521 event
= (struct mt76_connac_mcu_bss_event
*)skb
->data
;
523 if (band_idx
&& dev
->mt76
.phys
[MT_BAND1
])
524 mphy
= dev
->mt76
.phys
[MT_BAND1
];
526 mphy
= &dev
->mt76
.phy
;
528 if (event
->is_absent
)
529 ieee80211_stop_queues(mphy
->hw
);
531 ieee80211_wake_queues(mphy
->hw
);
535 mt7615_mcu_rx_unsolicited_event(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
537 struct mt7615_mcu_rxd
*rxd
= (struct mt7615_mcu_rxd
*)skb
->data
;
541 mt7615_mcu_rx_ext_event(dev
, skb
);
543 case MCU_EVENT_BSS_BEACON_LOSS
:
544 mt7615_mcu_beacon_loss_event(dev
, skb
);
547 mt7615_mcu_roc_event(dev
, skb
);
549 case MCU_EVENT_SCHED_SCAN_DONE
:
550 case MCU_EVENT_SCAN_DONE
:
551 mt7615_mcu_scan_event(dev
, skb
);
553 case MCU_EVENT_BSS_ABSENCE
:
554 mt7615_mcu_bss_event(dev
, skb
);
556 case MCU_EVENT_COREDUMP
:
557 mt76_connac_mcu_coredump_event(&dev
->mt76
, skb
,
566 void mt7615_mcu_rx_event(struct mt7615_dev
*dev
, struct sk_buff
*skb
)
568 struct mt7615_mcu_rxd
*rxd
= (struct mt7615_mcu_rxd
*)skb
->data
;
570 if (rxd
->ext_eid
== MCU_EXT_EVENT_THERMAL_PROTECT
||
571 rxd
->ext_eid
== MCU_EXT_EVENT_FW_LOG_2_HOST
||
572 rxd
->ext_eid
== MCU_EXT_EVENT_ASSERT_DUMP
||
573 rxd
->ext_eid
== MCU_EXT_EVENT_PS_SYNC
||
574 rxd
->eid
== MCU_EVENT_BSS_BEACON_LOSS
||
575 rxd
->eid
== MCU_EVENT_SCHED_SCAN_DONE
||
576 rxd
->eid
== MCU_EVENT_BSS_ABSENCE
||
577 rxd
->eid
== MCU_EVENT_SCAN_DONE
||
578 rxd
->eid
== MCU_EVENT_COREDUMP
||
579 rxd
->eid
== MCU_EVENT_ROC
||
581 mt7615_mcu_rx_unsolicited_event(dev
, skb
);
583 mt76_mcu_rx_event(&dev
->mt76
, skb
);
587 mt7615_mcu_muar_config(struct mt7615_dev
*dev
, struct ieee80211_vif
*vif
,
588 bool bssid
, bool enable
)
590 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
591 u32 idx
= mvif
->mt76
.omac_idx
- REPEATER_BSSID_START
;
592 u32 mask
= dev
->omac_mask
>> 32 & ~BIT(idx
);
593 const u8
*addr
= vif
->addr
;
605 .mode
= !!mask
|| enable
,
609 .index
= idx
* 2 + bssid
,
613 addr
= vif
->bss_conf
.bssid
;
616 ether_addr_copy(req
.addr
, addr
);
618 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(MUAR_UPDATE
),
619 &req
, sizeof(req
), true);
623 mt7615_mcu_add_dev(struct mt7615_phy
*phy
, struct ieee80211_vif
*vif
,
626 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
627 struct mt7615_dev
*dev
= phy
->dev
;
641 u8 omac_addr
[ETH_ALEN
];
645 .omac_idx
= mvif
->mt76
.omac_idx
,
646 .band_idx
= mvif
->mt76
.band_idx
,
647 .tlv_num
= cpu_to_le16(1),
651 .tag
= cpu_to_le16(DEV_INFO_ACTIVE
),
652 .len
= cpu_to_le16(sizeof(struct req_tlv
)),
654 .band_idx
= mvif
->mt76
.band_idx
,
658 if (mvif
->mt76
.omac_idx
>= REPEATER_BSSID_START
)
659 return mt7615_mcu_muar_config(dev
, vif
, false, enable
);
661 memcpy(data
.tlv
.omac_addr
, vif
->addr
, ETH_ALEN
);
662 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(DEV_INFO_UPDATE
),
663 &data
, sizeof(data
), true);
667 mt7615_mcu_add_beacon_offload(struct mt7615_dev
*dev
,
668 struct ieee80211_hw
*hw
,
669 struct ieee80211_vif
*vif
, bool enable
)
671 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
672 struct mt76_wcid
*wcid
= &dev
->mt76
.global_wcid
;
673 struct ieee80211_mutable_offsets offs
;
674 struct ieee80211_tx_info
*info
;
681 u8 need_pre_tbtt_int
;
687 /* bss color change */
691 .omac_idx
= mvif
->mt76
.omac_idx
,
693 .wlan_idx
= wcid
->idx
,
694 .band_idx
= mvif
->mt76
.band_idx
,
701 skb
= ieee80211_beacon_get_template(hw
, vif
, &offs
, 0);
705 if (skb
->len
> 512 - MT_TXD_SIZE
) {
706 dev_err(dev
->mt76
.dev
, "Bcn size limit exceed\n");
711 info
= IEEE80211_SKB_CB(skb
);
712 info
->hw_queue
|= FIELD_PREP(MT_TX_HW_QUEUE_PHY
, mvif
->mt76
.band_idx
);
714 mt7615_mac_write_txwi(dev
, (__le32
*)(req
.pkt
), skb
, wcid
, NULL
,
716 memcpy(req
.pkt
+ MT_TXD_SIZE
, skb
->data
, skb
->len
);
717 req
.pkt_len
= cpu_to_le16(MT_TXD_SIZE
+ skb
->len
);
718 req
.tim_ie_pos
= cpu_to_le16(MT_TXD_SIZE
+ offs
.tim_offset
);
719 if (offs
.cntdwn_counter_offs
[0]) {
722 csa_offs
= MT_TXD_SIZE
+ offs
.cntdwn_counter_offs
[0] - 4;
723 req
.csa_ie_pos
= cpu_to_le16(csa_offs
);
724 req
.csa_cnt
= skb
->data
[offs
.cntdwn_counter_offs
[0]];
729 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(BCN_OFFLOAD
), &req
,
734 mt7615_mcu_ctrl_pm_state(struct mt7615_dev
*dev
, int band
, int state
)
736 return mt76_connac_mcu_set_pm(&dev
->mt76
, band
, state
);
740 mt7615_mcu_add_bss(struct mt7615_phy
*phy
, struct ieee80211_vif
*vif
,
741 struct ieee80211_sta
*sta
, bool enable
)
743 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
744 struct mt7615_dev
*dev
= phy
->dev
;
747 if (mvif
->mt76
.omac_idx
>= REPEATER_BSSID_START
)
748 mt7615_mcu_muar_config(dev
, vif
, true, enable
);
750 skb
= mt76_connac_mcu_alloc_sta_req(&dev
->mt76
, &mvif
->mt76
, NULL
);
755 mt76_connac_mcu_bss_omac_tlv(skb
, vif
);
757 mt76_connac_mcu_bss_basic_tlv(skb
, vif
, sta
, phy
->mt76
,
758 mvif
->sta
.wcid
.idx
, enable
);
760 if (enable
&& mvif
->mt76
.omac_idx
>= EXT_BSSID_START
&&
761 mvif
->mt76
.omac_idx
< REPEATER_BSSID_START
)
762 mt76_connac_mcu_bss_ext_tlv(skb
, &mvif
->mt76
);
764 return mt76_mcu_skb_send_msg(&dev
->mt76
, skb
,
765 MCU_EXT_CMD(BSS_INFO_UPDATE
), true);
769 mt7615_mcu_wtbl_tx_ba(struct mt7615_dev
*dev
,
770 struct ieee80211_ampdu_params
*params
,
773 struct mt7615_sta
*msta
= (struct mt7615_sta
*)params
->sta
->drv_priv
;
774 struct mt7615_vif
*mvif
= msta
->vif
;
775 struct wtbl_req_hdr
*wtbl_hdr
;
776 struct sk_buff
*skb
= NULL
;
779 wtbl_hdr
= mt76_connac_mcu_alloc_wtbl_req(&dev
->mt76
, &msta
->wcid
,
780 WTBL_SET
, NULL
, &skb
);
781 if (IS_ERR(wtbl_hdr
))
782 return PTR_ERR(wtbl_hdr
);
784 mt76_connac_mcu_wtbl_ba_tlv(&dev
->mt76
, skb
, params
, enable
, true,
787 err
= mt76_mcu_skb_send_msg(&dev
->mt76
, skb
,
788 MCU_EXT_CMD(WTBL_UPDATE
), true);
792 skb
= mt76_connac_mcu_alloc_sta_req(&dev
->mt76
, &mvif
->mt76
,
797 mt76_connac_mcu_sta_ba_tlv(skb
, params
, enable
, true);
799 return mt76_mcu_skb_send_msg(&dev
->mt76
, skb
,
800 MCU_EXT_CMD(STA_REC_UPDATE
), true);
804 mt7615_mcu_wtbl_rx_ba(struct mt7615_dev
*dev
,
805 struct ieee80211_ampdu_params
*params
,
808 struct mt7615_sta
*msta
= (struct mt7615_sta
*)params
->sta
->drv_priv
;
809 struct mt7615_vif
*mvif
= msta
->vif
;
810 struct wtbl_req_hdr
*wtbl_hdr
;
814 skb
= mt76_connac_mcu_alloc_sta_req(&dev
->mt76
, &mvif
->mt76
,
819 mt76_connac_mcu_sta_ba_tlv(skb
, params
, enable
, false);
821 err
= mt76_mcu_skb_send_msg(&dev
->mt76
, skb
,
822 MCU_EXT_CMD(STA_REC_UPDATE
), true);
823 if (err
< 0 || !enable
)
827 wtbl_hdr
= mt76_connac_mcu_alloc_wtbl_req(&dev
->mt76
, &msta
->wcid
,
828 WTBL_SET
, NULL
, &skb
);
829 if (IS_ERR(wtbl_hdr
))
830 return PTR_ERR(wtbl_hdr
);
832 mt76_connac_mcu_wtbl_ba_tlv(&dev
->mt76
, skb
, params
, enable
, false,
835 return mt76_mcu_skb_send_msg(&dev
->mt76
, skb
,
836 MCU_EXT_CMD(WTBL_UPDATE
), true);
840 mt7615_mcu_wtbl_sta_add(struct mt7615_phy
*phy
, struct ieee80211_vif
*vif
,
841 struct ieee80211_sta
*sta
, bool enable
)
843 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
844 struct sk_buff
*skb
, *sskb
, *wskb
= NULL
;
845 struct ieee80211_link_sta
*link_sta
;
846 struct mt7615_dev
*dev
= phy
->dev
;
847 struct wtbl_req_hdr
*wtbl_hdr
;
848 struct mt7615_sta
*msta
;
849 bool new_entry
= true;
853 msta
= sta
? (struct mt7615_sta
*)sta
->drv_priv
: &mvif
->sta
;
854 link_sta
= sta
? &sta
->deflink
: NULL
;
856 sskb
= mt76_connac_mcu_alloc_sta_req(&dev
->mt76
, &mvif
->mt76
,
859 return PTR_ERR(sskb
);
865 mvif
->sta_added
= true;
867 conn_state
= enable
? CONN_STATE_PORT_SECURE
: CONN_STATE_DISCONNECT
;
868 mt76_connac_mcu_sta_basic_tlv(&dev
->mt76
, sskb
, vif
, link_sta
,
869 conn_state
, new_entry
);
871 mt76_connac_mcu_sta_tlv(phy
->mt76
, sskb
, sta
, vif
, 0,
872 MT76_STA_INFO_STATE_ASSOC
);
874 wtbl_hdr
= mt76_connac_mcu_alloc_wtbl_req(&dev
->mt76
, &msta
->wcid
,
875 WTBL_RESET_AND_SET
, NULL
,
877 if (IS_ERR(wtbl_hdr
))
878 return PTR_ERR(wtbl_hdr
);
881 mt76_connac_mcu_wtbl_generic_tlv(&dev
->mt76
, wskb
, vif
, sta
,
884 mt76_connac_mcu_wtbl_ht_tlv(&dev
->mt76
, wskb
, sta
,
885 NULL
, wtbl_hdr
, true, true);
886 mt76_connac_mcu_wtbl_hdr_trans_tlv(wskb
, vif
, &msta
->wcid
,
890 cmd
= enable
? MCU_EXT_CMD(WTBL_UPDATE
) : MCU_EXT_CMD(STA_REC_UPDATE
);
891 skb
= enable
? wskb
: sskb
;
893 err
= mt76_mcu_skb_send_msg(&dev
->mt76
, skb
, cmd
, true);
895 skb
= enable
? sskb
: wskb
;
901 cmd
= enable
? MCU_EXT_CMD(STA_REC_UPDATE
) : MCU_EXT_CMD(WTBL_UPDATE
);
902 skb
= enable
? sskb
: wskb
;
904 return mt76_mcu_skb_send_msg(&dev
->mt76
, skb
, cmd
, true);
908 mt7615_mcu_wtbl_update_hdr_trans(struct mt7615_dev
*dev
,
909 struct ieee80211_vif
*vif
,
910 struct ieee80211_sta
*sta
)
912 return mt76_connac_mcu_wtbl_update_hdr_trans(&dev
->mt76
, vif
, sta
);
915 static const struct mt7615_mcu_ops wtbl_update_ops
= {
916 .add_beacon_offload
= mt7615_mcu_add_beacon_offload
,
917 .set_pm_state
= mt7615_mcu_ctrl_pm_state
,
918 .add_dev_info
= mt7615_mcu_add_dev
,
919 .add_bss_info
= mt7615_mcu_add_bss
,
920 .add_tx_ba
= mt7615_mcu_wtbl_tx_ba
,
921 .add_rx_ba
= mt7615_mcu_wtbl_rx_ba
,
922 .sta_add
= mt7615_mcu_wtbl_sta_add
,
923 .set_drv_ctrl
= mt7615_mcu_drv_pmctrl
,
924 .set_fw_ctrl
= mt7615_mcu_fw_pmctrl
,
925 .set_sta_decap_offload
= mt7615_mcu_wtbl_update_hdr_trans
,
929 mt7615_mcu_sta_ba(struct mt7615_dev
*dev
,
930 struct ieee80211_ampdu_params
*params
,
931 bool enable
, bool tx
)
933 struct mt7615_sta
*msta
= (struct mt7615_sta
*)params
->sta
->drv_priv
;
934 struct mt7615_vif
*mvif
= msta
->vif
;
935 struct wtbl_req_hdr
*wtbl_hdr
;
936 struct tlv
*sta_wtbl
;
939 skb
= mt76_connac_mcu_alloc_sta_req(&dev
->mt76
, &mvif
->mt76
,
944 mt76_connac_mcu_sta_ba_tlv(skb
, params
, enable
, tx
);
946 sta_wtbl
= mt76_connac_mcu_add_tlv(skb
, STA_REC_WTBL
, sizeof(struct tlv
));
948 wtbl_hdr
= mt76_connac_mcu_alloc_wtbl_req(&dev
->mt76
, &msta
->wcid
,
949 WTBL_SET
, sta_wtbl
, &skb
);
950 if (IS_ERR(wtbl_hdr
))
951 return PTR_ERR(wtbl_hdr
);
953 mt76_connac_mcu_wtbl_ba_tlv(&dev
->mt76
, skb
, params
, enable
, tx
,
956 return mt76_mcu_skb_send_msg(&dev
->mt76
, skb
,
957 MCU_EXT_CMD(STA_REC_UPDATE
), true);
961 mt7615_mcu_sta_tx_ba(struct mt7615_dev
*dev
,
962 struct ieee80211_ampdu_params
*params
,
965 return mt7615_mcu_sta_ba(dev
, params
, enable
, true);
969 mt7615_mcu_sta_rx_ba(struct mt7615_dev
*dev
,
970 struct ieee80211_ampdu_params
*params
,
973 return mt7615_mcu_sta_ba(dev
, params
, enable
, false);
977 __mt7615_mcu_add_sta(struct mt76_phy
*phy
, struct ieee80211_vif
*vif
,
978 struct ieee80211_sta
*sta
, bool enable
, int cmd
,
981 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
982 struct mt76_sta_cmd_info info
= {
985 .offload_fw
= offload_fw
,
991 info
.wcid
= sta
? (struct mt76_wcid
*)sta
->drv_priv
: &mvif
->sta
.wcid
;
992 return mt76_connac_mcu_sta_cmd(phy
, &info
);
996 mt7615_mcu_add_sta(struct mt7615_phy
*phy
, struct ieee80211_vif
*vif
,
997 struct ieee80211_sta
*sta
, bool enable
)
999 return __mt7615_mcu_add_sta(phy
->mt76
, vif
, sta
, enable
,
1000 MCU_EXT_CMD(STA_REC_UPDATE
), false);
1004 mt7615_mcu_sta_update_hdr_trans(struct mt7615_dev
*dev
,
1005 struct ieee80211_vif
*vif
,
1006 struct ieee80211_sta
*sta
)
1008 struct mt7615_sta
*msta
= (struct mt7615_sta
*)sta
->drv_priv
;
1010 return mt76_connac_mcu_sta_update_hdr_trans(&dev
->mt76
,
1012 MCU_EXT_CMD(STA_REC_UPDATE
));
1015 static const struct mt7615_mcu_ops sta_update_ops
= {
1016 .add_beacon_offload
= mt7615_mcu_add_beacon_offload
,
1017 .set_pm_state
= mt7615_mcu_ctrl_pm_state
,
1018 .add_dev_info
= mt7615_mcu_add_dev
,
1019 .add_bss_info
= mt7615_mcu_add_bss
,
1020 .add_tx_ba
= mt7615_mcu_sta_tx_ba
,
1021 .add_rx_ba
= mt7615_mcu_sta_rx_ba
,
1022 .sta_add
= mt7615_mcu_add_sta
,
1023 .set_drv_ctrl
= mt7615_mcu_drv_pmctrl
,
1024 .set_fw_ctrl
= mt7615_mcu_fw_pmctrl
,
1025 .set_sta_decap_offload
= mt7615_mcu_sta_update_hdr_trans
,
1029 mt7615_mcu_uni_ctrl_pm_state(struct mt7615_dev
*dev
, int band
, int state
)
1035 mt7615_mcu_uni_add_beacon_offload(struct mt7615_dev
*dev
,
1036 struct ieee80211_hw
*hw
,
1037 struct ieee80211_vif
*vif
,
1040 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
1041 struct mt76_wcid
*wcid
= &dev
->mt76
.global_wcid
;
1042 struct ieee80211_mutable_offsets offs
;
1048 struct bcn_content_tlv
{
1054 /* 0: disable beacon offload
1055 * 1: enable beacon offload
1056 * 2: update probe respond offload
1059 /* 0: legacy format (TXD + payload)
1060 * 1: only cap field IE
1065 } __packed beacon_tlv
;
1068 .bss_idx
= mvif
->mt76
.idx
,
1071 .tag
= cpu_to_le16(UNI_BSS_INFO_BCN_CONTENT
),
1072 .len
= cpu_to_le16(sizeof(struct bcn_content_tlv
)),
1076 struct sk_buff
*skb
;
1081 skb
= ieee80211_beacon_get_template(mt76_hw(dev
), vif
, &offs
, 0);
1085 if (skb
->len
> 512 - MT_TXD_SIZE
) {
1086 dev_err(dev
->mt76
.dev
, "beacon size limit exceed\n");
1091 mt7615_mac_write_txwi(dev
, (__le32
*)(req
.beacon_tlv
.pkt
), skb
,
1092 wcid
, NULL
, 0, NULL
, 0, true);
1093 memcpy(req
.beacon_tlv
.pkt
+ MT_TXD_SIZE
, skb
->data
, skb
->len
);
1094 req
.beacon_tlv
.pkt_len
= cpu_to_le16(MT_TXD_SIZE
+ skb
->len
);
1095 req
.beacon_tlv
.tim_ie_pos
= cpu_to_le16(MT_TXD_SIZE
+ offs
.tim_offset
);
1097 if (offs
.cntdwn_counter_offs
[0]) {
1100 csa_offs
= MT_TXD_SIZE
+ offs
.cntdwn_counter_offs
[0] - 4;
1101 req
.beacon_tlv
.csa_ie_pos
= cpu_to_le16(csa_offs
);
1106 return mt76_mcu_send_msg(&dev
->mt76
, MCU_UNI_CMD(BSS_INFO_UPDATE
),
1107 &req
, sizeof(req
), true);
1111 mt7615_mcu_uni_add_dev(struct mt7615_phy
*phy
, struct ieee80211_vif
*vif
,
1114 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
1116 return mt76_connac_mcu_uni_add_dev(phy
->mt76
, &vif
->bss_conf
,
1117 &mvif
->sta
.wcid
, enable
);
1121 mt7615_mcu_uni_add_bss(struct mt7615_phy
*phy
, struct ieee80211_vif
*vif
,
1122 struct ieee80211_sta
*sta
, bool enable
)
1124 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
1126 return mt76_connac_mcu_uni_add_bss(phy
->mt76
, vif
, &mvif
->sta
.wcid
,
1131 mt7615_mcu_uni_add_sta(struct mt7615_phy
*phy
, struct ieee80211_vif
*vif
,
1132 struct ieee80211_sta
*sta
, bool enable
)
1134 return __mt7615_mcu_add_sta(phy
->mt76
, vif
, sta
, enable
,
1135 MCU_UNI_CMD(STA_REC_UPDATE
), true);
1139 mt7615_mcu_uni_tx_ba(struct mt7615_dev
*dev
,
1140 struct ieee80211_ampdu_params
*params
,
1143 struct mt7615_sta
*sta
= (struct mt7615_sta
*)params
->sta
->drv_priv
;
1145 return mt76_connac_mcu_sta_ba(&dev
->mt76
, &sta
->vif
->mt76
, params
,
1146 MCU_UNI_CMD(STA_REC_UPDATE
), enable
,
1151 mt7615_mcu_uni_rx_ba(struct mt7615_dev
*dev
,
1152 struct ieee80211_ampdu_params
*params
,
1155 struct mt7615_sta
*msta
= (struct mt7615_sta
*)params
->sta
->drv_priv
;
1156 struct mt7615_vif
*mvif
= msta
->vif
;
1157 struct wtbl_req_hdr
*wtbl_hdr
;
1158 struct tlv
*sta_wtbl
;
1159 struct sk_buff
*skb
;
1162 skb
= mt76_connac_mcu_alloc_sta_req(&dev
->mt76
, &mvif
->mt76
,
1165 return PTR_ERR(skb
);
1167 mt76_connac_mcu_sta_ba_tlv(skb
, params
, enable
, false);
1169 err
= mt76_mcu_skb_send_msg(&dev
->mt76
, skb
,
1170 MCU_UNI_CMD(STA_REC_UPDATE
), true);
1171 if (err
< 0 || !enable
)
1174 skb
= mt76_connac_mcu_alloc_sta_req(&dev
->mt76
, &mvif
->mt76
,
1177 return PTR_ERR(skb
);
1179 sta_wtbl
= mt76_connac_mcu_add_tlv(skb
, STA_REC_WTBL
,
1180 sizeof(struct tlv
));
1182 wtbl_hdr
= mt76_connac_mcu_alloc_wtbl_req(&dev
->mt76
, &msta
->wcid
,
1183 WTBL_SET
, sta_wtbl
, &skb
);
1184 if (IS_ERR(wtbl_hdr
))
1185 return PTR_ERR(wtbl_hdr
);
1187 mt76_connac_mcu_wtbl_ba_tlv(&dev
->mt76
, skb
, params
, enable
, false,
1188 sta_wtbl
, wtbl_hdr
);
1190 return mt76_mcu_skb_send_msg(&dev
->mt76
, skb
,
1191 MCU_UNI_CMD(STA_REC_UPDATE
), true);
1195 mt7615_mcu_sta_uni_update_hdr_trans(struct mt7615_dev
*dev
,
1196 struct ieee80211_vif
*vif
,
1197 struct ieee80211_sta
*sta
)
1199 struct mt7615_sta
*msta
= (struct mt7615_sta
*)sta
->drv_priv
;
1201 return mt76_connac_mcu_sta_update_hdr_trans(&dev
->mt76
,
1203 MCU_UNI_CMD(STA_REC_UPDATE
));
1206 static const struct mt7615_mcu_ops uni_update_ops
= {
1207 .add_beacon_offload
= mt7615_mcu_uni_add_beacon_offload
,
1208 .set_pm_state
= mt7615_mcu_uni_ctrl_pm_state
,
1209 .add_dev_info
= mt7615_mcu_uni_add_dev
,
1210 .add_bss_info
= mt7615_mcu_uni_add_bss
,
1211 .add_tx_ba
= mt7615_mcu_uni_tx_ba
,
1212 .add_rx_ba
= mt7615_mcu_uni_rx_ba
,
1213 .sta_add
= mt7615_mcu_uni_add_sta
,
1214 .set_drv_ctrl
= mt7615_mcu_lp_drv_pmctrl
,
1215 .set_fw_ctrl
= mt7615_mcu_fw_pmctrl
,
1216 .set_sta_decap_offload
= mt7615_mcu_sta_uni_update_hdr_trans
,
1219 int mt7615_mcu_restart(struct mt76_dev
*dev
)
1221 return mt76_mcu_send_msg(dev
, MCU_CMD(RESTART_DL_REQ
), NULL
, 0, true);
1223 EXPORT_SYMBOL_GPL(mt7615_mcu_restart
);
1225 static int mt7615_load_patch(struct mt7615_dev
*dev
, u32 addr
, const char *name
)
1227 const struct mt7615_patch_hdr
*hdr
;
1228 const struct firmware
*fw
= NULL
;
1231 ret
= firmware_request_nowarn(&fw
, name
, dev
->mt76
.dev
);
1235 if (!fw
|| !fw
->data
|| fw
->size
< sizeof(*hdr
)) {
1236 dev_err(dev
->mt76
.dev
, "Invalid firmware\n");
1241 sem
= mt76_connac_mcu_patch_sem_ctrl(&dev
->mt76
, true);
1245 case PATCH_NOT_DL_SEM_SUCCESS
:
1248 dev_err(dev
->mt76
.dev
, "Failed to get patch semaphore\n");
1253 hdr
= (const struct mt7615_patch_hdr
*)(fw
->data
);
1255 dev_info(dev
->mt76
.dev
, "HW/SW Version: 0x%x, Build Time: %.16s\n",
1256 be32_to_cpu(hdr
->hw_sw_ver
), hdr
->build_date
);
1258 len
= fw
->size
- sizeof(*hdr
);
1260 ret
= mt76_connac_mcu_init_download(&dev
->mt76
, addr
, len
,
1263 dev_err(dev
->mt76
.dev
, "Download request failed\n");
1267 ret
= mt76_mcu_send_firmware(&dev
->mt76
, MCU_CMD(FW_SCATTER
),
1268 fw
->data
+ sizeof(*hdr
), len
);
1270 dev_err(dev
->mt76
.dev
, "Failed to send firmware to device\n");
1274 ret
= mt76_connac_mcu_start_patch(&dev
->mt76
);
1276 dev_err(dev
->mt76
.dev
, "Failed to start patch\n");
1279 sem
= mt76_connac_mcu_patch_sem_ctrl(&dev
->mt76
, false);
1281 case PATCH_REL_SEM_SUCCESS
:
1285 dev_err(dev
->mt76
.dev
, "Failed to release patch semaphore\n");
1290 release_firmware(fw
);
1296 mt7615_mcu_send_ram_firmware(struct mt7615_dev
*dev
,
1297 const struct mt7615_fw_trailer
*hdr
,
1298 const u8
*data
, bool is_cr4
)
1300 int n_region
= is_cr4
? CR4_REGION_NUM
: N9_REGION_NUM
;
1301 int err
, i
, offset
= 0;
1302 u32 len
, addr
, mode
;
1304 for (i
= 0; i
< n_region
; i
++) {
1305 mode
= mt76_connac_mcu_gen_dl_mode(&dev
->mt76
,
1306 hdr
[i
].feature_set
, is_cr4
);
1307 len
= le32_to_cpu(hdr
[i
].len
) + IMG_CRC_LEN
;
1308 addr
= le32_to_cpu(hdr
[i
].addr
);
1310 err
= mt76_connac_mcu_init_download(&dev
->mt76
, addr
, len
,
1313 dev_err(dev
->mt76
.dev
, "Download request failed\n");
1317 err
= mt76_mcu_send_firmware(&dev
->mt76
, MCU_CMD(FW_SCATTER
),
1318 data
+ offset
, len
);
1320 dev_err(dev
->mt76
.dev
, "Failed to send firmware to device\n");
1330 static int mt7615_load_n9(struct mt7615_dev
*dev
, const char *name
)
1332 const struct mt7615_fw_trailer
*hdr
;
1333 const struct firmware
*fw
;
1336 ret
= request_firmware(&fw
, name
, dev
->mt76
.dev
);
1340 if (!fw
|| !fw
->data
|| fw
->size
< N9_REGION_NUM
* sizeof(*hdr
)) {
1341 dev_err(dev
->mt76
.dev
, "Invalid firmware\n");
1346 hdr
= (const struct mt7615_fw_trailer
*)(fw
->data
+ fw
->size
-
1347 N9_REGION_NUM
* sizeof(*hdr
));
1349 dev_info(dev
->mt76
.dev
, "N9 Firmware Version: %.10s, Build Time: %.15s\n",
1350 hdr
->fw_ver
, hdr
->build_date
);
1352 ret
= mt7615_mcu_send_ram_firmware(dev
, hdr
, fw
->data
, false);
1356 ret
= mt76_connac_mcu_start_firmware(&dev
->mt76
,
1357 le32_to_cpu(hdr
->addr
),
1360 dev_err(dev
->mt76
.dev
, "Failed to start N9 firmware\n");
1364 snprintf(dev
->mt76
.hw
->wiphy
->fw_version
,
1365 sizeof(dev
->mt76
.hw
->wiphy
->fw_version
),
1366 "%.10s-%.15s", hdr
->fw_ver
, hdr
->build_date
);
1368 if (!is_mt7615(&dev
->mt76
)) {
1369 dev
->fw_ver
= MT7615_FIRMWARE_V2
;
1370 dev
->mcu_ops
= &sta_update_ops
;
1372 dev
->fw_ver
= MT7615_FIRMWARE_V1
;
1373 dev
->mcu_ops
= &wtbl_update_ops
;
1377 release_firmware(fw
);
1381 static int mt7615_load_cr4(struct mt7615_dev
*dev
, const char *name
)
1383 const struct mt7615_fw_trailer
*hdr
;
1384 const struct firmware
*fw
;
1387 ret
= request_firmware(&fw
, name
, dev
->mt76
.dev
);
1391 if (!fw
|| !fw
->data
|| fw
->size
< CR4_REGION_NUM
* sizeof(*hdr
)) {
1392 dev_err(dev
->mt76
.dev
, "Invalid firmware\n");
1397 hdr
= (const struct mt7615_fw_trailer
*)(fw
->data
+ fw
->size
-
1398 CR4_REGION_NUM
* sizeof(*hdr
));
1400 dev_info(dev
->mt76
.dev
, "CR4 Firmware Version: %.10s, Build Time: %.15s\n",
1401 hdr
->fw_ver
, hdr
->build_date
);
1403 ret
= mt7615_mcu_send_ram_firmware(dev
, hdr
, fw
->data
, true);
1407 ret
= mt76_connac_mcu_start_firmware(&dev
->mt76
, 0,
1408 FW_START_WORKING_PDA_CR4
);
1410 dev_err(dev
->mt76
.dev
, "Failed to start CR4 firmware\n");
1415 release_firmware(fw
);
1420 static int mt7615_load_ram(struct mt7615_dev
*dev
)
1424 ret
= mt7615_load_n9(dev
, MT7615_FIRMWARE_N9
);
1428 return mt7615_load_cr4(dev
, MT7615_FIRMWARE_CR4
);
1431 static int mt7615_load_firmware(struct mt7615_dev
*dev
)
1436 val
= mt76_get_field(dev
, MT_TOP_MISC2
, MT_TOP_MISC2_FW_STATE
);
1438 if (val
!= FW_STATE_FW_DOWNLOAD
) {
1439 dev_err(dev
->mt76
.dev
, "Firmware is not ready for download\n");
1443 ret
= mt7615_load_patch(dev
, MT7615_PATCH_ADDRESS
, MT7615_ROM_PATCH
);
1447 ret
= mt7615_load_ram(dev
);
1451 if (!mt76_poll_msec(dev
, MT_TOP_MISC2
, MT_TOP_MISC2_FW_STATE
,
1452 FIELD_PREP(MT_TOP_MISC2_FW_STATE
,
1453 FW_STATE_RDY
), 500)) {
1454 dev_err(dev
->mt76
.dev
, "Timeout for initializing firmware\n");
1461 static int mt7622_load_firmware(struct mt7615_dev
*dev
)
1466 mt76_set(dev
, MT_WPDMA_GLO_CFG
, MT_WPDMA_GLO_CFG_BYPASS_TX_SCH
);
1468 val
= mt76_get_field(dev
, MT_TOP_OFF_RSV
, MT_TOP_OFF_RSV_FW_STATE
);
1469 if (val
!= FW_STATE_FW_DOWNLOAD
) {
1470 dev_err(dev
->mt76
.dev
, "Firmware is not ready for download\n");
1474 ret
= mt7615_load_patch(dev
, MT7622_PATCH_ADDRESS
, MT7622_ROM_PATCH
);
1478 ret
= mt7615_load_n9(dev
, MT7622_FIRMWARE_N9
);
1482 if (!mt76_poll_msec(dev
, MT_TOP_OFF_RSV
, MT_TOP_OFF_RSV_FW_STATE
,
1483 FIELD_PREP(MT_TOP_OFF_RSV_FW_STATE
,
1484 FW_STATE_NORMAL_TRX
), 1500)) {
1485 dev_err(dev
->mt76
.dev
, "Timeout for initializing firmware\n");
1489 mt76_clear(dev
, MT_WPDMA_GLO_CFG
, MT_WPDMA_GLO_CFG_BYPASS_TX_SCH
);
1494 int mt7615_mcu_fw_log_2_host(struct mt7615_dev
*dev
, u8 ctrl
)
1503 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(FW_LOG_2_HOST
),
1504 &data
, sizeof(data
), true);
1507 static int mt7615_mcu_cal_cache_apply(struct mt7615_dev
*dev
)
1513 .cache_enable
= true
1516 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(CAL_CACHE
), &data
,
1517 sizeof(data
), false);
1520 static int mt7663_load_n9(struct mt7615_dev
*dev
, const char *name
)
1522 u32 offset
= 0, override_addr
= 0, flag
= FW_START_DLYCAL
;
1523 const struct mt76_connac2_fw_trailer
*hdr
;
1524 const struct mt7663_fw_buf
*buf
;
1525 const struct firmware
*fw
;
1526 const u8
*base_addr
;
1529 ret
= request_firmware(&fw
, name
, dev
->mt76
.dev
);
1533 if (!fw
|| !fw
->data
|| fw
->size
< FW_V3_COMMON_TAILER_SIZE
) {
1534 dev_err(dev
->mt76
.dev
, "Invalid firmware\n");
1539 hdr
= (const void *)(fw
->data
+ fw
->size
- FW_V3_COMMON_TAILER_SIZE
);
1540 dev_info(dev
->mt76
.dev
, "N9 Firmware Version: %.10s, Build Time: %.15s\n",
1541 hdr
->fw_ver
, hdr
->build_date
);
1542 dev_info(dev
->mt76
.dev
, "Region number: 0x%x\n", hdr
->n_region
);
1544 base_addr
= fw
->data
+ fw
->size
- FW_V3_COMMON_TAILER_SIZE
;
1545 for (i
= 0; i
< hdr
->n_region
; i
++) {
1546 u32 shift
= (hdr
->n_region
- i
) * FW_V3_REGION_TAILER_SIZE
;
1547 u32 len
, addr
, mode
;
1549 dev_info(dev
->mt76
.dev
, "Parsing tailer Region: %d\n", i
);
1551 buf
= (const struct mt7663_fw_buf
*)(base_addr
- shift
);
1552 mode
= mt76_connac_mcu_gen_dl_mode(&dev
->mt76
,
1553 buf
->feature_set
, false);
1554 addr
= le32_to_cpu(buf
->img_dest_addr
);
1555 len
= le32_to_cpu(buf
->img_size
);
1557 ret
= mt76_connac_mcu_init_download(&dev
->mt76
, addr
, len
,
1560 dev_err(dev
->mt76
.dev
, "Download request failed\n");
1564 ret
= mt76_mcu_send_firmware(&dev
->mt76
, MCU_CMD(FW_SCATTER
),
1565 fw
->data
+ offset
, len
);
1567 dev_err(dev
->mt76
.dev
, "Failed to send firmware\n");
1571 offset
+= le32_to_cpu(buf
->img_size
);
1572 if (buf
->feature_set
& DL_MODE_VALID_RAM_ENTRY
) {
1573 override_addr
= le32_to_cpu(buf
->img_dest_addr
);
1574 dev_info(dev
->mt76
.dev
, "Region %d, override_addr = 0x%08x\n",
1580 flag
|= FW_START_OVERRIDE
;
1582 dev_info(dev
->mt76
.dev
, "override_addr = 0x%08x, option = %d\n",
1583 override_addr
, flag
);
1585 ret
= mt76_connac_mcu_start_firmware(&dev
->mt76
, override_addr
, flag
);
1587 dev_err(dev
->mt76
.dev
, "Failed to start N9 firmware\n");
1591 snprintf(dev
->mt76
.hw
->wiphy
->fw_version
,
1592 sizeof(dev
->mt76
.hw
->wiphy
->fw_version
),
1593 "%.10s-%.15s", hdr
->fw_ver
, hdr
->build_date
);
1596 release_firmware(fw
);
1602 mt7663_load_rom_patch(struct mt7615_dev
*dev
, const char **n9_firmware
)
1604 const char *selected_rom
, *secondary_rom
= MT7663_ROM_PATCH
;
1605 const char *primary_rom
= MT7663_OFFLOAD_ROM_PATCH
;
1608 if (!prefer_offload_fw
) {
1609 secondary_rom
= MT7663_OFFLOAD_ROM_PATCH
;
1610 primary_rom
= MT7663_ROM_PATCH
;
1612 selected_rom
= primary_rom
;
1614 ret
= mt7615_load_patch(dev
, MT7663_PATCH_ADDRESS
, primary_rom
);
1616 dev_info(dev
->mt76
.dev
, "%s not found, switching to %s",
1617 primary_rom
, secondary_rom
);
1618 ret
= mt7615_load_patch(dev
, MT7663_PATCH_ADDRESS
,
1621 dev_err(dev
->mt76
.dev
, "failed to load %s",
1625 selected_rom
= secondary_rom
;
1628 if (!strcmp(selected_rom
, MT7663_OFFLOAD_ROM_PATCH
)) {
1629 *n9_firmware
= MT7663_OFFLOAD_FIRMWARE_N9
;
1630 dev
->fw_ver
= MT7615_FIRMWARE_V3
;
1631 dev
->mcu_ops
= &uni_update_ops
;
1633 *n9_firmware
= MT7663_FIRMWARE_N9
;
1634 dev
->fw_ver
= MT7615_FIRMWARE_V2
;
1635 dev
->mcu_ops
= &sta_update_ops
;
1641 int __mt7663_load_firmware(struct mt7615_dev
*dev
)
1643 const char *n9_firmware
;
1646 ret
= mt76_get_field(dev
, MT_CONN_ON_MISC
, MT_TOP_MISC2_FW_N9_RDY
);
1648 dev_dbg(dev
->mt76
.dev
, "Firmware is already download\n");
1652 ret
= mt7663_load_rom_patch(dev
, &n9_firmware
);
1656 ret
= mt7663_load_n9(dev
, n9_firmware
);
1660 if (!mt76_poll_msec(dev
, MT_CONN_ON_MISC
, MT_TOP_MISC2_FW_N9_RDY
,
1661 MT_TOP_MISC2_FW_N9_RDY
, 1500)) {
1662 ret
= mt76_get_field(dev
, MT_CONN_ON_MISC
,
1663 MT7663_TOP_MISC2_FW_STATE
);
1664 dev_err(dev
->mt76
.dev
, "Timeout for initializing firmware\n");
1669 if (mt7615_firmware_offload(dev
))
1670 dev
->mt76
.hw
->wiphy
->wowlan
= &mt76_connac_wowlan_support
;
1671 #endif /* CONFIG_PM */
1673 dev_dbg(dev
->mt76
.dev
, "Firmware init done\n");
1677 EXPORT_SYMBOL_GPL(__mt7663_load_firmware
);
1679 static int mt7663_load_firmware(struct mt7615_dev
*dev
)
1683 mt76_set(dev
, MT_WPDMA_GLO_CFG
, MT_WPDMA_GLO_CFG_BYPASS_TX_SCH
);
1685 ret
= __mt7663_load_firmware(dev
);
1689 mt76_clear(dev
, MT_WPDMA_GLO_CFG
, MT_WPDMA_GLO_CFG_BYPASS_TX_SCH
);
1694 int mt7615_mcu_init(struct mt7615_dev
*dev
)
1696 static const struct mt76_mcu_ops mt7615_mcu_ops
= {
1697 .headroom
= sizeof(struct mt7615_mcu_txd
),
1698 .mcu_skb_send_msg
= mt7615_mcu_send_message
,
1699 .mcu_parse_response
= mt7615_mcu_parse_response
,
1703 dev
->mt76
.mcu_ops
= &mt7615_mcu_ops
,
1705 ret
= mt7615_mcu_drv_pmctrl(dev
);
1709 switch (mt76_chip(&dev
->mt76
)) {
1711 ret
= mt7622_load_firmware(dev
);
1714 ret
= mt7663_load_firmware(dev
);
1717 ret
= mt7615_load_firmware(dev
);
1723 mt76_queue_tx_cleanup(dev
, dev
->mt76
.q_mcu
[MT_MCUQ_FWDL
], false);
1724 dev_dbg(dev
->mt76
.dev
, "Firmware init done\n");
1725 set_bit(MT76_STATE_MCU_RUNNING
, &dev
->mphy
.state
);
1727 if (dev
->dbdc_support
) {
1728 ret
= mt7615_mcu_cal_cache_apply(dev
);
1733 return mt7615_mcu_fw_log_2_host(dev
, 0);
1735 EXPORT_SYMBOL_GPL(mt7615_mcu_init
);
1737 void mt7615_mcu_exit(struct mt7615_dev
*dev
)
1739 mt7615_mcu_restart(&dev
->mt76
);
1740 mt7615_mcu_set_fw_ctrl(dev
);
1741 skb_queue_purge(&dev
->mt76
.mcu
.res_q
);
1743 EXPORT_SYMBOL_GPL(mt7615_mcu_exit
);
1745 int mt7615_mcu_set_eeprom(struct mt7615_dev
*dev
)
1751 } __packed req_hdr
= {
1754 u8
*eep
= (u8
*)dev
->mt76
.eeprom
.data
;
1755 struct sk_buff
*skb
;
1756 int eep_len
, offset
;
1758 switch (mt76_chip(&dev
->mt76
)) {
1760 eep_len
= MT7622_EE_MAX
- MT_EE_NIC_CONF_0
;
1761 offset
= MT_EE_NIC_CONF_0
;
1764 eep_len
= MT7663_EE_MAX
- MT_EE_CHIP_ID
;
1765 req_hdr
.content_format
= 1;
1766 offset
= MT_EE_CHIP_ID
;
1769 eep_len
= MT7615_EE_MAX
- MT_EE_NIC_CONF_0
;
1770 offset
= MT_EE_NIC_CONF_0
;
1774 req_hdr
.len
= cpu_to_le16(eep_len
);
1776 skb
= mt76_mcu_msg_alloc(&dev
->mt76
, NULL
, sizeof(req_hdr
) + eep_len
);
1780 skb_put_data(skb
, &req_hdr
, sizeof(req_hdr
));
1781 skb_put_data(skb
, eep
+ offset
, eep_len
);
1783 return mt76_mcu_skb_send_msg(&dev
->mt76
, skb
,
1784 MCU_EXT_CMD(EFUSE_BUFFER_MODE
), true);
1787 int mt7615_mcu_set_wmm(struct mt7615_dev
*dev
, u8 queue
,
1788 const struct ieee80211_tx_queue_params
*params
)
1790 #define WMM_AIFS_SET BIT(0)
1791 #define WMM_CW_MIN_SET BIT(1)
1792 #define WMM_CW_MAX_SET BIT(2)
1793 #define WMM_TXOP_SET BIT(3)
1794 #define WMM_PARAM_SET (WMM_AIFS_SET | WMM_CW_MIN_SET | \
1795 WMM_CW_MAX_SET | WMM_TXOP_SET)
1808 .valid
= WMM_PARAM_SET
,
1809 .aifs
= params
->aifs
,
1811 .cw_max
= cpu_to_le16(10),
1812 .txop
= cpu_to_le16(params
->txop
),
1816 req
.cw_min
= fls(params
->cw_min
);
1818 req
.cw_max
= cpu_to_le16(fls(params
->cw_max
));
1820 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(EDCA_UPDATE
),
1821 &req
, sizeof(req
), true);
1824 int mt7615_mcu_set_dbdc(struct mt7615_dev
*dev
)
1826 struct mt7615_phy
*ext_phy
= mt7615_ext_phy(dev
);
1837 struct dbdc_entry entry
[64];
1839 .enable
= !!ext_phy
,
1846 #define ADD_DBDC_ENTRY(_type, _idx, _band) \
1848 req.entry[req.num].type = _type; \
1849 req.entry[req.num].index = _idx; \
1850 req.entry[req.num++].band = _band; \
1853 for (i
= 0; i
< 4; i
++) {
1854 bool band
= !!(ext_phy
->omac_mask
& BIT_ULL(i
));
1856 ADD_DBDC_ENTRY(DBDC_TYPE_BSS
, i
, band
);
1859 for (i
= 0; i
< 14; i
++) {
1860 bool band
= !!(ext_phy
->omac_mask
& BIT_ULL(0x11 + i
));
1862 ADD_DBDC_ENTRY(DBDC_TYPE_MBSS
, i
, band
);
1865 ADD_DBDC_ENTRY(DBDC_TYPE_MU
, 0, 1);
1867 for (i
= 0; i
< 3; i
++)
1868 ADD_DBDC_ENTRY(DBDC_TYPE_BF
, i
, 1);
1870 ADD_DBDC_ENTRY(DBDC_TYPE_WMM
, 0, 0);
1871 ADD_DBDC_ENTRY(DBDC_TYPE_WMM
, 1, 0);
1872 ADD_DBDC_ENTRY(DBDC_TYPE_WMM
, 2, 1);
1873 ADD_DBDC_ENTRY(DBDC_TYPE_WMM
, 3, 1);
1875 ADD_DBDC_ENTRY(DBDC_TYPE_MGMT
, 0, 0);
1876 ADD_DBDC_ENTRY(DBDC_TYPE_MGMT
, 1, 1);
1879 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(DBDC_CTRL
), &req
,
1883 int mt7615_mcu_set_fcc5_lpn(struct mt7615_dev
*dev
, int val
)
1889 .tag
= cpu_to_le16(0x1),
1890 .min_lpn
= cpu_to_le16(val
),
1893 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(SET_RADAR_TH
),
1894 &req
, sizeof(req
), true);
1897 int mt7615_mcu_set_pulse_th(struct mt7615_dev
*dev
,
1898 const struct mt7615_dfs_pulse
*pulse
)
1902 __le32 max_width
; /* us */
1903 __le32 max_pwr
; /* dbm */
1904 __le32 min_pwr
; /* dbm */
1905 __le32 min_stgr_pri
; /* us */
1906 __le32 max_stgr_pri
; /* us */
1907 __le32 min_cr_pri
; /* us */
1908 __le32 max_cr_pri
; /* us */
1910 .tag
= cpu_to_le16(0x3),
1911 #define __req_field(field) .field = cpu_to_le32(pulse->field)
1912 __req_field(max_width
),
1913 __req_field(max_pwr
),
1914 __req_field(min_pwr
),
1915 __req_field(min_stgr_pri
),
1916 __req_field(max_stgr_pri
),
1917 __req_field(min_cr_pri
),
1918 __req_field(max_cr_pri
),
1922 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(SET_RADAR_TH
),
1923 &req
, sizeof(req
), true);
1926 int mt7615_mcu_set_radar_th(struct mt7615_dev
*dev
, int index
,
1927 const struct mt7615_dfs_pattern
*pattern
)
1947 .tag
= cpu_to_le16(0x2),
1948 .radar_type
= cpu_to_le16(index
),
1949 #define __req_field_u8(field) .field = pattern->field
1950 #define __req_field_u32(field) .field = cpu_to_le32(pattern->field)
1951 __req_field_u8(enb
),
1952 __req_field_u8(stgr
),
1953 __req_field_u8(min_crpn
),
1954 __req_field_u8(max_crpn
),
1955 __req_field_u8(min_crpr
),
1956 __req_field_u8(min_pw
),
1957 __req_field_u8(max_pw
),
1958 __req_field_u32(min_pri
),
1959 __req_field_u32(max_pri
),
1960 __req_field_u8(min_crbn
),
1961 __req_field_u8(max_crbn
),
1962 __req_field_u8(min_stgpn
),
1963 __req_field_u8(max_stgpn
),
1964 __req_field_u8(min_stgpr
),
1965 #undef __req_field_u8
1966 #undef __req_field_u32
1969 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(SET_RADAR_TH
),
1970 &req
, sizeof(req
), true);
1973 int mt7615_mcu_rdd_send_pattern(struct mt7615_dev
*dev
)
1984 .pulse_num
= dev
->radar_pattern
.n_pulses
,
1986 u32 start_time
= ktime_to_ms(ktime_get_boottime());
1989 if (dev
->radar_pattern
.n_pulses
> ARRAY_SIZE(req
.pattern
))
1992 /* TODO: add some noise here */
1993 for (i
= 0; i
< dev
->radar_pattern
.n_pulses
; i
++) {
1994 u32 ts
= start_time
+ i
* dev
->radar_pattern
.period
;
1996 req
.pattern
[i
].width
= cpu_to_le16(dev
->radar_pattern
.width
);
1997 req
.pattern
[i
].power
= cpu_to_le16(dev
->radar_pattern
.power
);
1998 req
.pattern
[i
].start_time
= cpu_to_le32(ts
);
2001 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(SET_RDD_PATTERN
),
2002 &req
, sizeof(req
), false);
2005 static void mt7615_mcu_set_txpower_sku(struct mt7615_phy
*phy
, u8
*sku
)
2007 struct mt76_phy
*mphy
= phy
->mt76
;
2008 struct ieee80211_hw
*hw
= mphy
->hw
;
2009 struct mt76_power_limits limits
;
2010 s8
*limits_array
= (s8
*)&limits
;
2011 int n_chains
= hweight8(mphy
->antenna_mask
);
2012 int tx_power
= hw
->conf
.power_level
* 2;
2014 static const u8 sku_mapping
[] = {
2015 #define SKU_FIELD(_type, _field) \
2016 [MT_SKU_##_type] = offsetof(struct mt76_power_limits, _field)
2017 SKU_FIELD(CCK_1_2
, cck
[0]),
2018 SKU_FIELD(CCK_55_11
, cck
[2]),
2019 SKU_FIELD(OFDM_6_9
, ofdm
[0]),
2020 SKU_FIELD(OFDM_12_18
, ofdm
[2]),
2021 SKU_FIELD(OFDM_24_36
, ofdm
[4]),
2022 SKU_FIELD(OFDM_48
, ofdm
[6]),
2023 SKU_FIELD(OFDM_54
, ofdm
[7]),
2024 SKU_FIELD(HT20_0_8
, mcs
[0][0]),
2025 SKU_FIELD(HT20_32
, ofdm
[0]),
2026 SKU_FIELD(HT20_1_2_9_10
, mcs
[0][1]),
2027 SKU_FIELD(HT20_3_4_11_12
, mcs
[0][3]),
2028 SKU_FIELD(HT20_5_13
, mcs
[0][5]),
2029 SKU_FIELD(HT20_6_14
, mcs
[0][6]),
2030 SKU_FIELD(HT20_7_15
, mcs
[0][7]),
2031 SKU_FIELD(HT40_0_8
, mcs
[1][0]),
2032 SKU_FIELD(HT40_32
, ofdm
[0]),
2033 SKU_FIELD(HT40_1_2_9_10
, mcs
[1][1]),
2034 SKU_FIELD(HT40_3_4_11_12
, mcs
[1][3]),
2035 SKU_FIELD(HT40_5_13
, mcs
[1][5]),
2036 SKU_FIELD(HT40_6_14
, mcs
[1][6]),
2037 SKU_FIELD(HT40_7_15
, mcs
[1][7]),
2038 SKU_FIELD(VHT20_0
, mcs
[0][0]),
2039 SKU_FIELD(VHT20_1_2
, mcs
[0][1]),
2040 SKU_FIELD(VHT20_3_4
, mcs
[0][3]),
2041 SKU_FIELD(VHT20_5_6
, mcs
[0][5]),
2042 SKU_FIELD(VHT20_7
, mcs
[0][7]),
2043 SKU_FIELD(VHT20_8
, mcs
[0][8]),
2044 SKU_FIELD(VHT20_9
, mcs
[0][9]),
2045 SKU_FIELD(VHT40_0
, mcs
[1][0]),
2046 SKU_FIELD(VHT40_1_2
, mcs
[1][1]),
2047 SKU_FIELD(VHT40_3_4
, mcs
[1][3]),
2048 SKU_FIELD(VHT40_5_6
, mcs
[1][5]),
2049 SKU_FIELD(VHT40_7
, mcs
[1][7]),
2050 SKU_FIELD(VHT40_8
, mcs
[1][8]),
2051 SKU_FIELD(VHT40_9
, mcs
[1][9]),
2052 SKU_FIELD(VHT80_0
, mcs
[2][0]),
2053 SKU_FIELD(VHT80_1_2
, mcs
[2][1]),
2054 SKU_FIELD(VHT80_3_4
, mcs
[2][3]),
2055 SKU_FIELD(VHT80_5_6
, mcs
[2][5]),
2056 SKU_FIELD(VHT80_7
, mcs
[2][7]),
2057 SKU_FIELD(VHT80_8
, mcs
[2][8]),
2058 SKU_FIELD(VHT80_9
, mcs
[2][9]),
2059 SKU_FIELD(VHT160_0
, mcs
[3][0]),
2060 SKU_FIELD(VHT160_1_2
, mcs
[3][1]),
2061 SKU_FIELD(VHT160_3_4
, mcs
[3][3]),
2062 SKU_FIELD(VHT160_5_6
, mcs
[3][5]),
2063 SKU_FIELD(VHT160_7
, mcs
[3][7]),
2064 SKU_FIELD(VHT160_8
, mcs
[3][8]),
2065 SKU_FIELD(VHT160_9
, mcs
[3][9]),
2069 tx_power
= mt76_get_sar_power(mphy
, mphy
->chandef
.chan
, tx_power
);
2070 tx_power
-= mt76_tx_power_nss_delta(n_chains
);
2071 tx_power
= mt76_get_rate_power_limits(mphy
, mphy
->chandef
.chan
,
2073 mphy
->txpower_cur
= tx_power
;
2075 if (is_mt7663(mphy
->dev
)) {
2076 memset(sku
, tx_power
, MT_SKU_4SS_DELTA
+ 1);
2080 for (i
= 0; i
< MT_SKU_1SS_DELTA
; i
++)
2081 sku
[i
] = limits_array
[sku_mapping
[i
]];
2083 for (i
= 0; i
< 4; i
++) {
2086 if (i
< n_chains
- 1)
2087 delta
= mt76_tx_power_nss_delta(n_chains
) -
2088 mt76_tx_power_nss_delta(i
+ 1);
2089 sku
[MT_SKU_1SS_DELTA
+ i
] = delta
;
2093 static u8
mt7615_mcu_chan_bw(struct cfg80211_chan_def
*chandef
)
2095 static const u8 width_to_bw
[] = {
2096 [NL80211_CHAN_WIDTH_40
] = CMD_CBW_40MHZ
,
2097 [NL80211_CHAN_WIDTH_80
] = CMD_CBW_80MHZ
,
2098 [NL80211_CHAN_WIDTH_80P80
] = CMD_CBW_8080MHZ
,
2099 [NL80211_CHAN_WIDTH_160
] = CMD_CBW_160MHZ
,
2100 [NL80211_CHAN_WIDTH_5
] = CMD_CBW_5MHZ
,
2101 [NL80211_CHAN_WIDTH_10
] = CMD_CBW_10MHZ
,
2102 [NL80211_CHAN_WIDTH_20
] = CMD_CBW_20MHZ
,
2103 [NL80211_CHAN_WIDTH_20_NOHT
] = CMD_CBW_20MHZ
,
2106 if (chandef
->width
>= ARRAY_SIZE(width_to_bw
))
2109 return width_to_bw
[chandef
->width
];
2112 int mt7615_mcu_set_chan_info(struct mt7615_phy
*phy
, int cmd
)
2114 struct mt7615_dev
*dev
= phy
->dev
;
2115 struct cfg80211_chan_def
*chandef
= &phy
->mt76
->chandef
;
2116 int freq1
= chandef
->center_freq1
, freq2
= chandef
->center_freq2
;
2125 /* for 80+80 only */
2130 __le32 outband_freq
;
2136 .control_chan
= chandef
->chan
->hw_value
,
2137 .center_chan
= ieee80211_frequency_to_channel(freq1
),
2138 .tx_streams
= hweight8(phy
->mt76
->antenna_mask
),
2139 .rx_streams_mask
= phy
->mt76
->chainmask
,
2140 .center_chan2
= ieee80211_frequency_to_channel(freq2
),
2143 if (cmd
== MCU_EXT_CMD(SET_RX_PATH
) ||
2144 phy
->mt76
->hw
->conf
.flags
& IEEE80211_CONF_MONITOR
)
2145 req
.switch_reason
= CH_SWITCH_NORMAL
;
2146 else if (phy
->mt76
->offchannel
)
2147 req
.switch_reason
= CH_SWITCH_SCAN_BYPASS_DPD
;
2148 else if (!cfg80211_reg_can_beacon(phy
->mt76
->hw
->wiphy
, chandef
,
2150 req
.switch_reason
= CH_SWITCH_DFS
;
2152 req
.switch_reason
= CH_SWITCH_NORMAL
;
2154 req
.band_idx
= phy
!= &dev
->phy
;
2155 req
.bw
= mt7615_mcu_chan_bw(chandef
);
2157 if (mt76_testmode_enabled(phy
->mt76
))
2158 memset(req
.txpower_sku
, 0x3f, 49);
2160 mt7615_mcu_set_txpower_sku(phy
, req
.txpower_sku
);
2162 return mt76_mcu_send_msg(&dev
->mt76
, cmd
, &req
, sizeof(req
), true);
2165 int mt7615_mcu_get_temperature(struct mt7615_dev
*dev
)
2172 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(THERMAL_CTRL
),
2173 &req
, sizeof(req
), true);
2176 int mt7615_mcu_set_test_param(struct mt7615_dev
*dev
, u8 param
, bool test_mode
,
2188 .test_mode_en
= test_mode
,
2190 .value
= cpu_to_le32(val
),
2193 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(ATE_CTRL
),
2194 &req
, sizeof(req
), false);
2197 int mt7615_mcu_set_sku_en(struct mt7615_phy
*phy
, bool enable
)
2199 struct mt7615_dev
*dev
= phy
->dev
;
2207 .band_idx
= phy
!= &dev
->phy
,
2208 .sku_enable
= enable
,
2211 return mt76_mcu_send_msg(&dev
->mt76
,
2212 MCU_EXT_CMD(TX_POWER_FEATURE_CTRL
),
2213 &req
, sizeof(req
), true);
2216 static int mt7615_find_freq_idx(const u16
*freqs
, int n_freqs
, u16 cur
)
2220 for (i
= 0; i
< n_freqs
; i
++)
2221 if (cur
== freqs
[i
])
2227 static int mt7615_dcoc_freq_idx(u16 freq
, u8 bw
)
2229 static const u16 freq_list
[] = {
2230 4980, 5805, 5905, 5190,
2231 5230, 5270, 5310, 5350,
2232 5390, 5430, 5470, 5510,
2233 5550, 5590, 5630, 5670,
2234 5710, 5755, 5795, 5835,
2235 5875, 5210, 5290, 5370,
2236 5450, 5530, 5610, 5690,
2239 static const u16 freq_bw40
[] = {
2240 5190, 5230, 5270, 5310,
2241 5350, 5390, 5430, 5470,
2242 5510, 5550, 5590, 5630,
2243 5670, 5710, 5755, 5795,
2246 int offset_2g
= ARRAY_SIZE(freq_list
);
2253 return offset_2g
+ 1;
2255 return offset_2g
+ 2;
2257 return offset_2g
+ 3;
2261 case NL80211_CHAN_WIDTH_80
:
2262 case NL80211_CHAN_WIDTH_80P80
:
2263 case NL80211_CHAN_WIDTH_160
:
2266 idx
= mt7615_find_freq_idx(freq_bw40
, ARRAY_SIZE(freq_bw40
),
2269 freq
= freq_bw40
[idx
];
2273 idx
= mt7615_find_freq_idx(freq_bw40
, ARRAY_SIZE(freq_bw40
),
2276 freq
= freq_bw40
[idx
];
2280 case NL80211_CHAN_WIDTH_40
:
2281 idx
= mt7615_find_freq_idx(freq_bw40
, ARRAY_SIZE(freq_bw40
),
2290 return mt7615_find_freq_idx(freq_list
, ARRAY_SIZE(freq_list
), freq
);
2293 int mt7615_mcu_apply_rx_dcoc(struct mt7615_phy
*phy
)
2295 struct mt7615_dev
*dev
= phy
->dev
;
2296 struct cfg80211_chan_def
*chandef
= &phy
->mt76
->chandef
;
2297 int freq2
= chandef
->center_freq2
;
2301 u8 runtime_calibration
;
2314 __le32 sx0_i_lna
[4];
2315 __le32 sx0_q_lna
[4];
2317 __le32 sx2_i_lna
[4];
2318 __le32 sx2_q_lna
[4];
2323 .bw
= mt7615_mcu_chan_bw(chandef
),
2324 .band
= chandef
->center_freq1
> 4000,
2325 .dbdc_en
= !!dev
->mt76
.phys
[MT_BAND1
],
2327 u16 center_freq
= chandef
->center_freq1
;
2329 u8
*eep
= dev
->mt76
.eeprom
.data
;
2331 if (!(eep
[MT_EE_CALDATA_FLASH
] & MT_EE_CALDATA_FLASH_RX_CAL
))
2334 if (chandef
->width
== NL80211_CHAN_WIDTH_160
) {
2335 freq2
= center_freq
+ 40;
2340 req
.runtime_calibration
= 1;
2341 freq_idx
= mt7615_dcoc_freq_idx(center_freq
, chandef
->width
);
2345 memcpy(req
.dcoc_data
, eep
+ MT7615_EEPROM_DCOC_OFFSET
+
2346 freq_idx
* MT7615_EEPROM_DCOC_SIZE
,
2347 sizeof(req
.dcoc_data
));
2348 req
.runtime_calibration
= 0;
2351 req
.center_freq
= cpu_to_le16(center_freq
);
2352 ret
= mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(RXDCOC_CAL
), &req
,
2355 if ((chandef
->width
== NL80211_CHAN_WIDTH_80P80
||
2356 chandef
->width
== NL80211_CHAN_WIDTH_160
) && !req
.is_freq2
) {
2357 req
.is_freq2
= true;
2358 center_freq
= freq2
;
2365 static int mt7615_dpd_freq_idx(u16 freq
, u8 bw
)
2367 static const u16 freq_list
[] = {
2368 4920, 4940, 4960, 4980,
2369 5040, 5060, 5080, 5180,
2370 5200, 5220, 5240, 5260,
2371 5280, 5300, 5320, 5340,
2372 5360, 5380, 5400, 5420,
2373 5440, 5460, 5480, 5500,
2374 5520, 5540, 5560, 5580,
2375 5600, 5620, 5640, 5660,
2376 5680, 5700, 5720, 5745,
2377 5765, 5785, 5805, 5825,
2378 5845, 5865, 5885, 5905
2380 int offset_2g
= ARRAY_SIZE(freq_list
);
2387 return offset_2g
+ 1;
2389 return offset_2g
+ 2;
2392 if (bw
!= NL80211_CHAN_WIDTH_20
) {
2393 idx
= mt7615_find_freq_idx(freq_list
, ARRAY_SIZE(freq_list
),
2398 idx
= mt7615_find_freq_idx(freq_list
, ARRAY_SIZE(freq_list
),
2404 return mt7615_find_freq_idx(freq_list
, ARRAY_SIZE(freq_list
), freq
);
2408 int mt7615_mcu_apply_tx_dpd(struct mt7615_phy
*phy
)
2410 struct mt7615_dev
*dev
= phy
->dev
;
2411 struct cfg80211_chan_def
*chandef
= &phy
->mt76
->chandef
;
2412 int freq2
= chandef
->center_freq2
;
2416 u8 runtime_calibration
;
2444 .bw
= mt7615_mcu_chan_bw(chandef
),
2445 .band
= chandef
->center_freq1
> 4000,
2446 .dbdc_en
= !!dev
->mt76
.phys
[MT_BAND1
],
2448 u16 center_freq
= chandef
->center_freq1
;
2450 u8
*eep
= dev
->mt76
.eeprom
.data
;
2452 if (!(eep
[MT_EE_CALDATA_FLASH
] & MT_EE_CALDATA_FLASH_TX_DPD
))
2455 if (chandef
->width
== NL80211_CHAN_WIDTH_160
) {
2456 freq2
= center_freq
+ 40;
2461 req
.runtime_calibration
= 1;
2462 freq_idx
= mt7615_dpd_freq_idx(center_freq
, chandef
->width
);
2466 memcpy(&req
.dpd_data
, eep
+ MT7615_EEPROM_TXDPD_OFFSET
+
2467 freq_idx
* MT7615_EEPROM_TXDPD_SIZE
,
2468 sizeof(req
.dpd_data
));
2469 req
.runtime_calibration
= 0;
2472 req
.center_freq
= cpu_to_le16(center_freq
);
2473 ret
= mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(TXDPD_CAL
),
2474 &req
, sizeof(req
), true);
2476 if ((chandef
->width
== NL80211_CHAN_WIDTH_80P80
||
2477 chandef
->width
== NL80211_CHAN_WIDTH_160
) && !req
.is_freq2
) {
2478 req
.is_freq2
= true;
2479 center_freq
= freq2
;
2486 int mt7615_mcu_set_rx_hdr_trans_blacklist(struct mt7615_dev
*dev
)
2499 .etype
= cpu_to_le16(ETH_P_PAE
),
2502 return mt76_mcu_send_msg(&dev
->mt76
, MCU_EXT_CMD(RX_HDR_TRANS
),
2503 &req
, sizeof(req
), false);
2506 int mt7615_mcu_set_bss_pm(struct mt7615_dev
*dev
, struct ieee80211_vif
*vif
,
2509 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
2514 __le16 bcn_interval
;
2517 u8 bmc_delivered_ac
;
2518 u8 bmc_triggered_ac
;
2521 .bss_idx
= mvif
->mt76
.idx
,
2522 .aid
= cpu_to_le16(vif
->cfg
.aid
),
2523 .dtim_period
= vif
->bss_conf
.dtim_period
,
2524 .bcn_interval
= cpu_to_le16(vif
->bss_conf
.beacon_int
),
2530 .bss_idx
= mvif
->mt76
.idx
,
2534 if (vif
->type
!= NL80211_IFTYPE_STATION
)
2537 err
= mt76_mcu_send_msg(&dev
->mt76
, MCU_CE_CMD(SET_BSS_ABORT
),
2538 &req_hdr
, sizeof(req_hdr
), false);
2539 if (err
< 0 || !enable
)
2542 return mt76_mcu_send_msg(&dev
->mt76
, MCU_CE_CMD(SET_BSS_CONNECTED
),
2543 &req
, sizeof(req
), false);
2546 int mt7615_mcu_set_roc(struct mt7615_phy
*phy
, struct ieee80211_vif
*vif
,
2547 struct ieee80211_channel
*chan
, int duration
)
2549 struct mt7615_vif
*mvif
= (struct mt7615_vif
*)vif
->drv_priv
;
2550 struct mt7615_dev
*dev
= phy
->dev
;
2551 struct mt7615_roc_tlv req
= {
2552 .bss_idx
= mvif
->mt76
.idx
,
2554 .max_interval
= cpu_to_le32(duration
),
2555 .primary_chan
= chan
? chan
->hw_value
: 0,
2556 .band
= chan
? chan
->band
: 0,
2560 phy
->roc_grant
= false;
2562 return mt76_mcu_send_msg(&dev
->mt76
, MCU_CE_CMD(SET_ROC
),
2563 &req
, sizeof(req
), false);