1 /* SPDX-License-Identifier: ISC */
3 * Copyright (C) 2016 Felix Fietkau <nbd@nbd.name>
8 #define DMA_DUMMY_DATA ((void *)~0)
10 #define MT_RING_SIZE 0x10
12 #define MT_DMA_CTL_SD_LEN1 GENMASK(13, 0)
13 #define MT_DMA_CTL_LAST_SEC1 BIT(14)
14 #define MT_DMA_CTL_BURST BIT(15)
15 #define MT_DMA_CTL_SD_LEN0 GENMASK(29, 16)
16 #define MT_DMA_CTL_LAST_SEC0 BIT(30)
17 #define MT_DMA_CTL_DMA_DONE BIT(31)
18 #define MT_DMA_CTL_TO_HOST BIT(8)
19 #define MT_DMA_CTL_TO_HOST_A BIT(12)
20 #define MT_DMA_CTL_DROP BIT(14)
21 #define MT_DMA_CTL_TOKEN GENMASK(31, 16)
22 #define MT_DMA_CTL_SDP1_H GENMASK(19, 16)
23 #define MT_DMA_CTL_SDP0_H GENMASK(3, 0)
24 #define MT_DMA_CTL_WO_DROP BIT(8)
26 #define MT_DMA_PPE_CPU_REASON GENMASK(15, 11)
27 #define MT_DMA_PPE_ENTRY GENMASK(30, 16)
28 #define MT_DMA_INFO_DMA_FRAG BIT(9)
29 #define MT_DMA_INFO_PPE_VLD BIT(31)
31 #define MT_DMA_CTL_PN_CHK_FAIL BIT(13)
32 #define MT_DMA_CTL_VER_MASK BIT(7)
34 #define MT_DMA_RRO_EN BIT(13)
36 #define MT_DMA_WED_IND_CMD_CNT 8
37 #define MT_DMA_WED_IND_REASON GENMASK(15, 12)
39 #define MT_DMA_HDR_LEN 4
40 #define MT_RX_INFO_LEN 4
41 #define MT_FCE_INFO_LEN 4
42 #define MT_RX_RXWI_LEN 32
49 } __packed
__aligned(4);
51 struct mt76_wed_rro_desc
{
54 } __packed
__aligned(4);
63 enum mt76_mcu_evt_type
{
69 EVT_EVENT_CARRIER_DETECT_RSP
,
70 EVT_EVENT_DFS_DETECT_RSP
,
73 enum mt76_dma_wed_ind_reason
{
74 MT_DMA_WED_IND_REASON_NORMAL
,
75 MT_DMA_WED_IND_REASON_REPEAT
,
76 MT_DMA_WED_IND_REASON_OLDPKT
,
79 int mt76_dma_rx_poll(struct napi_struct
*napi
, int budget
);
80 void mt76_dma_attach(struct mt76_dev
*dev
);
81 void mt76_dma_cleanup(struct mt76_dev
*dev
);
82 int mt76_dma_rx_fill(struct mt76_dev
*dev
, struct mt76_queue
*q
,
84 void __mt76_dma_queue_reset(struct mt76_dev
*dev
, struct mt76_queue
*q
,
86 void mt76_dma_queue_reset(struct mt76_dev
*dev
, struct mt76_queue
*q
);
89 mt76_dma_reset_tx_queue(struct mt76_dev
*dev
, struct mt76_queue
*q
)
91 dev
->queue_ops
->reset_q(dev
, q
);
92 if (mtk_wed_device_active(&dev
->mmio
.wed
))
93 mt76_wed_dma_setup(dev
, q
, true);
97 mt76_dma_should_drop_buf(bool *drop
, u32 ctrl
, u32 buf1
, u32 info
)
102 *drop
= !!(ctrl
& (MT_DMA_CTL_TO_HOST_A
| MT_DMA_CTL_DROP
));
103 if (!(ctrl
& MT_DMA_CTL_VER_MASK
))
106 switch (FIELD_GET(MT_DMA_WED_IND_REASON
, buf1
)) {
107 case MT_DMA_WED_IND_REASON_REPEAT
:
110 case MT_DMA_WED_IND_REASON_OLDPKT
:
111 *drop
= !(info
& MT_DMA_INFO_DMA_FRAG
);
114 *drop
= !!(ctrl
& MT_DMA_CTL_PN_CHK_FAIL
);
119 static inline void *mt76_priv(struct net_device
*dev
)
121 struct mt76_dev
**priv
;
123 priv
= netdev_priv(dev
);