treewide: remove redundant IS_ERR() before error code check
[linux/fpc-iii.git] / drivers / net / wireless / quantenna / qtnfmac / shm_ipc.c
blobff678951d3b2453930495aa5434f8834bbce3c78
1 // SPDX-License-Identifier: GPL-2.0+
2 /* Copyright (c) 2015-2016 Quantenna Communications. All rights reserved. */
4 #include <linux/types.h>
5 #include <linux/io.h>
7 #include "shm_ipc.h"
9 #undef pr_fmt
10 #define pr_fmt(fmt) "qtnfmac shm_ipc: %s: " fmt, __func__
12 static bool qtnf_shm_ipc_has_new_data(struct qtnf_shm_ipc *ipc)
14 const u32 flags = readl(&ipc->shm_region->headroom.hdr.flags);
16 return (flags & QTNF_SHM_IPC_NEW_DATA);
19 static void qtnf_shm_handle_new_data(struct qtnf_shm_ipc *ipc)
21 size_t size;
22 bool rx_buff_ok = true;
23 struct qtnf_shm_ipc_region_header __iomem *shm_reg_hdr;
25 shm_reg_hdr = &ipc->shm_region->headroom.hdr;
27 size = readw(&shm_reg_hdr->data_len);
29 if (unlikely(size == 0 || size > QTN_IPC_MAX_DATA_SZ)) {
30 pr_err("wrong rx packet size: %zu\n", size);
31 rx_buff_ok = false;
34 if (likely(rx_buff_ok)) {
35 ipc->rx_packet_count++;
36 ipc->rx_callback.fn(ipc->rx_callback.arg,
37 ipc->shm_region->data, size);
40 writel(QTNF_SHM_IPC_ACK, &shm_reg_hdr->flags);
41 readl(&shm_reg_hdr->flags); /* flush PCIe write */
43 ipc->interrupt.fn(ipc->interrupt.arg);
46 static void qtnf_shm_ipc_irq_work(struct work_struct *work)
48 struct qtnf_shm_ipc *ipc = container_of(work, struct qtnf_shm_ipc,
49 irq_work);
51 while (qtnf_shm_ipc_has_new_data(ipc))
52 qtnf_shm_handle_new_data(ipc);
55 static void qtnf_shm_ipc_irq_inbound_handler(struct qtnf_shm_ipc *ipc)
57 u32 flags;
59 flags = readl(&ipc->shm_region->headroom.hdr.flags);
61 if (flags & QTNF_SHM_IPC_NEW_DATA)
62 queue_work(ipc->workqueue, &ipc->irq_work);
65 static void qtnf_shm_ipc_irq_outbound_handler(struct qtnf_shm_ipc *ipc)
67 u32 flags;
69 if (!READ_ONCE(ipc->waiting_for_ack))
70 return;
72 flags = readl(&ipc->shm_region->headroom.hdr.flags);
74 if (flags & QTNF_SHM_IPC_ACK) {
75 WRITE_ONCE(ipc->waiting_for_ack, 0);
76 complete(&ipc->tx_completion);
80 int qtnf_shm_ipc_init(struct qtnf_shm_ipc *ipc,
81 enum qtnf_shm_ipc_direction direction,
82 struct qtnf_shm_ipc_region __iomem *shm_region,
83 struct workqueue_struct *workqueue,
84 const struct qtnf_shm_ipc_int *interrupt,
85 const struct qtnf_shm_ipc_rx_callback *rx_callback)
87 BUILD_BUG_ON(offsetof(struct qtnf_shm_ipc_region, data) !=
88 QTN_IPC_REG_HDR_SZ);
89 BUILD_BUG_ON(sizeof(struct qtnf_shm_ipc_region) > QTN_IPC_REG_SZ);
91 ipc->shm_region = shm_region;
92 ipc->direction = direction;
93 ipc->interrupt = *interrupt;
94 ipc->rx_callback = *rx_callback;
95 ipc->tx_packet_count = 0;
96 ipc->rx_packet_count = 0;
97 ipc->workqueue = workqueue;
98 ipc->waiting_for_ack = 0;
99 ipc->tx_timeout_count = 0;
101 switch (direction) {
102 case QTNF_SHM_IPC_OUTBOUND:
103 ipc->irq_handler = qtnf_shm_ipc_irq_outbound_handler;
104 break;
105 case QTNF_SHM_IPC_INBOUND:
106 ipc->irq_handler = qtnf_shm_ipc_irq_inbound_handler;
107 break;
108 default:
109 return -EINVAL;
112 INIT_WORK(&ipc->irq_work, qtnf_shm_ipc_irq_work);
113 init_completion(&ipc->tx_completion);
115 return 0;
118 void qtnf_shm_ipc_free(struct qtnf_shm_ipc *ipc)
120 complete_all(&ipc->tx_completion);
123 int qtnf_shm_ipc_send(struct qtnf_shm_ipc *ipc, const u8 *buf, size_t size)
125 int ret = 0;
126 struct qtnf_shm_ipc_region_header __iomem *shm_reg_hdr;
128 shm_reg_hdr = &ipc->shm_region->headroom.hdr;
130 if (unlikely(size > QTN_IPC_MAX_DATA_SZ))
131 return -E2BIG;
133 ipc->tx_packet_count++;
135 writew(size, &shm_reg_hdr->data_len);
136 memcpy_toio(ipc->shm_region->data, buf, size);
138 /* sync previous writes before proceeding */
139 dma_wmb();
141 WRITE_ONCE(ipc->waiting_for_ack, 1);
143 /* sync previous memory write before announcing new data ready */
144 wmb();
146 writel(QTNF_SHM_IPC_NEW_DATA, &shm_reg_hdr->flags);
147 readl(&shm_reg_hdr->flags); /* flush PCIe write */
149 ipc->interrupt.fn(ipc->interrupt.arg);
151 if (!wait_for_completion_timeout(&ipc->tx_completion,
152 QTN_SHM_IPC_ACK_TIMEOUT)) {
153 ret = -ETIMEDOUT;
154 ipc->tx_timeout_count++;
155 pr_err("TX ACK timeout\n");
158 /* now we're not waiting for ACK even in case of timeout */
159 WRITE_ONCE(ipc->waiting_for_ack, 0);
161 return ret;