Merge tag 'pull-loongarch-20241016' of https://gitlab.com/gaosong/qemu into staging
[qemu/armbru.git] / net / queue.c
blobfb33856c188916a31e62bfd7a307a70d6fdc4c9e
1 /*
2 * Copyright (c) 2003-2008 Fabrice Bellard
3 * Copyright (c) 2009 Red Hat, Inc.
5 * Permission is hereby granted, free of charge, to any person obtaining a copy
6 * of this software and associated documentation files (the "Software"), to deal
7 * in the Software without restriction, including without limitation the rights
8 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9 * copies of the Software, and to permit persons to whom the Software is
10 * furnished to do so, subject to the following conditions:
12 * The above copyright notice and this permission notice shall be included in
13 * all copies or substantial portions of the Software.
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
21 * THE SOFTWARE.
24 #include "qemu/osdep.h"
25 #include "net/queue.h"
26 #include "qemu/queue.h"
27 #include "net/net.h"
29 /* The delivery handler may only return zero if it will call
30 * qemu_net_queue_flush() when it determines that it is once again able
31 * to deliver packets. It must also call qemu_net_queue_purge() in its
32 * cleanup path.
34 * If a sent callback is provided to send(), the caller must handle a
35 * zero return from the delivery handler by not sending any more packets
36 * until we have invoked the callback. Only in that case will we queue
37 * the packet.
39 * If a sent callback isn't provided, we just drop the packet to avoid
40 * unbounded queueing.
43 struct NetPacket {
44 QTAILQ_ENTRY(NetPacket) entry;
45 NetClientState *sender;
46 unsigned flags;
47 int size;
48 NetPacketSent *sent_cb;
49 uint8_t data[];
52 struct NetQueue {
53 void *opaque;
54 uint32_t nq_maxlen;
55 uint32_t nq_count;
56 NetQueueDeliverFunc *deliver;
58 QTAILQ_HEAD(, NetPacket) packets;
60 unsigned delivering : 1;
63 NetQueue *qemu_new_net_queue(NetQueueDeliverFunc *deliver, void *opaque)
65 NetQueue *queue;
67 queue = g_new0(NetQueue, 1);
69 queue->opaque = opaque;
70 queue->nq_maxlen = 10000;
71 queue->nq_count = 0;
72 queue->deliver = deliver;
74 QTAILQ_INIT(&queue->packets);
76 queue->delivering = 0;
78 return queue;
81 void qemu_del_net_queue(NetQueue *queue)
83 NetPacket *packet, *next;
85 QTAILQ_FOREACH_SAFE(packet, &queue->packets, entry, next) {
86 QTAILQ_REMOVE(&queue->packets, packet, entry);
87 g_free(packet);
90 g_free(queue);
93 static void qemu_net_queue_append(NetQueue *queue,
94 NetClientState *sender,
95 unsigned flags,
96 const uint8_t *buf,
97 size_t size,
98 NetPacketSent *sent_cb)
100 NetPacket *packet;
102 if (queue->nq_count >= queue->nq_maxlen && !sent_cb) {
103 return; /* drop if queue full and no callback */
105 packet = g_malloc(sizeof(NetPacket) + size);
106 packet->sender = sender;
107 packet->flags = flags;
108 packet->size = size;
109 packet->sent_cb = sent_cb;
110 memcpy(packet->data, buf, size);
112 queue->nq_count++;
113 QTAILQ_INSERT_TAIL(&queue->packets, packet, entry);
116 void qemu_net_queue_append_iov(NetQueue *queue,
117 NetClientState *sender,
118 unsigned flags,
119 const struct iovec *iov,
120 int iovcnt,
121 NetPacketSent *sent_cb)
123 NetPacket *packet;
124 size_t max_len = 0;
125 int i;
127 if (queue->nq_count >= queue->nq_maxlen && !sent_cb) {
128 return; /* drop if queue full and no callback */
130 for (i = 0; i < iovcnt; i++) {
131 max_len += iov[i].iov_len;
134 packet = g_malloc(sizeof(NetPacket) + max_len);
135 packet->sender = sender;
136 packet->sent_cb = sent_cb;
137 packet->flags = flags;
138 packet->size = 0;
140 for (i = 0; i < iovcnt; i++) {
141 size_t len = iov[i].iov_len;
143 memcpy(packet->data + packet->size, iov[i].iov_base, len);
144 packet->size += len;
147 queue->nq_count++;
148 QTAILQ_INSERT_TAIL(&queue->packets, packet, entry);
151 static ssize_t qemu_net_queue_deliver(NetQueue *queue,
152 NetClientState *sender,
153 unsigned flags,
154 const uint8_t *data,
155 size_t size)
157 ssize_t ret = -1;
158 struct iovec iov = {
159 .iov_base = (void *)data,
160 .iov_len = size
163 queue->delivering = 1;
164 ret = queue->deliver(sender, flags, &iov, 1, queue->opaque);
165 queue->delivering = 0;
167 return ret;
170 static ssize_t qemu_net_queue_deliver_iov(NetQueue *queue,
171 NetClientState *sender,
172 unsigned flags,
173 const struct iovec *iov,
174 int iovcnt)
176 ssize_t ret = -1;
178 queue->delivering = 1;
179 ret = queue->deliver(sender, flags, iov, iovcnt, queue->opaque);
180 queue->delivering = 0;
182 return ret;
185 ssize_t qemu_net_queue_receive(NetQueue *queue,
186 const uint8_t *data,
187 size_t size)
189 if (queue->delivering) {
190 return 0;
193 return qemu_net_queue_deliver(queue, NULL, 0, data, size);
196 ssize_t qemu_net_queue_send(NetQueue *queue,
197 NetClientState *sender,
198 unsigned flags,
199 const uint8_t *data,
200 size_t size,
201 NetPacketSent *sent_cb)
203 ssize_t ret;
205 if (queue->delivering || !qemu_can_send_packet(sender)) {
206 qemu_net_queue_append(queue, sender, flags, data, size, sent_cb);
207 return 0;
210 ret = qemu_net_queue_deliver(queue, sender, flags, data, size);
211 if (ret == 0) {
212 qemu_net_queue_append(queue, sender, flags, data, size, sent_cb);
213 return 0;
216 qemu_net_queue_flush(queue);
218 return ret;
221 ssize_t qemu_net_queue_send_iov(NetQueue *queue,
222 NetClientState *sender,
223 unsigned flags,
224 const struct iovec *iov,
225 int iovcnt,
226 NetPacketSent *sent_cb)
228 ssize_t ret;
230 if (queue->delivering || !qemu_can_send_packet(sender)) {
231 qemu_net_queue_append_iov(queue, sender, flags, iov, iovcnt, sent_cb);
232 return 0;
235 ret = qemu_net_queue_deliver_iov(queue, sender, flags, iov, iovcnt);
236 if (ret == 0) {
237 qemu_net_queue_append_iov(queue, sender, flags, iov, iovcnt, sent_cb);
238 return 0;
241 qemu_net_queue_flush(queue);
243 return ret;
246 void qemu_net_queue_purge(NetQueue *queue, NetClientState *from)
248 NetPacket *packet, *next;
250 QTAILQ_FOREACH_SAFE(packet, &queue->packets, entry, next) {
251 if (packet->sender == from) {
252 QTAILQ_REMOVE(&queue->packets, packet, entry);
253 queue->nq_count--;
254 if (packet->sent_cb) {
255 packet->sent_cb(packet->sender, 0);
257 g_free(packet);
262 bool qemu_net_queue_flush(NetQueue *queue)
264 if (queue->delivering)
265 return false;
267 while (!QTAILQ_EMPTY(&queue->packets)) {
268 NetPacket *packet;
269 int ret;
271 packet = QTAILQ_FIRST(&queue->packets);
272 QTAILQ_REMOVE(&queue->packets, packet, entry);
273 queue->nq_count--;
275 ret = qemu_net_queue_deliver(queue,
276 packet->sender,
277 packet->flags,
278 packet->data,
279 packet->size);
280 if (ret == 0) {
281 queue->nq_count++;
282 QTAILQ_INSERT_HEAD(&queue->packets, packet, entry);
283 return false;
286 if (packet->sent_cb) {
287 packet->sent_cb(packet->sender, ret);
290 g_free(packet);
292 return true;