milkymist: convert to memory API
[qemu/qmp-unstable.git] / net / queue.c
blob1ab5247a327d9918c0227cdc13232e8cd2fe4d06
1 /*
2 * Copyright (c) 2003-2008 Fabrice Bellard
3 * Copyright (c) 2009 Red Hat, Inc.
5 * Permission is hereby granted, free of charge, to any person obtaining a copy
6 * of this software and associated documentation files (the "Software"), to deal
7 * in the Software without restriction, including without limitation the rights
8 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9 * copies of the Software, and to permit persons to whom the Software is
10 * furnished to do so, subject to the following conditions:
12 * The above copyright notice and this permission notice shall be included in
13 * all copies or substantial portions of the Software.
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
21 * THE SOFTWARE.
24 #include "net/queue.h"
25 #include "qemu-queue.h"
27 /* The delivery handler may only return zero if it will call
28 * qemu_net_queue_flush() when it determines that it is once again able
29 * to deliver packets. It must also call qemu_net_queue_purge() in its
30 * cleanup path.
32 * If a sent callback is provided to send(), the caller must handle a
33 * zero return from the delivery handler by not sending any more packets
34 * until we have invoked the callback. Only in that case will we queue
35 * the packet.
37 * If a sent callback isn't provided, we just drop the packet to avoid
38 * unbounded queueing.
41 struct NetPacket {
42 QTAILQ_ENTRY(NetPacket) entry;
43 VLANClientState *sender;
44 unsigned flags;
45 int size;
46 NetPacketSent *sent_cb;
47 uint8_t data[0];
50 struct NetQueue {
51 NetPacketDeliver *deliver;
52 NetPacketDeliverIOV *deliver_iov;
53 void *opaque;
55 QTAILQ_HEAD(packets, NetPacket) packets;
57 unsigned delivering : 1;
60 NetQueue *qemu_new_net_queue(NetPacketDeliver *deliver,
61 NetPacketDeliverIOV *deliver_iov,
62 void *opaque)
64 NetQueue *queue;
66 queue = g_malloc0(sizeof(NetQueue));
68 queue->deliver = deliver;
69 queue->deliver_iov = deliver_iov;
70 queue->opaque = opaque;
72 QTAILQ_INIT(&queue->packets);
74 queue->delivering = 0;
76 return queue;
79 void qemu_del_net_queue(NetQueue *queue)
81 NetPacket *packet, *next;
83 QTAILQ_FOREACH_SAFE(packet, &queue->packets, entry, next) {
84 QTAILQ_REMOVE(&queue->packets, packet, entry);
85 g_free(packet);
88 g_free(queue);
91 static ssize_t qemu_net_queue_append(NetQueue *queue,
92 VLANClientState *sender,
93 unsigned flags,
94 const uint8_t *buf,
95 size_t size,
96 NetPacketSent *sent_cb)
98 NetPacket *packet;
100 packet = g_malloc(sizeof(NetPacket) + size);
101 packet->sender = sender;
102 packet->flags = flags;
103 packet->size = size;
104 packet->sent_cb = sent_cb;
105 memcpy(packet->data, buf, size);
107 QTAILQ_INSERT_TAIL(&queue->packets, packet, entry);
109 return size;
112 static ssize_t qemu_net_queue_append_iov(NetQueue *queue,
113 VLANClientState *sender,
114 unsigned flags,
115 const struct iovec *iov,
116 int iovcnt,
117 NetPacketSent *sent_cb)
119 NetPacket *packet;
120 size_t max_len = 0;
121 int i;
123 for (i = 0; i < iovcnt; i++) {
124 max_len += iov[i].iov_len;
127 packet = g_malloc(sizeof(NetPacket) + max_len);
128 packet->sender = sender;
129 packet->sent_cb = sent_cb;
130 packet->flags = flags;
131 packet->size = 0;
133 for (i = 0; i < iovcnt; i++) {
134 size_t len = iov[i].iov_len;
136 memcpy(packet->data + packet->size, iov[i].iov_base, len);
137 packet->size += len;
140 QTAILQ_INSERT_TAIL(&queue->packets, packet, entry);
142 return packet->size;
145 static ssize_t qemu_net_queue_deliver(NetQueue *queue,
146 VLANClientState *sender,
147 unsigned flags,
148 const uint8_t *data,
149 size_t size)
151 ssize_t ret = -1;
153 queue->delivering = 1;
154 ret = queue->deliver(sender, flags, data, size, queue->opaque);
155 queue->delivering = 0;
157 return ret;
160 static ssize_t qemu_net_queue_deliver_iov(NetQueue *queue,
161 VLANClientState *sender,
162 unsigned flags,
163 const struct iovec *iov,
164 int iovcnt)
166 ssize_t ret = -1;
168 queue->delivering = 1;
169 ret = queue->deliver_iov(sender, flags, iov, iovcnt, queue->opaque);
170 queue->delivering = 0;
172 return ret;
175 ssize_t qemu_net_queue_send(NetQueue *queue,
176 VLANClientState *sender,
177 unsigned flags,
178 const uint8_t *data,
179 size_t size,
180 NetPacketSent *sent_cb)
182 ssize_t ret;
184 if (queue->delivering) {
185 return qemu_net_queue_append(queue, sender, flags, data, size, NULL);
188 ret = qemu_net_queue_deliver(queue, sender, flags, data, size);
189 if (ret == 0) {
190 qemu_net_queue_append(queue, sender, flags, data, size, sent_cb);
191 return 0;
194 qemu_net_queue_flush(queue);
196 return ret;
199 ssize_t qemu_net_queue_send_iov(NetQueue *queue,
200 VLANClientState *sender,
201 unsigned flags,
202 const struct iovec *iov,
203 int iovcnt,
204 NetPacketSent *sent_cb)
206 ssize_t ret;
208 if (queue->delivering) {
209 return qemu_net_queue_append_iov(queue, sender, flags, iov, iovcnt, NULL);
212 ret = qemu_net_queue_deliver_iov(queue, sender, flags, iov, iovcnt);
213 if (ret == 0) {
214 qemu_net_queue_append_iov(queue, sender, flags, iov, iovcnt, sent_cb);
215 return 0;
218 qemu_net_queue_flush(queue);
220 return ret;
223 void qemu_net_queue_purge(NetQueue *queue, VLANClientState *from)
225 NetPacket *packet, *next;
227 QTAILQ_FOREACH_SAFE(packet, &queue->packets, entry, next) {
228 if (packet->sender == from) {
229 QTAILQ_REMOVE(&queue->packets, packet, entry);
230 g_free(packet);
235 void qemu_net_queue_flush(NetQueue *queue)
237 while (!QTAILQ_EMPTY(&queue->packets)) {
238 NetPacket *packet;
239 int ret;
241 packet = QTAILQ_FIRST(&queue->packets);
242 QTAILQ_REMOVE(&queue->packets, packet, entry);
244 ret = qemu_net_queue_deliver(queue,
245 packet->sender,
246 packet->flags,
247 packet->data,
248 packet->size);
249 if (ret == 0) {
250 QTAILQ_INSERT_HEAD(&queue->packets, packet, entry);
251 break;
254 if (packet->sent_cb) {
255 packet->sent_cb(packet->sender, ret);
258 g_free(packet);