Merge tag 'pull-loongarch-20241016' of https://gitlab.com/gaosong/qemu into staging
[qemu/armbru.git] / hw / net / spapr_llan.c
blob8af33d91b65e687db3238a0ae82ea408b97bd809
1 /*
2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
4 * PAPR Inter-VM Logical Lan, aka ibmveth
6 * Copyright (c) 2010,2011 David Gibson, IBM Corporation.
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
24 * THE SOFTWARE.
28 #include "qemu/osdep.h"
29 #include "qemu/log.h"
30 #include "qemu/module.h"
31 #include "net/net.h"
32 #include "migration/vmstate.h"
33 #include "hw/ppc/spapr.h"
34 #include "hw/ppc/spapr_vio.h"
35 #include "hw/qdev-properties.h"
36 #include "sysemu/sysemu.h"
37 #include "trace.h"
39 #include <libfdt.h>
40 #include "qom/object.h"
42 #define ETH_ALEN 6
43 #define MAX_PACKET_SIZE 65536
45 /* Compatibility flags for migration */
46 #define SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT 0
47 #define SPAPRVLAN_FLAG_RX_BUF_POOLS (1 << SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT)
50 * Virtual LAN device
53 typedef uint64_t vlan_bd_t;
55 #define VLAN_BD_VALID 0x8000000000000000ULL
56 #define VLAN_BD_TOGGLE 0x4000000000000000ULL
57 #define VLAN_BD_NO_CSUM 0x0200000000000000ULL
58 #define VLAN_BD_CSUM_GOOD 0x0100000000000000ULL
59 #define VLAN_BD_LEN_MASK 0x00ffffff00000000ULL
60 #define VLAN_BD_LEN(bd) (((bd) & VLAN_BD_LEN_MASK) >> 32)
61 #define VLAN_BD_ADDR_MASK 0x00000000ffffffffULL
62 #define VLAN_BD_ADDR(bd) ((bd) & VLAN_BD_ADDR_MASK)
64 #define VLAN_VALID_BD(addr, len) (VLAN_BD_VALID | \
65 (((len) << 32) & VLAN_BD_LEN_MASK) | \
66 (addr & VLAN_BD_ADDR_MASK))
68 #define VLAN_RXQC_TOGGLE 0x80
69 #define VLAN_RXQC_VALID 0x40
70 #define VLAN_RXQC_NO_CSUM 0x02
71 #define VLAN_RXQC_CSUM_GOOD 0x01
73 #define VLAN_RQ_ALIGNMENT 16
74 #define VLAN_RXQ_BD_OFF 0
75 #define VLAN_FILTER_BD_OFF 8
76 #define VLAN_RX_BDS_OFF 16
78 * The final 8 bytes of the buffer list is a counter of frames dropped
79 * because there was not a buffer in the buffer list capable of holding
80 * the frame. We must avoid it, or the operating system will report garbage
81 * for this statistic.
83 #define VLAN_RX_BDS_LEN (SPAPR_TCE_PAGE_SIZE - VLAN_RX_BDS_OFF - 8)
84 #define VLAN_MAX_BUFS (VLAN_RX_BDS_LEN / 8)
86 #define TYPE_VIO_SPAPR_VLAN_DEVICE "spapr-vlan"
87 OBJECT_DECLARE_SIMPLE_TYPE(SpaprVioVlan, VIO_SPAPR_VLAN_DEVICE)
89 #define RX_POOL_MAX_BDS 4096
90 #define RX_MAX_POOLS 5
92 typedef struct {
93 int32_t bufsize;
94 int32_t count;
95 vlan_bd_t bds[RX_POOL_MAX_BDS];
96 } RxBufPool;
98 struct SpaprVioVlan {
99 SpaprVioDevice sdev;
100 NICConf nicconf;
101 NICState *nic;
102 MACAddr perm_mac;
103 bool isopen;
104 hwaddr buf_list;
105 uint32_t add_buf_ptr, use_buf_ptr, rx_bufs;
106 hwaddr rxq_ptr;
107 QEMUTimer *rxp_timer;
108 uint32_t compat_flags; /* Compatibility flags for migration */
109 RxBufPool *rx_pool[RX_MAX_POOLS]; /* Receive buffer descriptor pools */
112 static bool spapr_vlan_can_receive(NetClientState *nc)
114 SpaprVioVlan *dev = qemu_get_nic_opaque(nc);
116 return dev->isopen && dev->rx_bufs > 0;
120 * The last 8 bytes of the receive buffer list page (that has been
121 * supplied by the guest with the H_REGISTER_LOGICAL_LAN call) contain
122 * a counter for frames that have been dropped because there was no
123 * suitable receive buffer available. This function is used to increase
124 * this counter by one.
126 static void spapr_vlan_record_dropped_rx_frame(SpaprVioVlan *dev)
128 uint64_t cnt;
130 cnt = vio_ldq(&dev->sdev, dev->buf_list + 4096 - 8);
131 vio_stq(&dev->sdev, dev->buf_list + 4096 - 8, cnt + 1);
135 * Get buffer descriptor from one of our receive buffer pools
137 static vlan_bd_t spapr_vlan_get_rx_bd_from_pool(SpaprVioVlan *dev,
138 size_t size)
140 vlan_bd_t bd;
141 int pool;
143 for (pool = 0; pool < RX_MAX_POOLS; pool++) {
144 if (dev->rx_pool[pool]->count > 0 &&
145 dev->rx_pool[pool]->bufsize >= size + 8) {
146 break;
149 if (pool == RX_MAX_POOLS) {
150 /* Failed to find a suitable buffer */
151 return 0;
155 trace_spapr_vlan_get_rx_bd_from_pool_found(pool,
156 dev->rx_pool[pool]->count,
157 dev->rx_bufs);
159 /* Remove the buffer from the pool */
160 dev->rx_pool[pool]->count--;
161 bd = dev->rx_pool[pool]->bds[dev->rx_pool[pool]->count];
162 dev->rx_pool[pool]->bds[dev->rx_pool[pool]->count] = 0;
164 return bd;
168 * Get buffer descriptor from the receive buffer list page that has been
169 * supplied by the guest with the H_REGISTER_LOGICAL_LAN call
171 static vlan_bd_t spapr_vlan_get_rx_bd_from_page(SpaprVioVlan *dev,
172 size_t size)
174 int buf_ptr = dev->use_buf_ptr;
175 vlan_bd_t bd;
177 do {
178 buf_ptr += 8;
179 if (buf_ptr >= VLAN_RX_BDS_LEN + VLAN_RX_BDS_OFF) {
180 buf_ptr = VLAN_RX_BDS_OFF;
183 bd = vio_ldq(&dev->sdev, dev->buf_list + buf_ptr);
185 trace_spapr_vlan_get_rx_bd_from_page(buf_ptr, (uint64_t)bd);
186 } while ((!(bd & VLAN_BD_VALID) || VLAN_BD_LEN(bd) < size + 8)
187 && buf_ptr != dev->use_buf_ptr);
189 if (!(bd & VLAN_BD_VALID) || VLAN_BD_LEN(bd) < size + 8) {
190 /* Failed to find a suitable buffer */
191 return 0;
194 /* Remove the buffer from the pool */
195 dev->use_buf_ptr = buf_ptr;
196 vio_stq(&dev->sdev, dev->buf_list + dev->use_buf_ptr, 0);
198 trace_spapr_vlan_get_rx_bd_from_page_found(dev->use_buf_ptr, dev->rx_bufs);
200 return bd;
203 static ssize_t spapr_vlan_receive(NetClientState *nc, const uint8_t *buf,
204 size_t size)
206 SpaprVioVlan *dev = qemu_get_nic_opaque(nc);
207 SpaprVioDevice *sdev = VIO_SPAPR_DEVICE(dev);
208 vlan_bd_t rxq_bd = vio_ldq(sdev, dev->buf_list + VLAN_RXQ_BD_OFF);
209 vlan_bd_t bd;
210 uint64_t handle;
211 uint8_t control;
213 trace_spapr_vlan_receive(sdev->qdev.id, dev->rx_bufs);
215 if (!dev->isopen) {
216 return -1;
219 if (!dev->rx_bufs) {
220 spapr_vlan_record_dropped_rx_frame(dev);
221 return 0;
224 if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
225 bd = spapr_vlan_get_rx_bd_from_pool(dev, size);
226 } else {
227 bd = spapr_vlan_get_rx_bd_from_page(dev, size);
229 if (!bd) {
230 spapr_vlan_record_dropped_rx_frame(dev);
231 return 0;
234 dev->rx_bufs--;
236 /* Transfer the packet data */
237 if (spapr_vio_dma_write(sdev, VLAN_BD_ADDR(bd) + 8, buf, size) < 0) {
238 return -1;
241 trace_spapr_vlan_receive_dma_completed();
243 /* Update the receive queue */
244 control = VLAN_RXQC_TOGGLE | VLAN_RXQC_VALID;
245 if (rxq_bd & VLAN_BD_TOGGLE) {
246 control ^= VLAN_RXQC_TOGGLE;
249 handle = vio_ldq(sdev, VLAN_BD_ADDR(bd));
250 vio_stq(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr + 8, handle);
251 vio_stl(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr + 4, size);
252 vio_sth(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr + 2, 8);
253 vio_stb(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr, control);
255 trace_spapr_vlan_receive_wrote(dev->rxq_ptr,
256 vio_ldq(sdev, VLAN_BD_ADDR(rxq_bd) +
257 dev->rxq_ptr),
258 vio_ldq(sdev, VLAN_BD_ADDR(rxq_bd) +
259 dev->rxq_ptr + 8));
261 dev->rxq_ptr += 16;
262 if (dev->rxq_ptr >= VLAN_BD_LEN(rxq_bd)) {
263 dev->rxq_ptr = 0;
264 vio_stq(sdev, dev->buf_list + VLAN_RXQ_BD_OFF, rxq_bd ^ VLAN_BD_TOGGLE);
267 if (sdev->signal_state & 1) {
268 spapr_vio_irq_pulse(sdev);
271 return size;
274 static NetClientInfo net_spapr_vlan_info = {
275 .type = NET_CLIENT_DRIVER_NIC,
276 .size = sizeof(NICState),
277 .can_receive = spapr_vlan_can_receive,
278 .receive = spapr_vlan_receive,
281 static void spapr_vlan_flush_rx_queue(void *opaque)
283 SpaprVioVlan *dev = opaque;
285 qemu_flush_queued_packets(qemu_get_queue(dev->nic));
288 static void spapr_vlan_reset_rx_pool(RxBufPool *rxp)
291 * Use INT_MAX as bufsize so that unused buffers are moved to the end
292 * of the list during the qsort in spapr_vlan_add_rxbuf_to_pool() later.
294 rxp->bufsize = INT_MAX;
295 rxp->count = 0;
296 memset(rxp->bds, 0, sizeof(rxp->bds));
299 static void spapr_vlan_reset(SpaprVioDevice *sdev)
301 SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
302 int i;
304 dev->buf_list = 0;
305 dev->rx_bufs = 0;
306 dev->isopen = 0;
308 if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
309 for (i = 0; i < RX_MAX_POOLS; i++) {
310 spapr_vlan_reset_rx_pool(dev->rx_pool[i]);
314 memcpy(&dev->nicconf.macaddr.a, &dev->perm_mac.a,
315 sizeof(dev->nicconf.macaddr.a));
316 qemu_format_nic_info_str(qemu_get_queue(dev->nic), dev->nicconf.macaddr.a);
319 static void spapr_vlan_realize(SpaprVioDevice *sdev, Error **errp)
321 SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
323 qemu_macaddr_default_if_unset(&dev->nicconf.macaddr);
325 memcpy(&dev->perm_mac.a, &dev->nicconf.macaddr.a, sizeof(dev->perm_mac.a));
327 dev->nic = qemu_new_nic(&net_spapr_vlan_info, &dev->nicconf,
328 object_get_typename(OBJECT(sdev)), sdev->qdev.id,
329 &sdev->qdev.mem_reentrancy_guard, dev);
330 qemu_format_nic_info_str(qemu_get_queue(dev->nic), dev->nicconf.macaddr.a);
332 dev->rxp_timer = timer_new_us(QEMU_CLOCK_VIRTUAL, spapr_vlan_flush_rx_queue,
333 dev);
336 static void spapr_vlan_instance_init(Object *obj)
338 SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(obj);
339 int i;
341 device_add_bootindex_property(obj, &dev->nicconf.bootindex,
342 "bootindex", "",
343 DEVICE(dev));
345 if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
346 for (i = 0; i < RX_MAX_POOLS; i++) {
347 dev->rx_pool[i] = g_new(RxBufPool, 1);
348 spapr_vlan_reset_rx_pool(dev->rx_pool[i]);
353 static void spapr_vlan_instance_finalize(Object *obj)
355 SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(obj);
356 int i;
358 if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
359 for (i = 0; i < RX_MAX_POOLS; i++) {
360 g_free(dev->rx_pool[i]);
361 dev->rx_pool[i] = NULL;
365 if (dev->rxp_timer) {
366 timer_free(dev->rxp_timer);
370 void spapr_vlan_create(SpaprVioBus *bus, NICInfo *nd)
372 DeviceState *dev;
374 dev = qdev_new("spapr-vlan");
376 qdev_set_nic_properties(dev, nd);
378 qdev_realize_and_unref(dev, &bus->bus, &error_fatal);
381 static int spapr_vlan_devnode(SpaprVioDevice *dev, void *fdt, int node_off)
383 SpaprVioVlan *vdev = VIO_SPAPR_VLAN_DEVICE(dev);
384 uint8_t padded_mac[8] = {0, 0};
385 int ret;
387 /* Some old phyp versions give the mac address in an 8-byte
388 * property. The kernel driver (before 3.10) has an insane workaround;
389 * rather than doing the obvious thing and checking the property
390 * length, it checks whether the first byte has 0b10 in the low
391 * bits. If a correct 6-byte property has a different first byte
392 * the kernel will get the wrong mac address, overrunning its
393 * buffer in the process (read only, thank goodness).
395 * Here we return a 6-byte address unless that would break a pre-3.10
396 * driver. In that case we return a padded 8-byte address to allow the old
397 * workaround to succeed. */
398 if ((vdev->nicconf.macaddr.a[0] & 0x3) == 0x2) {
399 ret = fdt_setprop(fdt, node_off, "local-mac-address",
400 &vdev->nicconf.macaddr, ETH_ALEN);
401 } else {
402 memcpy(&padded_mac[2], &vdev->nicconf.macaddr, ETH_ALEN);
403 ret = fdt_setprop(fdt, node_off, "local-mac-address",
404 padded_mac, sizeof(padded_mac));
406 if (ret < 0) {
407 return ret;
410 ret = fdt_setprop_cell(fdt, node_off, "ibm,mac-address-filters", 0);
411 if (ret < 0) {
412 return ret;
415 return 0;
418 static int check_bd(SpaprVioVlan *dev, vlan_bd_t bd,
419 target_ulong alignment)
421 if ((VLAN_BD_ADDR(bd) % alignment)
422 || (VLAN_BD_LEN(bd) % alignment)) {
423 return -1;
426 if (!spapr_vio_dma_valid(&dev->sdev, VLAN_BD_ADDR(bd),
427 VLAN_BD_LEN(bd), DMA_DIRECTION_FROM_DEVICE)
428 || !spapr_vio_dma_valid(&dev->sdev, VLAN_BD_ADDR(bd),
429 VLAN_BD_LEN(bd), DMA_DIRECTION_TO_DEVICE)) {
430 return -1;
433 return 0;
436 static target_ulong h_register_logical_lan(PowerPCCPU *cpu,
437 SpaprMachineState *spapr,
438 target_ulong opcode,
439 target_ulong *args)
441 target_ulong reg = args[0];
442 target_ulong buf_list = args[1];
443 target_ulong rec_queue = args[2];
444 target_ulong filter_list = args[3];
445 SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
446 SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
447 vlan_bd_t filter_list_bd;
449 if (!dev) {
450 return H_PARAMETER;
453 if (dev->isopen) {
454 hcall_dprintf("H_REGISTER_LOGICAL_LAN called twice without "
455 "H_FREE_LOGICAL_LAN\n");
456 return H_RESOURCE;
459 if (check_bd(dev, VLAN_VALID_BD(buf_list, SPAPR_TCE_PAGE_SIZE),
460 SPAPR_TCE_PAGE_SIZE) < 0) {
461 hcall_dprintf("Bad buf_list 0x" TARGET_FMT_lx "\n", buf_list);
462 return H_PARAMETER;
465 filter_list_bd = VLAN_VALID_BD(filter_list, SPAPR_TCE_PAGE_SIZE);
466 if (check_bd(dev, filter_list_bd, SPAPR_TCE_PAGE_SIZE) < 0) {
467 hcall_dprintf("Bad filter_list 0x" TARGET_FMT_lx "\n", filter_list);
468 return H_PARAMETER;
471 if (!(rec_queue & VLAN_BD_VALID)
472 || (check_bd(dev, rec_queue, VLAN_RQ_ALIGNMENT) < 0)) {
473 hcall_dprintf("Bad receive queue\n");
474 return H_PARAMETER;
477 dev->buf_list = buf_list;
478 sdev->signal_state = 0;
480 rec_queue &= ~VLAN_BD_TOGGLE;
482 /* Initialize the buffer list */
483 vio_stq(sdev, buf_list, rec_queue);
484 vio_stq(sdev, buf_list + 8, filter_list_bd);
485 spapr_vio_dma_set(sdev, buf_list + VLAN_RX_BDS_OFF, 0,
486 SPAPR_TCE_PAGE_SIZE - VLAN_RX_BDS_OFF);
487 dev->add_buf_ptr = VLAN_RX_BDS_OFF - 8;
488 dev->use_buf_ptr = VLAN_RX_BDS_OFF - 8;
489 dev->rx_bufs = 0;
490 dev->rxq_ptr = 0;
492 /* Initialize the receive queue */
493 spapr_vio_dma_set(sdev, VLAN_BD_ADDR(rec_queue), 0, VLAN_BD_LEN(rec_queue));
495 dev->isopen = 1;
496 qemu_flush_queued_packets(qemu_get_queue(dev->nic));
498 return H_SUCCESS;
502 static target_ulong h_free_logical_lan(PowerPCCPU *cpu,
503 SpaprMachineState *spapr,
504 target_ulong opcode, target_ulong *args)
506 target_ulong reg = args[0];
507 SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
508 SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
510 if (!dev) {
511 return H_PARAMETER;
514 if (!dev->isopen) {
515 hcall_dprintf("H_FREE_LOGICAL_LAN called without "
516 "H_REGISTER_LOGICAL_LAN\n");
517 return H_RESOURCE;
520 spapr_vlan_reset(sdev);
521 return H_SUCCESS;
525 * Used for qsort, this function compares two RxBufPools by size.
527 static int rx_pool_size_compare(const void *p1, const void *p2)
529 const RxBufPool *pool1 = *(RxBufPool **)p1;
530 const RxBufPool *pool2 = *(RxBufPool **)p2;
532 if (pool1->bufsize < pool2->bufsize) {
533 return -1;
535 return pool1->bufsize > pool2->bufsize;
539 * Search for a matching buffer pool with exact matching size,
540 * or return -1 if no matching pool has been found.
542 static int spapr_vlan_get_rx_pool_id(SpaprVioVlan *dev, int size)
544 int pool;
546 for (pool = 0; pool < RX_MAX_POOLS; pool++) {
547 if (dev->rx_pool[pool]->bufsize == size) {
548 return pool;
552 return -1;
556 * Enqueuing receive buffer by adding it to one of our receive buffer pools
558 static target_long spapr_vlan_add_rxbuf_to_pool(SpaprVioVlan *dev,
559 target_ulong buf)
561 int size = VLAN_BD_LEN(buf);
562 int pool;
564 pool = spapr_vlan_get_rx_pool_id(dev, size);
565 if (pool < 0) {
567 * No matching pool found? Try to use a new one. If the guest used all
568 * pools before, but changed the size of one pool in the meantime, we might
569 * need to recycle that pool here (if it's empty already). Thus scan
570 * all buffer pools now, starting with the last (likely empty) one.
572 for (pool = RX_MAX_POOLS - 1; pool >= 0 ; pool--) {
573 if (dev->rx_pool[pool]->count == 0) {
574 dev->rx_pool[pool]->bufsize = size;
576 * Sort pools by size so that spapr_vlan_receive()
577 * can later find the smallest buffer pool easily.
579 qsort(dev->rx_pool, RX_MAX_POOLS, sizeof(dev->rx_pool[0]),
580 rx_pool_size_compare);
581 pool = spapr_vlan_get_rx_pool_id(dev, size);
582 trace_spapr_vlan_add_rxbuf_to_pool_create(pool,
583 VLAN_BD_LEN(buf));
584 break;
588 /* Still no usable pool? Give up */
589 if (pool < 0 || dev->rx_pool[pool]->count >= RX_POOL_MAX_BDS) {
590 return H_RESOURCE;
593 trace_spapr_vlan_add_rxbuf_to_pool(pool, VLAN_BD_LEN(buf),
594 dev->rx_pool[pool]->count);
596 dev->rx_pool[pool]->bds[dev->rx_pool[pool]->count++] = buf;
598 return 0;
602 * This is the old way of enqueuing receive buffers: Add it to the rx queue
603 * page that has been supplied by the guest (which is quite limited in size).
605 static target_long spapr_vlan_add_rxbuf_to_page(SpaprVioVlan *dev,
606 target_ulong buf)
608 vlan_bd_t bd;
610 if (dev->rx_bufs >= VLAN_MAX_BUFS) {
611 return H_RESOURCE;
614 do {
615 dev->add_buf_ptr += 8;
616 if (dev->add_buf_ptr >= VLAN_RX_BDS_LEN + VLAN_RX_BDS_OFF) {
617 dev->add_buf_ptr = VLAN_RX_BDS_OFF;
620 bd = vio_ldq(&dev->sdev, dev->buf_list + dev->add_buf_ptr);
621 } while (bd & VLAN_BD_VALID);
623 vio_stq(&dev->sdev, dev->buf_list + dev->add_buf_ptr, buf);
625 trace_spapr_vlan_add_rxbuf_to_page(dev->add_buf_ptr, dev->rx_bufs, buf);
627 return 0;
630 static target_ulong h_add_logical_lan_buffer(PowerPCCPU *cpu,
631 SpaprMachineState *spapr,
632 target_ulong opcode,
633 target_ulong *args)
635 target_ulong reg = args[0];
636 target_ulong buf = args[1];
637 SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
638 SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
639 target_long ret;
641 trace_spapr_vlan_h_add_logical_lan_buffer(reg, buf);
643 if (!sdev) {
644 hcall_dprintf("Bad device\n");
645 return H_PARAMETER;
648 if ((check_bd(dev, buf, 4) < 0)
649 || (VLAN_BD_LEN(buf) < 16)) {
650 hcall_dprintf("Bad buffer enqueued\n");
651 return H_PARAMETER;
654 if (!dev->isopen) {
655 return H_RESOURCE;
658 if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
659 ret = spapr_vlan_add_rxbuf_to_pool(dev, buf);
660 } else {
661 ret = spapr_vlan_add_rxbuf_to_page(dev, buf);
663 if (ret) {
664 return ret;
667 dev->rx_bufs++;
670 * Give guest some more time to add additional RX buffers before we
671 * flush the receive queue, so that e.g. fragmented IP packets can
672 * be passed to the guest in one go later (instead of passing single
673 * fragments if there is only one receive buffer available).
675 timer_mod(dev->rxp_timer, qemu_clock_get_us(QEMU_CLOCK_VIRTUAL) + 500);
677 return H_SUCCESS;
680 static target_ulong h_send_logical_lan(PowerPCCPU *cpu,
681 SpaprMachineState *spapr,
682 target_ulong opcode, target_ulong *args)
684 target_ulong reg = args[0];
685 target_ulong *bufs = args + 1;
686 target_ulong continue_token = args[7];
687 SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
688 SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
689 unsigned total_len;
690 uint8_t *p;
691 g_autofree uint8_t *lbuf = NULL;
692 int i, nbufs;
693 int ret;
695 trace_spapr_vlan_h_send_logical_lan(reg, continue_token);
697 if (!sdev) {
698 return H_PARAMETER;
701 trace_spapr_vlan_h_send_logical_lan_rxbufs(dev->rx_bufs);
703 if (!dev->isopen) {
704 return H_DROPPED;
707 if (continue_token) {
708 return H_HARDWARE; /* FIXME actually handle this */
711 total_len = 0;
712 for (i = 0; i < 6; i++) {
713 trace_spapr_vlan_h_send_logical_lan_buf_desc(bufs[i]);
714 if (!(bufs[i] & VLAN_BD_VALID)) {
715 break;
717 total_len += VLAN_BD_LEN(bufs[i]);
720 nbufs = i;
721 trace_spapr_vlan_h_send_logical_lan_total(nbufs, total_len);
723 if (total_len == 0) {
724 return H_SUCCESS;
727 if (total_len > MAX_PACKET_SIZE) {
728 /* Don't let the guest force too large an allocation */
729 return H_RESOURCE;
732 lbuf = g_malloc(total_len);
733 p = lbuf;
734 for (i = 0; i < nbufs; i++) {
735 ret = spapr_vio_dma_read(sdev, VLAN_BD_ADDR(bufs[i]),
736 p, VLAN_BD_LEN(bufs[i]));
737 if (ret < 0) {
738 return ret;
741 p += VLAN_BD_LEN(bufs[i]);
744 qemu_send_packet(qemu_get_queue(dev->nic), lbuf, total_len);
746 return H_SUCCESS;
749 static target_ulong h_multicast_ctrl(PowerPCCPU *cpu, SpaprMachineState *spapr,
750 target_ulong opcode, target_ulong *args)
752 target_ulong reg = args[0];
753 SpaprVioDevice *dev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
755 if (!dev) {
756 return H_PARAMETER;
759 return H_SUCCESS;
762 static target_ulong h_change_logical_lan_mac(PowerPCCPU *cpu,
763 SpaprMachineState *spapr,
764 target_ulong opcode,
765 target_ulong *args)
767 target_ulong reg = args[0];
768 target_ulong macaddr = args[1];
769 SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
770 SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
771 int i;
773 if (!dev) {
774 hcall_dprintf("H_CHANGE_LOGICAL_LAN_MAC called when "
775 "no NIC is present\n");
776 return H_PARAMETER;
779 for (i = 0; i < ETH_ALEN; i++) {
780 dev->nicconf.macaddr.a[ETH_ALEN - i - 1] = macaddr & 0xff;
781 macaddr >>= 8;
784 qemu_format_nic_info_str(qemu_get_queue(dev->nic), dev->nicconf.macaddr.a);
786 return H_SUCCESS;
789 static Property spapr_vlan_properties[] = {
790 DEFINE_SPAPR_PROPERTIES(SpaprVioVlan, sdev),
791 DEFINE_NIC_PROPERTIES(SpaprVioVlan, nicconf),
792 DEFINE_PROP_BIT("use-rx-buffer-pools", SpaprVioVlan,
793 compat_flags, SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT, true),
794 DEFINE_PROP_END_OF_LIST(),
797 static bool spapr_vlan_rx_buffer_pools_needed(void *opaque)
799 SpaprVioVlan *dev = opaque;
801 return (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) != 0;
804 static const VMStateDescription vmstate_rx_buffer_pool = {
805 .name = "spapr_llan/rx_buffer_pool",
806 .version_id = 1,
807 .minimum_version_id = 1,
808 .needed = spapr_vlan_rx_buffer_pools_needed,
809 .fields = (const VMStateField[]) {
810 VMSTATE_INT32(bufsize, RxBufPool),
811 VMSTATE_INT32(count, RxBufPool),
812 VMSTATE_UINT64_ARRAY(bds, RxBufPool, RX_POOL_MAX_BDS),
813 VMSTATE_END_OF_LIST()
817 static const VMStateDescription vmstate_rx_pools = {
818 .name = "spapr_llan/rx_pools",
819 .version_id = 1,
820 .minimum_version_id = 1,
821 .needed = spapr_vlan_rx_buffer_pools_needed,
822 .fields = (const VMStateField[]) {
823 VMSTATE_ARRAY_OF_POINTER_TO_STRUCT(rx_pool, SpaprVioVlan,
824 RX_MAX_POOLS, 1,
825 vmstate_rx_buffer_pool, RxBufPool),
826 VMSTATE_END_OF_LIST()
830 static const VMStateDescription vmstate_spapr_llan = {
831 .name = "spapr_llan",
832 .version_id = 1,
833 .minimum_version_id = 1,
834 .fields = (const VMStateField[]) {
835 VMSTATE_SPAPR_VIO(sdev, SpaprVioVlan),
836 /* LLAN state */
837 VMSTATE_BOOL(isopen, SpaprVioVlan),
838 VMSTATE_UINT64(buf_list, SpaprVioVlan),
839 VMSTATE_UINT32(add_buf_ptr, SpaprVioVlan),
840 VMSTATE_UINT32(use_buf_ptr, SpaprVioVlan),
841 VMSTATE_UINT32(rx_bufs, SpaprVioVlan),
842 VMSTATE_UINT64(rxq_ptr, SpaprVioVlan),
844 VMSTATE_END_OF_LIST()
846 .subsections = (const VMStateDescription * const []) {
847 &vmstate_rx_pools,
848 NULL
852 static void spapr_vlan_class_init(ObjectClass *klass, void *data)
854 DeviceClass *dc = DEVICE_CLASS(klass);
855 SpaprVioDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass);
857 k->realize = spapr_vlan_realize;
858 k->reset = spapr_vlan_reset;
859 k->devnode = spapr_vlan_devnode;
860 k->dt_name = "l-lan";
861 k->dt_type = "network";
862 k->dt_compatible = "IBM,l-lan";
863 k->signal_mask = 0x1;
864 set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
865 device_class_set_props(dc, spapr_vlan_properties);
866 k->rtce_window_size = 0x10000000;
867 dc->vmsd = &vmstate_spapr_llan;
870 static const TypeInfo spapr_vlan_info = {
871 .name = TYPE_VIO_SPAPR_VLAN_DEVICE,
872 .parent = TYPE_VIO_SPAPR_DEVICE,
873 .instance_size = sizeof(SpaprVioVlan),
874 .class_init = spapr_vlan_class_init,
875 .instance_init = spapr_vlan_instance_init,
876 .instance_finalize = spapr_vlan_instance_finalize,
879 static void spapr_vlan_register_types(void)
881 spapr_register_hypercall(H_REGISTER_LOGICAL_LAN, h_register_logical_lan);
882 spapr_register_hypercall(H_FREE_LOGICAL_LAN, h_free_logical_lan);
883 spapr_register_hypercall(H_SEND_LOGICAL_LAN, h_send_logical_lan);
884 spapr_register_hypercall(H_ADD_LOGICAL_LAN_BUFFER,
885 h_add_logical_lan_buffer);
886 spapr_register_hypercall(H_MULTICAST_CTRL, h_multicast_ctrl);
887 spapr_register_hypercall(H_CHANGE_LOGICAL_LAN_MAC,
888 h_change_logical_lan_mac);
889 type_register_static(&spapr_vlan_info);
892 type_init(spapr_vlan_register_types)