2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
4 * PAPR Inter-VM Logical Lan, aka ibmveth
6 * Copyright (c) 2010,2011 David Gibson, IBM Corporation.
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
28 #include "qemu/osdep.h"
32 #include "qemu/module.h"
34 #include "migration/vmstate.h"
35 #include "hw/ppc/spapr.h"
36 #include "hw/ppc/spapr_vio.h"
37 #include "hw/qdev-properties.h"
38 #include "sysemu/sysemu.h"
44 #define MAX_PACKET_SIZE 65536
46 /* Compatibility flags for migration */
47 #define SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT 0
48 #define SPAPRVLAN_FLAG_RX_BUF_POOLS (1 << SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT)
54 typedef uint64_t vlan_bd_t
;
56 #define VLAN_BD_VALID 0x8000000000000000ULL
57 #define VLAN_BD_TOGGLE 0x4000000000000000ULL
58 #define VLAN_BD_NO_CSUM 0x0200000000000000ULL
59 #define VLAN_BD_CSUM_GOOD 0x0100000000000000ULL
60 #define VLAN_BD_LEN_MASK 0x00ffffff00000000ULL
61 #define VLAN_BD_LEN(bd) (((bd) & VLAN_BD_LEN_MASK) >> 32)
62 #define VLAN_BD_ADDR_MASK 0x00000000ffffffffULL
63 #define VLAN_BD_ADDR(bd) ((bd) & VLAN_BD_ADDR_MASK)
65 #define VLAN_VALID_BD(addr, len) (VLAN_BD_VALID | \
66 (((len) << 32) & VLAN_BD_LEN_MASK) | \
67 (addr & VLAN_BD_ADDR_MASK))
69 #define VLAN_RXQC_TOGGLE 0x80
70 #define VLAN_RXQC_VALID 0x40
71 #define VLAN_RXQC_NO_CSUM 0x02
72 #define VLAN_RXQC_CSUM_GOOD 0x01
74 #define VLAN_RQ_ALIGNMENT 16
75 #define VLAN_RXQ_BD_OFF 0
76 #define VLAN_FILTER_BD_OFF 8
77 #define VLAN_RX_BDS_OFF 16
79 * The final 8 bytes of the buffer list is a counter of frames dropped
80 * because there was not a buffer in the buffer list capable of holding
81 * the frame. We must avoid it, or the operating system will report garbage
84 #define VLAN_RX_BDS_LEN (SPAPR_TCE_PAGE_SIZE - VLAN_RX_BDS_OFF - 8)
85 #define VLAN_MAX_BUFS (VLAN_RX_BDS_LEN / 8)
87 #define TYPE_VIO_SPAPR_VLAN_DEVICE "spapr-vlan"
88 #define VIO_SPAPR_VLAN_DEVICE(obj) \
89 OBJECT_CHECK(SpaprVioVlan, (obj), TYPE_VIO_SPAPR_VLAN_DEVICE)
91 #define RX_POOL_MAX_BDS 4096
92 #define RX_MAX_POOLS 5
97 vlan_bd_t bds
[RX_POOL_MAX_BDS
];
100 typedef struct SpaprVioVlan
{
107 uint32_t add_buf_ptr
, use_buf_ptr
, rx_bufs
;
109 QEMUTimer
*rxp_timer
;
110 uint32_t compat_flags
; /* Compatibility flags for migration */
111 RxBufPool
*rx_pool
[RX_MAX_POOLS
]; /* Receive buffer descriptor pools */
114 static int spapr_vlan_can_receive(NetClientState
*nc
)
116 SpaprVioVlan
*dev
= qemu_get_nic_opaque(nc
);
118 return (dev
->isopen
&& dev
->rx_bufs
> 0);
122 * The last 8 bytes of the receive buffer list page (that has been
123 * supplied by the guest with the H_REGISTER_LOGICAL_LAN call) contain
124 * a counter for frames that have been dropped because there was no
125 * suitable receive buffer available. This function is used to increase
126 * this counter by one.
128 static void spapr_vlan_record_dropped_rx_frame(SpaprVioVlan
*dev
)
132 cnt
= vio_ldq(&dev
->sdev
, dev
->buf_list
+ 4096 - 8);
133 vio_stq(&dev
->sdev
, dev
->buf_list
+ 4096 - 8, cnt
+ 1);
137 * Get buffer descriptor from one of our receive buffer pools
139 static vlan_bd_t
spapr_vlan_get_rx_bd_from_pool(SpaprVioVlan
*dev
,
145 for (pool
= 0; pool
< RX_MAX_POOLS
; pool
++) {
146 if (dev
->rx_pool
[pool
]->count
> 0 &&
147 dev
->rx_pool
[pool
]->bufsize
>= size
+ 8) {
151 if (pool
== RX_MAX_POOLS
) {
152 /* Failed to find a suitable buffer */
157 trace_spapr_vlan_get_rx_bd_from_pool_found(pool
,
158 dev
->rx_pool
[pool
]->count
,
161 /* Remove the buffer from the pool */
162 dev
->rx_pool
[pool
]->count
--;
163 bd
= dev
->rx_pool
[pool
]->bds
[dev
->rx_pool
[pool
]->count
];
164 dev
->rx_pool
[pool
]->bds
[dev
->rx_pool
[pool
]->count
] = 0;
170 * Get buffer descriptor from the receive buffer list page that has been
171 * supplied by the guest with the H_REGISTER_LOGICAL_LAN call
173 static vlan_bd_t
spapr_vlan_get_rx_bd_from_page(SpaprVioVlan
*dev
,
176 int buf_ptr
= dev
->use_buf_ptr
;
181 if (buf_ptr
>= VLAN_RX_BDS_LEN
+ VLAN_RX_BDS_OFF
) {
182 buf_ptr
= VLAN_RX_BDS_OFF
;
185 bd
= vio_ldq(&dev
->sdev
, dev
->buf_list
+ buf_ptr
);
187 trace_spapr_vlan_get_rx_bd_from_page(buf_ptr
, (uint64_t)bd
);
188 } while ((!(bd
& VLAN_BD_VALID
) || VLAN_BD_LEN(bd
) < size
+ 8)
189 && buf_ptr
!= dev
->use_buf_ptr
);
191 if (!(bd
& VLAN_BD_VALID
) || VLAN_BD_LEN(bd
) < size
+ 8) {
192 /* Failed to find a suitable buffer */
196 /* Remove the buffer from the pool */
197 dev
->use_buf_ptr
= buf_ptr
;
198 vio_stq(&dev
->sdev
, dev
->buf_list
+ dev
->use_buf_ptr
, 0);
200 trace_spapr_vlan_get_rx_bd_from_page_found(dev
->use_buf_ptr
, dev
->rx_bufs
);
205 static ssize_t
spapr_vlan_receive(NetClientState
*nc
, const uint8_t *buf
,
208 SpaprVioVlan
*dev
= qemu_get_nic_opaque(nc
);
209 SpaprVioDevice
*sdev
= VIO_SPAPR_DEVICE(dev
);
210 vlan_bd_t rxq_bd
= vio_ldq(sdev
, dev
->buf_list
+ VLAN_RXQ_BD_OFF
);
215 trace_spapr_vlan_receive(sdev
->qdev
.id
, dev
->rx_bufs
);
222 spapr_vlan_record_dropped_rx_frame(dev
);
226 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
227 bd
= spapr_vlan_get_rx_bd_from_pool(dev
, size
);
229 bd
= spapr_vlan_get_rx_bd_from_page(dev
, size
);
232 spapr_vlan_record_dropped_rx_frame(dev
);
238 /* Transfer the packet data */
239 if (spapr_vio_dma_write(sdev
, VLAN_BD_ADDR(bd
) + 8, buf
, size
) < 0) {
243 trace_spapr_vlan_receive_dma_completed();
245 /* Update the receive queue */
246 control
= VLAN_RXQC_TOGGLE
| VLAN_RXQC_VALID
;
247 if (rxq_bd
& VLAN_BD_TOGGLE
) {
248 control
^= VLAN_RXQC_TOGGLE
;
251 handle
= vio_ldq(sdev
, VLAN_BD_ADDR(bd
));
252 vio_stq(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
+ 8, handle
);
253 vio_stl(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
+ 4, size
);
254 vio_sth(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
+ 2, 8);
255 vio_stb(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
, control
);
257 trace_spapr_vlan_receive_wrote(dev
->rxq_ptr
,
258 vio_ldq(sdev
, VLAN_BD_ADDR(rxq_bd
) +
260 vio_ldq(sdev
, VLAN_BD_ADDR(rxq_bd
) +
264 if (dev
->rxq_ptr
>= VLAN_BD_LEN(rxq_bd
)) {
266 vio_stq(sdev
, dev
->buf_list
+ VLAN_RXQ_BD_OFF
, rxq_bd
^ VLAN_BD_TOGGLE
);
269 if (sdev
->signal_state
& 1) {
270 qemu_irq_pulse(spapr_vio_qirq(sdev
));
276 static NetClientInfo net_spapr_vlan_info
= {
277 .type
= NET_CLIENT_DRIVER_NIC
,
278 .size
= sizeof(NICState
),
279 .can_receive
= spapr_vlan_can_receive
,
280 .receive
= spapr_vlan_receive
,
283 static void spapr_vlan_flush_rx_queue(void *opaque
)
285 SpaprVioVlan
*dev
= opaque
;
287 qemu_flush_queued_packets(qemu_get_queue(dev
->nic
));
290 static void spapr_vlan_reset_rx_pool(RxBufPool
*rxp
)
293 * Use INT_MAX as bufsize so that unused buffers are moved to the end
294 * of the list during the qsort in spapr_vlan_add_rxbuf_to_pool() later.
296 rxp
->bufsize
= INT_MAX
;
298 memset(rxp
->bds
, 0, sizeof(rxp
->bds
));
301 static void spapr_vlan_reset(SpaprVioDevice
*sdev
)
303 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
310 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
311 for (i
= 0; i
< RX_MAX_POOLS
; i
++) {
312 spapr_vlan_reset_rx_pool(dev
->rx_pool
[i
]);
316 memcpy(&dev
->nicconf
.macaddr
.a
, &dev
->perm_mac
.a
,
317 sizeof(dev
->nicconf
.macaddr
.a
));
318 qemu_format_nic_info_str(qemu_get_queue(dev
->nic
), dev
->nicconf
.macaddr
.a
);
321 static void spapr_vlan_realize(SpaprVioDevice
*sdev
, Error
**errp
)
323 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
325 qemu_macaddr_default_if_unset(&dev
->nicconf
.macaddr
);
327 memcpy(&dev
->perm_mac
.a
, &dev
->nicconf
.macaddr
.a
, sizeof(dev
->perm_mac
.a
));
329 dev
->nic
= qemu_new_nic(&net_spapr_vlan_info
, &dev
->nicconf
,
330 object_get_typename(OBJECT(sdev
)), sdev
->qdev
.id
, dev
);
331 qemu_format_nic_info_str(qemu_get_queue(dev
->nic
), dev
->nicconf
.macaddr
.a
);
333 dev
->rxp_timer
= timer_new_us(QEMU_CLOCK_VIRTUAL
, spapr_vlan_flush_rx_queue
,
337 static void spapr_vlan_instance_init(Object
*obj
)
339 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(obj
);
342 device_add_bootindex_property(obj
, &dev
->nicconf
.bootindex
,
346 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
347 for (i
= 0; i
< RX_MAX_POOLS
; i
++) {
348 dev
->rx_pool
[i
] = g_new(RxBufPool
, 1);
349 spapr_vlan_reset_rx_pool(dev
->rx_pool
[i
]);
354 static void spapr_vlan_instance_finalize(Object
*obj
)
356 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(obj
);
359 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
360 for (i
= 0; i
< RX_MAX_POOLS
; i
++) {
361 g_free(dev
->rx_pool
[i
]);
362 dev
->rx_pool
[i
] = NULL
;
366 if (dev
->rxp_timer
) {
367 timer_del(dev
->rxp_timer
);
368 timer_free(dev
->rxp_timer
);
372 void spapr_vlan_create(SpaprVioBus
*bus
, NICInfo
*nd
)
376 dev
= qdev_create(&bus
->bus
, "spapr-vlan");
378 qdev_set_nic_properties(dev
, nd
);
380 qdev_init_nofail(dev
);
383 static int spapr_vlan_devnode(SpaprVioDevice
*dev
, void *fdt
, int node_off
)
385 SpaprVioVlan
*vdev
= VIO_SPAPR_VLAN_DEVICE(dev
);
386 uint8_t padded_mac
[8] = {0, 0};
389 /* Some old phyp versions give the mac address in an 8-byte
390 * property. The kernel driver (before 3.10) has an insane workaround;
391 * rather than doing the obvious thing and checking the property
392 * length, it checks whether the first byte has 0b10 in the low
393 * bits. If a correct 6-byte property has a different first byte
394 * the kernel will get the wrong mac address, overrunning its
395 * buffer in the process (read only, thank goodness).
397 * Here we return a 6-byte address unless that would break a pre-3.10
398 * driver. In that case we return a padded 8-byte address to allow the old
399 * workaround to succeed. */
400 if ((vdev
->nicconf
.macaddr
.a
[0] & 0x3) == 0x2) {
401 ret
= fdt_setprop(fdt
, node_off
, "local-mac-address",
402 &vdev
->nicconf
.macaddr
, ETH_ALEN
);
404 memcpy(&padded_mac
[2], &vdev
->nicconf
.macaddr
, ETH_ALEN
);
405 ret
= fdt_setprop(fdt
, node_off
, "local-mac-address",
406 padded_mac
, sizeof(padded_mac
));
412 ret
= fdt_setprop_cell(fdt
, node_off
, "ibm,mac-address-filters", 0);
420 static int check_bd(SpaprVioVlan
*dev
, vlan_bd_t bd
,
421 target_ulong alignment
)
423 if ((VLAN_BD_ADDR(bd
) % alignment
)
424 || (VLAN_BD_LEN(bd
) % alignment
)) {
428 if (!spapr_vio_dma_valid(&dev
->sdev
, VLAN_BD_ADDR(bd
),
429 VLAN_BD_LEN(bd
), DMA_DIRECTION_FROM_DEVICE
)
430 || !spapr_vio_dma_valid(&dev
->sdev
, VLAN_BD_ADDR(bd
),
431 VLAN_BD_LEN(bd
), DMA_DIRECTION_TO_DEVICE
)) {
438 static target_ulong
h_register_logical_lan(PowerPCCPU
*cpu
,
439 SpaprMachineState
*spapr
,
443 target_ulong reg
= args
[0];
444 target_ulong buf_list
= args
[1];
445 target_ulong rec_queue
= args
[2];
446 target_ulong filter_list
= args
[3];
447 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
448 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
449 vlan_bd_t filter_list_bd
;
456 hcall_dprintf("H_REGISTER_LOGICAL_LAN called twice without "
457 "H_FREE_LOGICAL_LAN\n");
461 if (check_bd(dev
, VLAN_VALID_BD(buf_list
, SPAPR_TCE_PAGE_SIZE
),
462 SPAPR_TCE_PAGE_SIZE
) < 0) {
463 hcall_dprintf("Bad buf_list 0x" TARGET_FMT_lx
"\n", buf_list
);
467 filter_list_bd
= VLAN_VALID_BD(filter_list
, SPAPR_TCE_PAGE_SIZE
);
468 if (check_bd(dev
, filter_list_bd
, SPAPR_TCE_PAGE_SIZE
) < 0) {
469 hcall_dprintf("Bad filter_list 0x" TARGET_FMT_lx
"\n", filter_list
);
473 if (!(rec_queue
& VLAN_BD_VALID
)
474 || (check_bd(dev
, rec_queue
, VLAN_RQ_ALIGNMENT
) < 0)) {
475 hcall_dprintf("Bad receive queue\n");
479 dev
->buf_list
= buf_list
;
480 sdev
->signal_state
= 0;
482 rec_queue
&= ~VLAN_BD_TOGGLE
;
484 /* Initialize the buffer list */
485 vio_stq(sdev
, buf_list
, rec_queue
);
486 vio_stq(sdev
, buf_list
+ 8, filter_list_bd
);
487 spapr_vio_dma_set(sdev
, buf_list
+ VLAN_RX_BDS_OFF
, 0,
488 SPAPR_TCE_PAGE_SIZE
- VLAN_RX_BDS_OFF
);
489 dev
->add_buf_ptr
= VLAN_RX_BDS_OFF
- 8;
490 dev
->use_buf_ptr
= VLAN_RX_BDS_OFF
- 8;
494 /* Initialize the receive queue */
495 spapr_vio_dma_set(sdev
, VLAN_BD_ADDR(rec_queue
), 0, VLAN_BD_LEN(rec_queue
));
498 qemu_flush_queued_packets(qemu_get_queue(dev
->nic
));
504 static target_ulong
h_free_logical_lan(PowerPCCPU
*cpu
,
505 SpaprMachineState
*spapr
,
506 target_ulong opcode
, target_ulong
*args
)
508 target_ulong reg
= args
[0];
509 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
510 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
517 hcall_dprintf("H_FREE_LOGICAL_LAN called without "
518 "H_REGISTER_LOGICAL_LAN\n");
522 spapr_vlan_reset(sdev
);
527 * Used for qsort, this function compares two RxBufPools by size.
529 static int rx_pool_size_compare(const void *p1
, const void *p2
)
531 const RxBufPool
*pool1
= *(RxBufPool
**)p1
;
532 const RxBufPool
*pool2
= *(RxBufPool
**)p2
;
534 if (pool1
->bufsize
< pool2
->bufsize
) {
537 return pool1
->bufsize
> pool2
->bufsize
;
541 * Search for a matching buffer pool with exact matching size,
542 * or return -1 if no matching pool has been found.
544 static int spapr_vlan_get_rx_pool_id(SpaprVioVlan
*dev
, int size
)
548 for (pool
= 0; pool
< RX_MAX_POOLS
; pool
++) {
549 if (dev
->rx_pool
[pool
]->bufsize
== size
) {
558 * Enqueuing receive buffer by adding it to one of our receive buffer pools
560 static target_long
spapr_vlan_add_rxbuf_to_pool(SpaprVioVlan
*dev
,
563 int size
= VLAN_BD_LEN(buf
);
566 pool
= spapr_vlan_get_rx_pool_id(dev
, size
);
569 * No matching pool found? Try to use a new one. If the guest used all
570 * pools before, but changed the size of one pool in the meantime, we might
571 * need to recycle that pool here (if it's empty already). Thus scan
572 * all buffer pools now, starting with the last (likely empty) one.
574 for (pool
= RX_MAX_POOLS
- 1; pool
>= 0 ; pool
--) {
575 if (dev
->rx_pool
[pool
]->count
== 0) {
576 dev
->rx_pool
[pool
]->bufsize
= size
;
578 * Sort pools by size so that spapr_vlan_receive()
579 * can later find the smallest buffer pool easily.
581 qsort(dev
->rx_pool
, RX_MAX_POOLS
, sizeof(dev
->rx_pool
[0]),
582 rx_pool_size_compare
);
583 pool
= spapr_vlan_get_rx_pool_id(dev
, size
);
584 trace_spapr_vlan_add_rxbuf_to_pool_create(pool
,
590 /* Still no usable pool? Give up */
591 if (pool
< 0 || dev
->rx_pool
[pool
]->count
>= RX_POOL_MAX_BDS
) {
595 trace_spapr_vlan_add_rxbuf_to_pool(pool
, VLAN_BD_LEN(buf
),
596 dev
->rx_pool
[pool
]->count
);
598 dev
->rx_pool
[pool
]->bds
[dev
->rx_pool
[pool
]->count
++] = buf
;
604 * This is the old way of enqueuing receive buffers: Add it to the rx queue
605 * page that has been supplied by the guest (which is quite limited in size).
607 static target_long
spapr_vlan_add_rxbuf_to_page(SpaprVioVlan
*dev
,
612 if (dev
->rx_bufs
>= VLAN_MAX_BUFS
) {
617 dev
->add_buf_ptr
+= 8;
618 if (dev
->add_buf_ptr
>= VLAN_RX_BDS_LEN
+ VLAN_RX_BDS_OFF
) {
619 dev
->add_buf_ptr
= VLAN_RX_BDS_OFF
;
622 bd
= vio_ldq(&dev
->sdev
, dev
->buf_list
+ dev
->add_buf_ptr
);
623 } while (bd
& VLAN_BD_VALID
);
625 vio_stq(&dev
->sdev
, dev
->buf_list
+ dev
->add_buf_ptr
, buf
);
627 trace_spapr_vlan_add_rxbuf_to_page(dev
->add_buf_ptr
, dev
->rx_bufs
, buf
);
632 static target_ulong
h_add_logical_lan_buffer(PowerPCCPU
*cpu
,
633 SpaprMachineState
*spapr
,
637 target_ulong reg
= args
[0];
638 target_ulong buf
= args
[1];
639 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
640 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
643 trace_spapr_vlan_h_add_logical_lan_buffer(reg
, buf
);
646 hcall_dprintf("Bad device\n");
650 if ((check_bd(dev
, buf
, 4) < 0)
651 || (VLAN_BD_LEN(buf
) < 16)) {
652 hcall_dprintf("Bad buffer enqueued\n");
660 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
661 ret
= spapr_vlan_add_rxbuf_to_pool(dev
, buf
);
663 ret
= spapr_vlan_add_rxbuf_to_page(dev
, buf
);
672 * Give guest some more time to add additional RX buffers before we
673 * flush the receive queue, so that e.g. fragmented IP packets can
674 * be passed to the guest in one go later (instead of passing single
675 * fragments if there is only one receive buffer available).
677 timer_mod(dev
->rxp_timer
, qemu_clock_get_us(QEMU_CLOCK_VIRTUAL
) + 500);
682 static target_ulong
h_send_logical_lan(PowerPCCPU
*cpu
,
683 SpaprMachineState
*spapr
,
684 target_ulong opcode
, target_ulong
*args
)
686 target_ulong reg
= args
[0];
687 target_ulong
*bufs
= args
+ 1;
688 target_ulong continue_token
= args
[7];
689 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
690 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
696 trace_spapr_vlan_h_send_logical_lan(reg
, continue_token
);
702 trace_spapr_vlan_h_send_logical_lan_rxbufs(dev
->rx_bufs
);
708 if (continue_token
) {
709 return H_HARDWARE
; /* FIXME actually handle this */
713 for (i
= 0; i
< 6; i
++) {
714 trace_spapr_vlan_h_send_logical_lan_buf_desc(bufs
[i
]);
715 if (!(bufs
[i
] & VLAN_BD_VALID
)) {
718 total_len
+= VLAN_BD_LEN(bufs
[i
]);
722 trace_spapr_vlan_h_send_logical_lan_total(nbufs
, total_len
);
724 if (total_len
== 0) {
728 if (total_len
> MAX_PACKET_SIZE
) {
729 /* Don't let the guest force too large an allocation */
733 lbuf
= alloca(total_len
);
735 for (i
= 0; i
< nbufs
; i
++) {
736 ret
= spapr_vio_dma_read(sdev
, VLAN_BD_ADDR(bufs
[i
]),
737 p
, VLAN_BD_LEN(bufs
[i
]));
742 p
+= VLAN_BD_LEN(bufs
[i
]);
745 qemu_send_packet(qemu_get_queue(dev
->nic
), lbuf
, total_len
);
750 static target_ulong
h_multicast_ctrl(PowerPCCPU
*cpu
, SpaprMachineState
*spapr
,
751 target_ulong opcode
, target_ulong
*args
)
753 target_ulong reg
= args
[0];
754 SpaprVioDevice
*dev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
763 static target_ulong
h_change_logical_lan_mac(PowerPCCPU
*cpu
,
764 SpaprMachineState
*spapr
,
768 target_ulong reg
= args
[0];
769 target_ulong macaddr
= args
[1];
770 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
771 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
774 for (i
= 0; i
< ETH_ALEN
; i
++) {
775 dev
->nicconf
.macaddr
.a
[ETH_ALEN
- i
- 1] = macaddr
& 0xff;
779 qemu_format_nic_info_str(qemu_get_queue(dev
->nic
), dev
->nicconf
.macaddr
.a
);
784 static Property spapr_vlan_properties
[] = {
785 DEFINE_SPAPR_PROPERTIES(SpaprVioVlan
, sdev
),
786 DEFINE_NIC_PROPERTIES(SpaprVioVlan
, nicconf
),
787 DEFINE_PROP_BIT("use-rx-buffer-pools", SpaprVioVlan
,
788 compat_flags
, SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT
, true),
789 DEFINE_PROP_END_OF_LIST(),
792 static bool spapr_vlan_rx_buffer_pools_needed(void *opaque
)
794 SpaprVioVlan
*dev
= opaque
;
796 return (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) != 0;
799 static const VMStateDescription vmstate_rx_buffer_pool
= {
800 .name
= "spapr_llan/rx_buffer_pool",
802 .minimum_version_id
= 1,
803 .needed
= spapr_vlan_rx_buffer_pools_needed
,
804 .fields
= (VMStateField
[]) {
805 VMSTATE_INT32(bufsize
, RxBufPool
),
806 VMSTATE_INT32(count
, RxBufPool
),
807 VMSTATE_UINT64_ARRAY(bds
, RxBufPool
, RX_POOL_MAX_BDS
),
808 VMSTATE_END_OF_LIST()
812 static const VMStateDescription vmstate_rx_pools
= {
813 .name
= "spapr_llan/rx_pools",
815 .minimum_version_id
= 1,
816 .needed
= spapr_vlan_rx_buffer_pools_needed
,
817 .fields
= (VMStateField
[]) {
818 VMSTATE_ARRAY_OF_POINTER_TO_STRUCT(rx_pool
, SpaprVioVlan
,
820 vmstate_rx_buffer_pool
, RxBufPool
),
821 VMSTATE_END_OF_LIST()
825 static const VMStateDescription vmstate_spapr_llan
= {
826 .name
= "spapr_llan",
828 .minimum_version_id
= 1,
829 .fields
= (VMStateField
[]) {
830 VMSTATE_SPAPR_VIO(sdev
, SpaprVioVlan
),
832 VMSTATE_BOOL(isopen
, SpaprVioVlan
),
833 VMSTATE_UINT64(buf_list
, SpaprVioVlan
),
834 VMSTATE_UINT32(add_buf_ptr
, SpaprVioVlan
),
835 VMSTATE_UINT32(use_buf_ptr
, SpaprVioVlan
),
836 VMSTATE_UINT32(rx_bufs
, SpaprVioVlan
),
837 VMSTATE_UINT64(rxq_ptr
, SpaprVioVlan
),
839 VMSTATE_END_OF_LIST()
841 .subsections
= (const VMStateDescription
* []) {
847 static void spapr_vlan_class_init(ObjectClass
*klass
, void *data
)
849 DeviceClass
*dc
= DEVICE_CLASS(klass
);
850 SpaprVioDeviceClass
*k
= VIO_SPAPR_DEVICE_CLASS(klass
);
852 k
->realize
= spapr_vlan_realize
;
853 k
->reset
= spapr_vlan_reset
;
854 k
->devnode
= spapr_vlan_devnode
;
855 k
->dt_name
= "l-lan";
856 k
->dt_type
= "network";
857 k
->dt_compatible
= "IBM,l-lan";
858 k
->signal_mask
= 0x1;
859 set_bit(DEVICE_CATEGORY_NETWORK
, dc
->categories
);
860 dc
->props
= spapr_vlan_properties
;
861 k
->rtce_window_size
= 0x10000000;
862 dc
->vmsd
= &vmstate_spapr_llan
;
865 static const TypeInfo spapr_vlan_info
= {
866 .name
= TYPE_VIO_SPAPR_VLAN_DEVICE
,
867 .parent
= TYPE_VIO_SPAPR_DEVICE
,
868 .instance_size
= sizeof(SpaprVioVlan
),
869 .class_init
= spapr_vlan_class_init
,
870 .instance_init
= spapr_vlan_instance_init
,
871 .instance_finalize
= spapr_vlan_instance_finalize
,
874 static void spapr_vlan_register_types(void)
876 spapr_register_hypercall(H_REGISTER_LOGICAL_LAN
, h_register_logical_lan
);
877 spapr_register_hypercall(H_FREE_LOGICAL_LAN
, h_free_logical_lan
);
878 spapr_register_hypercall(H_SEND_LOGICAL_LAN
, h_send_logical_lan
);
879 spapr_register_hypercall(H_ADD_LOGICAL_LAN_BUFFER
,
880 h_add_logical_lan_buffer
);
881 spapr_register_hypercall(H_MULTICAST_CTRL
, h_multicast_ctrl
);
882 spapr_register_hypercall(H_CHANGE_LOGICAL_LAN_MAC
,
883 h_change_logical_lan_mac
);
884 type_register_static(&spapr_vlan_info
);
887 type_init(spapr_vlan_register_types
)