2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
4 * PAPR Inter-VM Logical Lan, aka ibmveth
6 * Copyright (c) 2010,2011 David Gibson, IBM Corporation.
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
28 #include "qemu/osdep.h"
30 #include "qemu/module.h"
32 #include "migration/vmstate.h"
33 #include "hw/ppc/spapr.h"
34 #include "hw/ppc/spapr_vio.h"
35 #include "hw/qdev-properties.h"
36 #include "sysemu/sysemu.h"
40 #include "qom/object.h"
43 #define MAX_PACKET_SIZE 65536
45 /* Compatibility flags for migration */
46 #define SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT 0
47 #define SPAPRVLAN_FLAG_RX_BUF_POOLS (1 << SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT)
53 typedef uint64_t vlan_bd_t
;
55 #define VLAN_BD_VALID 0x8000000000000000ULL
56 #define VLAN_BD_TOGGLE 0x4000000000000000ULL
57 #define VLAN_BD_NO_CSUM 0x0200000000000000ULL
58 #define VLAN_BD_CSUM_GOOD 0x0100000000000000ULL
59 #define VLAN_BD_LEN_MASK 0x00ffffff00000000ULL
60 #define VLAN_BD_LEN(bd) (((bd) & VLAN_BD_LEN_MASK) >> 32)
61 #define VLAN_BD_ADDR_MASK 0x00000000ffffffffULL
62 #define VLAN_BD_ADDR(bd) ((bd) & VLAN_BD_ADDR_MASK)
64 #define VLAN_VALID_BD(addr, len) (VLAN_BD_VALID | \
65 (((len) << 32) & VLAN_BD_LEN_MASK) | \
66 (addr & VLAN_BD_ADDR_MASK))
68 #define VLAN_RXQC_TOGGLE 0x80
69 #define VLAN_RXQC_VALID 0x40
70 #define VLAN_RXQC_NO_CSUM 0x02
71 #define VLAN_RXQC_CSUM_GOOD 0x01
73 #define VLAN_RQ_ALIGNMENT 16
74 #define VLAN_RXQ_BD_OFF 0
75 #define VLAN_FILTER_BD_OFF 8
76 #define VLAN_RX_BDS_OFF 16
78 * The final 8 bytes of the buffer list is a counter of frames dropped
79 * because there was not a buffer in the buffer list capable of holding
80 * the frame. We must avoid it, or the operating system will report garbage
83 #define VLAN_RX_BDS_LEN (SPAPR_TCE_PAGE_SIZE - VLAN_RX_BDS_OFF - 8)
84 #define VLAN_MAX_BUFS (VLAN_RX_BDS_LEN / 8)
86 #define TYPE_VIO_SPAPR_VLAN_DEVICE "spapr-vlan"
87 OBJECT_DECLARE_SIMPLE_TYPE(SpaprVioVlan
, VIO_SPAPR_VLAN_DEVICE
)
89 #define RX_POOL_MAX_BDS 4096
90 #define RX_MAX_POOLS 5
95 vlan_bd_t bds
[RX_POOL_MAX_BDS
];
105 uint32_t add_buf_ptr
, use_buf_ptr
, rx_bufs
;
107 QEMUTimer
*rxp_timer
;
108 uint32_t compat_flags
; /* Compatibility flags for migration */
109 RxBufPool
*rx_pool
[RX_MAX_POOLS
]; /* Receive buffer descriptor pools */
112 static bool spapr_vlan_can_receive(NetClientState
*nc
)
114 SpaprVioVlan
*dev
= qemu_get_nic_opaque(nc
);
116 return dev
->isopen
&& dev
->rx_bufs
> 0;
120 * The last 8 bytes of the receive buffer list page (that has been
121 * supplied by the guest with the H_REGISTER_LOGICAL_LAN call) contain
122 * a counter for frames that have been dropped because there was no
123 * suitable receive buffer available. This function is used to increase
124 * this counter by one.
126 static void spapr_vlan_record_dropped_rx_frame(SpaprVioVlan
*dev
)
130 cnt
= vio_ldq(&dev
->sdev
, dev
->buf_list
+ 4096 - 8);
131 vio_stq(&dev
->sdev
, dev
->buf_list
+ 4096 - 8, cnt
+ 1);
135 * Get buffer descriptor from one of our receive buffer pools
137 static vlan_bd_t
spapr_vlan_get_rx_bd_from_pool(SpaprVioVlan
*dev
,
143 for (pool
= 0; pool
< RX_MAX_POOLS
; pool
++) {
144 if (dev
->rx_pool
[pool
]->count
> 0 &&
145 dev
->rx_pool
[pool
]->bufsize
>= size
+ 8) {
149 if (pool
== RX_MAX_POOLS
) {
150 /* Failed to find a suitable buffer */
155 trace_spapr_vlan_get_rx_bd_from_pool_found(pool
,
156 dev
->rx_pool
[pool
]->count
,
159 /* Remove the buffer from the pool */
160 dev
->rx_pool
[pool
]->count
--;
161 bd
= dev
->rx_pool
[pool
]->bds
[dev
->rx_pool
[pool
]->count
];
162 dev
->rx_pool
[pool
]->bds
[dev
->rx_pool
[pool
]->count
] = 0;
168 * Get buffer descriptor from the receive buffer list page that has been
169 * supplied by the guest with the H_REGISTER_LOGICAL_LAN call
171 static vlan_bd_t
spapr_vlan_get_rx_bd_from_page(SpaprVioVlan
*dev
,
174 int buf_ptr
= dev
->use_buf_ptr
;
179 if (buf_ptr
>= VLAN_RX_BDS_LEN
+ VLAN_RX_BDS_OFF
) {
180 buf_ptr
= VLAN_RX_BDS_OFF
;
183 bd
= vio_ldq(&dev
->sdev
, dev
->buf_list
+ buf_ptr
);
185 trace_spapr_vlan_get_rx_bd_from_page(buf_ptr
, (uint64_t)bd
);
186 } while ((!(bd
& VLAN_BD_VALID
) || VLAN_BD_LEN(bd
) < size
+ 8)
187 && buf_ptr
!= dev
->use_buf_ptr
);
189 if (!(bd
& VLAN_BD_VALID
) || VLAN_BD_LEN(bd
) < size
+ 8) {
190 /* Failed to find a suitable buffer */
194 /* Remove the buffer from the pool */
195 dev
->use_buf_ptr
= buf_ptr
;
196 vio_stq(&dev
->sdev
, dev
->buf_list
+ dev
->use_buf_ptr
, 0);
198 trace_spapr_vlan_get_rx_bd_from_page_found(dev
->use_buf_ptr
, dev
->rx_bufs
);
203 static ssize_t
spapr_vlan_receive(NetClientState
*nc
, const uint8_t *buf
,
206 SpaprVioVlan
*dev
= qemu_get_nic_opaque(nc
);
207 SpaprVioDevice
*sdev
= VIO_SPAPR_DEVICE(dev
);
208 vlan_bd_t rxq_bd
= vio_ldq(sdev
, dev
->buf_list
+ VLAN_RXQ_BD_OFF
);
213 trace_spapr_vlan_receive(sdev
->qdev
.id
, dev
->rx_bufs
);
220 spapr_vlan_record_dropped_rx_frame(dev
);
224 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
225 bd
= spapr_vlan_get_rx_bd_from_pool(dev
, size
);
227 bd
= spapr_vlan_get_rx_bd_from_page(dev
, size
);
230 spapr_vlan_record_dropped_rx_frame(dev
);
236 /* Transfer the packet data */
237 if (spapr_vio_dma_write(sdev
, VLAN_BD_ADDR(bd
) + 8, buf
, size
) < 0) {
241 trace_spapr_vlan_receive_dma_completed();
243 /* Update the receive queue */
244 control
= VLAN_RXQC_TOGGLE
| VLAN_RXQC_VALID
;
245 if (rxq_bd
& VLAN_BD_TOGGLE
) {
246 control
^= VLAN_RXQC_TOGGLE
;
249 handle
= vio_ldq(sdev
, VLAN_BD_ADDR(bd
));
250 vio_stq(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
+ 8, handle
);
251 vio_stl(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
+ 4, size
);
252 vio_sth(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
+ 2, 8);
253 vio_stb(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
, control
);
255 trace_spapr_vlan_receive_wrote(dev
->rxq_ptr
,
256 vio_ldq(sdev
, VLAN_BD_ADDR(rxq_bd
) +
258 vio_ldq(sdev
, VLAN_BD_ADDR(rxq_bd
) +
262 if (dev
->rxq_ptr
>= VLAN_BD_LEN(rxq_bd
)) {
264 vio_stq(sdev
, dev
->buf_list
+ VLAN_RXQ_BD_OFF
, rxq_bd
^ VLAN_BD_TOGGLE
);
267 if (sdev
->signal_state
& 1) {
268 spapr_vio_irq_pulse(sdev
);
274 static NetClientInfo net_spapr_vlan_info
= {
275 .type
= NET_CLIENT_DRIVER_NIC
,
276 .size
= sizeof(NICState
),
277 .can_receive
= spapr_vlan_can_receive
,
278 .receive
= spapr_vlan_receive
,
281 static void spapr_vlan_flush_rx_queue(void *opaque
)
283 SpaprVioVlan
*dev
= opaque
;
285 qemu_flush_queued_packets(qemu_get_queue(dev
->nic
));
288 static void spapr_vlan_reset_rx_pool(RxBufPool
*rxp
)
291 * Use INT_MAX as bufsize so that unused buffers are moved to the end
292 * of the list during the qsort in spapr_vlan_add_rxbuf_to_pool() later.
294 rxp
->bufsize
= INT_MAX
;
296 memset(rxp
->bds
, 0, sizeof(rxp
->bds
));
299 static void spapr_vlan_reset(SpaprVioDevice
*sdev
)
301 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
308 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
309 for (i
= 0; i
< RX_MAX_POOLS
; i
++) {
310 spapr_vlan_reset_rx_pool(dev
->rx_pool
[i
]);
314 memcpy(&dev
->nicconf
.macaddr
.a
, &dev
->perm_mac
.a
,
315 sizeof(dev
->nicconf
.macaddr
.a
));
316 qemu_format_nic_info_str(qemu_get_queue(dev
->nic
), dev
->nicconf
.macaddr
.a
);
319 static void spapr_vlan_realize(SpaprVioDevice
*sdev
, Error
**errp
)
321 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
323 qemu_macaddr_default_if_unset(&dev
->nicconf
.macaddr
);
325 memcpy(&dev
->perm_mac
.a
, &dev
->nicconf
.macaddr
.a
, sizeof(dev
->perm_mac
.a
));
327 dev
->nic
= qemu_new_nic(&net_spapr_vlan_info
, &dev
->nicconf
,
328 object_get_typename(OBJECT(sdev
)), sdev
->qdev
.id
,
329 &sdev
->qdev
.mem_reentrancy_guard
, dev
);
330 qemu_format_nic_info_str(qemu_get_queue(dev
->nic
), dev
->nicconf
.macaddr
.a
);
332 dev
->rxp_timer
= timer_new_us(QEMU_CLOCK_VIRTUAL
, spapr_vlan_flush_rx_queue
,
336 static void spapr_vlan_instance_init(Object
*obj
)
338 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(obj
);
341 device_add_bootindex_property(obj
, &dev
->nicconf
.bootindex
,
345 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
346 for (i
= 0; i
< RX_MAX_POOLS
; i
++) {
347 dev
->rx_pool
[i
] = g_new(RxBufPool
, 1);
348 spapr_vlan_reset_rx_pool(dev
->rx_pool
[i
]);
353 static void spapr_vlan_instance_finalize(Object
*obj
)
355 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(obj
);
358 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
359 for (i
= 0; i
< RX_MAX_POOLS
; i
++) {
360 g_free(dev
->rx_pool
[i
]);
361 dev
->rx_pool
[i
] = NULL
;
365 if (dev
->rxp_timer
) {
366 timer_free(dev
->rxp_timer
);
370 void spapr_vlan_create(SpaprVioBus
*bus
, NICInfo
*nd
)
374 dev
= qdev_new("spapr-vlan");
376 qdev_set_nic_properties(dev
, nd
);
378 qdev_realize_and_unref(dev
, &bus
->bus
, &error_fatal
);
381 static int spapr_vlan_devnode(SpaprVioDevice
*dev
, void *fdt
, int node_off
)
383 SpaprVioVlan
*vdev
= VIO_SPAPR_VLAN_DEVICE(dev
);
384 uint8_t padded_mac
[8] = {0, 0};
387 /* Some old phyp versions give the mac address in an 8-byte
388 * property. The kernel driver (before 3.10) has an insane workaround;
389 * rather than doing the obvious thing and checking the property
390 * length, it checks whether the first byte has 0b10 in the low
391 * bits. If a correct 6-byte property has a different first byte
392 * the kernel will get the wrong mac address, overrunning its
393 * buffer in the process (read only, thank goodness).
395 * Here we return a 6-byte address unless that would break a pre-3.10
396 * driver. In that case we return a padded 8-byte address to allow the old
397 * workaround to succeed. */
398 if ((vdev
->nicconf
.macaddr
.a
[0] & 0x3) == 0x2) {
399 ret
= fdt_setprop(fdt
, node_off
, "local-mac-address",
400 &vdev
->nicconf
.macaddr
, ETH_ALEN
);
402 memcpy(&padded_mac
[2], &vdev
->nicconf
.macaddr
, ETH_ALEN
);
403 ret
= fdt_setprop(fdt
, node_off
, "local-mac-address",
404 padded_mac
, sizeof(padded_mac
));
410 ret
= fdt_setprop_cell(fdt
, node_off
, "ibm,mac-address-filters", 0);
418 static int check_bd(SpaprVioVlan
*dev
, vlan_bd_t bd
,
419 target_ulong alignment
)
421 if ((VLAN_BD_ADDR(bd
) % alignment
)
422 || (VLAN_BD_LEN(bd
) % alignment
)) {
426 if (!spapr_vio_dma_valid(&dev
->sdev
, VLAN_BD_ADDR(bd
),
427 VLAN_BD_LEN(bd
), DMA_DIRECTION_FROM_DEVICE
)
428 || !spapr_vio_dma_valid(&dev
->sdev
, VLAN_BD_ADDR(bd
),
429 VLAN_BD_LEN(bd
), DMA_DIRECTION_TO_DEVICE
)) {
436 static target_ulong
h_register_logical_lan(PowerPCCPU
*cpu
,
437 SpaprMachineState
*spapr
,
441 target_ulong reg
= args
[0];
442 target_ulong buf_list
= args
[1];
443 target_ulong rec_queue
= args
[2];
444 target_ulong filter_list
= args
[3];
445 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
446 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
447 vlan_bd_t filter_list_bd
;
454 hcall_dprintf("H_REGISTER_LOGICAL_LAN called twice without "
455 "H_FREE_LOGICAL_LAN\n");
459 if (check_bd(dev
, VLAN_VALID_BD(buf_list
, SPAPR_TCE_PAGE_SIZE
),
460 SPAPR_TCE_PAGE_SIZE
) < 0) {
461 hcall_dprintf("Bad buf_list 0x" TARGET_FMT_lx
"\n", buf_list
);
465 filter_list_bd
= VLAN_VALID_BD(filter_list
, SPAPR_TCE_PAGE_SIZE
);
466 if (check_bd(dev
, filter_list_bd
, SPAPR_TCE_PAGE_SIZE
) < 0) {
467 hcall_dprintf("Bad filter_list 0x" TARGET_FMT_lx
"\n", filter_list
);
471 if (!(rec_queue
& VLAN_BD_VALID
)
472 || (check_bd(dev
, rec_queue
, VLAN_RQ_ALIGNMENT
) < 0)) {
473 hcall_dprintf("Bad receive queue\n");
477 dev
->buf_list
= buf_list
;
478 sdev
->signal_state
= 0;
480 rec_queue
&= ~VLAN_BD_TOGGLE
;
482 /* Initialize the buffer list */
483 vio_stq(sdev
, buf_list
, rec_queue
);
484 vio_stq(sdev
, buf_list
+ 8, filter_list_bd
);
485 spapr_vio_dma_set(sdev
, buf_list
+ VLAN_RX_BDS_OFF
, 0,
486 SPAPR_TCE_PAGE_SIZE
- VLAN_RX_BDS_OFF
);
487 dev
->add_buf_ptr
= VLAN_RX_BDS_OFF
- 8;
488 dev
->use_buf_ptr
= VLAN_RX_BDS_OFF
- 8;
492 /* Initialize the receive queue */
493 spapr_vio_dma_set(sdev
, VLAN_BD_ADDR(rec_queue
), 0, VLAN_BD_LEN(rec_queue
));
496 qemu_flush_queued_packets(qemu_get_queue(dev
->nic
));
502 static target_ulong
h_free_logical_lan(PowerPCCPU
*cpu
,
503 SpaprMachineState
*spapr
,
504 target_ulong opcode
, target_ulong
*args
)
506 target_ulong reg
= args
[0];
507 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
508 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
515 hcall_dprintf("H_FREE_LOGICAL_LAN called without "
516 "H_REGISTER_LOGICAL_LAN\n");
520 spapr_vlan_reset(sdev
);
525 * Used for qsort, this function compares two RxBufPools by size.
527 static int rx_pool_size_compare(const void *p1
, const void *p2
)
529 const RxBufPool
*pool1
= *(RxBufPool
**)p1
;
530 const RxBufPool
*pool2
= *(RxBufPool
**)p2
;
532 if (pool1
->bufsize
< pool2
->bufsize
) {
535 return pool1
->bufsize
> pool2
->bufsize
;
539 * Search for a matching buffer pool with exact matching size,
540 * or return -1 if no matching pool has been found.
542 static int spapr_vlan_get_rx_pool_id(SpaprVioVlan
*dev
, int size
)
546 for (pool
= 0; pool
< RX_MAX_POOLS
; pool
++) {
547 if (dev
->rx_pool
[pool
]->bufsize
== size
) {
556 * Enqueuing receive buffer by adding it to one of our receive buffer pools
558 static target_long
spapr_vlan_add_rxbuf_to_pool(SpaprVioVlan
*dev
,
561 int size
= VLAN_BD_LEN(buf
);
564 pool
= spapr_vlan_get_rx_pool_id(dev
, size
);
567 * No matching pool found? Try to use a new one. If the guest used all
568 * pools before, but changed the size of one pool in the meantime, we might
569 * need to recycle that pool here (if it's empty already). Thus scan
570 * all buffer pools now, starting with the last (likely empty) one.
572 for (pool
= RX_MAX_POOLS
- 1; pool
>= 0 ; pool
--) {
573 if (dev
->rx_pool
[pool
]->count
== 0) {
574 dev
->rx_pool
[pool
]->bufsize
= size
;
576 * Sort pools by size so that spapr_vlan_receive()
577 * can later find the smallest buffer pool easily.
579 qsort(dev
->rx_pool
, RX_MAX_POOLS
, sizeof(dev
->rx_pool
[0]),
580 rx_pool_size_compare
);
581 pool
= spapr_vlan_get_rx_pool_id(dev
, size
);
582 trace_spapr_vlan_add_rxbuf_to_pool_create(pool
,
588 /* Still no usable pool? Give up */
589 if (pool
< 0 || dev
->rx_pool
[pool
]->count
>= RX_POOL_MAX_BDS
) {
593 trace_spapr_vlan_add_rxbuf_to_pool(pool
, VLAN_BD_LEN(buf
),
594 dev
->rx_pool
[pool
]->count
);
596 dev
->rx_pool
[pool
]->bds
[dev
->rx_pool
[pool
]->count
++] = buf
;
602 * This is the old way of enqueuing receive buffers: Add it to the rx queue
603 * page that has been supplied by the guest (which is quite limited in size).
605 static target_long
spapr_vlan_add_rxbuf_to_page(SpaprVioVlan
*dev
,
610 if (dev
->rx_bufs
>= VLAN_MAX_BUFS
) {
615 dev
->add_buf_ptr
+= 8;
616 if (dev
->add_buf_ptr
>= VLAN_RX_BDS_LEN
+ VLAN_RX_BDS_OFF
) {
617 dev
->add_buf_ptr
= VLAN_RX_BDS_OFF
;
620 bd
= vio_ldq(&dev
->sdev
, dev
->buf_list
+ dev
->add_buf_ptr
);
621 } while (bd
& VLAN_BD_VALID
);
623 vio_stq(&dev
->sdev
, dev
->buf_list
+ dev
->add_buf_ptr
, buf
);
625 trace_spapr_vlan_add_rxbuf_to_page(dev
->add_buf_ptr
, dev
->rx_bufs
, buf
);
630 static target_ulong
h_add_logical_lan_buffer(PowerPCCPU
*cpu
,
631 SpaprMachineState
*spapr
,
635 target_ulong reg
= args
[0];
636 target_ulong buf
= args
[1];
637 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
638 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
641 trace_spapr_vlan_h_add_logical_lan_buffer(reg
, buf
);
644 hcall_dprintf("Bad device\n");
648 if ((check_bd(dev
, buf
, 4) < 0)
649 || (VLAN_BD_LEN(buf
) < 16)) {
650 hcall_dprintf("Bad buffer enqueued\n");
658 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
659 ret
= spapr_vlan_add_rxbuf_to_pool(dev
, buf
);
661 ret
= spapr_vlan_add_rxbuf_to_page(dev
, buf
);
670 * Give guest some more time to add additional RX buffers before we
671 * flush the receive queue, so that e.g. fragmented IP packets can
672 * be passed to the guest in one go later (instead of passing single
673 * fragments if there is only one receive buffer available).
675 timer_mod(dev
->rxp_timer
, qemu_clock_get_us(QEMU_CLOCK_VIRTUAL
) + 500);
680 static target_ulong
h_send_logical_lan(PowerPCCPU
*cpu
,
681 SpaprMachineState
*spapr
,
682 target_ulong opcode
, target_ulong
*args
)
684 target_ulong reg
= args
[0];
685 target_ulong
*bufs
= args
+ 1;
686 target_ulong continue_token
= args
[7];
687 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
688 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
691 g_autofree
uint8_t *lbuf
= NULL
;
695 trace_spapr_vlan_h_send_logical_lan(reg
, continue_token
);
701 trace_spapr_vlan_h_send_logical_lan_rxbufs(dev
->rx_bufs
);
707 if (continue_token
) {
708 return H_HARDWARE
; /* FIXME actually handle this */
712 for (i
= 0; i
< 6; i
++) {
713 trace_spapr_vlan_h_send_logical_lan_buf_desc(bufs
[i
]);
714 if (!(bufs
[i
] & VLAN_BD_VALID
)) {
717 total_len
+= VLAN_BD_LEN(bufs
[i
]);
721 trace_spapr_vlan_h_send_logical_lan_total(nbufs
, total_len
);
723 if (total_len
== 0) {
727 if (total_len
> MAX_PACKET_SIZE
) {
728 /* Don't let the guest force too large an allocation */
732 lbuf
= g_malloc(total_len
);
734 for (i
= 0; i
< nbufs
; i
++) {
735 ret
= spapr_vio_dma_read(sdev
, VLAN_BD_ADDR(bufs
[i
]),
736 p
, VLAN_BD_LEN(bufs
[i
]));
741 p
+= VLAN_BD_LEN(bufs
[i
]);
744 qemu_send_packet(qemu_get_queue(dev
->nic
), lbuf
, total_len
);
749 static target_ulong
h_multicast_ctrl(PowerPCCPU
*cpu
, SpaprMachineState
*spapr
,
750 target_ulong opcode
, target_ulong
*args
)
752 target_ulong reg
= args
[0];
753 SpaprVioDevice
*dev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
762 static target_ulong
h_change_logical_lan_mac(PowerPCCPU
*cpu
,
763 SpaprMachineState
*spapr
,
767 target_ulong reg
= args
[0];
768 target_ulong macaddr
= args
[1];
769 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
770 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
774 hcall_dprintf("H_CHANGE_LOGICAL_LAN_MAC called when "
775 "no NIC is present\n");
779 for (i
= 0; i
< ETH_ALEN
; i
++) {
780 dev
->nicconf
.macaddr
.a
[ETH_ALEN
- i
- 1] = macaddr
& 0xff;
784 qemu_format_nic_info_str(qemu_get_queue(dev
->nic
), dev
->nicconf
.macaddr
.a
);
789 static Property spapr_vlan_properties
[] = {
790 DEFINE_SPAPR_PROPERTIES(SpaprVioVlan
, sdev
),
791 DEFINE_NIC_PROPERTIES(SpaprVioVlan
, nicconf
),
792 DEFINE_PROP_BIT("use-rx-buffer-pools", SpaprVioVlan
,
793 compat_flags
, SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT
, true),
794 DEFINE_PROP_END_OF_LIST(),
797 static bool spapr_vlan_rx_buffer_pools_needed(void *opaque
)
799 SpaprVioVlan
*dev
= opaque
;
801 return (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) != 0;
804 static const VMStateDescription vmstate_rx_buffer_pool
= {
805 .name
= "spapr_llan/rx_buffer_pool",
807 .minimum_version_id
= 1,
808 .needed
= spapr_vlan_rx_buffer_pools_needed
,
809 .fields
= (const VMStateField
[]) {
810 VMSTATE_INT32(bufsize
, RxBufPool
),
811 VMSTATE_INT32(count
, RxBufPool
),
812 VMSTATE_UINT64_ARRAY(bds
, RxBufPool
, RX_POOL_MAX_BDS
),
813 VMSTATE_END_OF_LIST()
817 static const VMStateDescription vmstate_rx_pools
= {
818 .name
= "spapr_llan/rx_pools",
820 .minimum_version_id
= 1,
821 .needed
= spapr_vlan_rx_buffer_pools_needed
,
822 .fields
= (const VMStateField
[]) {
823 VMSTATE_ARRAY_OF_POINTER_TO_STRUCT(rx_pool
, SpaprVioVlan
,
825 vmstate_rx_buffer_pool
, RxBufPool
),
826 VMSTATE_END_OF_LIST()
830 static const VMStateDescription vmstate_spapr_llan
= {
831 .name
= "spapr_llan",
833 .minimum_version_id
= 1,
834 .fields
= (const VMStateField
[]) {
835 VMSTATE_SPAPR_VIO(sdev
, SpaprVioVlan
),
837 VMSTATE_BOOL(isopen
, SpaprVioVlan
),
838 VMSTATE_UINT64(buf_list
, SpaprVioVlan
),
839 VMSTATE_UINT32(add_buf_ptr
, SpaprVioVlan
),
840 VMSTATE_UINT32(use_buf_ptr
, SpaprVioVlan
),
841 VMSTATE_UINT32(rx_bufs
, SpaprVioVlan
),
842 VMSTATE_UINT64(rxq_ptr
, SpaprVioVlan
),
844 VMSTATE_END_OF_LIST()
846 .subsections
= (const VMStateDescription
* const []) {
852 static void spapr_vlan_class_init(ObjectClass
*klass
, void *data
)
854 DeviceClass
*dc
= DEVICE_CLASS(klass
);
855 SpaprVioDeviceClass
*k
= VIO_SPAPR_DEVICE_CLASS(klass
);
857 k
->realize
= spapr_vlan_realize
;
858 k
->reset
= spapr_vlan_reset
;
859 k
->devnode
= spapr_vlan_devnode
;
860 k
->dt_name
= "l-lan";
861 k
->dt_type
= "network";
862 k
->dt_compatible
= "IBM,l-lan";
863 k
->signal_mask
= 0x1;
864 set_bit(DEVICE_CATEGORY_NETWORK
, dc
->categories
);
865 device_class_set_props(dc
, spapr_vlan_properties
);
866 k
->rtce_window_size
= 0x10000000;
867 dc
->vmsd
= &vmstate_spapr_llan
;
870 static const TypeInfo spapr_vlan_info
= {
871 .name
= TYPE_VIO_SPAPR_VLAN_DEVICE
,
872 .parent
= TYPE_VIO_SPAPR_DEVICE
,
873 .instance_size
= sizeof(SpaprVioVlan
),
874 .class_init
= spapr_vlan_class_init
,
875 .instance_init
= spapr_vlan_instance_init
,
876 .instance_finalize
= spapr_vlan_instance_finalize
,
879 static void spapr_vlan_register_types(void)
881 spapr_register_hypercall(H_REGISTER_LOGICAL_LAN
, h_register_logical_lan
);
882 spapr_register_hypercall(H_FREE_LOGICAL_LAN
, h_free_logical_lan
);
883 spapr_register_hypercall(H_SEND_LOGICAL_LAN
, h_send_logical_lan
);
884 spapr_register_hypercall(H_ADD_LOGICAL_LAN_BUFFER
,
885 h_add_logical_lan_buffer
);
886 spapr_register_hypercall(H_MULTICAST_CTRL
, h_multicast_ctrl
);
887 spapr_register_hypercall(H_CHANGE_LOGICAL_LAN_MAC
,
888 h_change_logical_lan_mac
);
889 type_register_static(&spapr_vlan_info
);
892 type_init(spapr_vlan_register_types
)