2 * QEMU NVM Express Virtual Namespace
4 * Copyright (c) 2019 CNEX Labs
5 * Copyright (c) 2020 Samsung Electronics
8 * Klaus Jensen <k.jensen@samsung.com>
10 * This work is licensed under the terms of the GNU GPL, version 2. See the
11 * COPYING file in the top-level directory.
15 #include "qemu/osdep.h"
16 #include "qemu/units.h"
17 #include "qemu/cutils.h"
18 #include "qemu/error-report.h"
19 #include "qapi/error.h"
20 #include "qemu/bitops.h"
21 #include "sysemu/sysemu.h"
22 #include "sysemu/block-backend.h"
27 #define MIN_DISCARD_GRANULARITY (4 * KiB)
28 #define NVME_DEFAULT_ZONE_SIZE (128 * MiB)
30 void nvme_ns_init_format(NvmeNamespace
*ns
)
32 NvmeIdNs
*id_ns
= &ns
->id_ns
;
37 ns
->lbaf
= id_ns
->lbaf
[NVME_ID_NS_FLBAS_INDEX(id_ns
->flbas
)];
38 ns
->lbasz
= 1 << ns
->lbaf
.ds
;
40 nlbas
= ns
->size
/ (ns
->lbasz
+ ns
->lbaf
.ms
);
42 id_ns
->nsze
= cpu_to_le64(nlbas
);
44 /* no thin provisioning */
45 id_ns
->ncap
= id_ns
->nsze
;
46 id_ns
->nuse
= id_ns
->ncap
;
48 ns
->moff
= nlbas
<< ns
->lbaf
.ds
;
50 npdg
= ns
->blkconf
.discard_granularity
/ ns
->lbasz
;
52 ret
= bdrv_get_info(blk_bs(ns
->blkconf
.blk
), &bdi
);
53 if (ret
>= 0 && bdi
.cluster_size
> ns
->blkconf
.discard_granularity
) {
54 npdg
= bdi
.cluster_size
/ ns
->lbasz
;
57 id_ns
->npda
= id_ns
->npdg
= npdg
- 1;
60 static int nvme_ns_init(NvmeNamespace
*ns
, Error
**errp
)
62 static uint64_t ns_count
;
63 NvmeIdNs
*id_ns
= &ns
->id_ns
;
64 NvmeIdNsNvm
*id_ns_nvm
= &ns
->id_ns_nvm
;
69 ns
->csi
= NVME_CSI_NVM
;
72 ns
->id_ns
.dlfeat
= 0x1;
74 /* support DULBE and I/O optimization fields */
75 id_ns
->nsfeat
|= (0x4 | 0x10);
77 if (ns
->params
.shared
) {
78 id_ns
->nmic
|= NVME_NMIC_NS_SHARED
;
81 /* Substitute a missing EUI-64 by an autogenerated one */
83 if (!ns
->params
.eui64
&& ns
->params
.eui64_default
) {
84 ns
->params
.eui64
= ns_count
+ NVME_EUI64_DEFAULT
;
88 id_ns
->mssrl
= cpu_to_le16(ns
->params
.mssrl
);
89 id_ns
->mcl
= cpu_to_le32(ns
->params
.mcl
);
90 id_ns
->msrc
= ns
->params
.msrc
;
91 id_ns
->eui64
= cpu_to_be64(ns
->params
.eui64
);
93 ds
= 31 - clz32(ns
->blkconf
.logical_block_size
);
96 id_ns
->mc
= NVME_ID_NS_MC_EXTENDED
| NVME_ID_NS_MC_SEPARATE
;
98 if (ms
&& ns
->params
.mset
) {
99 id_ns
->flbas
|= NVME_ID_NS_FLBAS_EXTENDED
;
103 id_ns
->dps
= ns
->params
.pi
;
104 if (ns
->params
.pi
&& ns
->params
.pil
) {
105 id_ns
->dps
|= NVME_ID_NS_DPS_FIRST_EIGHT
;
108 ns
->pif
= ns
->params
.pif
;
110 static const NvmeLBAF lbaf
[16] = {
112 [1] = { .ds
= 9, .ms
= 8 },
113 [2] = { .ds
= 9, .ms
= 16 },
114 [3] = { .ds
= 9, .ms
= 64 },
116 [5] = { .ds
= 12, .ms
= 8 },
117 [6] = { .ds
= 12, .ms
= 16 },
118 [7] = { .ds
= 12, .ms
= 64 },
123 memcpy(&id_ns
->lbaf
, &lbaf
, sizeof(lbaf
));
125 for (i
= 0; i
< ns
->nlbaf
; i
++) {
126 NvmeLBAF
*lbaf
= &id_ns
->lbaf
[i
];
127 if (lbaf
->ds
== ds
) {
128 if (lbaf
->ms
== ms
) {
135 /* add non-standard lba format */
136 id_ns
->lbaf
[ns
->nlbaf
].ds
= ds
;
137 id_ns
->lbaf
[ns
->nlbaf
].ms
= ms
;
144 id_ns_nvm
->elbaf
[i
] = (ns
->pif
& 0x3) << 7;
145 id_ns
->nlbaf
= ns
->nlbaf
- 1;
146 nvme_ns_init_format(ns
);
151 static int nvme_ns_init_blk(NvmeNamespace
*ns
, Error
**errp
)
155 if (!blkconf_blocksizes(&ns
->blkconf
, errp
)) {
159 read_only
= !blk_supports_write_perm(ns
->blkconf
.blk
);
160 if (!blkconf_apply_backend_options(&ns
->blkconf
, read_only
, false, errp
)) {
164 if (ns
->blkconf
.discard_granularity
== -1) {
165 ns
->blkconf
.discard_granularity
=
166 MAX(ns
->blkconf
.logical_block_size
, MIN_DISCARD_GRANULARITY
);
169 ns
->size
= blk_getlength(ns
->blkconf
.blk
);
171 error_setg_errno(errp
, -ns
->size
, "could not get blockdev size");
178 static int nvme_ns_zoned_check_calc_geometry(NvmeNamespace
*ns
, Error
**errp
)
180 uint64_t zone_size
, zone_cap
;
182 /* Make sure that the values of ZNS properties are sane */
183 if (ns
->params
.zone_size_bs
) {
184 zone_size
= ns
->params
.zone_size_bs
;
186 zone_size
= NVME_DEFAULT_ZONE_SIZE
;
188 if (ns
->params
.zone_cap_bs
) {
189 zone_cap
= ns
->params
.zone_cap_bs
;
191 zone_cap
= zone_size
;
193 if (zone_cap
> zone_size
) {
194 error_setg(errp
, "zone capacity %"PRIu64
"B exceeds "
195 "zone size %"PRIu64
"B", zone_cap
, zone_size
);
198 if (zone_size
< ns
->lbasz
) {
199 error_setg(errp
, "zone size %"PRIu64
"B too small, "
200 "must be at least %zuB", zone_size
, ns
->lbasz
);
203 if (zone_cap
< ns
->lbasz
) {
204 error_setg(errp
, "zone capacity %"PRIu64
"B too small, "
205 "must be at least %zuB", zone_cap
, ns
->lbasz
);
210 * Save the main zone geometry values to avoid
211 * calculating them later again.
213 ns
->zone_size
= zone_size
/ ns
->lbasz
;
214 ns
->zone_capacity
= zone_cap
/ ns
->lbasz
;
215 ns
->num_zones
= le64_to_cpu(ns
->id_ns
.nsze
) / ns
->zone_size
;
217 /* Do a few more sanity checks of ZNS properties */
218 if (!ns
->num_zones
) {
220 "insufficient drive capacity, must be at least the size "
221 "of one zone (%"PRIu64
"B)", zone_size
);
228 static void nvme_ns_zoned_init_state(NvmeNamespace
*ns
)
230 uint64_t start
= 0, zone_size
= ns
->zone_size
;
231 uint64_t capacity
= ns
->num_zones
* zone_size
;
235 ns
->zone_array
= g_new0(NvmeZone
, ns
->num_zones
);
236 if (ns
->params
.zd_extension_size
) {
237 ns
->zd_extensions
= g_malloc0(ns
->params
.zd_extension_size
*
241 QTAILQ_INIT(&ns
->exp_open_zones
);
242 QTAILQ_INIT(&ns
->imp_open_zones
);
243 QTAILQ_INIT(&ns
->closed_zones
);
244 QTAILQ_INIT(&ns
->full_zones
);
246 zone
= ns
->zone_array
;
247 for (i
= 0; i
< ns
->num_zones
; i
++, zone
++) {
248 if (start
+ zone_size
> capacity
) {
249 zone_size
= capacity
- start
;
251 zone
->d
.zt
= NVME_ZONE_TYPE_SEQ_WRITE
;
252 nvme_set_zone_state(zone
, NVME_ZONE_STATE_EMPTY
);
254 zone
->d
.zcap
= ns
->zone_capacity
;
255 zone
->d
.zslba
= start
;
261 ns
->zone_size_log2
= 0;
262 if (is_power_of_2(ns
->zone_size
)) {
263 ns
->zone_size_log2
= 63 - clz64(ns
->zone_size
);
267 static void nvme_ns_init_zoned(NvmeNamespace
*ns
)
269 NvmeIdNsZoned
*id_ns_z
;
272 nvme_ns_zoned_init_state(ns
);
274 id_ns_z
= g_new0(NvmeIdNsZoned
, 1);
276 /* MAR/MOR are zeroes-based, FFFFFFFFFh means no limit */
277 id_ns_z
->mar
= cpu_to_le32(ns
->params
.max_active_zones
- 1);
278 id_ns_z
->mor
= cpu_to_le32(ns
->params
.max_open_zones
- 1);
280 id_ns_z
->ozcs
= ns
->params
.cross_zone_read
?
281 NVME_ID_NS_ZONED_OZCS_RAZB
: 0x00;
283 for (i
= 0; i
<= ns
->id_ns
.nlbaf
; i
++) {
284 id_ns_z
->lbafe
[i
].zsze
= cpu_to_le64(ns
->zone_size
);
285 id_ns_z
->lbafe
[i
].zdes
=
286 ns
->params
.zd_extension_size
>> 6; /* Units of 64B */
289 if (ns
->params
.zrwas
) {
290 ns
->zns
.numzrwa
= ns
->params
.numzrwa
?
291 ns
->params
.numzrwa
: ns
->num_zones
;
293 ns
->zns
.zrwas
= ns
->params
.zrwas
>> ns
->lbaf
.ds
;
294 ns
->zns
.zrwafg
= ns
->params
.zrwafg
>> ns
->lbaf
.ds
;
296 id_ns_z
->ozcs
|= NVME_ID_NS_ZONED_OZCS_ZRWASUP
;
297 id_ns_z
->zrwacap
= NVME_ID_NS_ZONED_ZRWACAP_EXPFLUSHSUP
;
299 id_ns_z
->numzrwa
= cpu_to_le32(ns
->params
.numzrwa
);
300 id_ns_z
->zrwas
= cpu_to_le16(ns
->zns
.zrwas
);
301 id_ns_z
->zrwafg
= cpu_to_le16(ns
->zns
.zrwafg
);
304 id_ns_z
->ozcs
= cpu_to_le16(id_ns_z
->ozcs
);
306 ns
->csi
= NVME_CSI_ZONED
;
307 ns
->id_ns
.nsze
= cpu_to_le64(ns
->num_zones
* ns
->zone_size
);
308 ns
->id_ns
.ncap
= ns
->id_ns
.nsze
;
309 ns
->id_ns
.nuse
= ns
->id_ns
.ncap
;
312 * The device uses the BDRV_BLOCK_ZERO flag to determine the "deallocated"
313 * status of logical blocks. Since the spec defines that logical blocks
314 * SHALL be deallocated when then zone is in the Empty or Offline states,
315 * we can only support DULBE if the zone size is a multiple of the
318 if (ns
->zone_size
% (ns
->id_ns
.npdg
+ 1)) {
319 warn_report("the zone size (%"PRIu64
" blocks) is not a multiple of "
320 "the calculated deallocation granularity (%d blocks); "
321 "DULBE support disabled",
322 ns
->zone_size
, ns
->id_ns
.npdg
+ 1);
324 ns
->id_ns
.nsfeat
&= ~0x4;
327 ns
->id_ns_zoned
= id_ns_z
;
330 static void nvme_clear_zone(NvmeNamespace
*ns
, NvmeZone
*zone
)
334 zone
->w_ptr
= zone
->d
.wp
;
335 state
= nvme_get_zone_state(zone
);
336 if (zone
->d
.wp
!= zone
->d
.zslba
||
337 (zone
->d
.za
& NVME_ZA_ZD_EXT_VALID
)) {
338 if (state
!= NVME_ZONE_STATE_CLOSED
) {
339 trace_pci_nvme_clear_ns_close(state
, zone
->d
.zslba
);
340 nvme_set_zone_state(zone
, NVME_ZONE_STATE_CLOSED
);
342 nvme_aor_inc_active(ns
);
343 QTAILQ_INSERT_HEAD(&ns
->closed_zones
, zone
, entry
);
345 trace_pci_nvme_clear_ns_reset(state
, zone
->d
.zslba
);
346 if (zone
->d
.za
& NVME_ZA_ZRWA_VALID
) {
347 zone
->d
.za
&= ~NVME_ZA_ZRWA_VALID
;
350 nvme_set_zone_state(zone
, NVME_ZONE_STATE_EMPTY
);
355 * Close all the zones that are currently open.
357 static void nvme_zoned_ns_shutdown(NvmeNamespace
*ns
)
359 NvmeZone
*zone
, *next
;
361 QTAILQ_FOREACH_SAFE(zone
, &ns
->closed_zones
, entry
, next
) {
362 QTAILQ_REMOVE(&ns
->closed_zones
, zone
, entry
);
363 nvme_aor_dec_active(ns
);
364 nvme_clear_zone(ns
, zone
);
366 QTAILQ_FOREACH_SAFE(zone
, &ns
->imp_open_zones
, entry
, next
) {
367 QTAILQ_REMOVE(&ns
->imp_open_zones
, zone
, entry
);
368 nvme_aor_dec_open(ns
);
369 nvme_aor_dec_active(ns
);
370 nvme_clear_zone(ns
, zone
);
372 QTAILQ_FOREACH_SAFE(zone
, &ns
->exp_open_zones
, entry
, next
) {
373 QTAILQ_REMOVE(&ns
->exp_open_zones
, zone
, entry
);
374 nvme_aor_dec_open(ns
);
375 nvme_aor_dec_active(ns
);
376 nvme_clear_zone(ns
, zone
);
379 assert(ns
->nr_open_zones
== 0);
382 static NvmeRuHandle
*nvme_find_ruh_by_attr(NvmeEnduranceGroup
*endgrp
,
383 uint8_t ruha
, uint16_t *ruhid
)
385 for (uint16_t i
= 0; i
< endgrp
->fdp
.nruh
; i
++) {
386 NvmeRuHandle
*ruh
= &endgrp
->fdp
.ruhs
[i
];
388 if (ruh
->ruha
== ruha
) {
397 static bool nvme_ns_init_fdp(NvmeNamespace
*ns
, Error
**errp
)
399 NvmeEnduranceGroup
*endgrp
= ns
->endgrp
;
401 uint8_t lbafi
= NVME_ID_NS_FLBAS_INDEX(ns
->id_ns
.flbas
);
402 g_autofree
unsigned int *ruhids
= NULL
;
407 if (!ns
->params
.fdp
.ruhs
) {
409 ph
= ns
->fdp
.phs
= g_new(uint16_t, 1);
411 ruh
= nvme_find_ruh_by_attr(endgrp
, NVME_RUHA_CTRL
, ph
);
413 ruh
= nvme_find_ruh_by_attr(endgrp
, NVME_RUHA_UNUSED
, ph
);
415 error_setg(errp
, "no unused reclaim unit handles left");
419 ruh
->ruha
= NVME_RUHA_CTRL
;
421 ruh
->ruamw
= endgrp
->fdp
.runs
>> ns
->lbaf
.ds
;
423 for (uint16_t rg
= 0; rg
< endgrp
->fdp
.nrg
; rg
++) {
424 ruh
->rus
[rg
].ruamw
= ruh
->ruamw
;
426 } else if (ruh
->lbafi
!= lbafi
) {
427 error_setg(errp
, "lba format index of controller assigned "
428 "reclaim unit handle does not match namespace lba "
436 ruhid
= ruhids
= g_new0(unsigned int, endgrp
->fdp
.nruh
);
437 r
= p
= strdup(ns
->params
.fdp
.ruhs
);
439 /* parse the placement handle identifiers */
440 while ((token
= qemu_strsep(&p
, ";")) != NULL
) {
442 if (ns
->fdp
.nphs
> NVME_FDP_MAXPIDS
||
443 ns
->fdp
.nphs
== endgrp
->fdp
.nruh
) {
444 error_setg(errp
, "too many placement handles");
449 if (qemu_strtoui(token
, NULL
, 0, ruhid
++) < 0) {
450 error_setg(errp
, "cannot parse reclaim unit handle identifier");
458 ph
= ns
->fdp
.phs
= g_new(uint16_t, ns
->fdp
.nphs
);
462 /* verify the identifiers */
463 for (unsigned int i
= 0; i
< ns
->fdp
.nphs
; i
++, ruhid
++, ph
++) {
464 if (*ruhid
>= endgrp
->fdp
.nruh
) {
465 error_setg(errp
, "invalid reclaim unit handle identifier");
469 ruh
= &endgrp
->fdp
.ruhs
[*ruhid
];
472 case NVME_RUHA_UNUSED
:
473 ruh
->ruha
= NVME_RUHA_HOST
;
475 ruh
->ruamw
= endgrp
->fdp
.runs
>> ns
->lbaf
.ds
;
477 for (uint16_t rg
= 0; rg
< endgrp
->fdp
.nrg
; rg
++) {
478 ruh
->rus
[rg
].ruamw
= ruh
->ruamw
;
484 if (ruh
->lbafi
!= lbafi
) {
485 error_setg(errp
, "lba format index of host assigned"
486 "reclaim unit handle does not match namespace "
494 error_setg(errp
, "reclaim unit handle is controller assigned");
507 static int nvme_ns_check_constraints(NvmeNamespace
*ns
, Error
**errp
)
509 unsigned int pi_size
;
511 if (!ns
->blkconf
.blk
) {
512 error_setg(errp
, "block backend not configured");
517 if (ns
->params
.pi
> NVME_ID_NS_DPS_TYPE_3
) {
518 error_setg(errp
, "invalid 'pi' value");
522 switch (ns
->params
.pif
) {
523 case NVME_PI_GUARD_16
:
526 case NVME_PI_GUARD_64
:
530 error_setg(errp
, "invalid 'pif'");
534 if (ns
->params
.ms
< pi_size
) {
535 error_setg(errp
, "at least %u bytes of metadata required to "
536 "enable protection information", pi_size
);
541 if (ns
->params
.nsid
> NVME_MAX_NAMESPACES
) {
542 error_setg(errp
, "invalid namespace id (must be between 0 and %d)",
543 NVME_MAX_NAMESPACES
);
547 if (ns
->params
.zoned
&& ns
->endgrp
&& ns
->endgrp
->fdp
.enabled
) {
548 error_setg(errp
, "cannot be a zoned- in an FDP configuration");
552 if (ns
->params
.zoned
) {
553 if (ns
->params
.max_active_zones
) {
554 if (ns
->params
.max_open_zones
> ns
->params
.max_active_zones
) {
555 error_setg(errp
, "max_open_zones (%u) exceeds "
556 "max_active_zones (%u)", ns
->params
.max_open_zones
,
557 ns
->params
.max_active_zones
);
561 if (!ns
->params
.max_open_zones
) {
562 ns
->params
.max_open_zones
= ns
->params
.max_active_zones
;
566 if (ns
->params
.zd_extension_size
) {
567 if (ns
->params
.zd_extension_size
& 0x3f) {
568 error_setg(errp
, "zone descriptor extension size must be a "
572 if ((ns
->params
.zd_extension_size
>> 6) > 0xff) {
574 "zone descriptor extension size is too large");
579 if (ns
->params
.zrwas
) {
580 if (ns
->params
.zrwas
% ns
->blkconf
.logical_block_size
) {
581 error_setg(errp
, "zone random write area size (zoned.zrwas "
582 "%"PRIu64
") must be a multiple of the logical "
583 "block size (logical_block_size %"PRIu32
")",
584 ns
->params
.zrwas
, ns
->blkconf
.logical_block_size
);
588 if (ns
->params
.zrwafg
== -1) {
589 ns
->params
.zrwafg
= ns
->blkconf
.logical_block_size
;
592 if (ns
->params
.zrwas
% ns
->params
.zrwafg
) {
593 error_setg(errp
, "zone random write area size (zoned.zrwas "
594 "%"PRIu64
") must be a multiple of the zone random "
595 "write area flush granularity (zoned.zrwafg, "
596 "%"PRIu64
")", ns
->params
.zrwas
, ns
->params
.zrwafg
);
600 if (ns
->params
.max_active_zones
) {
601 if (ns
->params
.numzrwa
> ns
->params
.max_active_zones
) {
602 error_setg(errp
, "number of zone random write area "
603 "resources (zoned.numzrwa, %d) must be less "
604 "than or equal to maximum active resources "
605 "(zoned.max_active_zones, %d)",
607 ns
->params
.max_active_zones
);
617 int nvme_ns_setup(NvmeNamespace
*ns
, Error
**errp
)
619 if (nvme_ns_check_constraints(ns
, errp
)) {
623 if (nvme_ns_init_blk(ns
, errp
)) {
627 if (nvme_ns_init(ns
, errp
)) {
630 if (ns
->params
.zoned
) {
631 if (nvme_ns_zoned_check_calc_geometry(ns
, errp
) != 0) {
634 nvme_ns_init_zoned(ns
);
637 if (ns
->endgrp
&& ns
->endgrp
->fdp
.enabled
) {
638 if (!nvme_ns_init_fdp(ns
, errp
)) {
646 void nvme_ns_drain(NvmeNamespace
*ns
)
648 blk_drain(ns
->blkconf
.blk
);
651 void nvme_ns_shutdown(NvmeNamespace
*ns
)
653 blk_flush(ns
->blkconf
.blk
);
654 if (ns
->params
.zoned
) {
655 nvme_zoned_ns_shutdown(ns
);
659 void nvme_ns_cleanup(NvmeNamespace
*ns
)
661 if (ns
->params
.zoned
) {
662 g_free(ns
->id_ns_zoned
);
663 g_free(ns
->zone_array
);
664 g_free(ns
->zd_extensions
);
667 if (ns
->endgrp
&& ns
->endgrp
->fdp
.enabled
) {
672 static void nvme_ns_unrealize(DeviceState
*dev
)
674 NvmeNamespace
*ns
= NVME_NS(dev
);
677 nvme_ns_shutdown(ns
);
681 static void nvme_ns_realize(DeviceState
*dev
, Error
**errp
)
683 NvmeNamespace
*ns
= NVME_NS(dev
);
684 BusState
*s
= qdev_get_parent_bus(dev
);
685 NvmeCtrl
*n
= NVME(s
->parent
);
686 NvmeSubsystem
*subsys
= n
->subsys
;
687 uint32_t nsid
= ns
->params
.nsid
;
691 /* If no subsys, the ns cannot be attached to more than one ctrl. */
692 ns
->params
.shared
= false;
693 if (ns
->params
.detached
) {
694 error_setg(errp
, "detached requires that the nvme device is "
695 "linked to an nvme-subsys device");
700 * If this namespace belongs to a subsystem (through a link on the
701 * controller device), reparent the device.
703 if (!qdev_set_parent_bus(dev
, &subsys
->bus
.parent_bus
, errp
)) {
707 ns
->endgrp
= &subsys
->endgrp
;
710 if (nvme_ns_setup(ns
, errp
)) {
715 for (i
= 1; i
<= NVME_MAX_NAMESPACES
; i
++) {
716 if (nvme_ns(n
, i
) || nvme_subsys_ns(subsys
, i
)) {
720 nsid
= ns
->params
.nsid
= i
;
725 error_setg(errp
, "no free namespace id");
729 if (nvme_ns(n
, nsid
) || nvme_subsys_ns(subsys
, nsid
)) {
730 error_setg(errp
, "namespace id '%d' already allocated", nsid
);
736 subsys
->namespaces
[nsid
] = ns
;
738 ns
->id_ns
.endgid
= cpu_to_le16(0x1);
740 if (ns
->params
.detached
) {
744 if (ns
->params
.shared
) {
745 for (i
= 0; i
< ARRAY_SIZE(subsys
->ctrls
); i
++) {
746 NvmeCtrl
*ctrl
= subsys
->ctrls
[i
];
748 if (ctrl
&& ctrl
!= SUBSYS_SLOT_RSVD
) {
749 nvme_attach_ns(ctrl
, ns
);
758 nvme_attach_ns(n
, ns
);
761 static Property nvme_ns_props
[] = {
762 DEFINE_BLOCK_PROPERTIES(NvmeNamespace
, blkconf
),
763 DEFINE_PROP_BOOL("detached", NvmeNamespace
, params
.detached
, false),
764 DEFINE_PROP_BOOL("shared", NvmeNamespace
, params
.shared
, true),
765 DEFINE_PROP_UINT32("nsid", NvmeNamespace
, params
.nsid
, 0),
766 DEFINE_PROP_UUID_NODEFAULT("uuid", NvmeNamespace
, params
.uuid
),
767 DEFINE_PROP_UINT64("eui64", NvmeNamespace
, params
.eui64
, 0),
768 DEFINE_PROP_UINT16("ms", NvmeNamespace
, params
.ms
, 0),
769 DEFINE_PROP_UINT8("mset", NvmeNamespace
, params
.mset
, 0),
770 DEFINE_PROP_UINT8("pi", NvmeNamespace
, params
.pi
, 0),
771 DEFINE_PROP_UINT8("pil", NvmeNamespace
, params
.pil
, 0),
772 DEFINE_PROP_UINT8("pif", NvmeNamespace
, params
.pif
, 0),
773 DEFINE_PROP_UINT16("mssrl", NvmeNamespace
, params
.mssrl
, 128),
774 DEFINE_PROP_UINT32("mcl", NvmeNamespace
, params
.mcl
, 128),
775 DEFINE_PROP_UINT8("msrc", NvmeNamespace
, params
.msrc
, 127),
776 DEFINE_PROP_BOOL("zoned", NvmeNamespace
, params
.zoned
, false),
777 DEFINE_PROP_SIZE("zoned.zone_size", NvmeNamespace
, params
.zone_size_bs
,
778 NVME_DEFAULT_ZONE_SIZE
),
779 DEFINE_PROP_SIZE("zoned.zone_capacity", NvmeNamespace
, params
.zone_cap_bs
,
781 DEFINE_PROP_BOOL("zoned.cross_read", NvmeNamespace
,
782 params
.cross_zone_read
, false),
783 DEFINE_PROP_UINT32("zoned.max_active", NvmeNamespace
,
784 params
.max_active_zones
, 0),
785 DEFINE_PROP_UINT32("zoned.max_open", NvmeNamespace
,
786 params
.max_open_zones
, 0),
787 DEFINE_PROP_UINT32("zoned.descr_ext_size", NvmeNamespace
,
788 params
.zd_extension_size
, 0),
789 DEFINE_PROP_UINT32("zoned.numzrwa", NvmeNamespace
, params
.numzrwa
, 0),
790 DEFINE_PROP_SIZE("zoned.zrwas", NvmeNamespace
, params
.zrwas
, 0),
791 DEFINE_PROP_SIZE("zoned.zrwafg", NvmeNamespace
, params
.zrwafg
, -1),
792 DEFINE_PROP_BOOL("eui64-default", NvmeNamespace
, params
.eui64_default
,
794 DEFINE_PROP_STRING("fdp.ruhs", NvmeNamespace
, params
.fdp
.ruhs
),
795 DEFINE_PROP_END_OF_LIST(),
798 static void nvme_ns_class_init(ObjectClass
*oc
, void *data
)
800 DeviceClass
*dc
= DEVICE_CLASS(oc
);
802 set_bit(DEVICE_CATEGORY_STORAGE
, dc
->categories
);
804 dc
->bus_type
= TYPE_NVME_BUS
;
805 dc
->realize
= nvme_ns_realize
;
806 dc
->unrealize
= nvme_ns_unrealize
;
807 device_class_set_props(dc
, nvme_ns_props
);
808 dc
->desc
= "Virtual NVMe namespace";
811 static void nvme_ns_instance_init(Object
*obj
)
813 NvmeNamespace
*ns
= NVME_NS(obj
);
814 char *bootindex
= g_strdup_printf("/namespace@%d,0", ns
->params
.nsid
);
816 device_add_bootindex_property(obj
, &ns
->bootindex
, "bootindex",
817 bootindex
, DEVICE(obj
));
822 static const TypeInfo nvme_ns_info
= {
823 .name
= TYPE_NVME_NS
,
824 .parent
= TYPE_DEVICE
,
825 .class_init
= nvme_ns_class_init
,
826 .instance_size
= sizeof(NvmeNamespace
),
827 .instance_init
= nvme_ns_instance_init
,
830 static void nvme_ns_register_types(void)
832 type_register_static(&nvme_ns_info
);
835 type_init(nvme_ns_register_types
)