1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright (C) 2019 Arrikto, Inc. All Rights Reserved.
8 #include <linux/slab.h>
9 #include <linux/rwsem.h>
10 #include <linux/bitops.h>
11 #include <linux/bitmap.h>
12 #include <linux/device-mapper.h>
14 #include "persistent-data/dm-bitset.h"
15 #include "persistent-data/dm-space-map.h"
16 #include "persistent-data/dm-block-manager.h"
17 #include "persistent-data/dm-transaction-manager.h"
19 #include "dm-clone-metadata.h"
21 #define DM_MSG_PREFIX "clone metadata"
23 #define SUPERBLOCK_LOCATION 0
24 #define SUPERBLOCK_MAGIC 0x8af27f64
25 #define SUPERBLOCK_CSUM_XOR 257649492
27 #define DM_CLONE_MAX_CONCURRENT_LOCKS 5
31 /* Min and max dm-clone metadata versions supported */
32 #define DM_CLONE_MIN_METADATA_VERSION 1
33 #define DM_CLONE_MAX_METADATA_VERSION 1
36 * On-disk metadata layout
38 struct superblock_disk
{
47 __u8 metadata_space_map_root
[SPACE_MAP_ROOT_SIZE
];
56 * Region and Dirty bitmaps.
58 * dm-clone logically splits the source and destination devices in regions of
59 * fixed size. The destination device's regions are gradually hydrated, i.e.,
60 * we copy (clone) the source's regions to the destination device. Eventually,
61 * all regions will get hydrated and all I/O will be served from the
64 * We maintain an on-disk bitmap which tracks the state of each of the
65 * destination device's regions, i.e., whether they are hydrated or not.
67 * To save constantly doing look ups on disk we keep an in core copy of the
68 * on-disk bitmap, the region_map.
70 * In order to track which regions are hydrated during a metadata transaction,
71 * we use a second set of bitmaps, the dmap (dirty bitmap), which includes two
72 * bitmaps, namely dirty_regions and dirty_words. The dirty_regions bitmap
73 * tracks the regions that got hydrated during the current metadata
74 * transaction. The dirty_words bitmap tracks the dirty words, i.e. longs, of
75 * the dirty_regions bitmap.
77 * This allows us to precisely track the regions that were hydrated during the
78 * current metadata transaction and update the metadata accordingly, when we
79 * commit the current transaction. This is important because dm-clone should
80 * only commit the metadata of regions that were properly flushed to the
81 * destination device beforehand. Otherwise, in case of a crash, we could end
82 * up with a corrupted dm-clone device.
84 * When a region finishes hydrating dm-clone calls
85 * dm_clone_set_region_hydrated(), or for discard requests
86 * dm_clone_cond_set_range(), which sets the corresponding bits in region_map
89 * During a metadata commit we scan dmap->dirty_words and dmap->dirty_regions
90 * and update the on-disk metadata accordingly. Thus, we don't have to flush to
91 * disk the whole region_map. We can just flush the dirty region_map bits.
93 * We use the helper dmap->dirty_words bitmap, which is smaller than the
94 * original region_map, to reduce the amount of memory accesses during a
95 * metadata commit. Moreover, as dm-bitset also accesses the on-disk bitmap in
96 * 64-bit word granularity, the dirty_words bitmap helps us avoid useless disk
99 * We could update directly the on-disk bitmap, when dm-clone calls either
100 * dm_clone_set_region_hydrated() or dm_clone_cond_set_range(), buts this
101 * inserts significant metadata I/O overhead in dm-clone's I/O path. Also, as
102 * these two functions don't block, we can call them in interrupt context,
103 * e.g., in a hooked overwrite bio's completion routine, and further reduce the
104 * I/O completion latency.
106 * We maintain two dirty bitmap sets. During a metadata commit we atomically
107 * swap the currently used dmap with the unused one. This allows the metadata
108 * update functions to run concurrently with an ongoing commit.
111 unsigned long *dirty_words
;
112 unsigned long *dirty_regions
;
113 unsigned int changed
;
116 struct dm_clone_metadata
{
117 /* The metadata block device */
118 struct block_device
*bdev
;
120 sector_t target_size
;
121 sector_t region_size
;
122 unsigned long nr_regions
;
123 unsigned long nr_words
;
125 /* Spinlock protecting the region and dirty bitmaps. */
126 spinlock_t bitmap_lock
;
127 struct dirty_map dmap
[2];
128 struct dirty_map
*current_dmap
;
130 /* Protected by lock */
131 struct dirty_map
*committing_dmap
;
134 * In core copy of the on-disk bitmap to save constantly doing look ups
137 unsigned long *region_map
;
139 /* Protected by bitmap_lock */
140 unsigned int read_only
;
142 struct dm_block_manager
*bm
;
143 struct dm_space_map
*sm
;
144 struct dm_transaction_manager
*tm
;
146 struct rw_semaphore lock
;
148 struct dm_disk_bitset bitset_info
;
149 dm_block_t bitset_root
;
152 * Reading the space map root can fail, so we read it into this
153 * buffer before the superblock is locked and updated.
155 __u8 metadata_space_map_root
[SPACE_MAP_ROOT_SIZE
];
157 bool hydration_done
:1;
161 /*---------------------------------------------------------------------------*/
164 * Superblock validation.
166 static void sb_prepare_for_write(const struct dm_block_validator
*v
,
167 struct dm_block
*b
, size_t sb_block_size
)
169 struct superblock_disk
*sb
;
172 sb
= dm_block_data(b
);
173 sb
->blocknr
= cpu_to_le64(dm_block_location(b
));
175 csum
= dm_bm_checksum(&sb
->flags
, sb_block_size
- sizeof(__le32
),
176 SUPERBLOCK_CSUM_XOR
);
177 sb
->csum
= cpu_to_le32(csum
);
180 static int sb_check(const struct dm_block_validator
*v
, struct dm_block
*b
,
181 size_t sb_block_size
)
183 struct superblock_disk
*sb
;
184 u32 csum
, metadata_version
;
186 sb
= dm_block_data(b
);
188 if (dm_block_location(b
) != le64_to_cpu(sb
->blocknr
)) {
189 DMERR("Superblock check failed: blocknr %llu, expected %llu",
190 le64_to_cpu(sb
->blocknr
),
191 (unsigned long long)dm_block_location(b
));
195 if (le64_to_cpu(sb
->magic
) != SUPERBLOCK_MAGIC
) {
196 DMERR("Superblock check failed: magic %llu, expected %llu",
197 le64_to_cpu(sb
->magic
),
198 (unsigned long long)SUPERBLOCK_MAGIC
);
202 csum
= dm_bm_checksum(&sb
->flags
, sb_block_size
- sizeof(__le32
),
203 SUPERBLOCK_CSUM_XOR
);
204 if (sb
->csum
!= cpu_to_le32(csum
)) {
205 DMERR("Superblock check failed: checksum %u, expected %u",
206 csum
, le32_to_cpu(sb
->csum
));
210 /* Check metadata version */
211 metadata_version
= le32_to_cpu(sb
->version
);
212 if (metadata_version
< DM_CLONE_MIN_METADATA_VERSION
||
213 metadata_version
> DM_CLONE_MAX_METADATA_VERSION
) {
214 DMERR("Clone metadata version %u found, but only versions between %u and %u supported.",
215 metadata_version
, DM_CLONE_MIN_METADATA_VERSION
,
216 DM_CLONE_MAX_METADATA_VERSION
);
223 static const struct dm_block_validator sb_validator
= {
224 .name
= "superblock",
225 .prepare_for_write
= sb_prepare_for_write
,
230 * Check if the superblock is formatted or not. We consider the superblock to
231 * be formatted in case we find non-zero bytes in it.
233 static int __superblock_all_zeroes(struct dm_block_manager
*bm
, bool *formatted
)
236 unsigned int i
, nr_words
;
237 struct dm_block
*sblock
;
238 __le64
*data_le
, zero
= cpu_to_le64(0);
241 * We don't use a validator here because the superblock could be all
244 r
= dm_bm_read_lock(bm
, SUPERBLOCK_LOCATION
, NULL
, &sblock
);
246 DMERR("Failed to read_lock superblock");
250 data_le
= dm_block_data(sblock
);
253 /* This assumes that the block size is a multiple of 8 bytes */
254 BUG_ON(dm_bm_block_size(bm
) % sizeof(__le64
));
255 nr_words
= dm_bm_block_size(bm
) / sizeof(__le64
);
256 for (i
= 0; i
< nr_words
; i
++) {
257 if (data_le
[i
] != zero
) {
263 dm_bm_unlock(sblock
);
268 /*---------------------------------------------------------------------------*/
271 * Low-level metadata handling.
273 static inline int superblock_read_lock(struct dm_clone_metadata
*cmd
,
274 struct dm_block
**sblock
)
276 return dm_bm_read_lock(cmd
->bm
, SUPERBLOCK_LOCATION
, &sb_validator
, sblock
);
279 static inline int superblock_write_lock_zero(struct dm_clone_metadata
*cmd
,
280 struct dm_block
**sblock
)
282 return dm_bm_write_lock_zero(cmd
->bm
, SUPERBLOCK_LOCATION
, &sb_validator
, sblock
);
285 static int __copy_sm_root(struct dm_clone_metadata
*cmd
)
290 r
= dm_sm_root_size(cmd
->sm
, &root_size
);
294 return dm_sm_copy_root(cmd
->sm
, &cmd
->metadata_space_map_root
, root_size
);
297 /* Save dm-clone metadata in superblock */
298 static void __prepare_superblock(struct dm_clone_metadata
*cmd
,
299 struct superblock_disk
*sb
)
301 sb
->flags
= cpu_to_le32(0UL);
303 /* FIXME: UUID is currently unused */
304 memset(sb
->uuid
, 0, sizeof(sb
->uuid
));
306 sb
->magic
= cpu_to_le64(SUPERBLOCK_MAGIC
);
307 sb
->version
= cpu_to_le32(DM_CLONE_MAX_METADATA_VERSION
);
309 /* Save the metadata space_map root */
310 memcpy(&sb
->metadata_space_map_root
, &cmd
->metadata_space_map_root
,
311 sizeof(cmd
->metadata_space_map_root
));
313 sb
->region_size
= cpu_to_le64(cmd
->region_size
);
314 sb
->target_size
= cpu_to_le64(cmd
->target_size
);
315 sb
->bitset_root
= cpu_to_le64(cmd
->bitset_root
);
318 static int __open_metadata(struct dm_clone_metadata
*cmd
)
321 struct dm_block
*sblock
;
322 struct superblock_disk
*sb
;
324 r
= superblock_read_lock(cmd
, &sblock
);
327 DMERR("Failed to read_lock superblock");
331 sb
= dm_block_data(sblock
);
333 /* Verify that target_size and region_size haven't changed. */
334 if (cmd
->region_size
!= le64_to_cpu(sb
->region_size
) ||
335 cmd
->target_size
!= le64_to_cpu(sb
->target_size
)) {
336 DMERR("Region and/or target size don't match the ones in metadata");
341 r
= dm_tm_open_with_sm(cmd
->bm
, SUPERBLOCK_LOCATION
,
342 sb
->metadata_space_map_root
,
343 sizeof(sb
->metadata_space_map_root
),
347 DMERR("dm_tm_open_with_sm failed");
351 dm_disk_bitset_init(cmd
->tm
, &cmd
->bitset_info
);
352 cmd
->bitset_root
= le64_to_cpu(sb
->bitset_root
);
355 dm_bm_unlock(sblock
);
360 static int __format_metadata(struct dm_clone_metadata
*cmd
)
363 struct dm_block
*sblock
;
364 struct superblock_disk
*sb
;
366 r
= dm_tm_create_with_sm(cmd
->bm
, SUPERBLOCK_LOCATION
, &cmd
->tm
, &cmd
->sm
);
368 DMERR("Failed to create transaction manager");
372 dm_disk_bitset_init(cmd
->tm
, &cmd
->bitset_info
);
374 r
= dm_bitset_empty(&cmd
->bitset_info
, &cmd
->bitset_root
);
376 DMERR("Failed to create empty on-disk bitset");
380 r
= dm_bitset_resize(&cmd
->bitset_info
, cmd
->bitset_root
, 0,
381 cmd
->nr_regions
, false, &cmd
->bitset_root
);
383 DMERR("Failed to resize on-disk bitset to %lu entries", cmd
->nr_regions
);
387 /* Flush to disk all blocks, except the superblock */
388 r
= dm_tm_pre_commit(cmd
->tm
);
390 DMERR("dm_tm_pre_commit failed");
394 r
= __copy_sm_root(cmd
);
396 DMERR("__copy_sm_root failed");
400 r
= superblock_write_lock_zero(cmd
, &sblock
);
402 DMERR("Failed to write_lock superblock");
406 sb
= dm_block_data(sblock
);
407 __prepare_superblock(cmd
, sb
);
408 r
= dm_tm_commit(cmd
->tm
, sblock
);
410 DMERR("Failed to commit superblock");
417 dm_sm_destroy(cmd
->sm
);
418 dm_tm_destroy(cmd
->tm
);
423 static int __open_or_format_metadata(struct dm_clone_metadata
*cmd
, bool may_format_device
)
426 bool formatted
= false;
428 r
= __superblock_all_zeroes(cmd
->bm
, &formatted
);
433 return may_format_device
? __format_metadata(cmd
) : -EPERM
;
435 return __open_metadata(cmd
);
438 static int __create_persistent_data_structures(struct dm_clone_metadata
*cmd
,
439 bool may_format_device
)
443 /* Create block manager */
444 cmd
->bm
= dm_block_manager_create(cmd
->bdev
,
445 DM_CLONE_METADATA_BLOCK_SIZE
<< SECTOR_SHIFT
,
446 DM_CLONE_MAX_CONCURRENT_LOCKS
);
447 if (IS_ERR(cmd
->bm
)) {
448 DMERR("Failed to create block manager");
449 return PTR_ERR(cmd
->bm
);
452 r
= __open_or_format_metadata(cmd
, may_format_device
);
454 dm_block_manager_destroy(cmd
->bm
);
459 static void __destroy_persistent_data_structures(struct dm_clone_metadata
*cmd
)
461 dm_sm_destroy(cmd
->sm
);
462 dm_tm_destroy(cmd
->tm
);
463 dm_block_manager_destroy(cmd
->bm
);
466 /*---------------------------------------------------------------------------*/
468 static int __dirty_map_init(struct dirty_map
*dmap
, unsigned long nr_words
,
469 unsigned long nr_regions
)
473 dmap
->dirty_words
= kvzalloc(bitmap_size(nr_words
), GFP_KERNEL
);
474 if (!dmap
->dirty_words
)
477 dmap
->dirty_regions
= kvzalloc(bitmap_size(nr_regions
), GFP_KERNEL
);
478 if (!dmap
->dirty_regions
) {
479 kvfree(dmap
->dirty_words
);
486 static void __dirty_map_exit(struct dirty_map
*dmap
)
488 kvfree(dmap
->dirty_words
);
489 kvfree(dmap
->dirty_regions
);
492 static int dirty_map_init(struct dm_clone_metadata
*cmd
)
494 if (__dirty_map_init(&cmd
->dmap
[0], cmd
->nr_words
, cmd
->nr_regions
)) {
495 DMERR("Failed to allocate dirty bitmap");
499 if (__dirty_map_init(&cmd
->dmap
[1], cmd
->nr_words
, cmd
->nr_regions
)) {
500 DMERR("Failed to allocate dirty bitmap");
501 __dirty_map_exit(&cmd
->dmap
[0]);
505 cmd
->current_dmap
= &cmd
->dmap
[0];
506 cmd
->committing_dmap
= NULL
;
511 static void dirty_map_exit(struct dm_clone_metadata
*cmd
)
513 __dirty_map_exit(&cmd
->dmap
[0]);
514 __dirty_map_exit(&cmd
->dmap
[1]);
517 static int __load_bitset_in_core(struct dm_clone_metadata
*cmd
)
521 struct dm_bitset_cursor c
;
523 /* Flush bitset cache */
524 r
= dm_bitset_flush(&cmd
->bitset_info
, cmd
->bitset_root
, &cmd
->bitset_root
);
528 r
= dm_bitset_cursor_begin(&cmd
->bitset_info
, cmd
->bitset_root
, cmd
->nr_regions
, &c
);
533 __assign_bit(i
, cmd
->region_map
, dm_bitset_cursor_get_value(&c
));
535 if (i
>= (cmd
->nr_regions
- 1))
538 r
= dm_bitset_cursor_next(&c
);
544 dm_bitset_cursor_end(&c
);
549 struct dm_clone_metadata
*dm_clone_metadata_open(struct block_device
*bdev
,
550 sector_t target_size
,
551 sector_t region_size
)
554 struct dm_clone_metadata
*cmd
;
556 cmd
= kzalloc(sizeof(*cmd
), GFP_KERNEL
);
558 DMERR("Failed to allocate memory for dm-clone metadata");
559 return ERR_PTR(-ENOMEM
);
563 cmd
->target_size
= target_size
;
564 cmd
->region_size
= region_size
;
565 cmd
->nr_regions
= dm_sector_div_up(cmd
->target_size
, cmd
->region_size
);
566 cmd
->nr_words
= BITS_TO_LONGS(cmd
->nr_regions
);
568 init_rwsem(&cmd
->lock
);
569 spin_lock_init(&cmd
->bitmap_lock
);
571 cmd
->fail_io
= false;
572 cmd
->hydration_done
= false;
574 cmd
->region_map
= kvmalloc(bitmap_size(cmd
->nr_regions
), GFP_KERNEL
);
575 if (!cmd
->region_map
) {
576 DMERR("Failed to allocate memory for region bitmap");
581 r
= __create_persistent_data_structures(cmd
, true);
583 goto out_with_region_map
;
585 r
= __load_bitset_in_core(cmd
);
587 DMERR("Failed to load on-disk region map");
591 r
= dirty_map_init(cmd
);
595 if (bitmap_full(cmd
->region_map
, cmd
->nr_regions
))
596 cmd
->hydration_done
= true;
601 __destroy_persistent_data_structures(cmd
);
604 kvfree(cmd
->region_map
);
612 void dm_clone_metadata_close(struct dm_clone_metadata
*cmd
)
615 __destroy_persistent_data_structures(cmd
);
618 kvfree(cmd
->region_map
);
622 bool dm_clone_is_hydration_done(struct dm_clone_metadata
*cmd
)
624 return cmd
->hydration_done
;
627 bool dm_clone_is_region_hydrated(struct dm_clone_metadata
*cmd
, unsigned long region_nr
)
629 return dm_clone_is_hydration_done(cmd
) || test_bit(region_nr
, cmd
->region_map
);
632 bool dm_clone_is_range_hydrated(struct dm_clone_metadata
*cmd
,
633 unsigned long start
, unsigned long nr_regions
)
637 if (dm_clone_is_hydration_done(cmd
))
640 bit
= find_next_zero_bit(cmd
->region_map
, cmd
->nr_regions
, start
);
642 return (bit
>= (start
+ nr_regions
));
645 unsigned int dm_clone_nr_of_hydrated_regions(struct dm_clone_metadata
*cmd
)
647 return bitmap_weight(cmd
->region_map
, cmd
->nr_regions
);
650 unsigned long dm_clone_find_next_unhydrated_region(struct dm_clone_metadata
*cmd
,
653 return find_next_zero_bit(cmd
->region_map
, cmd
->nr_regions
, start
);
656 static int __update_metadata_word(struct dm_clone_metadata
*cmd
,
657 unsigned long *dirty_regions
,
661 unsigned long index
= word
* BITS_PER_LONG
;
662 unsigned long max_index
= min(cmd
->nr_regions
, (word
+ 1) * BITS_PER_LONG
);
664 while (index
< max_index
) {
665 if (test_bit(index
, dirty_regions
)) {
666 r
= dm_bitset_set_bit(&cmd
->bitset_info
, cmd
->bitset_root
,
667 index
, &cmd
->bitset_root
);
669 DMERR("dm_bitset_set_bit failed");
672 __clear_bit(index
, dirty_regions
);
680 static int __metadata_commit(struct dm_clone_metadata
*cmd
)
683 struct dm_block
*sblock
;
684 struct superblock_disk
*sb
;
686 /* Flush bitset cache */
687 r
= dm_bitset_flush(&cmd
->bitset_info
, cmd
->bitset_root
, &cmd
->bitset_root
);
689 DMERR("dm_bitset_flush failed");
693 /* Flush to disk all blocks, except the superblock */
694 r
= dm_tm_pre_commit(cmd
->tm
);
696 DMERR("dm_tm_pre_commit failed");
700 /* Save the space map root in cmd->metadata_space_map_root */
701 r
= __copy_sm_root(cmd
);
703 DMERR("__copy_sm_root failed");
707 /* Lock the superblock */
708 r
= superblock_write_lock_zero(cmd
, &sblock
);
710 DMERR("Failed to write_lock superblock");
714 /* Save the metadata in superblock */
715 sb
= dm_block_data(sblock
);
716 __prepare_superblock(cmd
, sb
);
718 /* Unlock superblock and commit it to disk */
719 r
= dm_tm_commit(cmd
->tm
, sblock
);
721 DMERR("Failed to commit superblock");
726 * FIXME: Find a more efficient way to check if the hydration is done.
728 if (bitmap_full(cmd
->region_map
, cmd
->nr_regions
))
729 cmd
->hydration_done
= true;
734 static int __flush_dmap(struct dm_clone_metadata
*cmd
, struct dirty_map
*dmap
)
741 word
= find_next_bit(dmap
->dirty_words
, cmd
->nr_words
, word
);
743 if (word
== cmd
->nr_words
)
746 r
= __update_metadata_word(cmd
, dmap
->dirty_regions
, word
);
751 __clear_bit(word
, dmap
->dirty_words
);
753 } while (word
< cmd
->nr_words
);
755 r
= __metadata_commit(cmd
);
760 /* Update the changed flag */
761 spin_lock_irq(&cmd
->bitmap_lock
);
763 spin_unlock_irq(&cmd
->bitmap_lock
);
768 int dm_clone_metadata_pre_commit(struct dm_clone_metadata
*cmd
)
771 struct dirty_map
*dmap
, *next_dmap
;
773 down_write(&cmd
->lock
);
775 if (cmd
->fail_io
|| dm_bm_is_read_only(cmd
->bm
)) {
780 /* Get current dirty bitmap */
781 dmap
= cmd
->current_dmap
;
783 /* Get next dirty bitmap */
784 next_dmap
= (dmap
== &cmd
->dmap
[0]) ? &cmd
->dmap
[1] : &cmd
->dmap
[0];
787 * The last commit failed, so we don't have a clean dirty-bitmap to
790 if (WARN_ON(next_dmap
->changed
|| cmd
->committing_dmap
)) {
795 /* Swap dirty bitmaps */
796 spin_lock_irq(&cmd
->bitmap_lock
);
797 cmd
->current_dmap
= next_dmap
;
798 spin_unlock_irq(&cmd
->bitmap_lock
);
800 /* Set old dirty bitmap as currently committing */
801 cmd
->committing_dmap
= dmap
;
803 up_write(&cmd
->lock
);
808 int dm_clone_metadata_commit(struct dm_clone_metadata
*cmd
)
812 down_write(&cmd
->lock
);
814 if (cmd
->fail_io
|| dm_bm_is_read_only(cmd
->bm
))
817 if (WARN_ON(!cmd
->committing_dmap
)) {
822 r
= __flush_dmap(cmd
, cmd
->committing_dmap
);
824 /* Clear committing dmap */
825 cmd
->committing_dmap
= NULL
;
828 up_write(&cmd
->lock
);
833 int dm_clone_set_region_hydrated(struct dm_clone_metadata
*cmd
, unsigned long region_nr
)
836 struct dirty_map
*dmap
;
837 unsigned long word
, flags
;
839 if (unlikely(region_nr
>= cmd
->nr_regions
)) {
840 DMERR("Region %lu out of range (total number of regions %lu)",
841 region_nr
, cmd
->nr_regions
);
845 word
= region_nr
/ BITS_PER_LONG
;
847 spin_lock_irqsave(&cmd
->bitmap_lock
, flags
);
849 if (cmd
->read_only
) {
854 dmap
= cmd
->current_dmap
;
856 __set_bit(word
, dmap
->dirty_words
);
857 __set_bit(region_nr
, dmap
->dirty_regions
);
858 __set_bit(region_nr
, cmd
->region_map
);
862 spin_unlock_irqrestore(&cmd
->bitmap_lock
, flags
);
867 int dm_clone_cond_set_range(struct dm_clone_metadata
*cmd
, unsigned long start
,
868 unsigned long nr_regions
)
871 struct dirty_map
*dmap
;
872 unsigned long word
, region_nr
;
874 if (unlikely(start
>= cmd
->nr_regions
|| (start
+ nr_regions
) < start
||
875 (start
+ nr_regions
) > cmd
->nr_regions
)) {
876 DMERR("Invalid region range: start %lu, nr_regions %lu (total number of regions %lu)",
877 start
, nr_regions
, cmd
->nr_regions
);
881 spin_lock_irq(&cmd
->bitmap_lock
);
883 if (cmd
->read_only
) {
888 dmap
= cmd
->current_dmap
;
889 for (region_nr
= start
; region_nr
< (start
+ nr_regions
); region_nr
++) {
890 if (!test_bit(region_nr
, cmd
->region_map
)) {
891 word
= region_nr
/ BITS_PER_LONG
;
892 __set_bit(word
, dmap
->dirty_words
);
893 __set_bit(region_nr
, dmap
->dirty_regions
);
894 __set_bit(region_nr
, cmd
->region_map
);
899 spin_unlock_irq(&cmd
->bitmap_lock
);
905 * WARNING: This must not be called concurrently with either
906 * dm_clone_set_region_hydrated() or dm_clone_cond_set_range(), as it changes
907 * cmd->region_map without taking the cmd->bitmap_lock spinlock. The only
908 * exception is after setting the metadata to read-only mode, using
909 * dm_clone_metadata_set_read_only().
911 * We don't take the spinlock because __load_bitset_in_core() does I/O, so it
914 int dm_clone_reload_in_core_bitset(struct dm_clone_metadata
*cmd
)
918 down_write(&cmd
->lock
);
923 r
= __load_bitset_in_core(cmd
);
925 up_write(&cmd
->lock
);
930 bool dm_clone_changed_this_transaction(struct dm_clone_metadata
*cmd
)
935 spin_lock_irqsave(&cmd
->bitmap_lock
, flags
);
936 r
= cmd
->dmap
[0].changed
|| cmd
->dmap
[1].changed
;
937 spin_unlock_irqrestore(&cmd
->bitmap_lock
, flags
);
942 int dm_clone_metadata_abort(struct dm_clone_metadata
*cmd
)
946 down_write(&cmd
->lock
);
948 if (cmd
->fail_io
|| dm_bm_is_read_only(cmd
->bm
))
951 __destroy_persistent_data_structures(cmd
);
953 r
= __create_persistent_data_structures(cmd
, false);
955 /* If something went wrong we can neither write nor read the metadata */
959 up_write(&cmd
->lock
);
964 void dm_clone_metadata_set_read_only(struct dm_clone_metadata
*cmd
)
966 down_write(&cmd
->lock
);
968 spin_lock_irq(&cmd
->bitmap_lock
);
970 spin_unlock_irq(&cmd
->bitmap_lock
);
973 dm_bm_set_read_only(cmd
->bm
);
975 up_write(&cmd
->lock
);
978 void dm_clone_metadata_set_read_write(struct dm_clone_metadata
*cmd
)
980 down_write(&cmd
->lock
);
982 spin_lock_irq(&cmd
->bitmap_lock
);
984 spin_unlock_irq(&cmd
->bitmap_lock
);
987 dm_bm_set_read_write(cmd
->bm
);
989 up_write(&cmd
->lock
);
992 int dm_clone_get_free_metadata_block_count(struct dm_clone_metadata
*cmd
,
997 down_read(&cmd
->lock
);
1000 r
= dm_sm_get_nr_free(cmd
->sm
, result
);
1002 up_read(&cmd
->lock
);
1007 int dm_clone_get_metadata_dev_size(struct dm_clone_metadata
*cmd
,
1012 down_read(&cmd
->lock
);
1015 r
= dm_sm_get_nr_blocks(cmd
->sm
, result
);
1017 up_read(&cmd
->lock
);