1 // SPDX-License-Identifier: GPL-2.0
3 * Copyright (C) 2016 CNEX Labs
4 * Initial release: Javier Gonzalez <javier@cnexlabs.com>
5 * Matias Bjorling <matias@cnexlabs.com>
7 * This program is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU General Public License version
9 * 2 as published by the Free Software Foundation.
11 * This program is distributed in the hope that it will be useful, but
12 * WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * General Public License for more details.
16 * pblk-map.c - pblk's lba-ppa mapping strategy
22 static int pblk_map_page_data(struct pblk
*pblk
, unsigned int sentry
,
23 struct ppa_addr
*ppa_list
,
24 unsigned long *lun_bitmap
,
26 unsigned int valid_secs
)
28 struct pblk_line
*line
= pblk_line_get_data(pblk
);
29 struct pblk_emeta
*emeta
;
30 struct pblk_w_ctx
*w_ctx
;
33 int nr_secs
= pblk
->min_write_pgs
;
39 if (pblk_line_is_full(line
)) {
40 struct pblk_line
*prev_line
= line
;
42 /* If we cannot allocate a new line, make sure to store metadata
43 * on current line and then fail
45 line
= pblk_line_replace_data(pblk
);
46 pblk_line_close_meta(pblk
, prev_line
);
49 pblk_pipeline_stop(pblk
);
56 lba_list
= emeta_to_lbas(pblk
, emeta
->buf
);
58 paddr
= pblk_alloc_page(pblk
, line
, nr_secs
);
60 for (i
= 0; i
< nr_secs
; i
++, paddr
++) {
61 struct pblk_sec_meta
*meta
= pblk_get_meta(pblk
, meta_list
, i
);
62 __le64 addr_empty
= cpu_to_le64(ADDR_EMPTY
);
64 /* ppa to be sent to the device */
65 ppa_list
[i
] = addr_to_gen_ppa(pblk
, paddr
, line
->id
);
67 /* Write context for target bio completion on write buffer. Note
68 * that the write buffer is protected by the sync backpointer,
69 * and a single writer thread have access to each specific entry
70 * at a time. Thus, it is safe to modify the context for the
71 * entry we are setting up for submission without taking any
72 * lock or memory barrier.
76 atomic_inc(&line
->sec_to_update
);
77 w_ctx
= pblk_rb_w_ctx(&pblk
->rwb
, sentry
+ i
);
78 w_ctx
->ppa
= ppa_list
[i
];
79 meta
->lba
= cpu_to_le64(w_ctx
->lba
);
80 lba_list
[paddr
] = cpu_to_le64(w_ctx
->lba
);
81 if (lba_list
[paddr
] != addr_empty
)
82 line
->nr_valid_lbas
++;
84 atomic64_inc(&pblk
->pad_wa
);
86 lba_list
[paddr
] = addr_empty
;
87 meta
->lba
= addr_empty
;
88 __pblk_map_invalidate(pblk
, line
, paddr
);
92 pblk_down_rq(pblk
, ppa_list
[0], lun_bitmap
);
96 int pblk_map_rq(struct pblk
*pblk
, struct nvm_rq
*rqd
, unsigned int sentry
,
97 unsigned long *lun_bitmap
, unsigned int valid_secs
,
100 void *meta_list
= pblk_get_meta_for_writes(pblk
, rqd
);
102 struct ppa_addr
*ppa_list
= nvm_rq_to_ppa_list(rqd
);
103 unsigned int map_secs
;
104 int min
= pblk
->min_write_pgs
;
108 for (i
= off
; i
< rqd
->nr_ppas
; i
+= min
) {
109 map_secs
= (i
+ min
> valid_secs
) ? (valid_secs
% min
) : min
;
110 meta_buffer
= pblk_get_meta(pblk
, meta_list
, i
);
112 ret
= pblk_map_page_data(pblk
, sentry
+ i
, &ppa_list
[i
],
113 lun_bitmap
, meta_buffer
, map_secs
);
121 /* only if erase_ppa is set, acquire erase semaphore */
122 int pblk_map_erase_rq(struct pblk
*pblk
, struct nvm_rq
*rqd
,
123 unsigned int sentry
, unsigned long *lun_bitmap
,
124 unsigned int valid_secs
, struct ppa_addr
*erase_ppa
)
126 struct nvm_tgt_dev
*dev
= pblk
->dev
;
127 struct nvm_geo
*geo
= &dev
->geo
;
128 struct pblk_line_meta
*lm
= &pblk
->lm
;
129 void *meta_list
= pblk_get_meta_for_writes(pblk
, rqd
);
131 struct ppa_addr
*ppa_list
= nvm_rq_to_ppa_list(rqd
);
132 struct pblk_line
*e_line
, *d_line
;
133 unsigned int map_secs
;
134 int min
= pblk
->min_write_pgs
;
139 for (i
= 0; i
< rqd
->nr_ppas
; i
+= min
) {
140 map_secs
= (i
+ min
> valid_secs
) ? (valid_secs
% min
) : min
;
141 meta_buffer
= pblk_get_meta(pblk
, meta_list
, i
);
143 ret
= pblk_map_page_data(pblk
, sentry
+ i
, &ppa_list
[i
],
144 lun_bitmap
, meta_buffer
, map_secs
);
148 erase_lun
= pblk_ppa_to_pos(geo
, ppa_list
[i
]);
150 /* line can change after page map. We might also be writing the
153 e_line
= pblk_line_get_erase(pblk
);
155 return pblk_map_rq(pblk
, rqd
, sentry
, lun_bitmap
,
156 valid_secs
, i
+ min
);
158 spin_lock(&e_line
->lock
);
159 if (!test_bit(erase_lun
, e_line
->erase_bitmap
)) {
160 set_bit(erase_lun
, e_line
->erase_bitmap
);
161 atomic_dec(&e_line
->left_eblks
);
163 *erase_ppa
= ppa_list
[i
];
164 erase_ppa
->a
.blk
= e_line
->id
;
165 erase_ppa
->a
.reserved
= 0;
167 spin_unlock(&e_line
->lock
);
169 /* Avoid evaluating e_line->left_eblks */
170 return pblk_map_rq(pblk
, rqd
, sentry
, lun_bitmap
,
171 valid_secs
, i
+ min
);
173 spin_unlock(&e_line
->lock
);
176 d_line
= pblk_line_get_data(pblk
);
178 /* line can change after page map. We might also be writing the
181 e_line
= pblk_line_get_erase(pblk
);
185 /* Erase blocks that are bad in this line but might not be in next */
186 if (unlikely(pblk_ppa_empty(*erase_ppa
)) &&
187 bitmap_weight(d_line
->blk_bitmap
, lm
->blk_per_line
)) {
191 bit
= find_next_bit(d_line
->blk_bitmap
,
192 lm
->blk_per_line
, bit
+ 1);
193 if (bit
>= lm
->blk_per_line
)
196 spin_lock(&e_line
->lock
);
197 if (test_bit(bit
, e_line
->erase_bitmap
)) {
198 spin_unlock(&e_line
->lock
);
201 spin_unlock(&e_line
->lock
);
203 set_bit(bit
, e_line
->erase_bitmap
);
204 atomic_dec(&e_line
->left_eblks
);
205 *erase_ppa
= pblk
->luns
[bit
].bppa
; /* set ch and lun */
206 erase_ppa
->a
.blk
= e_line
->id
;