printf: Remove unused 'bprintf'
[drm/drm-misc.git] / fs / squashfs / block.c
blob2dc730800f448d8cb44f2d5c4e625e607d1faf2f
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Squashfs - a compressed read only filesystem for Linux
5 * Copyright (c) 2002, 2003, 2004, 2005, 2006, 2007, 2008
6 * Phillip Lougher <phillip@squashfs.org.uk>
8 * block.c
9 */
12 * This file implements the low-level routines to read and decompress
13 * datablocks and metadata blocks.
16 #include <linux/blkdev.h>
17 #include <linux/fs.h>
18 #include <linux/vfs.h>
19 #include <linux/slab.h>
20 #include <linux/pagemap.h>
21 #include <linux/string.h>
22 #include <linux/bio.h>
24 #include "squashfs_fs.h"
25 #include "squashfs_fs_sb.h"
26 #include "squashfs.h"
27 #include "decompressor.h"
28 #include "page_actor.h"
31 * Returns the amount of bytes copied to the page actor.
33 static int copy_bio_to_actor(struct bio *bio,
34 struct squashfs_page_actor *actor,
35 int offset, int req_length)
37 void *actor_addr;
38 struct bvec_iter_all iter_all = {};
39 struct bio_vec *bvec = bvec_init_iter_all(&iter_all);
40 int copied_bytes = 0;
41 int actor_offset = 0;
43 squashfs_actor_nobuff(actor);
44 actor_addr = squashfs_first_page(actor);
46 if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all)))
47 return 0;
49 while (copied_bytes < req_length) {
50 int bytes_to_copy = min_t(int, bvec->bv_len - offset,
51 PAGE_SIZE - actor_offset);
53 bytes_to_copy = min_t(int, bytes_to_copy,
54 req_length - copied_bytes);
55 if (!IS_ERR(actor_addr))
56 memcpy(actor_addr + actor_offset, bvec_virt(bvec) +
57 offset, bytes_to_copy);
59 actor_offset += bytes_to_copy;
60 copied_bytes += bytes_to_copy;
61 offset += bytes_to_copy;
63 if (actor_offset >= PAGE_SIZE) {
64 actor_addr = squashfs_next_page(actor);
65 if (!actor_addr)
66 break;
67 actor_offset = 0;
69 if (offset >= bvec->bv_len) {
70 if (!bio_next_segment(bio, &iter_all))
71 break;
72 offset = 0;
75 squashfs_finish_page(actor);
76 return copied_bytes;
79 static int squashfs_bio_read_cached(struct bio *fullbio,
80 struct address_space *cache_mapping, u64 index, int length,
81 u64 read_start, u64 read_end, int page_count)
83 struct page *head_to_cache = NULL, *tail_to_cache = NULL;
84 struct block_device *bdev = fullbio->bi_bdev;
85 int start_idx = 0, end_idx = 0;
86 struct bvec_iter_all iter_all;
87 struct bio *bio = NULL;
88 struct bio_vec *bv;
89 int idx = 0;
90 int err = 0;
92 bio_for_each_segment_all(bv, fullbio, iter_all) {
93 struct page *page = bv->bv_page;
95 if (page->mapping == cache_mapping) {
96 idx++;
97 continue;
101 * We only use this when the device block size is the same as
102 * the page size, so read_start and read_end cover full pages.
104 * Compare these to the original required index and length to
105 * only cache pages which were requested partially, since these
106 * are the ones which are likely to be needed when reading
107 * adjacent blocks.
109 if (idx == 0 && index != read_start)
110 head_to_cache = page;
111 else if (idx == page_count - 1 && index + length != read_end)
112 tail_to_cache = page;
114 if (!bio || idx != end_idx) {
115 struct bio *new = bio_alloc_clone(bdev, fullbio,
116 GFP_NOIO, &fs_bio_set);
118 if (bio) {
119 bio_trim(bio, start_idx * PAGE_SECTORS,
120 (end_idx - start_idx) * PAGE_SECTORS);
121 bio_chain(bio, new);
122 submit_bio(bio);
125 bio = new;
126 start_idx = idx;
129 idx++;
130 end_idx = idx;
133 if (bio) {
134 bio_trim(bio, start_idx * PAGE_SECTORS,
135 (end_idx - start_idx) * PAGE_SECTORS);
136 err = submit_bio_wait(bio);
137 bio_put(bio);
140 if (err)
141 return err;
143 if (head_to_cache) {
144 int ret = add_to_page_cache_lru(head_to_cache, cache_mapping,
145 read_start >> PAGE_SHIFT,
146 GFP_NOIO);
148 if (!ret) {
149 SetPageUptodate(head_to_cache);
150 unlock_page(head_to_cache);
155 if (tail_to_cache) {
156 int ret = add_to_page_cache_lru(tail_to_cache, cache_mapping,
157 (read_end >> PAGE_SHIFT) - 1,
158 GFP_NOIO);
160 if (!ret) {
161 SetPageUptodate(tail_to_cache);
162 unlock_page(tail_to_cache);
166 return 0;
169 static struct page *squashfs_get_cache_page(struct address_space *mapping,
170 pgoff_t index)
172 struct page *page;
174 if (!mapping)
175 return NULL;
177 page = find_get_page(mapping, index);
178 if (!page)
179 return NULL;
181 if (!PageUptodate(page)) {
182 put_page(page);
183 return NULL;
186 return page;
189 static int squashfs_bio_read(struct super_block *sb, u64 index, int length,
190 struct bio **biop, int *block_offset)
192 struct squashfs_sb_info *msblk = sb->s_fs_info;
193 struct address_space *cache_mapping = msblk->cache_mapping;
194 const u64 read_start = round_down(index, msblk->devblksize);
195 const sector_t block = read_start >> msblk->devblksize_log2;
196 const u64 read_end = round_up(index + length, msblk->devblksize);
197 const sector_t block_end = read_end >> msblk->devblksize_log2;
198 int offset = read_start - round_down(index, PAGE_SIZE);
199 int total_len = (block_end - block) << msblk->devblksize_log2;
200 const int page_count = DIV_ROUND_UP(total_len + offset, PAGE_SIZE);
201 int error, i;
202 struct bio *bio;
204 bio = bio_kmalloc(page_count, GFP_NOIO);
205 if (!bio)
206 return -ENOMEM;
207 bio_init(bio, sb->s_bdev, bio->bi_inline_vecs, page_count, REQ_OP_READ);
208 bio->bi_iter.bi_sector = block * (msblk->devblksize >> SECTOR_SHIFT);
210 for (i = 0; i < page_count; ++i) {
211 unsigned int len =
212 min_t(unsigned int, PAGE_SIZE - offset, total_len);
213 pgoff_t index = (read_start >> PAGE_SHIFT) + i;
214 struct page *page;
216 page = squashfs_get_cache_page(cache_mapping, index);
217 if (!page)
218 page = alloc_page(GFP_NOIO);
220 if (!page) {
221 error = -ENOMEM;
222 goto out_free_bio;
226 * Use the __ version to avoid merging since we need each page
227 * to be separate when we check for and avoid cached pages.
229 __bio_add_page(bio, page, len, offset);
230 offset = 0;
231 total_len -= len;
234 if (cache_mapping)
235 error = squashfs_bio_read_cached(bio, cache_mapping, index,
236 length, read_start, read_end,
237 page_count);
238 else
239 error = submit_bio_wait(bio);
240 if (error)
241 goto out_free_bio;
243 *biop = bio;
244 *block_offset = index & ((1 << msblk->devblksize_log2) - 1);
245 return 0;
247 out_free_bio:
248 bio_free_pages(bio);
249 bio_uninit(bio);
250 kfree(bio);
251 return error;
255 * Read and decompress a metadata block or datablock. Length is non-zero
256 * if a datablock is being read (the size is stored elsewhere in the
257 * filesystem), otherwise the length is obtained from the first two bytes of
258 * the metadata block. A bit in the length field indicates if the block
259 * is stored uncompressed in the filesystem (usually because compression
260 * generated a larger block - this does occasionally happen with compression
261 * algorithms).
263 int squashfs_read_data(struct super_block *sb, u64 index, int length,
264 u64 *next_index, struct squashfs_page_actor *output)
266 struct squashfs_sb_info *msblk = sb->s_fs_info;
267 struct bio *bio = NULL;
268 int compressed;
269 int res;
270 int offset;
272 if (length) {
274 * Datablock.
276 compressed = SQUASHFS_COMPRESSED_BLOCK(length);
277 length = SQUASHFS_COMPRESSED_SIZE_BLOCK(length);
278 TRACE("Block @ 0x%llx, %scompressed size %d, src size %d\n",
279 index, compressed ? "" : "un", length, output->length);
280 } else {
282 * Metadata block.
284 const u8 *data;
285 struct bvec_iter_all iter_all = {};
286 struct bio_vec *bvec = bvec_init_iter_all(&iter_all);
288 if (index + 2 > msblk->bytes_used) {
289 res = -EIO;
290 goto out;
292 res = squashfs_bio_read(sb, index, 2, &bio, &offset);
293 if (res)
294 goto out;
296 if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) {
297 res = -EIO;
298 goto out_free_bio;
300 /* Extract the length of the metadata block */
301 data = bvec_virt(bvec);
302 length = data[offset];
303 if (offset < bvec->bv_len - 1) {
304 length |= data[offset + 1] << 8;
305 } else {
306 if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) {
307 res = -EIO;
308 goto out_free_bio;
310 data = bvec_virt(bvec);
311 length |= data[0] << 8;
313 bio_free_pages(bio);
314 bio_uninit(bio);
315 kfree(bio);
317 compressed = SQUASHFS_COMPRESSED(length);
318 length = SQUASHFS_COMPRESSED_SIZE(length);
319 index += 2;
321 TRACE("Block @ 0x%llx, %scompressed size %d\n", index - 2,
322 compressed ? "" : "un", length);
324 if (length <= 0 || length > output->length ||
325 (index + length) > msblk->bytes_used) {
326 res = -EIO;
327 goto out;
330 if (next_index)
331 *next_index = index + length;
333 res = squashfs_bio_read(sb, index, length, &bio, &offset);
334 if (res)
335 goto out;
337 if (compressed) {
338 if (!msblk->stream) {
339 res = -EIO;
340 goto out_free_bio;
342 res = msblk->thread_ops->decompress(msblk, bio, offset, length, output);
343 } else {
344 res = copy_bio_to_actor(bio, output, offset, length);
347 out_free_bio:
348 bio_free_pages(bio);
349 bio_uninit(bio);
350 kfree(bio);
351 out:
352 if (res < 0) {
353 ERROR("Failed to read block 0x%llx: %d\n", index, res);
354 if (msblk->panic_on_errors)
355 panic("squashfs read failed");
358 return res;