1 #include <linux/capability.h>
2 #include <linux/blkdev.h>
3 #include <linux/export.h>
5 #include <linux/blkpg.h>
6 #include <linux/hdreg.h>
7 #include <linux/backing-dev.h>
9 #include <linux/blktrace_api.h>
10 #include <asm/uaccess.h>
12 static int blkpg_ioctl(struct block_device
*bdev
, struct blkpg_ioctl_arg __user
*arg
)
14 struct block_device
*bdevp
;
16 struct hd_struct
*part
, *lpart
;
17 struct blkpg_ioctl_arg a
;
18 struct blkpg_partition p
;
19 struct disk_part_iter piter
;
20 long long start
, length
;
23 if (!capable(CAP_SYS_ADMIN
))
25 if (copy_from_user(&a
, arg
, sizeof(struct blkpg_ioctl_arg
)))
27 if (copy_from_user(&p
, a
.data
, sizeof(struct blkpg_partition
)))
30 if (bdev
!= bdev
->bd_contains
)
36 case BLKPG_ADD_PARTITION
:
38 length
= p
.length
>> 9;
39 /* check for fit in a hd_struct */
40 if (sizeof(sector_t
) == sizeof(long) &&
41 sizeof(long long) > sizeof(long)) {
42 long pstart
= start
, plength
= length
;
43 if (pstart
!= start
|| plength
!= length
44 || pstart
< 0 || plength
< 0 || partno
> 65535)
48 mutex_lock(&bdev
->bd_mutex
);
51 disk_part_iter_init(&piter
, disk
,
52 DISK_PITER_INCL_EMPTY
);
53 while ((part
= disk_part_iter_next(&piter
))) {
54 if (!(start
+ length
<= part
->start_sect
||
55 start
>= part
->start_sect
+ part
->nr_sects
)) {
56 disk_part_iter_exit(&piter
);
57 mutex_unlock(&bdev
->bd_mutex
);
61 disk_part_iter_exit(&piter
);
64 part
= add_partition(disk
, partno
, start
, length
,
65 ADDPART_FLAG_NONE
, NULL
);
66 mutex_unlock(&bdev
->bd_mutex
);
67 return PTR_ERR_OR_ZERO(part
);
68 case BLKPG_DEL_PARTITION
:
69 part
= disk_get_part(disk
, partno
);
73 bdevp
= bdget(part_devt(part
));
78 mutex_lock(&bdevp
->bd_mutex
);
79 if (bdevp
->bd_openers
) {
80 mutex_unlock(&bdevp
->bd_mutex
);
86 invalidate_bdev(bdevp
);
88 mutex_lock_nested(&bdev
->bd_mutex
, 1);
89 delete_partition(disk
, partno
);
90 mutex_unlock(&bdev
->bd_mutex
);
91 mutex_unlock(&bdevp
->bd_mutex
);
95 case BLKPG_RESIZE_PARTITION
:
97 /* new length of partition in bytes */
98 length
= p
.length
>> 9;
99 /* check for fit in a hd_struct */
100 if (sizeof(sector_t
) == sizeof(long) &&
101 sizeof(long long) > sizeof(long)) {
102 long pstart
= start
, plength
= length
;
103 if (pstart
!= start
|| plength
!= length
104 || pstart
< 0 || plength
< 0)
107 part
= disk_get_part(disk
, partno
);
110 bdevp
= bdget(part_devt(part
));
115 mutex_lock(&bdevp
->bd_mutex
);
116 mutex_lock_nested(&bdev
->bd_mutex
, 1);
117 if (start
!= part
->start_sect
) {
118 mutex_unlock(&bdevp
->bd_mutex
);
119 mutex_unlock(&bdev
->bd_mutex
);
125 disk_part_iter_init(&piter
, disk
,
126 DISK_PITER_INCL_EMPTY
);
127 while ((lpart
= disk_part_iter_next(&piter
))) {
128 if (lpart
->partno
!= partno
&&
129 !(start
+ length
<= lpart
->start_sect
||
130 start
>= lpart
->start_sect
+ lpart
->nr_sects
)
132 disk_part_iter_exit(&piter
);
133 mutex_unlock(&bdevp
->bd_mutex
);
134 mutex_unlock(&bdev
->bd_mutex
);
140 disk_part_iter_exit(&piter
);
141 part_nr_sects_write(part
, (sector_t
)length
);
142 i_size_write(bdevp
->bd_inode
, p
.length
);
143 mutex_unlock(&bdevp
->bd_mutex
);
144 mutex_unlock(&bdev
->bd_mutex
);
154 * This is an exported API for the block driver, and will not
155 * acquire bd_mutex. This API should be used in case that
156 * caller has held bd_mutex already.
158 int __blkdev_reread_part(struct block_device
*bdev
)
160 struct gendisk
*disk
= bdev
->bd_disk
;
162 if (!disk_part_scan_enabled(disk
) || bdev
!= bdev
->bd_contains
)
164 if (!capable(CAP_SYS_ADMIN
))
167 lockdep_assert_held(&bdev
->bd_mutex
);
169 return rescan_partitions(disk
, bdev
);
171 EXPORT_SYMBOL(__blkdev_reread_part
);
174 * This is an exported API for the block driver, and will
175 * try to acquire bd_mutex. If bd_mutex has been held already
176 * in current context, please call __blkdev_reread_part().
178 * Make sure the held locks in current context aren't required
179 * in open()/close() handler and I/O path for avoiding ABBA deadlock:
180 * - bd_mutex is held before calling block driver's open/close
182 * - reading partition table may submit I/O to the block device
184 int blkdev_reread_part(struct block_device
*bdev
)
188 mutex_lock(&bdev
->bd_mutex
);
189 res
= __blkdev_reread_part(bdev
);
190 mutex_unlock(&bdev
->bd_mutex
);
194 EXPORT_SYMBOL(blkdev_reread_part
);
196 static int blk_ioctl_discard(struct block_device
*bdev
, uint64_t start
,
197 uint64_t len
, int secure
)
199 unsigned long flags
= 0;
208 if (start
+ len
> (i_size_read(bdev
->bd_inode
) >> 9))
211 flags
|= BLKDEV_DISCARD_SECURE
;
212 return blkdev_issue_discard(bdev
, start
, len
, GFP_KERNEL
, flags
);
215 static int blk_ioctl_zeroout(struct block_device
*bdev
, uint64_t start
,
225 if (start
+ len
> (i_size_read(bdev
->bd_inode
) >> 9))
228 return blkdev_issue_zeroout(bdev
, start
, len
, GFP_KERNEL
, false);
231 static int put_ushort(unsigned long arg
, unsigned short val
)
233 return put_user(val
, (unsigned short __user
*)arg
);
236 static int put_int(unsigned long arg
, int val
)
238 return put_user(val
, (int __user
*)arg
);
241 static int put_uint(unsigned long arg
, unsigned int val
)
243 return put_user(val
, (unsigned int __user
*)arg
);
246 static int put_long(unsigned long arg
, long val
)
248 return put_user(val
, (long __user
*)arg
);
251 static int put_ulong(unsigned long arg
, unsigned long val
)
253 return put_user(val
, (unsigned long __user
*)arg
);
256 static int put_u64(unsigned long arg
, u64 val
)
258 return put_user(val
, (u64 __user
*)arg
);
261 int __blkdev_driver_ioctl(struct block_device
*bdev
, fmode_t mode
,
262 unsigned cmd
, unsigned long arg
)
264 struct gendisk
*disk
= bdev
->bd_disk
;
266 if (disk
->fops
->ioctl
)
267 return disk
->fops
->ioctl(bdev
, mode
, cmd
, arg
);
272 * For the record: _GPL here is only because somebody decided to slap it
273 * on the previous export. Sheer idiocy, since it wasn't copyrightable
274 * at all and could be open-coded without any exports by anybody who cares.
276 EXPORT_SYMBOL_GPL(__blkdev_driver_ioctl
);
279 * Is it an unrecognized ioctl? The correct returns are either
280 * ENOTTY (final) or ENOIOCTLCMD ("I don't know this one, try a
281 * fallback"). ENOIOCTLCMD gets turned into ENOTTY by the ioctl
282 * code before returning.
284 * Confused drivers sometimes return EINVAL, which is wrong. It
285 * means "I understood the ioctl command, but the parameters to
288 * We should aim to just fix the broken drivers, the EINVAL case
291 static inline int is_unrecognized_ioctl(int ret
)
293 return ret
== -EINVAL
||
299 * always keep this in sync with compat_blkdev_ioctl()
301 int blkdev_ioctl(struct block_device
*bdev
, fmode_t mode
, unsigned cmd
,
304 struct gendisk
*disk
= bdev
->bd_disk
;
305 struct backing_dev_info
*bdi
;
308 unsigned int max_sectors
;
312 if (!capable(CAP_SYS_ADMIN
))
315 ret
= __blkdev_driver_ioctl(bdev
, mode
, cmd
, arg
);
316 if (!is_unrecognized_ioctl(ret
))
320 invalidate_bdev(bdev
);
324 ret
= __blkdev_driver_ioctl(bdev
, mode
, cmd
, arg
);
325 if (!is_unrecognized_ioctl(ret
))
327 if (!capable(CAP_SYS_ADMIN
))
329 if (get_user(n
, (int __user
*)(arg
)))
331 set_device_ro(bdev
, n
);
335 case BLKSECDISCARD
: {
338 if (!(mode
& FMODE_WRITE
))
341 if (copy_from_user(range
, (void __user
*)arg
, sizeof(range
)))
344 return blk_ioctl_discard(bdev
, range
[0], range
[1],
345 cmd
== BLKSECDISCARD
);
350 if (!(mode
& FMODE_WRITE
))
353 if (copy_from_user(range
, (void __user
*)arg
, sizeof(range
)))
356 return blk_ioctl_zeroout(bdev
, range
[0], range
[1]);
360 struct hd_geometry geo
;
364 if (!disk
->fops
->getgeo
)
368 * We need to set the startsect first, the driver may
369 * want to override it.
371 memset(&geo
, 0, sizeof(geo
));
372 geo
.start
= get_start_sect(bdev
);
373 ret
= disk
->fops
->getgeo(bdev
, &geo
);
376 if (copy_to_user((struct hd_geometry __user
*)arg
, &geo
,
385 bdi
= blk_get_backing_dev_info(bdev
);
386 return put_long(arg
, (bdi
->ra_pages
* PAGE_CACHE_SIZE
) / 512);
388 return put_int(arg
, bdev_read_only(bdev
) != 0);
389 case BLKBSZGET
: /* get block device soft block size (cf. BLKSSZGET) */
390 return put_int(arg
, block_size(bdev
));
391 case BLKSSZGET
: /* get block device logical block size */
392 return put_int(arg
, bdev_logical_block_size(bdev
));
393 case BLKPBSZGET
: /* get block device physical block size */
394 return put_uint(arg
, bdev_physical_block_size(bdev
));
396 return put_uint(arg
, bdev_io_min(bdev
));
398 return put_uint(arg
, bdev_io_opt(bdev
));
400 return put_int(arg
, bdev_alignment_offset(bdev
));
401 case BLKDISCARDZEROES
:
402 return put_uint(arg
, bdev_discard_zeroes_data(bdev
));
404 max_sectors
= min_t(unsigned int, USHRT_MAX
,
405 queue_max_sectors(bdev_get_queue(bdev
)));
406 return put_ushort(arg
, max_sectors
);
408 return put_ushort(arg
, !blk_queue_nonrot(bdev_get_queue(bdev
)));
411 if(!capable(CAP_SYS_ADMIN
))
413 bdi
= blk_get_backing_dev_info(bdev
);
414 bdi
->ra_pages
= (arg
* 512) / PAGE_CACHE_SIZE
;
417 /* set the logical block size */
418 if (!capable(CAP_SYS_ADMIN
))
422 if (get_user(n
, (int __user
*) arg
))
424 if (!(mode
& FMODE_EXCL
)) {
426 if (blkdev_get(bdev
, mode
| FMODE_EXCL
, &bdev
) < 0)
429 ret
= set_blocksize(bdev
, n
);
430 if (!(mode
& FMODE_EXCL
))
431 blkdev_put(bdev
, mode
| FMODE_EXCL
);
434 ret
= blkpg_ioctl(bdev
, (struct blkpg_ioctl_arg __user
*) arg
);
437 ret
= blkdev_reread_part(bdev
);
440 size
= i_size_read(bdev
->bd_inode
);
441 if ((size
>> 9) > ~0UL)
443 return put_ulong(arg
, size
>> 9);
445 return put_u64(arg
, i_size_read(bdev
->bd_inode
));
449 case BLKTRACETEARDOWN
:
450 ret
= blk_trace_ioctl(bdev
, cmd
, (char __user
*) arg
);
453 ret
= __blkdev_driver_ioctl(bdev
, mode
, cmd
, arg
);
457 EXPORT_SYMBOL_GPL(blkdev_ioctl
);