1 // SPDX-License-Identifier: GPL-2.0
3 * (C) 2001 Clemson University and The University of Chicago
5 * Changes by Acxiom Corporation to add protocol version to kernel
6 * communication, Copyright Acxiom Corporation, 2005.
8 * See COPYING in top-level directory.
12 #include "orangefs-kernel.h"
13 #include "orangefs-dev-proto.h"
14 #include "orangefs-bufmap.h"
15 #include "orangefs-debugfs.h"
17 #include <linux/debugfs.h>
18 #include <linux/slab.h>
20 /* this file implements the /dev/pvfs2-req device node */
22 uint32_t orangefs_userspace_version
;
24 static int open_access_count
;
26 static DEFINE_MUTEX(devreq_mutex
);
28 #define DUMP_DEVICE_ERROR() \
30 gossip_err("*****************************************************\n");\
31 gossip_err("ORANGEFS Device Error: You cannot open the device file "); \
32 gossip_err("\n/dev/%s more than once. Please make sure that\nthere " \
33 "are no ", ORANGEFS_REQDEVICE_NAME); \
34 gossip_err("instances of a program using this device\ncurrently " \
35 "running. (You must verify this!)\n"); \
36 gossip_err("For example, you can use the lsof program as follows:\n");\
37 gossip_err("'lsof | grep %s' (run this as root)\n", \
38 ORANGEFS_REQDEVICE_NAME); \
39 gossip_err(" open_access_count = %d\n", open_access_count); \
40 gossip_err("*****************************************************\n");\
43 static int hash_func(__u64 tag
, int table_size
)
45 return do_div(tag
, (unsigned int)table_size
);
48 static void orangefs_devreq_add_op(struct orangefs_kernel_op_s
*op
)
50 int index
= hash_func(op
->tag
, hash_table_size
);
52 list_add_tail(&op
->list
, &orangefs_htable_ops_in_progress
[index
]);
56 * find the op with this tag and remove it from the in progress
59 static struct orangefs_kernel_op_s
*orangefs_devreq_remove_op(__u64 tag
)
61 struct orangefs_kernel_op_s
*op
, *next
;
64 index
= hash_func(tag
, hash_table_size
);
66 spin_lock(&orangefs_htable_ops_in_progress_lock
);
67 list_for_each_entry_safe(op
,
69 &orangefs_htable_ops_in_progress
[index
],
71 if (op
->tag
== tag
&& !op_state_purged(op
) &&
72 !op_state_given_up(op
)) {
73 list_del_init(&op
->list
);
74 spin_unlock(&orangefs_htable_ops_in_progress_lock
);
79 spin_unlock(&orangefs_htable_ops_in_progress_lock
);
83 /* Returns whether any FS are still pending remounted */
84 static int mark_all_pending_mounts(void)
87 struct orangefs_sb_info_s
*orangefs_sb
= NULL
;
89 spin_lock(&orangefs_superblocks_lock
);
90 list_for_each_entry(orangefs_sb
, &orangefs_superblocks
, list
) {
91 /* All of these file system require a remount */
92 orangefs_sb
->mount_pending
= 1;
95 spin_unlock(&orangefs_superblocks_lock
);
100 * Determine if a given file system needs to be remounted or not
101 * Returns -1 on error
102 * 0 if already mounted
105 static int fs_mount_pending(__s32 fsid
)
107 int mount_pending
= -1;
108 struct orangefs_sb_info_s
*orangefs_sb
= NULL
;
110 spin_lock(&orangefs_superblocks_lock
);
111 list_for_each_entry(orangefs_sb
, &orangefs_superblocks
, list
) {
112 if (orangefs_sb
->fs_id
== fsid
) {
113 mount_pending
= orangefs_sb
->mount_pending
;
117 spin_unlock(&orangefs_superblocks_lock
);
118 return mount_pending
;
121 static int orangefs_devreq_open(struct inode
*inode
, struct file
*file
)
125 /* in order to ensure that the filesystem driver sees correct UIDs */
126 if (file
->f_cred
->user_ns
!= &init_user_ns
) {
127 gossip_err("%s: device cannot be opened outside init_user_ns\n",
132 if (!(file
->f_flags
& O_NONBLOCK
)) {
133 gossip_err("%s: device cannot be opened in blocking mode\n",
138 gossip_debug(GOSSIP_DEV_DEBUG
, "client-core: opening device\n");
139 mutex_lock(&devreq_mutex
);
141 if (open_access_count
== 0) {
142 open_access_count
= 1;
147 mutex_unlock(&devreq_mutex
);
151 gossip_debug(GOSSIP_DEV_DEBUG
,
152 "pvfs2-client-core: open device complete (ret = %d)\n",
157 /* Function for read() callers into the device */
158 static ssize_t
orangefs_devreq_read(struct file
*file
,
160 size_t count
, loff_t
*offset
)
162 struct orangefs_kernel_op_s
*op
, *temp
;
163 __s32 proto_ver
= ORANGEFS_KERNEL_PROTO_VERSION
;
164 static __s32 magic
= ORANGEFS_DEVREQ_MAGIC
;
165 struct orangefs_kernel_op_s
*cur_op
;
168 /* We do not support blocking IO. */
169 if (!(file
->f_flags
& O_NONBLOCK
)) {
170 gossip_err("%s: blocking read from client-core.\n",
176 * The client will do an ioctl to find MAX_DEV_REQ_UPSIZE, then
177 * always read with that size buffer.
179 if (count
!= MAX_DEV_REQ_UPSIZE
) {
180 gossip_err("orangefs: client-core tried to read wrong size\n");
184 /* Check for an empty list before locking. */
185 if (list_empty(&orangefs_request_list
))
190 /* Get next op (if any) from top of list. */
191 spin_lock(&orangefs_request_list_lock
);
192 list_for_each_entry_safe(op
, temp
, &orangefs_request_list
, list
) {
194 /* This lock is held past the end of the loop when we break. */
195 spin_lock(&op
->lock
);
196 if (unlikely(op_state_purged(op
) || op_state_given_up(op
))) {
197 spin_unlock(&op
->lock
);
201 fsid
= fsid_of_op(op
);
202 if (fsid
!= ORANGEFS_FS_ID_NULL
) {
204 /* Skip ops whose filesystem needs to be mounted. */
205 ret
= fs_mount_pending(fsid
);
207 gossip_debug(GOSSIP_DEV_DEBUG
,
208 "%s: mount pending, skipping op tag "
212 get_opname_string(op
));
213 spin_unlock(&op
->lock
);
216 * Skip ops whose filesystem we don't know about unless
217 * it is being mounted or unmounted. It is possible for
218 * a filesystem we don't know about to be unmounted if
219 * it fails to mount in the kernel after userspace has
220 * been sent the mount request.
222 /* XXX: is there a better way to detect this? */
223 } else if (ret
== -1 &&
225 ORANGEFS_VFS_OP_FS_MOUNT
||
227 ORANGEFS_VFS_OP_GETATTR
||
229 ORANGEFS_VFS_OP_FS_UMOUNT
)) {
230 gossip_debug(GOSSIP_DEV_DEBUG
,
231 "orangefs: skipping op tag %llu %s\n",
232 llu(op
->tag
), get_opname_string(op
));
234 "orangefs: ERROR: fs_mount_pending %d\n",
236 spin_unlock(&op
->lock
);
241 * Either this op does not pertain to a filesystem, is mounting
242 * a filesystem, or pertains to a mounted filesystem. Let it
250 * At this point we either have a valid op and can continue or have not
251 * found an op and must ask the client to try again later.
254 spin_unlock(&orangefs_request_list_lock
);
258 gossip_debug(GOSSIP_DEV_DEBUG
, "%s: reading op tag %llu %s\n",
261 get_opname_string(cur_op
));
264 * Such an op should never be on the list in the first place. If so, we
267 if (op_state_in_progress(cur_op
) || op_state_serviced(cur_op
)) {
268 gossip_err("orangefs: ERROR: Current op already queued.\n");
269 list_del_init(&cur_op
->list
);
270 spin_unlock(&cur_op
->lock
);
271 spin_unlock(&orangefs_request_list_lock
);
275 list_del_init(&cur_op
->list
);
276 spin_unlock(&orangefs_request_list_lock
);
278 spin_unlock(&cur_op
->lock
);
280 /* Push the upcall out. */
281 ret
= copy_to_user(buf
, &proto_ver
, sizeof(__s32
));
284 ret
= copy_to_user(buf
+sizeof(__s32
), &magic
, sizeof(__s32
));
287 ret
= copy_to_user(buf
+2 * sizeof(__s32
), &cur_op
->tag
, sizeof(__u64
));
290 ret
= copy_to_user(buf
+2*sizeof(__s32
)+sizeof(__u64
), &cur_op
->upcall
,
291 sizeof(struct orangefs_upcall_s
));
295 spin_lock(&orangefs_htable_ops_in_progress_lock
);
296 spin_lock(&cur_op
->lock
);
297 if (unlikely(op_state_given_up(cur_op
))) {
298 spin_unlock(&cur_op
->lock
);
299 spin_unlock(&orangefs_htable_ops_in_progress_lock
);
300 complete(&cur_op
->waitq
);
305 * Set the operation to be in progress and move it between lists since
306 * it has been sent to the client.
308 set_op_state_inprogress(cur_op
);
309 gossip_debug(GOSSIP_DEV_DEBUG
,
310 "%s: 1 op:%s: op_state:%d: process:%s:\n",
312 get_opname_string(cur_op
),
315 orangefs_devreq_add_op(cur_op
);
316 spin_unlock(&cur_op
->lock
);
317 spin_unlock(&orangefs_htable_ops_in_progress_lock
);
319 /* The client only asks to read one size buffer. */
320 return MAX_DEV_REQ_UPSIZE
;
323 * We were unable to copy the op data to the client. Put the op back in
324 * list. If client has crashed, the op will be purged later when the
325 * device is released.
327 gossip_err("orangefs: Failed to copy data to user space\n");
328 spin_lock(&orangefs_request_list_lock
);
329 spin_lock(&cur_op
->lock
);
330 if (likely(!op_state_given_up(cur_op
))) {
331 set_op_state_waiting(cur_op
);
332 gossip_debug(GOSSIP_DEV_DEBUG
,
333 "%s: 2 op:%s: op_state:%d: process:%s:\n",
335 get_opname_string(cur_op
),
338 list_add(&cur_op
->list
, &orangefs_request_list
);
339 spin_unlock(&cur_op
->lock
);
341 spin_unlock(&cur_op
->lock
);
342 complete(&cur_op
->waitq
);
344 spin_unlock(&orangefs_request_list_lock
);
349 * Function for writev() callers into the device.
351 * Userspace should have written:
355 * - struct orangefs_downcall_s
356 * - trailer buffer (in the case of READDIR operations)
358 static ssize_t
orangefs_devreq_write_iter(struct kiocb
*iocb
,
359 struct iov_iter
*iter
)
362 struct orangefs_kernel_op_s
*op
= NULL
;
368 int total
= ret
= iov_iter_count(iter
);
369 int downcall_size
= sizeof(struct orangefs_downcall_s
);
370 int head_size
= sizeof(head
);
372 gossip_debug(GOSSIP_DEV_DEBUG
, "%s: total:%d: ret:%zd:\n",
377 if (total
< MAX_DEV_REQ_DOWNSIZE
) {
378 gossip_err("%s: total:%d: must be at least:%u:\n",
381 (unsigned int) MAX_DEV_REQ_DOWNSIZE
);
385 if (!copy_from_iter_full(&head
, head_size
, iter
)) {
386 gossip_err("%s: failed to copy head.\n", __func__
);
390 if (head
.version
< ORANGEFS_MINIMUM_USERSPACE_VERSION
) {
391 gossip_err("%s: userspace claims version"
392 "%d, minimum version required: %d.\n",
395 ORANGEFS_MINIMUM_USERSPACE_VERSION
);
399 if (head
.magic
!= ORANGEFS_DEVREQ_MAGIC
) {
400 gossip_err("Error: Device magic number does not match.\n");
404 if (!orangefs_userspace_version
) {
405 orangefs_userspace_version
= head
.version
;
406 } else if (orangefs_userspace_version
!= head
.version
) {
407 gossip_err("Error: userspace version changes\n");
411 /* remove the op from the in progress hash table */
412 op
= orangefs_devreq_remove_op(head
.tag
);
414 gossip_debug(GOSSIP_DEV_DEBUG
,
415 "%s: No one's waiting for tag %llu\n",
416 __func__
, llu(head
.tag
));
420 if (!copy_from_iter_full(&op
->downcall
, downcall_size
, iter
)) {
421 gossip_err("%s: failed to copy downcall.\n", __func__
);
425 if (op
->downcall
.status
)
429 * We've successfully peeled off the head and the downcall.
430 * Something has gone awry if total doesn't equal the
431 * sum of head_size, downcall_size and trailer_size.
433 if ((head_size
+ downcall_size
+ op
->downcall
.trailer_size
) != total
) {
434 gossip_err("%s: funky write, head_size:%d"
435 ": downcall_size:%d: trailer_size:%lld"
436 ": total size:%d:\n",
440 op
->downcall
.trailer_size
,
445 /* Only READDIR operations should have trailers. */
446 if ((op
->downcall
.type
!= ORANGEFS_VFS_OP_READDIR
) &&
447 (op
->downcall
.trailer_size
!= 0)) {
448 gossip_err("%s: %x operation with trailer.",
454 /* READDIR operations should always have trailers. */
455 if ((op
->downcall
.type
== ORANGEFS_VFS_OP_READDIR
) &&
456 (op
->downcall
.trailer_size
== 0)) {
457 gossip_err("%s: %x operation with no trailer.",
463 if (op
->downcall
.type
!= ORANGEFS_VFS_OP_READDIR
)
466 op
->downcall
.trailer_buf
= vmalloc(op
->downcall
.trailer_size
);
467 if (!op
->downcall
.trailer_buf
)
470 memset(op
->downcall
.trailer_buf
, 0, op
->downcall
.trailer_size
);
471 if (!copy_from_iter_full(op
->downcall
.trailer_buf
,
472 op
->downcall
.trailer_size
, iter
)) {
473 gossip_err("%s: failed to copy trailer.\n", __func__
);
474 vfree(op
->downcall
.trailer_buf
);
480 * Return to vfs waitqueue, and back to service_operation
481 * through wait_for_matching_downcall.
483 spin_lock(&op
->lock
);
484 if (unlikely(op_is_cancel(op
))) {
485 spin_unlock(&op
->lock
);
487 } else if (unlikely(op_state_given_up(op
))) {
488 spin_unlock(&op
->lock
);
489 complete(&op
->waitq
);
491 set_op_state_serviced(op
);
492 gossip_debug(GOSSIP_DEV_DEBUG
,
493 "%s: op:%s: op_state:%d: process:%s:\n",
495 get_opname_string(op
),
498 spin_unlock(&op
->lock
);
503 op
->downcall
.status
= -(ORANGEFS_ERROR_BIT
| 9);
508 op
->downcall
.status
= -(ORANGEFS_ERROR_BIT
| 8);
514 * NOTE: gets called when the last reference to this device is dropped.
515 * Using the open_access_count variable, we enforce a reference count
516 * on this file so that it can be opened by only one process at a time.
517 * the devreq_mutex is used to make sure all i/o has completed
518 * before we call orangefs_bufmap_finalize, and similar such tricky
521 static int orangefs_devreq_release(struct inode
*inode
, struct file
*file
)
525 gossip_debug(GOSSIP_DEV_DEBUG
,
526 "%s:pvfs2-client-core: exiting, closing device\n",
529 mutex_lock(&devreq_mutex
);
530 orangefs_bufmap_finalize();
532 open_access_count
= -1;
534 unmounted
= mark_all_pending_mounts();
535 gossip_debug(GOSSIP_DEV_DEBUG
, "ORANGEFS Device Close: Filesystem(s) %s\n",
536 (unmounted
? "UNMOUNTED" : "MOUNTED"));
539 purge_inprogress_ops();
541 orangefs_bufmap_run_down();
543 gossip_debug(GOSSIP_DEV_DEBUG
,
544 "pvfs2-client-core: device close complete\n");
545 open_access_count
= 0;
546 orangefs_userspace_version
= 0;
547 mutex_unlock(&devreq_mutex
);
551 int is_daemon_in_service(void)
556 * What this function does is checks if client-core is alive
557 * based on the access count we maintain on the device.
559 mutex_lock(&devreq_mutex
);
560 in_service
= open_access_count
== 1 ? 0 : -EIO
;
561 mutex_unlock(&devreq_mutex
);
565 bool __is_daemon_in_service(void)
567 return open_access_count
== 1;
570 static inline long check_ioctl_command(unsigned int command
)
572 /* Check for valid ioctl codes */
573 if (_IOC_TYPE(command
) != ORANGEFS_DEV_MAGIC
) {
574 gossip_err("device ioctl magic numbers don't match! Did you rebuild pvfs2-client-core/libpvfs2? [cmd %x, magic %x != %x]\n",
580 /* and valid ioctl commands */
581 if (_IOC_NR(command
) >= ORANGEFS_DEV_MAXNR
|| _IOC_NR(command
) <= 0) {
582 gossip_err("Invalid ioctl command number [%d >= %d]\n",
583 _IOC_NR(command
), ORANGEFS_DEV_MAXNR
);
589 static long dispatch_ioctl_command(unsigned int command
, unsigned long arg
)
591 static __s32 magic
= ORANGEFS_DEVREQ_MAGIC
;
592 static __s32 max_up_size
= MAX_DEV_REQ_UPSIZE
;
593 static __s32 max_down_size
= MAX_DEV_REQ_DOWNSIZE
;
594 struct ORANGEFS_dev_map_desc user_desc
;
596 int upstream_kmod
= 1;
597 struct orangefs_sb_info_s
*orangefs_sb
;
599 /* mtmoore: add locking here */
602 case ORANGEFS_DEV_GET_MAGIC
:
603 return ((put_user(magic
, (__s32 __user
*) arg
) == -EFAULT
) ?
606 case ORANGEFS_DEV_GET_MAX_UPSIZE
:
607 return ((put_user(max_up_size
,
608 (__s32 __user
*) arg
) == -EFAULT
) ?
611 case ORANGEFS_DEV_GET_MAX_DOWNSIZE
:
612 return ((put_user(max_down_size
,
613 (__s32 __user
*) arg
) == -EFAULT
) ?
616 case ORANGEFS_DEV_MAP
:
617 ret
= copy_from_user(&user_desc
,
618 (struct ORANGEFS_dev_map_desc __user
*)
620 sizeof(struct ORANGEFS_dev_map_desc
));
621 /* WTF -EIO and not -EFAULT? */
622 return ret
? -EIO
: orangefs_bufmap_initialize(&user_desc
);
623 case ORANGEFS_DEV_REMOUNT_ALL
:
624 gossip_debug(GOSSIP_DEV_DEBUG
,
625 "%s: got ORANGEFS_DEV_REMOUNT_ALL\n",
629 * remount all mounted orangefs volumes to regain the lost
630 * dynamic mount tables (if any) -- NOTE: this is done
631 * without keeping the superblock list locked due to the
632 * upcall/downcall waiting. also, the request mutex is
633 * used to ensure that no operations will be serviced until
634 * all of the remounts are serviced (to avoid ops between
637 ret
= mutex_lock_interruptible(&orangefs_request_mutex
);
640 gossip_debug(GOSSIP_DEV_DEBUG
,
641 "%s: priority remount in progress\n",
643 spin_lock(&orangefs_superblocks_lock
);
644 list_for_each_entry(orangefs_sb
, &orangefs_superblocks
, list
) {
646 * We have to drop the spinlock, so entries can be
647 * removed. They can't be freed, though, so we just
648 * keep the forward pointers and zero the back ones -
649 * that way we can get to the rest of the list.
651 if (!orangefs_sb
->list
.prev
)
653 gossip_debug(GOSSIP_DEV_DEBUG
,
654 "%s: Remounting SB %p\n",
658 spin_unlock(&orangefs_superblocks_lock
);
659 ret
= orangefs_remount(orangefs_sb
);
660 spin_lock(&orangefs_superblocks_lock
);
662 gossip_debug(GOSSIP_DEV_DEBUG
,
663 "SB %p remount failed\n",
668 spin_unlock(&orangefs_superblocks_lock
);
669 gossip_debug(GOSSIP_DEV_DEBUG
,
670 "%s: priority remount complete\n",
672 mutex_unlock(&orangefs_request_mutex
);
675 case ORANGEFS_DEV_UPSTREAM
:
676 ret
= copy_to_user((void __user
*)arg
,
678 sizeof(upstream_kmod
));
685 case ORANGEFS_DEV_CLIENT_MASK
:
686 return orangefs_debugfs_new_client_mask((void __user
*)arg
);
687 case ORANGEFS_DEV_CLIENT_STRING
:
688 return orangefs_debugfs_new_client_string((void __user
*)arg
);
689 case ORANGEFS_DEV_DEBUG
:
690 return orangefs_debugfs_new_debug((void __user
*)arg
);
697 static long orangefs_devreq_ioctl(struct file
*file
,
698 unsigned int command
, unsigned long arg
)
702 /* Check for properly constructed commands */
703 ret
= check_ioctl_command(command
);
707 return (int)dispatch_ioctl_command(command
, arg
);
710 #ifdef CONFIG_COMPAT /* CONFIG_COMPAT is in .config */
712 /* Compat structure for the ORANGEFS_DEV_MAP ioctl */
713 struct ORANGEFS_dev_map_desc32
{
720 static unsigned long translate_dev_map26(unsigned long args
, long *error
)
722 struct ORANGEFS_dev_map_desc32 __user
*p32
= (void __user
*)args
;
724 * Depending on the architecture, allocate some space on the
725 * user-call-stack based on our expected layout.
727 struct ORANGEFS_dev_map_desc __user
*p
=
728 compat_alloc_user_space(sizeof(*p
));
732 /* get the ptr from the 32 bit user-space */
733 if (get_user(addr
, &p32
->ptr
))
735 /* try to put that into a 64-bit layout */
736 if (put_user(compat_ptr(addr
), &p
->ptr
))
738 /* copy the remaining fields */
739 if (copy_in_user(&p
->total_size
, &p32
->total_size
, sizeof(__s32
)))
741 if (copy_in_user(&p
->size
, &p32
->size
, sizeof(__s32
)))
743 if (copy_in_user(&p
->count
, &p32
->count
, sizeof(__s32
)))
745 return (unsigned long)p
;
752 * 32 bit user-space apps' ioctl handlers when kernel modules
753 * is compiled as a 64 bit one
755 static long orangefs_devreq_compat_ioctl(struct file
*filp
, unsigned int cmd
,
759 unsigned long arg
= args
;
761 /* Check for properly constructed commands */
762 ret
= check_ioctl_command(cmd
);
765 if (cmd
== ORANGEFS_DEV_MAP
) {
767 * convert the arguments to what we expect internally
770 arg
= translate_dev_map26(args
, &ret
);
772 gossip_err("Could not translate dev map\n");
776 /* no other ioctl requires translation */
777 return dispatch_ioctl_command(cmd
, arg
);
780 #endif /* CONFIG_COMPAT is in .config */
782 /* the assigned character device major number */
783 static int orangefs_dev_major
;
786 * Initialize orangefs device specific state:
787 * Must be called at module load time only
789 int orangefs_dev_init(void)
791 /* register orangefs-req device */
792 orangefs_dev_major
= register_chrdev(0,
793 ORANGEFS_REQDEVICE_NAME
,
794 &orangefs_devreq_file_operations
);
795 if (orangefs_dev_major
< 0) {
796 gossip_debug(GOSSIP_DEV_DEBUG
,
797 "Failed to register /dev/%s (error %d)\n",
798 ORANGEFS_REQDEVICE_NAME
, orangefs_dev_major
);
799 return orangefs_dev_major
;
802 gossip_debug(GOSSIP_DEV_DEBUG
,
803 "*** /dev/%s character device registered ***\n",
804 ORANGEFS_REQDEVICE_NAME
);
805 gossip_debug(GOSSIP_DEV_DEBUG
, "'mknod /dev/%s c %d 0'.\n",
806 ORANGEFS_REQDEVICE_NAME
, orangefs_dev_major
);
810 void orangefs_dev_cleanup(void)
812 unregister_chrdev(orangefs_dev_major
, ORANGEFS_REQDEVICE_NAME
);
813 gossip_debug(GOSSIP_DEV_DEBUG
,
814 "*** /dev/%s character device unregistered ***\n",
815 ORANGEFS_REQDEVICE_NAME
);
818 static __poll_t
orangefs_devreq_poll(struct file
*file
,
819 struct poll_table_struct
*poll_table
)
821 __poll_t poll_revent_mask
= 0;
823 poll_wait(file
, &orangefs_request_list_waitq
, poll_table
);
825 if (!list_empty(&orangefs_request_list
))
826 poll_revent_mask
|= EPOLLIN
;
827 return poll_revent_mask
;
830 const struct file_operations orangefs_devreq_file_operations
= {
831 .owner
= THIS_MODULE
,
832 .read
= orangefs_devreq_read
,
833 .write_iter
= orangefs_devreq_write_iter
,
834 .open
= orangefs_devreq_open
,
835 .release
= orangefs_devreq_release
,
836 .unlocked_ioctl
= orangefs_devreq_ioctl
,
838 #ifdef CONFIG_COMPAT /* CONFIG_COMPAT is in .config */
839 .compat_ioctl
= orangefs_devreq_compat_ioctl
,
841 .poll
= orangefs_devreq_poll