1 // SPDX-License-Identifier: GPL-2.0-or-later
5 * Code which implements the kernel side of a minimal userspace
6 * interface to our DLM. This file handles the virtual file system
7 * used for communication with userspace. Credit should go to ramfs,
8 * which was a template for the fs side of this module.
10 * Copyright (C) 2003, 2004 Oracle. All rights reserved.
13 /* Simple VFS hooks based on: */
15 * Resizable simple ram filesystem for Linux.
17 * Copyright (C) 2000 Linus Torvalds.
18 * 2000 Transmeta Corp.
21 #include <linux/module.h>
23 #include <linux/fs_context.h>
24 #include <linux/pagemap.h>
25 #include <linux/types.h>
26 #include <linux/slab.h>
27 #include <linux/highmem.h>
28 #include <linux/init.h>
29 #include <linux/string.h>
30 #include <linux/backing-dev.h>
31 #include <linux/poll.h>
33 #include <linux/uaccess.h>
35 #include "../stackglue.h"
38 #define MLOG_MASK_PREFIX ML_DLMFS
39 #include "../cluster/masklog.h"
42 static const struct super_operations dlmfs_ops
;
43 static const struct file_operations dlmfs_file_operations
;
44 static const struct inode_operations dlmfs_dir_inode_operations
;
45 static const struct inode_operations dlmfs_root_inode_operations
;
46 static const struct inode_operations dlmfs_file_inode_operations
;
47 static struct kmem_cache
*dlmfs_inode_cache
;
49 struct workqueue_struct
*user_dlm_worker
;
54 * These are the ABI capabilities of dlmfs.
56 * Over time, dlmfs has added some features that were not part of the
57 * initial ABI. Unfortunately, some of these features are not detectable
58 * via standard usage. For example, Linux's default poll always returns
59 * EPOLLIN, so there is no way for a caller of poll(2) to know when dlmfs
60 * added poll support. Instead, we provide this list of new capabilities.
62 * Capabilities is a read-only attribute. We do it as a module parameter
63 * so we can discover it whether dlmfs is built in, loaded, or even not
66 * The ABI features are local to this machine's dlmfs mount. This is
67 * distinct from the locking protocol, which is concerned with inter-node
71 * - bast : EPOLLIN against the file descriptor of a held lock
72 * signifies a bast fired on the lock.
74 #define DLMFS_CAPABILITIES "bast stackglue"
75 static int param_set_dlmfs_capabilities(const char *val
,
76 const struct kernel_param
*kp
)
78 printk(KERN_ERR
"%s: readonly parameter\n", kp
->name
);
81 static int param_get_dlmfs_capabilities(char *buffer
,
82 const struct kernel_param
*kp
)
84 return sysfs_emit(buffer
, DLMFS_CAPABILITIES
);
86 module_param_call(capabilities
, param_set_dlmfs_capabilities
,
87 param_get_dlmfs_capabilities
, NULL
, 0444);
88 MODULE_PARM_DESC(capabilities
, DLMFS_CAPABILITIES
);
92 * decodes a set of open flags into a valid lock level and a set of flags.
93 * returns < 0 if we have invalid flags
94 * flags which mean something to us:
95 * O_RDONLY -> PRMODE level
96 * O_WRONLY -> EXMODE level
98 * O_NONBLOCK -> NOQUEUE
100 static int dlmfs_decode_open_flags(int open_flags
,
104 if (open_flags
& (O_WRONLY
|O_RDWR
))
105 *level
= DLM_LOCK_EX
;
107 *level
= DLM_LOCK_PR
;
110 if (open_flags
& O_NONBLOCK
)
111 *flags
|= DLM_LKF_NOQUEUE
;
116 static int dlmfs_file_open(struct inode
*inode
,
119 int status
, level
, flags
;
120 struct dlmfs_filp_private
*fp
= NULL
;
121 struct dlmfs_inode_private
*ip
;
123 if (S_ISDIR(inode
->i_mode
))
126 mlog(0, "open called on inode %lu, flags 0x%x\n", inode
->i_ino
,
129 status
= dlmfs_decode_open_flags(file
->f_flags
, &level
, &flags
);
133 /* We don't want to honor O_APPEND at read/write time as it
134 * doesn't make sense for LVB writes. */
135 file
->f_flags
&= ~O_APPEND
;
137 fp
= kmalloc(sizeof(*fp
), GFP_NOFS
);
142 fp
->fp_lock_level
= level
;
146 status
= user_dlm_cluster_lock(&ip
->ip_lockres
, level
, flags
);
148 /* this is a strange error to return here but I want
149 * to be able userspace to be able to distinguish a
150 * valid lock request from one that simply couldn't be
152 if (flags
& DLM_LKF_NOQUEUE
&& status
== -EAGAIN
)
158 file
->private_data
= fp
;
163 static int dlmfs_file_release(struct inode
*inode
,
167 struct dlmfs_inode_private
*ip
= DLMFS_I(inode
);
168 struct dlmfs_filp_private
*fp
= file
->private_data
;
170 if (S_ISDIR(inode
->i_mode
))
173 mlog(0, "close called on inode %lu\n", inode
->i_ino
);
176 level
= fp
->fp_lock_level
;
177 if (level
!= DLM_LOCK_IV
)
178 user_dlm_cluster_unlock(&ip
->ip_lockres
, level
);
181 file
->private_data
= NULL
;
188 * We do ->setattr() just to override size changes. Our size is the size
189 * of the LVB and nothing else.
191 static int dlmfs_file_setattr(struct mnt_idmap
*idmap
,
192 struct dentry
*dentry
, struct iattr
*attr
)
195 struct inode
*inode
= d_inode(dentry
);
197 attr
->ia_valid
&= ~ATTR_SIZE
;
198 error
= setattr_prepare(&nop_mnt_idmap
, dentry
, attr
);
202 setattr_copy(&nop_mnt_idmap
, inode
, attr
);
203 mark_inode_dirty(inode
);
207 static __poll_t
dlmfs_file_poll(struct file
*file
, poll_table
*wait
)
210 struct inode
*inode
= file_inode(file
);
211 struct dlmfs_inode_private
*ip
= DLMFS_I(inode
);
213 poll_wait(file
, &ip
->ip_lockres
.l_event
, wait
);
215 spin_lock(&ip
->ip_lockres
.l_lock
);
216 if (ip
->ip_lockres
.l_flags
& USER_LOCK_BLOCKED
)
217 event
= EPOLLIN
| EPOLLRDNORM
;
218 spin_unlock(&ip
->ip_lockres
.l_lock
);
223 static ssize_t
dlmfs_file_read(struct file
*file
,
228 char lvb
[DLM_LVB_LEN
];
230 if (!user_dlm_read_lvb(file_inode(file
), lvb
))
233 return simple_read_from_buffer(buf
, count
, ppos
, lvb
, sizeof(lvb
));
236 static ssize_t
dlmfs_file_write(struct file
*filp
,
237 const char __user
*buf
,
241 char lvb_buf
[DLM_LVB_LEN
];
243 struct inode
*inode
= file_inode(filp
);
245 mlog(0, "inode %lu, count = %zu, *ppos = %llu\n",
246 inode
->i_ino
, count
, *ppos
);
248 if (*ppos
>= DLM_LVB_LEN
)
251 /* don't write past the lvb */
252 if (count
> DLM_LVB_LEN
- *ppos
)
253 count
= DLM_LVB_LEN
- *ppos
;
258 bytes_left
= copy_from_user(lvb_buf
, buf
, count
);
261 user_dlm_write_lvb(inode
, lvb_buf
, count
);
263 *ppos
= *ppos
+ count
;
264 mlog(0, "wrote %zu bytes\n", count
);
268 static void dlmfs_init_once(void *foo
)
270 struct dlmfs_inode_private
*ip
=
271 (struct dlmfs_inode_private
*) foo
;
274 ip
->ip_parent
= NULL
;
276 inode_init_once(&ip
->ip_vfs_inode
);
279 static struct inode
*dlmfs_alloc_inode(struct super_block
*sb
)
281 struct dlmfs_inode_private
*ip
;
283 ip
= alloc_inode_sb(sb
, dlmfs_inode_cache
, GFP_NOFS
);
287 return &ip
->ip_vfs_inode
;
290 static void dlmfs_free_inode(struct inode
*inode
)
292 kmem_cache_free(dlmfs_inode_cache
, DLMFS_I(inode
));
295 static void dlmfs_evict_inode(struct inode
*inode
)
298 struct dlmfs_inode_private
*ip
;
299 struct user_lock_res
*lockres
;
304 mlog(0, "inode %lu\n", inode
->i_ino
);
307 lockres
= &ip
->ip_lockres
;
309 if (S_ISREG(inode
->i_mode
)) {
310 spin_lock(&lockres
->l_lock
);
311 teardown
= !!(lockres
->l_flags
& USER_LOCK_IN_TEARDOWN
);
312 spin_unlock(&lockres
->l_lock
);
314 status
= user_dlm_destroy_lock(lockres
);
322 mlog(0, "we're a directory, ip->ip_conn = 0x%p\n", ip
->ip_conn
);
323 /* we must be a directory. If required, lets unregister the
324 * dlm context now. */
326 user_dlm_unregister(ip
->ip_conn
);
328 ip
->ip_parent
= NULL
;
332 static struct inode
*dlmfs_get_root_inode(struct super_block
*sb
)
334 struct inode
*inode
= new_inode(sb
);
335 umode_t mode
= S_IFDIR
| 0755;
338 inode
->i_ino
= get_next_ino();
339 inode_init_owner(&nop_mnt_idmap
, inode
, NULL
, mode
);
340 simple_inode_init_ts(inode
);
343 inode
->i_fop
= &simple_dir_operations
;
344 inode
->i_op
= &dlmfs_root_inode_operations
;
350 static struct inode
*dlmfs_get_inode(struct inode
*parent
,
351 struct dentry
*dentry
,
354 struct super_block
*sb
= parent
->i_sb
;
355 struct inode
* inode
= new_inode(sb
);
356 struct dlmfs_inode_private
*ip
;
361 inode
->i_ino
= get_next_ino();
362 inode_init_owner(&nop_mnt_idmap
, inode
, parent
, mode
);
363 simple_inode_init_ts(inode
);
366 ip
->ip_conn
= DLMFS_I(parent
)->ip_conn
;
368 switch (mode
& S_IFMT
) {
370 /* for now we don't support anything other than
371 * directories and regular files. */
375 inode
->i_op
= &dlmfs_file_inode_operations
;
376 inode
->i_fop
= &dlmfs_file_operations
;
378 i_size_write(inode
, DLM_LVB_LEN
);
380 user_dlm_lock_res_init(&ip
->ip_lockres
, dentry
);
382 /* released at clear_inode time, this insures that we
383 * get to drop the dlm reference on each lock *before*
384 * we call the unregister code for releasing parent
386 ip
->ip_parent
= igrab(parent
);
387 BUG_ON(!ip
->ip_parent
);
390 inode
->i_op
= &dlmfs_dir_inode_operations
;
391 inode
->i_fop
= &simple_dir_operations
;
393 /* directory inodes start off with i_nlink ==
394 * 2 (for "." entry) */
402 * File creation. Allocate an inode, and we're done..
405 static int dlmfs_mkdir(struct mnt_idmap
* idmap
,
407 struct dentry
* dentry
,
411 struct inode
*inode
= NULL
;
412 const struct qstr
*domain
= &dentry
->d_name
;
413 struct dlmfs_inode_private
*ip
;
414 struct ocfs2_cluster_connection
*conn
;
416 mlog(0, "mkdir %.*s\n", domain
->len
, domain
->name
);
418 /* verify that we have a proper domain */
419 if (domain
->len
>= GROUP_NAME_MAX
) {
421 mlog(ML_ERROR
, "invalid domain name for directory.\n");
425 inode
= dlmfs_get_inode(dir
, dentry
, mode
| S_IFDIR
);
434 conn
= user_dlm_register(domain
);
436 status
= PTR_ERR(conn
);
437 mlog(ML_ERROR
, "Error %d could not register domain \"%.*s\"\n",
438 status
, domain
->len
, domain
->name
);
444 d_instantiate(dentry
, inode
);
445 dget(dentry
); /* Extra count - pin the dentry in core */
454 static int dlmfs_create(struct mnt_idmap
*idmap
,
456 struct dentry
*dentry
,
462 const struct qstr
*name
= &dentry
->d_name
;
464 mlog(0, "create %.*s\n", name
->len
, name
->name
);
466 /* verify name is valid and doesn't contain any dlm reserved
468 if (name
->len
>= USER_DLM_LOCK_ID_MAX_LEN
||
469 name
->name
[0] == '$') {
471 mlog(ML_ERROR
, "invalid lock name, %.*s\n", name
->len
,
476 inode
= dlmfs_get_inode(dir
, dentry
, mode
| S_IFREG
);
483 d_instantiate(dentry
, inode
);
484 dget(dentry
); /* Extra count - pin the dentry in core */
489 static int dlmfs_unlink(struct inode
*dir
,
490 struct dentry
*dentry
)
493 struct inode
*inode
= d_inode(dentry
);
495 mlog(0, "unlink inode %lu\n", inode
->i_ino
);
497 /* if there are no current holders, or none that are waiting
498 * to acquire a lock, this basically destroys our lockres. */
499 status
= user_dlm_destroy_lock(&DLMFS_I(inode
)->ip_lockres
);
501 mlog(ML_ERROR
, "unlink %pd, error %d from destroy\n",
505 status
= simple_unlink(dir
, dentry
);
510 static int dlmfs_fill_super(struct super_block
*sb
, struct fs_context
*fc
)
512 sb
->s_maxbytes
= MAX_LFS_FILESIZE
;
513 sb
->s_blocksize
= PAGE_SIZE
;
514 sb
->s_blocksize_bits
= PAGE_SHIFT
;
515 sb
->s_magic
= DLMFS_MAGIC
;
516 sb
->s_op
= &dlmfs_ops
;
517 sb
->s_root
= d_make_root(dlmfs_get_root_inode(sb
));
523 static const struct file_operations dlmfs_file_operations
= {
524 .open
= dlmfs_file_open
,
525 .release
= dlmfs_file_release
,
526 .poll
= dlmfs_file_poll
,
527 .read
= dlmfs_file_read
,
528 .write
= dlmfs_file_write
,
529 .llseek
= default_llseek
,
532 static const struct inode_operations dlmfs_dir_inode_operations
= {
533 .create
= dlmfs_create
,
534 .lookup
= simple_lookup
,
535 .unlink
= dlmfs_unlink
,
538 /* this way we can restrict mkdir to only the toplevel of the fs. */
539 static const struct inode_operations dlmfs_root_inode_operations
= {
540 .lookup
= simple_lookup
,
541 .mkdir
= dlmfs_mkdir
,
542 .rmdir
= simple_rmdir
,
545 static const struct super_operations dlmfs_ops
= {
546 .statfs
= simple_statfs
,
547 .alloc_inode
= dlmfs_alloc_inode
,
548 .free_inode
= dlmfs_free_inode
,
549 .evict_inode
= dlmfs_evict_inode
,
550 .drop_inode
= generic_delete_inode
,
553 static const struct inode_operations dlmfs_file_inode_operations
= {
554 .getattr
= simple_getattr
,
555 .setattr
= dlmfs_file_setattr
,
558 static int dlmfs_get_tree(struct fs_context
*fc
)
560 return get_tree_nodev(fc
, dlmfs_fill_super
);
563 static const struct fs_context_operations dlmfs_context_ops
= {
564 .get_tree
= dlmfs_get_tree
,
567 static int dlmfs_init_fs_context(struct fs_context
*fc
)
569 fc
->ops
= &dlmfs_context_ops
;
574 static struct file_system_type dlmfs_fs_type
= {
575 .owner
= THIS_MODULE
,
576 .name
= "ocfs2_dlmfs",
577 .kill_sb
= kill_litter_super
,
578 .init_fs_context
= dlmfs_init_fs_context
,
580 MODULE_ALIAS_FS("ocfs2_dlmfs");
582 static int __init
init_dlmfs_fs(void)
585 int cleanup_inode
= 0, cleanup_worker
= 0;
587 dlmfs_inode_cache
= kmem_cache_create("dlmfs_inode_cache",
588 sizeof(struct dlmfs_inode_private
),
589 0, (SLAB_HWCACHE_ALIGN
|SLAB_RECLAIM_ACCOUNT
|
592 if (!dlmfs_inode_cache
) {
598 user_dlm_worker
= alloc_workqueue("user_dlm", WQ_MEM_RECLAIM
, 0);
599 if (!user_dlm_worker
) {
605 user_dlm_set_locking_protocol();
606 status
= register_filesystem(&dlmfs_fs_type
);
610 kmem_cache_destroy(dlmfs_inode_cache
);
612 destroy_workqueue(user_dlm_worker
);
614 printk("OCFS2 User DLM kernel interface loaded\n");
618 static void __exit
exit_dlmfs_fs(void)
620 unregister_filesystem(&dlmfs_fs_type
);
622 destroy_workqueue(user_dlm_worker
);
625 * Make sure all delayed rcu free inodes are flushed before we
629 kmem_cache_destroy(dlmfs_inode_cache
);
633 MODULE_AUTHOR("Oracle");
634 MODULE_LICENSE("GPL");
635 MODULE_DESCRIPTION("OCFS2 DLM-Filesystem");
637 module_init(init_dlmfs_fs
)
638 module_exit(exit_dlmfs_fs
)