4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
21 /* Portions Copyright 2007 Shivakumar GN */
23 * Copyright 2008 Sun Microsystems, Inc. All rights reserved.
24 * Use is subject to license terms.
27 * Copyright (c) 2017 by Delphix. All rights reserved.
30 #include <sys/types.h>
31 #include <sys/cmn_err.h>
32 #include <sys/debug.h>
33 #include <sys/dirent.h>
36 #include <sys/mutex.h>
37 #include <sys/sysmacros.h>
38 #include <sys/systm.h>
39 #include <sys/sunddi.h>
41 #include <sys/vmsystm.h>
43 #include <sys/vnode.h>
46 #include <vm/seg_vn.h>
51 * Generic pseudo-filesystem routines.
53 * There are significant similarities between the implementation of certain file
54 * system entry points across different filesystems. While one could attempt to
55 * "choke up on the bat" and incorporate common functionality into a VOP
56 * preamble or postamble, such an approach is limited in the benefit it can
57 * provide. In this file we instead define a toolkit of routines which can be
58 * called from a filesystem (with in-kernel pseudo-filesystems being the focus
59 * of the exercise) in a more component-like fashion.
61 * There are three basic classes of routines:
63 * 1) Lowlevel support routines
65 * These routines are designed to play a support role for existing
66 * pseudo-filesystems (such as procfs). They simplify common tasks,
67 * without forcing the filesystem to hand over management to GFS. The
68 * routines covered are:
77 * 2) Complete GFS management
79 * These routines take a more active role in management of the
80 * pseudo-filesystem. They handle the relationship between vnode private
81 * data and VFS data, as well as the relationship between vnodes in the
82 * directory hierarchy.
84 * In order to use these interfaces, the first member of every private
85 * v_data must be a gfs_file_t or a gfs_dir_t. This hands over all control
102 * 3) Single File pseudo-filesystems
104 * This routine creates a rooted file to be overlayed ontop of another
105 * file in the physical filespace.
107 * Note that the parent is NULL (actually the vfs), but there is nothing
108 * technically keeping such a file from utilizing the "Complete GFS
109 * management" set of routines.
111 * gfs_root_create_file()
115 * Low level directory routines
117 * These routines provide some simple abstractions for reading directories.
118 * They are designed to be used by existing pseudo filesystems (namely procfs)
119 * that already have a complicated management infrastructure.
123 * gfs_get_parent_ino: used to obtain a parent inode number and the
124 * inode number of the given vnode in preparation for calling gfs_readdir_init.
127 gfs_get_parent_ino(vnode_t
*dvp
, cred_t
*cr
, caller_context_t
*ct
,
128 ino64_t
*pino
, ino64_t
*ino
)
131 gfs_dir_t
*dp
= dvp
->v_data
;
134 *ino
= dp
->gfsd_file
.gfs_ino
;
135 parent
= dp
->gfsd_file
.gfs_parent
;
137 if (parent
== NULL
) {
138 *pino
= *ino
; /* root of filesystem */
139 } else if (dvp
->v_flag
& V_XATTRDIR
) {
142 va
.va_mask
= AT_NODEID
;
143 error
= fop_getattr(parent
, &va
, 0, cr
, ct
);
146 *pino
= va
.va_nodeid
;
148 *pino
= ((gfs_file_t
*)(parent
->v_data
))->gfs_ino
;
155 * gfs_readdir_init: initiate a generic readdir
156 * st - a pointer to an uninitialized gfs_readdir_state_t structure
157 * name_max - the directory's maximum file name length
158 * ureclen - the exported file-space record length (1 for non-legacy FSs)
159 * uiop - the uiop passed to readdir
160 * parent - the parent directory's inode
161 * self - this directory's inode
162 * flags - flags from fop_readdir
164 * Returns 0 or a non-zero errno.
166 * Typical fop_readdir usage of gfs_readdir_*:
168 * if ((error = gfs_readdir_init(...)) != 0)
171 * while ((error = gfs_readdir_pred(..., &voffset)) != 0) {
172 * if (!consumer_entry_at(voffset))
173 * voffset = consumer_next_entry(voffset);
174 * if (consumer_eof(voffset)) {
178 * if ((error = gfs_readdir_emit(..., voffset,
179 * consumer_ino(voffset), consumer_name(voffset))) != 0)
182 * return (gfs_readdir_fini(..., error, eofp, eof));
184 * As you can see, a zero result from gfs_readdir_pred() or
185 * gfs_readdir_emit() indicates that processing should continue,
186 * whereas a non-zero result indicates that the loop should terminate.
187 * Most consumers need do nothing more than let gfs_readdir_fini()
188 * determine what the cause of failure was and return the appropriate
192 gfs_readdir_init(gfs_readdir_state_t
*st
, int name_max
, int ureclen
,
193 uio_t
*uiop
, ino64_t parent
, ino64_t self
, int flags
)
197 if (uiop
->uio_loffset
< 0 || uiop
->uio_resid
<= 0 ||
198 (uiop
->uio_loffset
% ureclen
) != 0)
201 st
->grd_ureclen
= ureclen
;
202 st
->grd_oresid
= uiop
->uio_resid
;
203 st
->grd_namlen
= name_max
;
204 if (flags
& V_RDDIR_ENTFLAGS
)
205 dirent_size
= EDIRENT_RECLEN(st
->grd_namlen
);
207 dirent_size
= DIRENT64_RECLEN(st
->grd_namlen
);
208 st
->grd_dirent
= kmem_zalloc(dirent_size
, KM_SLEEP
);
209 st
->grd_parent
= parent
;
211 st
->grd_flags
= flags
;
217 * gfs_readdir_emit_int: internal routine to emit directory entry
219 * st - the current readdir state, which must have d_ino/ed_ino
220 * and d_name/ed_name set
221 * uiop - caller-supplied uio pointer
222 * next - the offset of the next entry
225 gfs_readdir_emit_int(gfs_readdir_state_t
*st
, uio_t
*uiop
, offset_t next
)
231 if (st
->grd_flags
& V_RDDIR_ENTFLAGS
) {
232 edp
= st
->grd_dirent
;
233 reclen
= EDIRENT_RECLEN(strlen(edp
->ed_name
));
236 reclen
= DIRENT64_RECLEN(strlen(dp
->d_name
));
239 if (reclen
> uiop
->uio_resid
) {
241 * Error if no entries were returned yet
243 if (uiop
->uio_resid
== st
->grd_oresid
)
248 if (st
->grd_flags
& V_RDDIR_ENTFLAGS
) {
250 edp
->ed_reclen
= (ushort_t
)reclen
;
253 dp
->d_reclen
= (ushort_t
)reclen
;
256 if (uiomove((caddr_t
)st
->grd_dirent
, reclen
, UIO_READ
, uiop
))
259 uiop
->uio_loffset
= next
;
265 * gfs_readdir_emit: emit a directory entry
266 * voff - the virtual offset (obtained from gfs_readdir_pred)
267 * ino - the entry's inode
268 * name - the entry's name
269 * eflags - value for ed_eflags (if processing edirent_t)
271 * Returns a 0 on success, a non-zero errno on failure, or -1 if the
272 * readdir loop should terminate. A non-zero result (either errno or
273 * -1) from this function is typically passed directly to
274 * gfs_readdir_fini().
277 gfs_readdir_emit(gfs_readdir_state_t
*st
, uio_t
*uiop
, offset_t voff
,
278 ino64_t ino
, const char *name
, int eflags
)
280 offset_t off
= (voff
+ 2) * st
->grd_ureclen
;
282 if (st
->grd_flags
& V_RDDIR_ENTFLAGS
) {
283 edirent_t
*edp
= st
->grd_dirent
;
286 (void) strncpy(edp
->ed_name
, name
, st
->grd_namlen
);
287 edp
->ed_eflags
= eflags
;
289 dirent64_t
*dp
= st
->grd_dirent
;
292 (void) strncpy(dp
->d_name
, name
, st
->grd_namlen
);
296 * Inter-entry offsets are invalid, so we assume a record size of
297 * grd_ureclen and explicitly set the offset appropriately.
299 return (gfs_readdir_emit_int(st
, uiop
, off
+ st
->grd_ureclen
));
303 * gfs_readdir_emitn: like gfs_readdir_emit(), but takes an integer
304 * instead of a string for the entry's name.
307 gfs_readdir_emitn(gfs_readdir_state_t
*st
, uio_t
*uiop
, offset_t voff
,
308 ino64_t ino
, unsigned long num
)
313 return (gfs_readdir_emit(st
, uiop
, voff
, ino
, buf
, 0));
317 * gfs_readdir_pred: readdir loop predicate
318 * voffp - a pointer in which the next virtual offset should be stored
320 * Returns a 0 on success, a non-zero errno on failure, or -1 if the
321 * readdir loop should terminate. A non-zero result (either errno or
322 * -1) from this function is typically passed directly to
323 * gfs_readdir_fini().
326 gfs_readdir_pred(gfs_readdir_state_t
*st
, uio_t
*uiop
, offset_t
*voffp
)
332 if (uiop
->uio_resid
<= 0)
335 off
= uiop
->uio_loffset
/ st
->grd_ureclen
;
338 if ((error
= gfs_readdir_emit(st
, uiop
, voff
, st
->grd_self
,
341 } else if (off
== 1) {
342 if ((error
= gfs_readdir_emit(st
, uiop
, voff
, st
->grd_parent
,
354 * gfs_readdir_fini: generic readdir cleanup
355 * error - if positive, an error to return
356 * eofp - the eofp passed to readdir
357 * eof - the eof value
359 * Returns a 0 on success, a non-zero errno on failure. This result
360 * should be returned from readdir.
363 gfs_readdir_fini(gfs_readdir_state_t
*st
, int error
, int *eofp
, int eof
)
367 if (st
->grd_flags
& V_RDDIR_ENTFLAGS
)
368 dirent_size
= EDIRENT_RECLEN(st
->grd_namlen
);
370 dirent_size
= DIRENT64_RECLEN(st
->grd_namlen
);
371 kmem_free(st
->grd_dirent
, dirent_size
);
382 * Performs a basic check for "." and ".." directory entries.
385 gfs_lookup_dot(vnode_t
**vpp
, vnode_t
*dvp
, vnode_t
*pvp
, const char *nm
)
387 if (*nm
== '\0' || strcmp(nm
, ".") == 0) {
391 } else if (strcmp(nm
, "..") == 0) {
393 ASSERT(dvp
->v_flag
& VROOT
);
407 * gfs_file_create(): create a new GFS file
409 * size - size of private data structure (v_data)
410 * pvp - parent vnode (GFS directory)
411 * ops - vnode operations vector
413 * In order to use this interface, the parent vnode must have been created by
414 * gfs_dir_create(), and the private data stored in v_data must have a
415 * 'gfs_file_t' as its first field.
417 * Given these constraints, this routine will automatically:
419 * - Allocate v_data for the vnode
420 * - Initialize necessary fields in the vnode
424 gfs_file_create(size_t size
, struct vnode
*pvp
, const struct vnodeops
*ops
)
430 * Allocate vnode and internal data structure
432 fp
= kmem_zalloc(size
, KM_SLEEP
);
433 vp
= vn_alloc(KM_SLEEP
);
436 * Set up various pointers
439 fp
->gfs_parent
= pvp
;
442 fp
->gfs_type
= GFS_FILE
;
445 * Initialize vnode and hold parent.
449 VN_SET_VFS_TYPE_DEV(vp
, pvp
->v_vfsp
, VREG
, 0);
457 * gfs_dir_create: creates a new directory in the parent
459 * size - size of private data structure (v_data)
460 * pvp - parent vnode (GFS directory)
461 * ops - vnode operations vector
462 * entries - NULL-terminated list of static entries (if any)
463 * maxlen - maximum length of a directory entry
464 * readdir_cb - readdir callback (see gfs_dir_readdir)
465 * inode_cb - inode callback (see gfs_dir_readdir)
466 * lookup_cb - lookup callback (see gfs_dir_lookup)
468 * In order to use this function, the first member of the private vnode
469 * structure (v_data) must be a gfs_dir_t. For each directory, there are
470 * static entries, defined when the structure is initialized, and dynamic
471 * entries, retrieved through callbacks.
473 * If a directory has static entries, then it must supply a inode callback,
474 * which will compute the inode number based on the parent and the index.
475 * For a directory with dynamic entries, the caller must supply a readdir
476 * callback and a lookup callback. If a static lookup fails, we fall back to
477 * the supplied lookup callback, if any.
479 * This function also performs the same initialization as gfs_file_create().
482 gfs_dir_create(size_t struct_size
, struct vnode
*pvp
,
483 const struct vnodeops
*ops
,
484 gfs_dirent_t
*entries
, gfs_inode_cb inode_cb
, int maxlen
,
485 gfs_readdir_cb readdir_cb
, gfs_lookup_cb lookup_cb
)
491 vp
= gfs_file_create(struct_size
, pvp
, ops
);
495 dp
->gfsd_file
.gfs_type
= GFS_DIR
;
496 dp
->gfsd_maxlen
= maxlen
;
498 if (entries
!= NULL
) {
499 for (de
= entries
; de
->gfse_name
!= NULL
; de
++)
502 dp
->gfsd_static
= kmem_alloc(
503 dp
->gfsd_nstatic
* sizeof (gfs_dirent_t
), KM_SLEEP
);
504 bcopy(entries
, dp
->gfsd_static
,
505 dp
->gfsd_nstatic
* sizeof (gfs_dirent_t
));
508 dp
->gfsd_readdir
= readdir_cb
;
509 dp
->gfsd_lookup
= lookup_cb
;
510 dp
->gfsd_inode
= inode_cb
;
512 mutex_init(&dp
->gfsd_lock
, NULL
, MUTEX_DEFAULT
, NULL
);
518 * gfs_root_create(): create a root vnode for a GFS filesystem
520 * Similar to gfs_dir_create(), this creates a root vnode for a filesystem. The
521 * only difference is that it takes a vfs_t instead of a vnode_t as its parent.
524 gfs_root_create(size_t size
, struct vfs
*vfsp
,
525 const struct vnodeops
*ops
, ino64_t ino
,
526 gfs_dirent_t
*entries
, gfs_inode_cb inode_cb
, int maxlen
,
527 gfs_readdir_cb readdir_cb
, gfs_lookup_cb lookup_cb
)
529 struct vnode
*vp
= gfs_dir_create(size
, NULL
, ops
, entries
, inode_cb
,
530 maxlen
, readdir_cb
, lookup_cb
);
532 /* Manually set the inode */
533 ((gfs_file_t
*)vp
->v_data
)->gfs_ino
= ino
;
536 VN_SET_VFS_TYPE_DEV(vp
, vfsp
, VDIR
, 0);
537 vp
->v_flag
|= VROOT
| VNOCACHE
| VNOMAP
| VNOSWAP
| VNOMOUNT
;
543 * gfs_root_create_file(): create a root vnode for a GFS file as a filesystem
545 * Similar to gfs_root_create(), this creates a root vnode for a file to
546 * be the pseudo-filesystem.
549 gfs_root_create_file(size_t size
, struct vfs
*vfsp
,
550 const struct vnodeops
*ops
, ino64_t ino
)
552 struct vnode
*vp
= gfs_file_create(size
, NULL
, ops
);
554 ((gfs_file_t
*)vp
->v_data
)->gfs_ino
= ino
;
557 VN_SET_VFS_TYPE_DEV(vp
, vfsp
, VREG
, 0);
558 vp
->v_flag
|= VROOT
| VNOCACHE
| VNOMAP
| VNOSWAP
| VNOMOUNT
;
564 * gfs_file_inactive()
566 * Called from the fop_inactive() routine. If necessary, this routine will
567 * remove the given vnode from the parent directory and clean up any references
570 * If the vnode was not removed (due to a race with vget), then NULL is
571 * returned. Otherwise, a pointer to the private data is returned.
574 gfs_file_inactive(vnode_t
*vp
)
577 gfs_dirent_t
*ge
= NULL
;
578 gfs_file_t
*fp
= vp
->v_data
;
579 gfs_dir_t
*dp
= NULL
;
582 if (fp
->gfs_parent
== NULL
|| (vp
->v_flag
& V_XATTRDIR
))
585 dp
= fp
->gfs_parent
->v_data
;
588 * First, see if this vnode is cached in the parent.
593 * Find it in the set of static entries.
595 for (i
= 0; i
< dp
->gfsd_nstatic
; i
++) {
596 ge
= &dp
->gfsd_static
[i
];
598 if (ge
->gfse_vnode
== vp
)
603 * If 'ge' is NULL, then it is a dynamic entry.
608 if (vp
->v_flag
& V_XATTRDIR
) {
609 mutex_enter(&fp
->gfs_parent
->v_lock
);
611 mutex_enter(&vp
->v_lock
);
612 if (vp
->v_count
== 1) {
614 * Really remove this vnode
619 * If this was a statically cached entry, simply set the
620 * cached vnode to NULL.
622 ge
->gfse_vnode
= NULL
;
624 if (vp
->v_flag
& V_XATTRDIR
) {
625 fp
->gfs_parent
->v_xattrdir
= NULL
;
626 mutex_exit(&fp
->gfs_parent
->v_lock
);
628 mutex_exit(&vp
->v_lock
);
631 * Free vnode and release parent
633 if (fp
->gfs_parent
) {
637 VN_RELE(fp
->gfs_parent
);
639 ASSERT(vp
->v_vfsp
!= NULL
);
640 VFS_RELE(vp
->v_vfsp
);
646 mutex_exit(&vp
->v_lock
);
647 if (vp
->v_flag
& V_XATTRDIR
) {
648 mutex_exit(&fp
->gfs_parent
->v_lock
);
660 * Same as above, but for directories.
663 gfs_dir_inactive(vnode_t
*vp
)
667 ASSERT(vp
->v_type
== VDIR
);
669 if ((dp
= gfs_file_inactive(vp
)) != NULL
) {
670 mutex_destroy(&dp
->gfsd_lock
);
671 if (dp
->gfsd_nstatic
)
672 kmem_free(dp
->gfsd_static
,
673 dp
->gfsd_nstatic
* sizeof (gfs_dirent_t
));
680 * gfs_dir_lookup_dynamic()
682 * This routine looks up the provided name amongst the dynamic entries
683 * in the gfs directory and returns the corresponding vnode, if found.
685 * The gfs directory is expected to be locked by the caller prior to
686 * calling this function. The directory will be unlocked during the
687 * execution of this function, but will be locked upon return from the
688 * function. This function returns 0 on success, non-zero on error.
690 * The dynamic lookups are performed by invoking the lookup
691 * callback, which is passed to this function as the first argument.
692 * The arguments to the callback are:
694 * int gfs_lookup_cb(vnode_t *pvp, const char *nm, vnode_t **vpp, cred_t *cr,
695 * int flags, int *deflgs, pathname_t *rpnp);
699 * vpp - pointer to resulting vnode
700 * cr - pointer to cred
701 * flags - flags value from lookup request
702 * ignored here; currently only used to request
703 * insensitive lookups
704 * direntflgs - output parameter, directory entry flags
705 * ignored here; currently only used to indicate a lookup
706 * has more than one possible match when case is not considered
707 * realpnp - output parameter, real pathname
708 * ignored here; when lookup was performed case-insensitively,
709 * this field contains the "real" name of the file.
711 * Returns 0 on success, non-zero on error.
714 gfs_dir_lookup_dynamic(gfs_lookup_cb callback
, gfs_dir_t
*dp
,
715 const char *nm
, vnode_t
*dvp
, vnode_t
**vpp
, cred_t
*cr
, int flags
,
716 int *direntflags
, pathname_t
*realpnp
)
722 ASSERT(GFS_DIR_LOCKED(dp
));
725 * Drop the directory lock, as the lookup routine
726 * will need to allocate memory, or otherwise deadlock on this
730 ret
= callback(dvp
, nm
, vpp
, &ino
, cr
, flags
, direntflags
, realpnp
);
734 * The callback for extended attributes returns a vnode
735 * with v_data from an underlying fs.
737 if (ret
== 0 && !IS_XATTRDIR(dvp
)) {
738 fp
= (gfs_file_t
*)((*vpp
)->v_data
);
747 * gfs_dir_lookup_static()
749 * This routine looks up the provided name amongst the static entries
750 * in the gfs directory and returns the corresponding vnode, if found.
751 * The first argument to the function is a pointer to the comparison
752 * function this function should use to decide if names are a match.
754 * If a match is found, and GFS_CACHE_VNODE is set and the vnode
755 * exists, we simply return the existing vnode. Otherwise, we call
756 * the static entry's callback routine, caching the result if
757 * necessary. If the idx pointer argument is non-NULL, we use it to
758 * return the index of the matching static entry.
760 * The gfs directory is expected to be locked by the caller prior to calling
761 * this function. The directory may be unlocked during the execution of
762 * this function, but will be locked upon return from the function.
764 * This function returns 0 if a match is found, ENOENT if not.
767 gfs_dir_lookup_static(int (*compare
)(const char *, const char *),
768 gfs_dir_t
*dp
, const char *nm
, vnode_t
*dvp
, int *idx
,
769 vnode_t
**vpp
, pathname_t
*rpnp
)
775 ASSERT(GFS_DIR_LOCKED(dp
));
778 * Search static entries.
780 for (i
= 0; i
< dp
->gfsd_nstatic
; i
++) {
781 ge
= &dp
->gfsd_static
[i
];
783 if (compare(ge
->gfse_name
, nm
) == 0) {
785 (void) strlcpy(rpnp
->pn_buf
, ge
->gfse_name
,
788 if (ge
->gfse_vnode
) {
789 ASSERT(ge
->gfse_flags
& GFS_CACHE_VNODE
);
796 * We drop the directory lock, as the constructor will
797 * need to do KM_SLEEP allocations. If we return from
798 * the constructor only to find that a parallel
799 * operation has completed, and GFS_CACHE_VNODE is set
800 * for this entry, we discard the result in favor of
804 vp
= ge
->gfse_ctor(dvp
);
807 ((gfs_file_t
*)vp
->v_data
)->gfs_index
= i
;
809 /* Set the inode according to the callback. */
810 ((gfs_file_t
*)vp
->v_data
)->gfs_ino
=
811 dp
->gfsd_inode(dvp
, i
);
813 if (ge
->gfse_flags
& GFS_CACHE_VNODE
) {
814 if (ge
->gfse_vnode
== NULL
) {
818 * A parallel constructor beat us to it;
819 * return existing vnode. We have to be
820 * careful because we can't release the
821 * current vnode while holding the
822 * directory lock; its inactive routine
823 * will try to lock this directory.
849 * Looks up the given name in the directory and returns the corresponding
852 * First, we search statically defined entries, if any, with a call to
853 * gfs_dir_lookup_static(). If no static entry is found, and we have
854 * a callback function we try a dynamic lookup via gfs_dir_lookup_dynamic().
856 * This function returns 0 on success, non-zero on error.
859 gfs_dir_lookup(vnode_t
*dvp
, const char *nm
, vnode_t
**vpp
, cred_t
*cr
,
860 int flags
, int *direntflags
, pathname_t
*realpnp
)
862 gfs_dir_t
*dp
= dvp
->v_data
;
864 vnode_t
*dynvp
= NULL
;
866 int (*compare
)(const char *, const char *);
869 ASSERT(dvp
->v_type
== VDIR
);
871 if (gfs_lookup_dot(vpp
, dvp
, dp
->gfsd_file
.gfs_parent
, nm
) == 0)
874 casecheck
= (flags
& FIGNORECASE
) != 0 && direntflags
!= NULL
;
875 if (vfs_has_feature(dvp
->v_vfsp
, VFSFT_NOCASESENSITIVE
) ||
876 (flags
& FIGNORECASE
))
877 compare
= strcasecmp
;
883 error
= gfs_dir_lookup_static(compare
, dp
, nm
, dvp
, &idx
, &vp
, realpnp
);
885 if (vp
&& casecheck
) {
889 for (i
= idx
+ 1; i
< dp
->gfsd_nstatic
; i
++) {
890 ge
= &dp
->gfsd_static
[i
];
892 if (strcasecmp(ge
->gfse_name
, nm
) == 0) {
893 *direntflags
|= ED_CASE_CONFLICT
;
899 if ((error
|| casecheck
) && dp
->gfsd_lookup
)
900 error
= gfs_dir_lookup_dynamic(dp
->gfsd_lookup
, dp
, nm
, dvp
,
901 &dynvp
, cr
, flags
, direntflags
, vp
? NULL
: realpnp
);
904 /* static and dynamic entries are case-insensitive conflict */
906 *direntflags
|= ED_CASE_CONFLICT
;
908 } else if (vp
== NULL
) {
910 } else if (error
== ENOENT
) {
925 * gfs_dir_readdir: does a readdir() on the given directory
927 * dvp - directory vnode
928 * uiop - uio structure
930 * data - arbitrary data passed to readdir callback
932 * This routine does all the readdir() dirty work. Even so, the caller must
933 * supply two callbacks in order to get full compatibility.
935 * If the directory contains static entries, an inode callback must be
936 * specified. This avoids having to create every vnode and call fop_getattr()
937 * when reading the directory. This function has the following arguments:
939 * ino_t gfs_inode_cb(vnode_t *vp, int index);
941 * vp - vnode for the directory
942 * index - index in original gfs_dirent_t array
944 * Returns the inode number for the given entry.
946 * For directories with dynamic entries, a readdir callback must be provided.
947 * This is significantly more complex, thanks to the particulars of
950 * int gfs_readdir_cb(vnode_t *vp, void *dp, int *eofp,
951 * offset_t *off, offset_t *nextoff, void *data, int flags)
953 * vp - directory vnode
954 * dp - directory entry, sized according to maxlen given to
955 * gfs_dir_create(). callback must fill in d_name and
956 * d_ino (if a dirent64_t), or ed_name, ed_ino, and ed_eflags
957 * (if an edirent_t). edirent_t is used if V_RDDIR_ENTFLAGS
959 * eofp - callback must set to 1 when EOF has been reached
960 * off - on entry, the last offset read from the directory. Callback
961 * must set to the offset of the current entry, typically left
963 * nextoff - callback must set to offset of next entry. Typically
965 * data - caller-supplied data
966 * flags - fop_readdir flags
968 * Return 0 on success, or error on failure.
971 gfs_dir_readdir(vnode_t
*dvp
, uio_t
*uiop
, int *eofp
, void *data
, cred_t
*cr
,
972 caller_context_t
*ct
, int flags
)
974 gfs_readdir_state_t gstate
;
978 gfs_dir_t
*dp
= dvp
->v_data
;
980 error
= gfs_get_parent_ino(dvp
, cr
, ct
, &pino
, &ino
);
984 if ((error
= gfs_readdir_init(&gstate
, dp
->gfsd_maxlen
, 1, uiop
,
985 pino
, ino
, flags
)) != 0)
988 while ((error
= gfs_readdir_pred(&gstate
, uiop
, &off
)) == 0 &&
991 if (off
>= 0 && off
< dp
->gfsd_nstatic
) {
992 ino
= dp
->gfsd_inode(dvp
, off
);
994 if ((error
= gfs_readdir_emit(&gstate
, uiop
,
995 off
, ino
, dp
->gfsd_static
[off
].gfse_name
, 0))
999 } else if (dp
->gfsd_readdir
) {
1000 off
-= dp
->gfsd_nstatic
;
1002 if ((error
= dp
->gfsd_readdir(dvp
,
1003 gstate
.grd_dirent
, &eof
, &off
, &next
,
1004 data
, flags
)) != 0 || eof
)
1007 off
+= dp
->gfsd_nstatic
+ 2;
1008 next
+= dp
->gfsd_nstatic
+ 2;
1010 if ((error
= gfs_readdir_emit_int(&gstate
, uiop
,
1015 * Offset is beyond the end of the static entries, and
1016 * we have no dynamic entries. Set EOF.
1022 return (gfs_readdir_fini(&gstate
, error
, eofp
, eof
));
1027 * gfs_vop_lookup: fop_lookup() entry point
1029 * For use directly in vnode ops table. Given a GFS directory, calls
1030 * gfs_dir_lookup() as necessary.
1034 gfs_vop_lookup(vnode_t
*dvp
, char *nm
, vnode_t
**vpp
, pathname_t
*pnp
,
1035 int flags
, vnode_t
*rdir
, cred_t
*cr
, caller_context_t
*ct
,
1036 int *direntflags
, pathname_t
*realpnp
)
1038 return (gfs_dir_lookup(dvp
, nm
, vpp
, cr
, flags
, direntflags
, realpnp
));
1042 * gfs_vop_readdir: fop_readdir() entry point
1044 * For use directly in vnode ops table. Given a GFS directory, calls
1045 * gfs_dir_readdir() as necessary.
1049 gfs_vop_readdir(vnode_t
*vp
, uio_t
*uiop
, cred_t
*cr
, int *eofp
,
1050 caller_context_t
*ct
, int flags
)
1052 return (gfs_dir_readdir(vp
, uiop
, eofp
, NULL
, cr
, ct
, flags
));
1057 * gfs_vop_map: fop_map() entry point
1059 * Convenient routine for handling pseudo-files that wish to allow mmap() calls.
1060 * This function only works for readonly files, and uses the read function for
1061 * the vnode to fill in the data. The mapped data is immediately faulted in and
1062 * filled with the necessary data during this call; there are no getpage() or
1063 * putpage() routines.
1067 gfs_vop_map(vnode_t
*vp
, offset_t off
, struct as
*as
, caddr_t
*addrp
,
1068 size_t len
, uchar_t prot
, uchar_t maxprot
, uint_t flags
, cred_t
*cred
,
1069 caller_context_t
*ct
)
1072 ssize_t resid
= len
;
1075 * Check for bad parameters
1081 if (vp
->v_flag
& VNOMAP
)
1085 if ((long)off
< 0 || (long)(off
+ len
) < 0)
1087 if (vp
->v_type
!= VREG
)
1089 if ((prot
& (PROT_EXEC
| PROT_WRITE
)) != 0)
1093 * Find appropriate address if needed, otherwise clear address range.
1096 rv
= choose_addr(as
, addrp
, len
, off
, ADDR_VACALIGN
, flags
);
1105 rv
= as_map(as
, *addrp
, len
, segvn_create
, zfod_argsp
);
1111 * Fill with data from read()
1113 rv
= vn_rdwr(UIO_READ
, vp
, *addrp
, len
, off
, UIO_USERSPACE
,
1114 0, (rlim64_t
)0, cred
, &resid
);
1116 if (rv
== 0 && resid
!= 0)
1121 (void) as_unmap(as
, *addrp
, len
);
1129 * gfs_vop_inactive: fop_inactive() entry point
1131 * Given a vnode that is a GFS file or directory, call gfs_file_inactive() or
1132 * gfs_dir_inactive() as necessary, and kmem_free()s associated private data.
1136 gfs_vop_inactive(vnode_t
*vp
, cred_t
*cr
, caller_context_t
*ct
)
1138 gfs_file_t
*fp
= vp
->v_data
;
1141 if (fp
->gfs_type
== GFS_DIR
)
1142 data
= gfs_dir_inactive(vp
);
1144 data
= gfs_file_inactive(vp
);
1147 kmem_free(data
, fp
->gfs_size
);