Current Path : /usr/src/sys/fs/coda/ |
FreeBSD hs32.drive.ne.jp 9.1-RELEASE FreeBSD 9.1-RELEASE #1: Wed Jan 14 12:18:08 JST 2015 root@hs32.drive.ne.jp:/sys/amd64/compile/hs32 amd64 |
Current File : //usr/src/sys/fs/coda/coda_vnops.c |
/*- * Coda: an Experimental Distributed File System * Release 3.1 * * Copyright (c) 1987-1998 Carnegie Mellon University * All Rights Reserved * * Permission to use, copy, modify and distribute this software and its * documentation is hereby granted, provided that both the copyright * notice and this permission notice appear in all copies of the * software, derivative works or modified versions, and any portions * thereof, and that both notices appear in supporting documentation, and * that credit is given to Carnegie Mellon University in all documents * and publicity pertaining to direct or indirect use of this code or its * derivatives. * * CODA IS AN EXPERIMENTAL SOFTWARE SYSTEM AND IS KNOWN TO HAVE BUGS, * SOME OF WHICH MAY HAVE SERIOUS CONSEQUENCES. CARNEGIE MELLON ALLOWS * FREE USE OF THIS SOFTWARE IN ITS "AS IS" CONDITION. CARNEGIE MELLON * DISCLAIMS ANY LIABILITY OF ANY KIND FOR ANY DAMAGES WHATSOEVER * RESULTING DIRECTLY OR INDIRECTLY FROM THE USE OF THIS SOFTWARE OR OF * ANY DERIVATIVE WORK. * * Carnegie Mellon encourages users of this software to return any * improvements or extensions that they make, and to grant Carnegie * Mellon the rights to redistribute these changes without encumbrance. * * @(#) src/sys/coda/coda_vnops.c,v 1.1.1.1 1998/08/29 21:14:52 rvb Exp $ */ /* * Mach Operating System * Copyright (c) 1990 Carnegie-Mellon University * Copyright (c) 1989 Carnegie-Mellon University * All rights reserved. The CMU software License Agreement specifies * the terms and conditions for use and redistribution. */ /* * This code was written for the Coda filesystem at Carnegie Mellon * University. Contributers include David Steere, James Kistler, and * M. Satyanarayanan. */ #include <sys/cdefs.h> __FBSDID("$FreeBSD: release/9.1.0/sys/fs/coda/coda_vnops.c 215548 2010-11-19 21:17:34Z kib $"); #include <sys/param.h> #include <sys/systm.h> #include <sys/acct.h> #include <sys/errno.h> #include <sys/fcntl.h> #include <sys/kernel.h> #include <sys/lock.h> #include <sys/malloc.h> #include <sys/file.h> /* Must come after sys/malloc.h */ #include <sys/mount.h> #include <sys/mutex.h> #include <sys/namei.h> #include <sys/proc.h> #include <sys/uio.h> #include <sys/unistd.h> #include <vm/vm.h> #include <vm/vm_object.h> #include <vm/vm_extern.h> #include <fs/coda/coda.h> #include <fs/coda/cnode.h> #include <fs/coda/coda_vnops.h> #include <fs/coda/coda_venus.h> #include <fs/coda/coda_opstats.h> #include <fs/coda/coda_subr.h> #include <fs/coda/coda_pioctl.h> /* * These flags select various performance enhancements. */ static int coda_attr_cache = 1; /* Set to cache attributes. */ static int coda_symlink_cache = 1; /* Set to cache symbolic links. */ static int coda_access_cache = 1; /* Set to cache some access checks. */ /* * Structure to keep track of vfs calls. */ static struct coda_op_stats coda_vnodeopstats[CODA_VNODEOPS_SIZE]; #define MARK_ENTRY(op) (coda_vnodeopstats[op].entries++) #define MARK_INT_SAT(op) (coda_vnodeopstats[op].sat_intrn++) #define MARK_INT_FAIL(op) (coda_vnodeopstats[op].unsat_intrn++) #define MARK_INT_GEN(op) (coda_vnodeopstats[op].gen_intrn++) /* * What we are delaying for in printf. */ int coda_printf_delay = 0; /* In microseconds */ int coda_vnop_print_entry = 0; static int coda_lockdebug = 0; /* * Some FreeBSD details: * * codadev_modevent is called at boot time or module load time. */ #define ENTRY do { \ if (coda_vnop_print_entry) \ myprintf(("Entered %s\n", __func__)); \ } while (0) /* * Definition of the vnode operation vector. */ struct vop_vector coda_vnodeops = { .vop_default = &default_vnodeops, .vop_cachedlookup = coda_lookup, /* uncached lookup */ .vop_lookup = vfs_cache_lookup, /* namecache lookup */ .vop_create = coda_create, /* create */ .vop_open = coda_open, /* open */ .vop_close = coda_close, /* close */ .vop_access = coda_access, /* access */ .vop_getattr = coda_getattr, /* getattr */ .vop_setattr = coda_setattr, /* setattr */ .vop_read = coda_read, /* read */ .vop_write = coda_write, /* write */ .vop_ioctl = coda_ioctl, /* ioctl */ .vop_fsync = coda_fsync, /* fsync */ .vop_remove = coda_remove, /* remove */ .vop_link = coda_link, /* link */ .vop_rename = coda_rename, /* rename */ .vop_mkdir = coda_mkdir, /* mkdir */ .vop_rmdir = coda_rmdir, /* rmdir */ .vop_symlink = coda_symlink, /* symlink */ .vop_readdir = coda_readdir, /* readdir */ .vop_readlink = coda_readlink, /* readlink */ .vop_inactive = coda_inactive, /* inactive */ .vop_reclaim = coda_reclaim, /* reclaim */ .vop_lock1 = coda_lock, /* lock */ .vop_unlock = coda_unlock, /* unlock */ .vop_bmap = VOP_EOPNOTSUPP, /* bmap */ .vop_print = VOP_NULL, /* print */ .vop_islocked = coda_islocked, /* islocked */ .vop_pathconf = coda_pathconf, /* pathconf */ .vop_poll = vop_stdpoll, .vop_getpages = vop_stdgetpages, /* pager intf.*/ .vop_putpages = vop_stdputpages, /* pager intf.*/ .vop_getwritemount = vop_stdgetwritemount, #if 0 /* missing */ .vop_cachedlookup = ufs_lookup, .vop_whiteout = ufs_whiteout, #endif }; static void coda_print_vattr(struct vattr *attr); int coda_vnodeopstats_init(void) { int i; for(i=0; i<CODA_VNODEOPS_SIZE; i++) { coda_vnodeopstats[i].opcode = i; coda_vnodeopstats[i].entries = 0; coda_vnodeopstats[i].sat_intrn = 0; coda_vnodeopstats[i].unsat_intrn = 0; coda_vnodeopstats[i].gen_intrn = 0; } return (0); } /* * coda_open calls Venus which returns an open file descriptor the cache file * holding the data. We get the vnode while we are still in the context of * the venus process in coda_psdev.c. This vnode is then passed back to the * caller and opened. */ int coda_open(struct vop_open_args *ap) { /* * FreeBSD can pass the O_EXCL flag in mode, even though the check * has already happened. Venus defensively assumes that if open is * passed the EXCL, it must be a bug. We strip the flag here. */ /* true args */ struct vnode **vpp = &(ap->a_vp); struct cnode *cp = VTOC(*vpp); int flag = ap->a_mode & (~O_EXCL); struct ucred *cred = ap->a_cred; struct thread *td = ap->a_td; /* locals */ int error; struct vnode *vp; MARK_ENTRY(CODA_OPEN_STATS); /* * Check for open of control file. */ if (IS_CTL_VP(*vpp)) { /* XXX */ /* if (WRITEABLE(flag)) */ if (flag & (FWRITE | O_TRUNC | O_CREAT | O_EXCL)) { MARK_INT_FAIL(CODA_OPEN_STATS); return (EACCES); } MARK_INT_SAT(CODA_OPEN_STATS); return (0); } error = venus_open(vtomi((*vpp)), &cp->c_fid, flag, cred, td->td_proc, &vp); if (error) return (error); CODADEBUG(CODA_OPEN, myprintf(("open: vp %p result %d\n", vp, error));); /* * Save the vnode pointer for the cache file. */ if (cp->c_ovp == NULL) { cp->c_ovp = vp; } else { if (cp->c_ovp != vp) panic("coda_open: cp->c_ovp != ITOV(ip)"); } cp->c_ocount++; /* * Flush the attribute cached if writing the file. */ if (flag & FWRITE) { cp->c_owrite++; cp->c_flags &= ~C_VATTR; } /* * Open the cache file. */ vn_lock(vp, LK_EXCLUSIVE | LK_RETRY); error = VOP_OPEN(vp, flag, cred, td, NULL); if (error) { VOP_UNLOCK(vp, 0); printf("coda_open: VOP_OPEN on container failed %d\n", error); return (error); } (*vpp)->v_object = vp->v_object; VOP_UNLOCK(vp, 0); return (0); } /* * Close the cache file used for I/O and notify Venus. */ int coda_close(struct vop_close_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); int flag = ap->a_fflag; struct ucred *cred = ap->a_cred; struct thread *td = ap->a_td; /* locals */ int error; MARK_ENTRY(CODA_CLOSE_STATS); /* * Check for close of control file. */ if (IS_CTL_VP(vp)) { MARK_INT_SAT(CODA_CLOSE_STATS); return (0); } if (cp->c_ovp) { vn_lock(cp->c_ovp, LK_EXCLUSIVE | LK_RETRY); /* Do errors matter here? */ VOP_CLOSE(cp->c_ovp, flag, cred, td); vput(cp->c_ovp); } #ifdef CODA_VERBOSE else printf("coda_close: NO container vp %p/cp %p\n", vp, cp); #endif if (--cp->c_ocount == 0) cp->c_ovp = NULL; /* * File was opened for write. */ if (flag & FWRITE) --cp->c_owrite; if (!IS_UNMOUNTING(cp)) error = venus_close(vtomi(vp), &cp->c_fid, flag, cred, td->td_proc); else error = ENODEV; CODADEBUG(CODA_CLOSE, myprintf(("close: result %d\n",error));); return (error); } int coda_read(struct vop_read_args *ap) { ENTRY; return (coda_rdwr(ap->a_vp, ap->a_uio, UIO_READ, ap->a_ioflag, ap->a_cred, ap->a_uio->uio_td)); } int coda_write(struct vop_write_args *ap) { ENTRY; return (coda_rdwr(ap->a_vp, ap->a_uio, UIO_WRITE, ap->a_ioflag, ap->a_cred, ap->a_uio->uio_td)); } int coda_rdwr(struct vnode *vp, struct uio *uiop, enum uio_rw rw, int ioflag, struct ucred *cred, struct thread *td) { /* upcall decl */ /* NOTE: container file operation!!! */ /* locals */ struct cnode *cp = VTOC(vp); struct vnode *cfvp = cp->c_ovp; int opened_internally = 0; int error = 0; MARK_ENTRY(CODA_RDWR_STATS); CODADEBUG(CODA_RDWR, myprintf(("coda_rdwr(%d, %p, %zd, %lld, %d)\n", rw, (void *)uiop->uio_iov->iov_base, uiop->uio_resid, (long long)uiop->uio_offset, uiop->uio_segflg));); /* * Check for rdwr of control object. */ if (IS_CTL_VP(vp)) { MARK_INT_FAIL(CODA_RDWR_STATS); return (EINVAL); } /* * If file is not already open this must be a page {read,write} * request and we should open it internally. */ if (cfvp == NULL) { opened_internally = 1; MARK_INT_GEN(CODA_OPEN_STATS); error = VOP_OPEN(vp, (rw == UIO_READ ? FREAD : FWRITE), cred, td, NULL); #ifdef CODA_VERBOSE printf("coda_rdwr: Internally Opening %p\n", vp); #endif if (error) { printf("coda_rdwr: VOP_OPEN on container failed " "%d\n", error); return (error); } cfvp = cp->c_ovp; } /* * Have UFS handle the call. */ CODADEBUG(CODA_RDWR, myprintf(("indirect rdwr: fid = %s, refcnt = " "%d\n", coda_f2s(&cp->c_fid), CTOV(cp)->v_usecount));); vn_lock(cfvp, LK_EXCLUSIVE | LK_RETRY); if (rw == UIO_READ) { error = VOP_READ(cfvp, uiop, ioflag, cred); } else { error = VOP_WRITE(cfvp, uiop, ioflag, cred); /* * ufs_write updates the vnode_pager_setsize for the * vnode/object. * * XXX: Since we now share vm objects between layers, this is * probably unnecessary. */ { struct vattr attr; if (VOP_GETATTR(cfvp, &attr, cred) == 0) vnode_pager_setsize(vp, attr.va_size); } } VOP_UNLOCK(cfvp, 0); if (error) MARK_INT_FAIL(CODA_RDWR_STATS); else MARK_INT_SAT(CODA_RDWR_STATS); /* * Do an internal close if necessary. */ if (opened_internally) { MARK_INT_GEN(CODA_CLOSE_STATS); (void)VOP_CLOSE(vp, (rw == UIO_READ ? FREAD : FWRITE), cred, td); } /* * Invalidate cached attributes if writing. */ if (rw == UIO_WRITE) cp->c_flags &= ~C_VATTR; return (error); } int coda_ioctl(struct vop_ioctl_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; int com = ap->a_command; caddr_t data = ap->a_data; int flag = ap->a_fflag; struct ucred *cred = ap->a_cred; struct thread *td = ap->a_td; /* locals */ int error; struct vnode *tvp; struct nameidata ndp; struct PioctlData *iap = (struct PioctlData *)data; MARK_ENTRY(CODA_IOCTL_STATS); CODADEBUG(CODA_IOCTL, myprintf(("in coda_ioctl on %s\n", iap->path));); /* * Don't check for operation on a dying object, for ctlvp it * shouldn't matter. * * Must be control object to succeed. */ if (!IS_CTL_VP(vp)) { MARK_INT_FAIL(CODA_IOCTL_STATS); CODADEBUG(CODA_IOCTL, myprintf(("coda_ioctl error: vp != " "ctlvp"));); return (EOPNOTSUPP); } /* * Look up the pathname. * * Should we use the name cache here? It would get it from lookupname * sooner or later anyway, right? */ NDINIT(&ndp, LOOKUP, (iap->follow ? FOLLOW : NOFOLLOW), UIO_USERSPACE, iap->path, td); error = namei(&ndp); tvp = ndp.ni_vp; if (error) { MARK_INT_FAIL(CODA_IOCTL_STATS); CODADEBUG(CODA_IOCTL, myprintf(("coda_ioctl error: lookup " "returns %d\n", error));); return (error); } /* * Make sure this is a coda style cnode, but it may be a different * vfsp. */ if (tvp->v_op != &coda_vnodeops) { vrele(tvp); NDFREE(&ndp, NDF_ONLY_PNBUF); MARK_INT_FAIL(CODA_IOCTL_STATS); CODADEBUG(CODA_IOCTL, myprintf(("coda_ioctl error: %s not a coda object\n", iap->path));); return (EINVAL); } if (iap->vi.in_size > VC_MAXDATASIZE || iap->vi.out_size > VC_MAXDATASIZE) { NDFREE(&ndp, 0); return (EINVAL); } error = venus_ioctl(vtomi(tvp), &((VTOC(tvp))->c_fid), com, flag, data, cred, td->td_proc); if (error) MARK_INT_FAIL(CODA_IOCTL_STATS); else CODADEBUG(CODA_IOCTL, myprintf(("Ioctl returns %d \n", error));); vrele(tvp); NDFREE(&ndp, NDF_ONLY_PNBUF); return (error); } /* * To reduce the cost of a user-level venus;we cache attributes in the * kernel. Each cnode has storage allocated for an attribute. If c_vattr is * valid, return a reference to it. Otherwise, get the attributes from venus * and store them in the cnode. There is some question if this method is a * security leak. But I think that in order to make this call, the user must * have done a lookup and opened the file, and therefore should already have * access. */ int coda_getattr(struct vop_getattr_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); struct vattr *vap = ap->a_vap; struct ucred *cred = ap->a_cred; /* locals */ struct vnode *convp; int error, size; MARK_ENTRY(CODA_GETATTR_STATS); if (IS_UNMOUNTING(cp)) return (ENODEV); /* * Check for getattr of control object. */ if (IS_CTL_VP(vp)) { MARK_INT_FAIL(CODA_GETATTR_STATS); return (ENOENT); } /* * Check to see if the attributes have already been cached. */ if (VALID_VATTR(cp)) { CODADEBUG(CODA_GETATTR, myprintf(("attr cache hit: %s\n", coda_f2s(&cp->c_fid)));); CODADEBUG(CODA_GETATTR, if (!(codadebug & ~CODA_GETATTR)) coda_print_vattr(&cp->c_vattr);); *vap = cp->c_vattr; MARK_INT_SAT(CODA_GETATTR_STATS); return (0); } error = venus_getattr(vtomi(vp), &cp->c_fid, cred, vap); if (!error) { CODADEBUG(CODA_GETATTR, myprintf(("getattr miss %s: result " "%d\n", coda_f2s(&cp->c_fid), error));); CODADEBUG(CODA_GETATTR, if (!(codadebug & ~CODA_GETATTR)) coda_print_vattr(vap);); /* * XXX: Since we now share vm objects between layers, this is * probably unnecessary. */ size = vap->va_size; convp = cp->c_ovp; if (convp != NULL) vnode_pager_setsize(convp, size); /* * If not open for write, store attributes in cnode. */ if ((cp->c_owrite == 0) && (coda_attr_cache)) { cp->c_vattr = *vap; cp->c_flags |= C_VATTR; } } return (error); } int coda_setattr(struct vop_setattr_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); struct vattr *vap = ap->a_vap; struct ucred *cred = ap->a_cred; /* locals */ struct vnode *convp; int error, size; MARK_ENTRY(CODA_SETATTR_STATS); /* * Check for setattr of control object. */ if (IS_CTL_VP(vp)) { MARK_INT_FAIL(CODA_SETATTR_STATS); return (ENOENT); } if (codadebug & CODADBGMSK(CODA_SETATTR)) coda_print_vattr(vap); error = venus_setattr(vtomi(vp), &cp->c_fid, vap, cred); if (!error) cp->c_flags &= ~(C_VATTR | C_ACCCACHE); /* * XXX: Since we now share vm objects between layers, this is * probably unnecessary. * * XXX: Shouldn't we only be doing this "set" if C_VATTR remains * valid after venus_setattr()? */ size = vap->va_size; convp = cp->c_ovp; if (size != VNOVAL && convp != NULL) vnode_pager_setsize(convp, size); CODADEBUG(CODA_SETATTR, myprintf(("setattr %d\n", error));); return (error); } int coda_access(struct vop_access_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); accmode_t accmode = ap->a_accmode; struct ucred *cred = ap->a_cred; struct thread *td = ap->a_td; /* locals */ int error; MARK_ENTRY(CODA_ACCESS_STATS); /* * Check for access of control object. Only read access is allowed * on it. */ if (IS_CTL_VP(vp)) { /* * Bogus hack - all will be marked as successes. */ MARK_INT_SAT(CODA_ACCESS_STATS); return (((accmode & VREAD) && !(accmode & (VWRITE | VEXEC))) ? 0 : EACCES); } /* * We maintain a one-entry LRU positive access cache with each cnode. * In principle we could also track negative results, and for more * than one uid, but we don't yet. Venus is responsible for * invalidating this cache as required. */ if (coda_access_cache && VALID_ACCCACHE(cp) && (cred->cr_uid == cp->c_cached_uid) && (accmode & cp->c_cached_mode) == accmode) { MARK_INT_SAT(CODA_ACCESS_STATS); return (0); } error = venus_access(vtomi(vp), &cp->c_fid, accmode, cred, td->td_proc); if (error == 0 && coda_access_cache) { /*- * When we have a new successful request, we consider three * cases: * * - No initialized access cache, in which case cache the * result. * - Cached result for a different user, in which case we * replace the entry. * - Cached result for the same user, in which case we add * any newly granted rights to the cached mode. * * XXXRW: If we ever move to something more interesting than * uid-based token lookup, we'll need to change this. */ cp->c_flags |= C_ACCCACHE; if (cp->c_cached_uid != cred->cr_uid) { cp->c_cached_mode = accmode; cp->c_cached_uid = cred->cr_uid; } else cp->c_cached_mode |= accmode; } return (error); } int coda_readlink(struct vop_readlink_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); struct uio *uiop = ap->a_uio; struct ucred *cred = ap->a_cred; struct thread *td = ap->a_uio->uio_td; /* locals */ int error; char *str; int len; MARK_ENTRY(CODA_READLINK_STATS); /* * Check for readlink of control object. */ if (IS_CTL_VP(vp)) { MARK_INT_FAIL(CODA_READLINK_STATS); return (ENOENT); } if ((coda_symlink_cache) && (VALID_SYMLINK(cp))) { /* * Symlink was cached. */ uiop->uio_rw = UIO_READ; error = uiomove(cp->c_symlink, (int)cp->c_symlen, uiop); if (error) MARK_INT_FAIL(CODA_READLINK_STATS); else MARK_INT_SAT(CODA_READLINK_STATS); return (error); } error = venus_readlink(vtomi(vp), &cp->c_fid, cred, td != NULL ? td->td_proc : NULL, &str, &len); if (!error) { uiop->uio_rw = UIO_READ; error = uiomove(str, len, uiop); if (coda_symlink_cache) { cp->c_symlink = str; cp->c_symlen = len; cp->c_flags |= C_SYMLINK; } else CODA_FREE(str, len); } CODADEBUG(CODA_READLINK, myprintf(("in readlink result %d\n", error));); return (error); } int coda_fsync(struct vop_fsync_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); struct thread *td = ap->a_td; /* locals */ struct vnode *convp = cp->c_ovp; int error; MARK_ENTRY(CODA_FSYNC_STATS); /* * Check for fsync on an unmounting object. * * XXX: Is this comment true on FreeBSD? It seems likely, since * unmounting is fairly non-atomic. * * The NetBSD kernel, in it's infinite wisdom, can try to fsync after * an unmount has been initiated. This is a Bad Thing, which we have * to avoid. Not a legitimate failure for stats. */ if (IS_UNMOUNTING(cp)) return (ENODEV); /* * Check for fsync of control object. */ if (IS_CTL_VP(vp)) { MARK_INT_SAT(CODA_FSYNC_STATS); return (0); } if (convp != NULL) { vn_lock(convp, LK_EXCLUSIVE | LK_RETRY); VOP_FSYNC(convp, MNT_WAIT, td); VOP_UNLOCK(convp, 0); } /* * We see fsyncs with usecount == 1 then usecount == 0. For now we * ignore them. */ #if 0 VI_LOCK(vp); if (!vp->v_usecount) { printf("coda_fsync on vnode %p with %d usecount. " "c_flags = %x (%x)\n", vp, vp->v_usecount, cp->c_flags, cp->c_flags&C_PURGING); } VI_UNLOCK(vp); #endif /* * We can expect fsync on any vnode at all if venus is purging it. * Venus can't very well answer the fsync request, now can it? * Hopefully, it won't have to, because hopefully, venus preserves * the (possibly untrue) invariant that it never purges an open * vnode. Hopefully. */ if (cp->c_flags & C_PURGING) return (0); /* XXX: needs research */ return (0); error = venus_fsync(vtomi(vp), &cp->c_fid, td->td_proc); CODADEBUG(CODA_FSYNC, myprintf(("in fsync result %d\n", error));); return (error); } int coda_inactive(struct vop_inactive_args *ap) { /* * XXX - at the moment, inactive doesn't look at cred, and doesn't * have a proc pointer. Oops. */ /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); struct ucred *cred __attribute__((unused)) = NULL; struct thread *td __attribute__((unused)) = curthread; /* upcall decl */ /* locals */ /* * We don't need to send inactive to venus - DCS. */ MARK_ENTRY(CODA_INACTIVE_STATS); CODADEBUG(CODA_INACTIVE, myprintf(("in inactive, %s, vfsp %p\n", coda_f2s(&cp->c_fid), vp->v_mount));); vp->v_object = NULL; /* * If an array has been allocated to hold the symlink, deallocate it. */ if ((coda_symlink_cache) && (VALID_SYMLINK(cp))) { if (cp->c_symlink == NULL) panic("coda_inactive: null symlink pointer in cnode"); CODA_FREE(cp->c_symlink, cp->c_symlen); cp->c_flags &= ~C_SYMLINK; cp->c_symlen = 0; } /* * Remove it from the table so it can't be found. */ coda_unsave(cp); if ((struct coda_mntinfo *)(vp->v_mount->mnt_data) == NULL) { myprintf(("Help! vfsp->vfs_data was NULL, but vnode %p " "wasn't dying\n", vp)); panic("badness in coda_inactive\n"); } if (IS_UNMOUNTING(cp)) { #ifdef DEBUG printf("coda_inactive: IS_UNMOUNTING use %d: vp %p, cp %p\n", vrefcnt(vp), vp, cp); if (cp->c_ovp != NULL) printf("coda_inactive: cp->ovp != NULL use %d: vp " "%p, cp %p\n", vrefcnt(vp), vp, cp); #endif } else vgone(vp); MARK_INT_SAT(CODA_INACTIVE_STATS); return (0); } /* * Remote filesystem operations having to do with directory manipulation. */ /* * In FreeBSD, lookup returns the vnode locked. */ int coda_lookup(struct vop_cachedlookup_args *ap) { /* true args */ struct vnode *dvp = ap->a_dvp; struct cnode *dcp = VTOC(dvp); struct vnode **vpp = ap->a_vpp; /* * It looks as though ap->a_cnp->ni_cnd->cn_nameptr holds the rest of * the string to xlate, and that we must try to get at least * ap->a_cnp->ni_cnd->cn_namelen of those characters to macth. I * could be wrong. */ struct componentname *cnp = ap->a_cnp; struct ucred *cred = cnp->cn_cred; struct thread *td = cnp->cn_thread; /* locals */ struct cnode *cp; const char *nm = cnp->cn_nameptr; int len = cnp->cn_namelen; struct CodaFid VFid; int vtype; int error = 0; MARK_ENTRY(CODA_LOOKUP_STATS); CODADEBUG(CODA_LOOKUP, myprintf(("lookup: %s in %s\n", nm, coda_f2s(&dcp->c_fid)));); /* * Check for lookup of control object. */ if (IS_CTL_NAME(dvp, nm, len)) { *vpp = coda_ctlvp; vref(*vpp); MARK_INT_SAT(CODA_LOOKUP_STATS); goto exit; } if (len+1 > CODA_MAXNAMLEN) { MARK_INT_FAIL(CODA_LOOKUP_STATS); CODADEBUG(CODA_LOOKUP, myprintf(("name too long: lookup, " "%s (%s)\n", coda_f2s(&dcp->c_fid), nm));); *vpp = NULL; error = EINVAL; goto exit; } error = venus_lookup(vtomi(dvp), &dcp->c_fid, nm, len, cred, td->td_proc, &VFid, &vtype); if (error) { MARK_INT_FAIL(CODA_LOOKUP_STATS); CODADEBUG(CODA_LOOKUP, myprintf(("lookup error on %s " "(%s)%d\n", coda_f2s(&dcp->c_fid), nm, error));); *vpp = NULL; } else { MARK_INT_SAT(CODA_LOOKUP_STATS); CODADEBUG(CODA_LOOKUP, myprintf(("lookup: %s type %o " "result %d\n", coda_f2s(&VFid), vtype, error));); cp = make_coda_node(&VFid, dvp->v_mount, vtype); *vpp = CTOV(cp); /* * Enter the new vnode in the namecache only if the top bit * isn't set. * * And don't enter a new vnode for an invalid one! */ if (!(vtype & CODA_NOCACHE) && (cnp->cn_flags & MAKEENTRY)) cache_enter(dvp, *vpp, cnp); } exit: /* * If we are creating, and this was the last name to be looked up, * and the error was ENOENT, then there really shouldn't be an error * and we can make the leaf NULL and return success. Since this is * supposed to work under Mach as well as FreeBSD, we're leaving this * fn wrapped. We also must tell lookup/namei that we need to save * the last component of the name. (Create will have to free the * name buffer later...lucky us...). */ if (((cnp->cn_nameiop == CREATE) || (cnp->cn_nameiop == RENAME)) && (cnp->cn_flags & ISLASTCN) && (error == ENOENT)) { error = EJUSTRETURN; cnp->cn_flags |= SAVENAME; *ap->a_vpp = NULL; } /* * If we are removing, and we are at the last element, and we found * it, then we need to keep the name around so that the removal will * go ahead as planned. Unfortunately, this will probably also lock * the to-be-removed vnode, which may or may not be a good idea. * I'll have to look at the bits of coda_remove to make sure. We'll * only save the name if we did in fact find the name, otherwise * coda_remove won't have a chance to free the pathname. */ if ((cnp->cn_nameiop == DELETE) && (cnp->cn_flags & ISLASTCN) && !error) cnp->cn_flags |= SAVENAME; /* * If the lookup went well, we need to (potentially?) unlock the * parent, and lock the child. We are only responsible for checking * to see if the parent is supposed to be unlocked before we return. * We must always lock the child (provided there is one, and (the * parent isn't locked or it isn't the same as the parent.) Simple, * huh? We can never leave the parent locked unless we are ISLASTCN. */ if (!error || (error == EJUSTRETURN)) { if (cnp->cn_flags & ISDOTDOT) { VOP_UNLOCK(dvp, 0); /* * The parent is unlocked. As long as there is a * child, lock it without bothering to check anything * else. */ if (*ap->a_vpp) vn_lock(*ap->a_vpp, LK_EXCLUSIVE | LK_RETRY); vn_lock(dvp, LK_RETRY|LK_EXCLUSIVE); } else { /* * The parent is locked, and may be the same as the * child. If different, go ahead and lock it. */ if (*ap->a_vpp && (*ap->a_vpp != dvp)) vn_lock(*ap->a_vpp, LK_EXCLUSIVE | LK_RETRY); } } else { /* * If the lookup failed, we need to ensure that the leaf is * NULL. * * Don't change any locking? */ *ap->a_vpp = NULL; } return (error); } /*ARGSUSED*/ int coda_create(struct vop_create_args *ap) { /* true args */ struct vnode *dvp = ap->a_dvp; struct cnode *dcp = VTOC(dvp); struct vattr *va = ap->a_vap; int exclusive = 1; int mode = ap->a_vap->va_mode; struct vnode **vpp = ap->a_vpp; struct componentname *cnp = ap->a_cnp; struct ucred *cred = cnp->cn_cred; struct thread *td = cnp->cn_thread; /* locals */ int error; struct cnode *cp; const char *nm = cnp->cn_nameptr; int len = cnp->cn_namelen; struct CodaFid VFid; struct vattr attr; MARK_ENTRY(CODA_CREATE_STATS); /* * All creates are exclusive XXX. * * I'm assuming the 'mode' argument is the file mode bits XXX. * * Check for create of control object. */ if (IS_CTL_NAME(dvp, nm, len)) { *vpp = (struct vnode *)0; MARK_INT_FAIL(CODA_CREATE_STATS); return (EACCES); } error = venus_create(vtomi(dvp), &dcp->c_fid, nm, len, exclusive, mode, va, cred, td->td_proc, &VFid, &attr); if (!error) { /* * If this is an exclusive create, panic if the file already * exists. * * Venus should have detected the file and reported EEXIST. */ if ((exclusive == 1) && (coda_find(&VFid) != NULL)) panic("cnode existed for newly created file!"); cp = make_coda_node(&VFid, dvp->v_mount, attr.va_type); *vpp = CTOV(cp); /* * Update va to reflect the new attributes. */ (*va) = attr; /* * Update the attribute cache and mark it as valid. */ if (coda_attr_cache) { VTOC(*vpp)->c_vattr = attr; VTOC(*vpp)->c_flags |= C_VATTR; } /* * Invalidate the parent's attr cache, the modification time * has changed. */ VTOC(dvp)->c_flags &= ~C_VATTR; cache_enter(dvp, *vpp, cnp); CODADEBUG(CODA_CREATE, myprintf(("create: %s, result %d\n", coda_f2s(&VFid), error));); } else { *vpp = (struct vnode *)0; CODADEBUG(CODA_CREATE, myprintf(("create error %d\n", error));); } if (!error) { if (cnp->cn_flags & MAKEENTRY) cache_enter(dvp, *vpp, cnp); if (cnp->cn_flags & LOCKLEAF) vn_lock(*ap->a_vpp, LK_EXCLUSIVE | LK_RETRY); } else if (error == ENOENT) { /* * XXXRW: We only enter a negative entry if ENOENT is * returned, not other errors. But will Venus invalidate dvp * properly in all cases when new files appear via the * network rather than a local operation? */ if (cnp->cn_flags & MAKEENTRY) cache_enter(dvp, NULL, cnp); } return (error); } int coda_remove(struct vop_remove_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct vnode *dvp = ap->a_dvp; struct cnode *cp = VTOC(dvp); struct componentname *cnp = ap->a_cnp; struct ucred *cred = cnp->cn_cred; struct thread *td = cnp->cn_thread; /* locals */ int error; const char *nm = cnp->cn_nameptr; int len = cnp->cn_namelen; #if 0 struct cnode *tp; #endif MARK_ENTRY(CODA_REMOVE_STATS); CODADEBUG(CODA_REMOVE, myprintf(("remove: %s in %s\n", nm, coda_f2s(&cp->c_fid)));); /* * Check for remove of control object. */ if (IS_CTL_NAME(dvp, nm, len)) { MARK_INT_FAIL(CODA_REMOVE_STATS); return (ENOENT); } /* * Invalidate the parent's attr cache, the modification time has * changed. We don't yet know if the last reference to the file is * being removed, but we do know the reference count on the child has * changed, so invalidate its attr cache also. */ VTOC(dvp)->c_flags &= ~C_VATTR; VTOC(vp)->c_flags &= ~(C_VATTR | C_ACCCACHE); error = venus_remove(vtomi(dvp), &cp->c_fid, nm, len, cred, td->td_proc); cache_purge(vp); CODADEBUG(CODA_REMOVE, myprintf(("in remove result %d\n",error));); return (error); } int coda_link(struct vop_link_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); struct vnode *tdvp = ap->a_tdvp; struct cnode *tdcp = VTOC(tdvp); struct componentname *cnp = ap->a_cnp; struct ucred *cred = cnp->cn_cred; struct thread *td = cnp->cn_thread; /* locals */ int error; const char *nm = cnp->cn_nameptr; int len = cnp->cn_namelen; MARK_ENTRY(CODA_LINK_STATS); if (codadebug & CODADBGMSK(CODA_LINK)) { myprintf(("nb_link: vp fid: %s\n", coda_f2s(&cp->c_fid))); myprintf(("nb_link: tdvp fid: %s)\n", coda_f2s(&tdcp->c_fid))); } if (codadebug & CODADBGMSK(CODA_LINK)) { myprintf(("link: vp fid: %s\n", coda_f2s(&cp->c_fid))); myprintf(("link: tdvp fid: %s\n", coda_f2s(&tdcp->c_fid))); } /* * Check for link to/from control object. */ if (IS_CTL_NAME(tdvp, nm, len) || IS_CTL_VP(vp)) { MARK_INT_FAIL(CODA_LINK_STATS); return (EACCES); } error = venus_link(vtomi(vp), &cp->c_fid, &tdcp->c_fid, nm, len, cred, td->td_proc); /* * Invalidate the parent's attr cache, the modification time has * changed. */ VTOC(tdvp)->c_flags &= ~C_VATTR; VTOC(vp)->c_flags &= ~C_VATTR; CODADEBUG(CODA_LINK, myprintf(("in link result %d\n",error));); return (error); } int coda_rename(struct vop_rename_args *ap) { /* true args */ struct vnode *fvp = ap->a_fvp; struct vnode *tvp = ap->a_tvp; struct vnode *odvp = ap->a_fdvp; struct cnode *odcp = VTOC(odvp); struct componentname *fcnp = ap->a_fcnp; struct vnode *ndvp = ap->a_tdvp; struct cnode *ndcp = VTOC(ndvp); struct componentname *tcnp = ap->a_tcnp; struct ucred *cred = fcnp->cn_cred; struct thread *td = fcnp->cn_thread; /* true args */ int error; const char *fnm = fcnp->cn_nameptr; int flen = fcnp->cn_namelen; const char *tnm = tcnp->cn_nameptr; int tlen = tcnp->cn_namelen; MARK_ENTRY(CODA_RENAME_STATS); /* * Check for rename involving control object. */ if (IS_CTL_NAME(odvp, fnm, flen) || IS_CTL_NAME(ndvp, tnm, tlen)) { MARK_INT_FAIL(CODA_RENAME_STATS); return (EACCES); } /* * Remove the entries for both source and target directories, which * should catch references to the children. Perhaps we could purge * less? */ cache_purge(odvp); cache_purge(ndvp); /* * Invalidate parent directories as modification times have changed. * Invalidate access cache on renamed file as rights may have * changed. */ VTOC(odvp)->c_flags &= ~C_VATTR; VTOC(ndvp)->c_flags &= ~C_VATTR; VTOC(fvp)->c_flags &= ~C_ACCCACHE; if (flen+1 > CODA_MAXNAMLEN) { MARK_INT_FAIL(CODA_RENAME_STATS); error = EINVAL; goto exit; } if (tlen+1 > CODA_MAXNAMLEN) { MARK_INT_FAIL(CODA_RENAME_STATS); error = EINVAL; goto exit; } error = venus_rename(vtomi(odvp), &odcp->c_fid, &ndcp->c_fid, fnm, flen, tnm, tlen, cred, td->td_proc); exit: CODADEBUG(CODA_RENAME, myprintf(("in rename result %d\n",error));); /* * Update namecache to reflect that the names of various objects may * have changed (or gone away entirely). */ cache_purge(fvp); cache_purge(tvp); /* * Release parents first, then children. */ vrele(odvp); if (tvp) { if (tvp == ndvp) vrele(ndvp); else vput(ndvp); vput(tvp); } else vput(ndvp); vrele(fvp); return (error); } int coda_mkdir(struct vop_mkdir_args *ap) { /* true args */ struct vnode *dvp = ap->a_dvp; struct cnode *dcp = VTOC(dvp); struct componentname *cnp = ap->a_cnp; struct vattr *va = ap->a_vap; struct vnode **vpp = ap->a_vpp; struct ucred *cred = cnp->cn_cred; struct thread *td = cnp->cn_thread; /* locals */ int error; const char *nm = cnp->cn_nameptr; int len = cnp->cn_namelen; struct cnode *cp; struct CodaFid VFid; struct vattr ova; MARK_ENTRY(CODA_MKDIR_STATS); /* * Check for mkdir of target object. */ if (IS_CTL_NAME(dvp, nm, len)) { *vpp = (struct vnode *)0; MARK_INT_FAIL(CODA_MKDIR_STATS); return (EACCES); } if (len+1 > CODA_MAXNAMLEN) { *vpp = (struct vnode *)0; MARK_INT_FAIL(CODA_MKDIR_STATS); return (EACCES); } error = venus_mkdir(vtomi(dvp), &dcp->c_fid, nm, len, va, cred, td->td_proc, &VFid, &ova); if (!error) { if (coda_find(&VFid) != NULL) panic("cnode existed for newly created directory!"); cp = make_coda_node(&VFid, dvp->v_mount, va->va_type); *vpp = CTOV(cp); /* * Enter the new vnode in the Name Cache. */ if (cnp->cn_flags & MAKEENTRY) cache_enter(dvp, *vpp, cnp); /* * Update the attr cache and mark as valid. */ if (coda_attr_cache) { VTOC(*vpp)->c_vattr = ova; VTOC(*vpp)->c_flags |= C_VATTR; } /* * Invalidate the parent's attr cache, the modification time * has changed. */ VTOC(dvp)->c_flags &= ~C_VATTR; vn_lock(*vpp, LK_EXCLUSIVE | LK_RETRY); CODADEBUG( CODA_MKDIR, myprintf(("mkdir: %s result %d\n", coda_f2s(&VFid), error));); } else { *vpp = NULL; CODADEBUG(CODA_MKDIR, myprintf(("mkdir error %d\n",error));); } return (error); } int coda_rmdir(struct vop_rmdir_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct vnode *dvp = ap->a_dvp; struct cnode *dcp = VTOC(dvp); struct componentname *cnp = ap->a_cnp; struct ucred *cred = cnp->cn_cred; struct thread *td = cnp->cn_thread; /* true args */ int error; const char *nm = cnp->cn_nameptr; int len = cnp->cn_namelen; #if 0 struct cnode *cp; #endif MARK_ENTRY(CODA_RMDIR_STATS); /* * Check for rmdir of control object. */ if (IS_CTL_NAME(dvp, nm, len)) { MARK_INT_FAIL(CODA_RMDIR_STATS); return (ENOENT); } /* * Possibly somewhat conservative purging, perhaps we just need to * purge vp? */ cache_purge(dvp); cache_purge(vp); /* * Invalidate the parent's attr cache, the modification time has * changed. */ dcp->c_flags &= ~C_VATTR; error = venus_rmdir(vtomi(dvp), &dcp->c_fid, nm, len, cred, td->td_proc); CODADEBUG(CODA_RMDIR, myprintf(("in rmdir result %d\n", error));); return (error); } int coda_symlink(struct vop_symlink_args *ap) { /* true args */ struct vnode *tdvp = ap->a_dvp; struct cnode *tdcp = VTOC(tdvp); struct componentname *cnp = ap->a_cnp; struct vattr *tva = ap->a_vap; char *path = ap->a_target; struct ucred *cred = cnp->cn_cred; struct thread *td = cnp->cn_thread; struct vnode **vpp = ap->a_vpp; /* locals */ int error; /*- * XXX I'm assuming the following things about coda_symlink's * arguments: * t(foo) is the new name/parent/etc being created. * lname is the contents of the new symlink. */ char *nm = cnp->cn_nameptr; int len = cnp->cn_namelen; int plen = strlen(path); /* * Here's the strategy for the moment: perform the symlink, then do a * lookup to grab the resulting vnode. I know this requires two * communications with Venus for a new sybolic link, but that's the * way the ball bounces. I don't yet want to change the way the Mach * symlink works. When Mach support is deprecated, we should change * symlink so that the common case returns the resultant vnode in a * vpp argument. */ MARK_ENTRY(CODA_SYMLINK_STATS); /* * Check for symlink of control object. */ if (IS_CTL_NAME(tdvp, nm, len)) { MARK_INT_FAIL(CODA_SYMLINK_STATS); return (EACCES); } if (plen+1 > CODA_MAXPATHLEN) { MARK_INT_FAIL(CODA_SYMLINK_STATS); return (EINVAL); } if (len+1 > CODA_MAXNAMLEN) { MARK_INT_FAIL(CODA_SYMLINK_STATS); error = EINVAL; goto exit; } error = venus_symlink(vtomi(tdvp), &tdcp->c_fid, path, plen, nm, len, tva, cred, td->td_proc); /* * Invalidate the parent's attr cache, the modification time has * changed. */ tdcp->c_flags &= ~C_VATTR; if (error == 0) error = VOP_LOOKUP(tdvp, vpp, cnp); exit: CODADEBUG(CODA_SYMLINK, myprintf(("in symlink result %d\n",error));); return (error); } /* * Read directory entries. * * XXX: This forwards the operator straight to the cache vnode using * VOP_READDIR(), rather than calling venus_readdir(). Why? */ int coda_readdir(struct vop_readdir_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); struct uio *uiop = ap->a_uio; struct ucred *cred = ap->a_cred; int *eofflag = ap->a_eofflag; u_long **cookies = ap->a_cookies; int *ncookies = ap->a_ncookies; struct thread *td = ap->a_uio->uio_td; /* upcall decl */ /* locals */ int error = 0; int opened_internally = 0; MARK_ENTRY(CODA_READDIR_STATS); CODADEBUG(CODA_READDIR, myprintf(("coda_readdir(%p, %zd, %lld, %d)\n", (void *)uiop->uio_iov->iov_base, uiop->uio_resid, (long long)uiop->uio_offset, uiop->uio_segflg));); /* * Check for readdir of control object. */ if (IS_CTL_VP(vp)) { MARK_INT_FAIL(CODA_READDIR_STATS); return (ENOENT); } /* * If directory is not already open do an "internal open" on it. * * XXX: Why would this happen? For files, there's memory mapping, * execution, and other kernel access paths such as ktrace. For * directories, it is less clear. */ if (cp->c_ovp == NULL) { opened_internally = 1; MARK_INT_GEN(CODA_OPEN_STATS); error = VOP_OPEN(vp, FREAD, cred, td, NULL); printf("coda_readdir: Internally Opening %p\n", vp); if (error) { printf("coda_readdir: VOP_OPEN on container failed " "%d\n", error); return (error); } } /* * Have UFS handle the call. */ CODADEBUG(CODA_READDIR, myprintf(("indirect readdir: fid = %s, " "refcnt = %d\n", coda_f2s(&cp->c_fid), vp->v_usecount));); vn_lock(cp->c_ovp, LK_SHARED | LK_RETRY); error = VOP_READDIR(cp->c_ovp, uiop, cred, eofflag, ncookies, cookies); VOP_UNLOCK(cp->c_ovp, 0); if (error) MARK_INT_FAIL(CODA_READDIR_STATS); else MARK_INT_SAT(CODA_READDIR_STATS); /* * Do an "internal close" if necessary. */ if (opened_internally) { MARK_INT_GEN(CODA_CLOSE_STATS); (void)VOP_CLOSE(vp, FREAD, cred, td); } return (error); } int coda_reclaim(struct vop_reclaim_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); /* upcall decl */ /* locals */ /* * Forced unmount/flush will let vnodes with non-zero use be * destroyed! */ ENTRY; if (IS_UNMOUNTING(cp)) { #ifdef DEBUG if (VTOC(vp)->c_ovp) { if (IS_UNMOUNTING(cp)) printf("coda_reclaim: c_ovp not void: vp " "%p, cp %p\n", vp, cp); } #endif } cache_purge(vp); coda_free(VTOC(vp)); vp->v_data = NULL; vp->v_object = NULL; return (0); } int coda_lock(struct vop_lock1_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); /* upcall decl */ /* locals */ ENTRY; if ((ap->a_flags & LK_INTERLOCK) == 0) { VI_LOCK(vp); ap->a_flags |= LK_INTERLOCK; } if (coda_lockdebug) myprintf(("Attempting lock on %s\n", coda_f2s(&cp->c_fid))); return (vop_stdlock(ap)); } int coda_unlock(struct vop_unlock_args *ap) { /* true args */ struct vnode *vp = ap->a_vp; struct cnode *cp = VTOC(vp); /* upcall decl */ /* locals */ ENTRY; if (coda_lockdebug) myprintf(("Attempting unlock on %s\n", coda_f2s(&cp->c_fid))); return (vop_stdunlock(ap)); } int coda_islocked(struct vop_islocked_args *ap) { /* true args */ ENTRY; return (vop_stdislocked(ap)); } static void coda_print_vattr(struct vattr *attr) { char *typestr; switch (attr->va_type) { case VNON: typestr = "VNON"; break; case VREG: typestr = "VREG"; break; case VDIR: typestr = "VDIR"; break; case VBLK: typestr = "VBLK"; break; case VCHR: typestr = "VCHR"; break; case VLNK: typestr = "VLNK"; break; case VSOCK: typestr = "VSCK"; break; case VFIFO: typestr = "VFFO"; break; case VBAD: typestr = "VBAD"; break; default: typestr = "????"; break; } myprintf(("attr: type %s mode %d uid %d gid %d fsid %d rdev %d\n", typestr, (int)attr->va_mode, (int)attr->va_uid, (int)attr->va_gid, (int)attr->va_fsid, (int)attr->va_rdev)); myprintf((" fileid %d nlink %d size %d blocksize %d bytes %d\n", (int)attr->va_fileid, (int)attr->va_nlink, (int)attr->va_size, (int)attr->va_blocksize,(int)attr->va_bytes)); myprintf((" gen %ld flags %ld vaflags %d\n", attr->va_gen, attr->va_flags, attr->va_vaflags)); myprintf((" atime sec %d nsec %d\n", (int)attr->va_atime.tv_sec, (int)attr->va_atime.tv_nsec)); myprintf((" mtime sec %d nsec %d\n", (int)attr->va_mtime.tv_sec, (int)attr->va_mtime.tv_nsec)); myprintf((" ctime sec %d nsec %d\n", (int)attr->va_ctime.tv_sec, (int)attr->va_ctime.tv_nsec)); } /* * How to print a ucred. */ void coda_print_cred(struct ucred *cred) { int i; myprintf(("ref %d\tuid %d\n",cred->cr_ref,cred->cr_uid)); for (i=0; i < cred->cr_ngroups; i++) myprintf(("\tgroup %d: (%d)\n",i,cred->cr_groups[i])); myprintf(("\n")); } /* * Return a vnode for the given fid. If no cnode exists for this fid create * one and put it in a table hashed by coda_f2i(). If the cnode for this fid * is already in the table return it (ref count is incremented by coda_find. * The cnode will be flushed from the table when coda_inactive calls * coda_unsave. */ struct cnode * make_coda_node(struct CodaFid *fid, struct mount *vfsp, short type) { struct cnode *cp; struct vnode *vp; int err; /* * XXXRW: This really needs a moderate amount of reworking. We need * to properly tolerate failures of getnewvnode() and insmntque(), * and callers need to be able to accept an error back from * make_coda_node. There may also be more general issues in how we * handle forced unmount. Finally, if/when Coda loses its dependency * on Giant, the ordering of this needs rethinking. */ cp = coda_find(fid); if (cp != NULL) { vref(CTOV(cp)); return (cp); } cp = coda_alloc(); cp->c_fid = *fid; err = getnewvnode("coda", vfsp, &coda_vnodeops, &vp); if (err) panic("coda: getnewvnode returned error %d\n", err); vp->v_data = cp; vp->v_type = type; cp->c_vnode = vp; coda_save(cp); err = insmntque(vp, vfsp); if (err != 0) printf("coda: insmntque failed: error %d", err); return (cp); } int coda_pathconf(struct vop_pathconf_args *ap) { switch (ap->a_name) { case _PC_NAME_MAX: *ap->a_retval = CODA_MAXNAMLEN; return (0); case _PC_PATH_MAX: *ap->a_retval = CODA_MAXPATHLEN; return (0); default: return (vop_stdpathconf(ap)); } }