vp 172 arch/i386/i386/vm_machdep.c cpu_coredump(struct proc *p, struct vnode *vp, struct ucred *cred,
vp 198 arch/i386/i386/vm_machdep.c error = vn_rdwr(UIO_WRITE, vp, (caddr_t)&cseg, chdr->c_seghdrsize,
vp 204 arch/i386/i386/vm_machdep.c error = vn_rdwr(UIO_WRITE, vp, (caddr_t)&md_core, sizeof(md_core),
vp 480 arch/i386/stand/installboot/installboot.c char *vp;
vp 505 arch/i386/stand/installboot/installboot.c vp = proto + nl->n_value;
vp 509 arch/i386/stand/installboot/installboot.c lp = (u_int32_t *) vp;
vp 516 arch/i386/stand/installboot/installboot.c wp = (u_int16_t *) vp;
vp 523 arch/i386/stand/installboot/installboot.c bp = (u_int8_t *) vp;
vp 61 compat/common/compat_dir.c struct vnode *vp;
vp 67 compat/common/compat_dir.c vp = (struct vnode *)fp->f_data;
vp 69 compat/common/compat_dir.c if (vp->v_type != VDIR)
vp 72 compat/common/compat_dir.c if ((error = VOP_GETATTR(vp, &va, fp->f_cred, curproc)) != 0)
vp 78 compat/common/compat_dir.c error = vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, curproc);
vp 98 compat/common/compat_dir.c error = VOP_READDIR(vp, &auio, fp->f_cred, &eofflag, &ncookies,
vp 169 compat/common/compat_dir.c VOP_UNLOCK(vp, 0, curproc);
vp 192 compat/common/vfs_syscalls_35.c struct vnode *vp;
vp 205 compat/common/vfs_syscalls_35.c if ((error = VFS_FHTOVP(mp, &fh.fh_fid, &vp)))
vp 207 compat/common/vfs_syscalls_35.c error = vn_stat(vp, &ub, p);
vp 208 compat/common/vfs_syscalls_35.c vput(vp);
vp 340 compat/common/vfs_syscalls_43.c struct vnode *vp;
vp 357 compat/common/vfs_syscalls_43.c vp = (struct vnode *)fp->f_data;
vp 358 compat/common/vfs_syscalls_43.c if (vp->v_type != VDIR) {
vp 371 compat/common/vfs_syscalls_43.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 374 compat/common/vfs_syscalls_43.c if (vp->v_mount->mnt_maxsymlinklen <= 0) {
vp 375 compat/common/vfs_syscalls_43.c error = VOP_READDIR(vp, &auio, fp->f_cred, &eofflag,
vp 392 compat/common/vfs_syscalls_43.c error = VOP_READDIR(vp, &kuio, fp->f_cred, &eofflag,
vp 429 compat/common/vfs_syscalls_43.c VOP_UNLOCK(vp, 0, p);
vp 142 compat/freebsd/freebsd_misc.c struct vnode *vp;
vp 150 compat/freebsd/freebsd_misc.c vp = (struct vnode *)fp->f_data;
vp 1058 compat/hpux/hpux_compat.c struct vnode *vp;
vp 1135 compat/hpux/hpux_compat.c vp = nd.ni_vp;
vp 1137 compat/hpux/hpux_compat.c if (VOP_ACCESS(vp, VREAD, cred, p) == 0)
vp 1139 compat/hpux/hpux_compat.c if (vn_writechk(vp) == 0 && VOP_ACCESS(vp, VWRITE, cred, p) == 0)
vp 1141 compat/hpux/hpux_compat.c if (VOP_ACCESS(vp, VEXEC, cred, p) == 0)
vp 1143 compat/hpux/hpux_compat.c vput(vp);
vp 1340 compat/hpux/hpux_compat.c struct vnode *vp;
vp 1362 compat/hpux/hpux_compat.c vp = nd.ni_vp;
vp 1363 compat/hpux/hpux_compat.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 1366 compat/hpux/hpux_compat.c error = VOP_SETATTR(vp, &vattr, nd.ni_cnd.cn_cred, p);
vp 1367 compat/hpux/hpux_compat.c vput(vp);
vp 215 compat/hpux/hpux_file.c struct vnode *vp;
vp 262 compat/hpux/hpux_file.c vp = (struct vnode *)fp->f_data;
vp 286 compat/hpux/hpux_file.c error = VOP_ADVLOCK(vp, (caddr_t)p, F_SETLK, &fl, flg);
vp 295 compat/hpux/hpux_file.c error = VOP_ADVLOCK(vp, (caddr_t)p, F_SETLK, &fl, flg);
vp 299 compat/hpux/hpux_file.c error = VOP_ADVLOCK(vp, (caddr_t)p, F_UNLCK, &fl,
vp 315 compat/hpux/hpux_file.c vp = (struct vnode *)fp->f_data;
vp 332 compat/hpux/hpux_file.c VOP_ADVLOCK(vp, (caddr_t)p, F_GETLK, &fl, F_POSIX)))
vp 288 compat/ibcs2/ibcs2_exec.c coff_find_section(p, vp, fp, sh, s_type)
vp 290 compat/ibcs2/ibcs2_exec.c struct vnode *vp;
vp 301 compat/ibcs2/ibcs2_exec.c error = vn_rdwr(UIO_READ, vp, (caddr_t) sh,
vp 250 compat/ibcs2/ibcs2_fcntl.c register struct vnode *vp;
vp 261 compat/ibcs2/ibcs2_fcntl.c vp = nd.ni_vp;
vp 272 compat/ibcs2/ibcs2_fcntl.c if ((flags & VWRITE) == 0 || (error = vn_writechk(vp)) == 0)
vp 273 compat/ibcs2/ibcs2_fcntl.c error = VOP_ACCESS(vp, flags, cred, p);
vp 275 compat/ibcs2/ibcs2_fcntl.c vput(vp);
vp 456 compat/ibcs2/ibcs2_misc.c struct vnode *vp;
vp 471 compat/ibcs2/ibcs2_misc.c vp = (struct vnode *)fp->f_data;
vp 472 compat/ibcs2/ibcs2_misc.c if (vp->v_type != VDIR) {
vp 531 compat/linux/linux_exec.c struct vnode *vp;
vp 545 compat/linux/linux_exec.c vp = ni.ni_vp;
vp 547 compat/linux/linux_exec.c if ((error = vn_rdwr(UIO_READ, vp, (caddr_t) &hdr, LINUX_AOUT_HDR_SIZE,
vp 550 compat/linux/linux_exec.c vrele(vp);
vp 555 compat/linux/linux_exec.c vrele(vp);
vp 568 compat/linux/linux_exec.c if ((hdr.a_text != 0 || hdr.a_data != 0) && vp->v_writecount != 0) {
vp 569 compat/linux/linux_exec.c vrele(vp);
vp 572 compat/linux/linux_exec.c vn_marktext(vp);
vp 578 compat/linux/linux_exec.c hdr.a_text + hdr.a_data, taddr, vp, LINUX_N_TXTOFF(hdr, magic),
vp 597 compat/linux/linux_exec.c vrele(vp);
vp 319 compat/linux/linux_file.c struct vnode *vp;
vp 402 compat/linux/linux_file.c vp = (struct vnode *)fp->f_data;
vp 403 compat/linux/linux_file.c if (vp->v_type != VCHR)
vp 406 compat/linux/linux_file.c error = VOP_GETATTR(vp, &va, p->p_ucred, p);
vp 1318 compat/linux/linux_socket.c struct vnode *vp;
vp 1329 compat/linux/linux_socket.c vp = (struct vnode *)fp->f_data;
vp 1330 compat/linux/linux_socket.c isdev = vp->v_type == VCHR;
vp 577 compat/ossaudio/ossaudio.c struct vnode *vp;
vp 585 compat/ossaudio/ossaudio.c vp = (struct vnode *)fp->f_data;
vp 586 compat/ossaudio/ossaudio.c if (vp->v_type != VCHR)
vp 588 compat/ossaudio/ossaudio.c if (VOP_GETATTR(vp, &va, p->p_ucred, p))
vp 503 compat/sunos/sunos_ioctl.c struct vnode *vp;
vp 506 compat/sunos/sunos_ioctl.c vp = (struct vnode *)fp->f_data;
vp 507 compat/sunos/sunos_ioctl.c if (error == EIO && vp != NULL &&
vp 508 compat/sunos/sunos_ioctl.c vp->v_type == VCHR && major(vp->v_rdev) == 21)
vp 425 compat/sunos/sunos_misc.c struct vnode *vp;
vp 433 compat/sunos/sunos_misc.c vp = (struct vnode *)fp->f_data;
vp 435 compat/sunos/sunos_misc.c if (vp->v_type != VDIR) {
vp 587 compat/sunos/sunos_misc.c register struct vnode *vp;
vp 595 compat/sunos/sunos_misc.c vp = (struct vnode *)fp->f_data;
vp 596 compat/sunos/sunos_misc.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 597 compat/sunos/sunos_misc.c if (vp->v_type != VDIR)
vp 600 compat/sunos/sunos_misc.c error = VOP_ACCESS(vp, VEXEC, p->p_ucred, p);
vp 601 compat/sunos/sunos_misc.c VOP_UNLOCK(vp, 0, p);
vp 606 compat/sunos/sunos_misc.c VREF(vp);
vp 609 compat/sunos/sunos_misc.c fdp->fd_rdir = vp;
vp 260 compat/svr4/svr4_fcntl.c struct vnode *vp;
vp 271 compat/svr4/svr4_fcntl.c vp = (struct vnode *)fp->f_data;
vp 272 compat/svr4/svr4_fcntl.c if (fp->f_type != DTYPE_VNODE || vp->v_type == VFIFO)
vp 277 compat/svr4/svr4_fcntl.c if ((error = VOP_GETATTR(vp, &vattr, p->p_ucred, p)) != 0)
vp 423 compat/svr4/svr4_misc.c struct vnode *vp;
vp 432 compat/svr4/svr4_misc.c vp = (struct vnode *) fp->f_data;
vp 433 compat/svr4/svr4_misc.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 434 compat/svr4/svr4_misc.c if (vp->v_type != VDIR)
vp 437 compat/svr4/svr4_misc.c error = VOP_ACCESS(vp, VEXEC, p->p_ucred, p);
vp 438 compat/svr4/svr4_misc.c VOP_UNLOCK(vp, 0, p);
vp 443 compat/svr4/svr4_misc.c VREF(vp);
vp 446 compat/svr4/svr4_misc.c fdp->fd_rdir = vp;
vp 58 ddb/db_variables.c struct db_variable *vp;
vp 62 ddb/db_variables.c for (vp = db_vars; vp < db_evars; vp++) {
vp 63 ddb/db_variables.c if (!strcmp(db_tok_string, vp->name)) {
vp 64 ddb/db_variables.c *varp = vp;
vp 68 ddb/db_variables.c for (vp = db_regs; vp < db_eregs; vp++) {
vp 69 ddb/db_variables.c if (!strcmp(db_tok_string, vp->name)) {
vp 70 ddb/db_variables.c *varp = vp;
vp 83 ddb/db_variables.c struct db_variable *vp;
vp 85 ddb/db_variables.c if (!db_find_variable(&vp))
vp 88 ddb/db_variables.c db_read_variable(vp, valuep);
vp 96 ddb/db_variables.c struct db_variable *vp;
vp 98 ddb/db_variables.c if (!db_find_variable(&vp))
vp 101 ddb/db_variables.c db_write_variable(vp, &value);
vp 108 ddb/db_variables.c db_read_variable(struct db_variable *vp, db_expr_t *valuep)
vp 110 ddb/db_variables.c int (*func)(struct db_variable *, db_expr_t *, int) = vp->fcn;
vp 113 ddb/db_variables.c *valuep = *(vp->valuep);
vp 115 ddb/db_variables.c (*func)(vp, valuep, DB_VAR_GET);
vp 119 ddb/db_variables.c db_write_variable(struct db_variable *vp, db_expr_t *valuep)
vp 121 ddb/db_variables.c int (*func)(struct db_variable *, db_expr_t *, int) = vp->fcn;
vp 124 ddb/db_variables.c *(vp->valuep) = *valuep;
vp 126 ddb/db_variables.c (*func)(vp, valuep, DB_VAR_SET);
vp 134 ddb/db_variables.c struct db_variable *vp;
vp 142 ddb/db_variables.c if (!db_find_variable(&vp)) {
vp 160 ddb/db_variables.c db_write_variable(vp, &value);
vp 272 dev/ccd.c struct vnode *vp;
vp 306 dev/ccd.c vp = ccd->ccd_vpp[ix];
vp 308 dev/ccd.c ci->ci_vp = vp;
vp 329 dev/ccd.c if ((error = VOP_GETATTR(vp, &va, p->p_ucred, p)) != 0) {
vp 342 dev/ccd.c error = VOP_IOCTL(vp, DIOCGPART, (caddr_t)&dpart,
vp 1309 dev/ccd.c struct vnode *vp;
vp 1319 dev/ccd.c vp = nd.ni_vp;
vp 1321 dev/ccd.c if (vp->v_usecount > 1) {
vp 1322 dev/ccd.c VOP_UNLOCK(vp, 0, p);
vp 1323 dev/ccd.c (void)vn_close(vp, FREAD|FWRITE, p->p_ucred, p);
vp 1327 dev/ccd.c if ((error = VOP_GETATTR(vp, &va, p->p_ucred, p)) != 0) {
vp 1330 dev/ccd.c VOP_UNLOCK(vp, 0, p);
vp 1331 dev/ccd.c (void)vn_close(vp, FREAD|FWRITE, p->p_ucred, p);
vp 1337 dev/ccd.c VOP_UNLOCK(vp, 0, p);
vp 1338 dev/ccd.c (void)vn_close(vp, FREAD|FWRITE, p->p_ucred, p);
vp 1343 dev/ccd.c CCD_DCALL(CCDB_VNODE, vprint("ccdlookup: vnode info", vp));
vp 1346 dev/ccd.c VOP_UNLOCK(vp, 0, p);
vp 1347 dev/ccd.c *vpp = vp;
vp 67 dev/clock_subr.h #define todr_getcal(ct, vp) ((*(ct)->todr_gettime)(ct, vp))
vp 90 dev/cons.c struct vnode *vp;
vp 106 dev/cons.c if (vfinddev(dev, VCHR, &vp) && vcount(vp))
vp 530 dev/ic/bt485.c bt485_update(vp)
vp 531 dev/ic/bt485.c void *vp;
vp 533 dev/ic/bt485.c struct bt485data *data = vp;
vp 422 dev/ic/ibm561.c ibm561_update(vp)
vp 423 dev/ic/ibm561.c void *vp;
vp 425 dev/ic/ibm561.c struct ibm561data *data = (struct ibm561data *)vp;
vp 129 dev/ic/isp_target.c void * *vp;
vp 145 dev/ic/isp_target.c unp.vp = vptr;
vp 234 dev/ic/mk48txx.c mk48txx_getcal(handle, vp)
vp 236 dev/ic/mk48txx.c int *vp;
vp 250 dev/ic/mk48txx.c mk48txx_get_nvram_size(handle, vp)
vp 252 dev/ic/mk48txx.c bus_size_t *vp;
vp 255 dev/ic/mk48txx.c *vp = mk->mk_nvramsz;
vp 264 dev/pci/ami_pci.c const struct ami_pci_vendor *vp;
vp 267 dev/pci/ami_pci.c for (vp = ami_pci_vendors;
vp 268 dev/pci/ami_pci.c vp->id && vp->id != (csr & 0xffff); vp++);
vp 269 dev/pci/ami_pci.c if (vp->id)
vp 270 dev/pci/ami_pci.c snprintf(modelbuf, sizeof(modelbuf), "%s %x", vp->name,
vp 509 dev/pci/if_che.c che_read_flash_reg(struct cheg_softc *sc, size_t bcnt, int cont, u_int32_t *vp)
vp 520 dev/pci/if_che.c *vp = che_read(sc, CHE_REG_SF_DATA);
vp 1176 dev/pci/pciide.c const struct pciide_vendor_desc *vp;
vp 1179 dev/pci/pciide.c for (i = 0, vp = pciide_vendors;
vp 1181 dev/pci/pciide.c vp++, i++)
vp 1182 dev/pci/pciide.c if (PCI_VENDOR(id) == vp->ide_vendor)
vp 1188 dev/pci/pciide.c for (pp = vp->ide_products, i = 0; i < vp->ide_nproducts; pp++, i++)
vp 1192 dev/pci/pciide.c if (i == vp->ide_nproducts)
vp 3026 dev/pcmcia/if_ray.c ray_read_region(struct ray_softc *sc, bus_size_t off, void *vp, size_t c)
vp 3032 dev/pcmcia/if_ray.c p = vp;
vp 3072 dev/pcmcia/if_ray.c bus_space_read_region_1(sc->sc_memt, sc->sc_memh, off, vp, c);
vp 3084 dev/pcmcia/if_ray.c ray_write_region(struct ray_softc *sc, bus_size_t off, void *vp, size_t c)
vp 3090 dev/pcmcia/if_ray.c p = vp;
vp 3129 dev/pcmcia/if_ray.c bus_space_write_region_1(sc->sc_memt, sc->sc_memh, off, vp, c);
vp 100 dev/raidframe/rf_copyback.c struct vnode *vp;
vp 138 dev/raidframe/rf_copyback.c vp = raidPtr->raid_cinfo[frow][fcol].ci_vp;
vp 140 dev/raidframe/rf_copyback.c rf_close_component(raidPtr, vp, ac);
vp 150 dev/raidframe/rf_copyback.c retcode = raidlookup(raidPtr->Disks[frow][fcol].devname, proc, &vp);
vp 169 dev/raidframe/rf_copyback.c if ((retcode = VOP_GETATTR(vp, &va, proc->p_ucred, proc)) != 0)
vp 173 dev/raidframe/rf_copyback.c retcode = VOP_IOCTL(vp, DIOCGPART, (caddr_t) &dpart, FREAD,
vp 183 dev/raidframe/rf_copyback.c raidPtr->raid_cinfo[frow][fcol].ci_vp = vp;
vp 516 dev/raidframe/rf_disks.c raidPtr->raid_cinfo[r][c].ci_vp = ac->vp;
vp 607 dev/raidframe/rf_disks.c VOP_CLOSE(ac->vp, FREAD, NOCRED, 0);
vp 608 dev/raidframe/rf_disks.c vput(ac->vp);
vp 609 dev/raidframe/rf_disks.c ac->vp = NULL;
vp 645 dev/raidframe/rf_disks.c struct vnode *vp;
vp 666 dev/raidframe/rf_disks.c error = raidlookup(diskPtr->devname, curproc, &vp);
vp 678 dev/raidframe/rf_disks.c if ((error = VOP_GETATTR(vp, &va, proc->p_ucred, proc)) != 0) {
vp 681 dev/raidframe/rf_disks.c error = VOP_IOCTL(vp, DIOCGPART, (caddr_t) & dpart, FREAD,
vp 691 dev/raidframe/rf_disks.c raidPtr->raid_cinfo[row][col].ci_vp = vp;
vp 190 dev/raidframe/rf_openbsd.h struct vnode *vp; /* Master Vnode Pointer. */
vp 2186 dev/raidframe/rf_openbsdkintf.c struct vnode *vp;
vp 2197 dev/raidframe/rf_openbsdkintf.c vp = nd.ni_vp;
vp 2198 dev/raidframe/rf_openbsdkintf.c if (vp->v_usecount > 1) {
vp 2199 dev/raidframe/rf_openbsdkintf.c VOP_UNLOCK(vp, 0, p);
vp 2200 dev/raidframe/rf_openbsdkintf.c (void)vn_close(vp, FREAD|FWRITE, p->p_ucred, p);
vp 2203 dev/raidframe/rf_openbsdkintf.c if ((error = VOP_GETATTR(vp, &va, p->p_ucred, p)) != 0) {
vp 2204 dev/raidframe/rf_openbsdkintf.c VOP_UNLOCK(vp, 0, p);
vp 2205 dev/raidframe/rf_openbsdkintf.c (void)vn_close(vp, FREAD|FWRITE, p->p_ucred, p);
vp 2210 dev/raidframe/rf_openbsdkintf.c VOP_UNLOCK(vp, 0, p);
vp 2211 dev/raidframe/rf_openbsdkintf.c (void)vn_close(vp, FREAD|FWRITE, p->p_ucred, p);
vp 2214 dev/raidframe/rf_openbsdkintf.c VOP_UNLOCK(vp, 0, p);
vp 2215 dev/raidframe/rf_openbsdkintf.c *vpp = vp;
vp 2558 dev/raidframe/rf_openbsdkintf.c rf_close_component(RF_Raid_t *raidPtr, struct vnode *vp, int auto_configured)
vp 2562 dev/raidframe/rf_openbsdkintf.c if (vp != NULL) {
vp 2565 dev/raidframe/rf_openbsdkintf.c VOP_CLOSE(vp, FREAD | FWRITE, NOCRED, p);
vp 2566 dev/raidframe/rf_openbsdkintf.c vrele(vp);
vp 2569 dev/raidframe/rf_openbsdkintf.c (void) vn_close(vp, FREAD | FWRITE, p->p_ucred, p);
vp 2580 dev/raidframe/rf_openbsdkintf.c struct vnode *vp;
vp 2590 dev/raidframe/rf_openbsdkintf.c vp = raidPtr->raid_cinfo[r][c].ci_vp;
vp 2592 dev/raidframe/rf_openbsdkintf.c rf_close_component(raidPtr, vp, acd);
vp 2599 dev/raidframe/rf_openbsdkintf.c vp = raidPtr->raid_cinfo[0][raidPtr->numCol + r].ci_vp;
vp 2601 dev/raidframe/rf_openbsdkintf.c rf_close_component(raidPtr, vp, acd);
vp 2704 dev/raidframe/rf_openbsdkintf.c struct vnode *vp;
vp 2743 dev/raidframe/rf_openbsdkintf.c if (bdevvp(dev, &vp))
vp 2746 dev/raidframe/rf_openbsdkintf.c error = VOP_OPEN(vp, FREAD, NOCRED, 0);
vp 2753 dev/raidframe/rf_openbsdkintf.c vput(vp);
vp 2758 dev/raidframe/rf_openbsdkintf.c error = VOP_IOCTL(vp, DIOCGDINFO, (caddr_t)&label,
vp 2773 dev/raidframe/rf_openbsdkintf.c VOP_CLOSE(vp, FREAD | FWRITE, NOCRED, 0);
vp 2774 dev/raidframe/rf_openbsdkintf.c vrele(vp);
vp 2793 dev/raidframe/rf_openbsdkintf.c if (bdevvp(dev, &vp))
vp 2796 dev/raidframe/rf_openbsdkintf.c error = VOP_OPEN(vp, FREAD, NOCRED, 0);
vp 2799 dev/raidframe/rf_openbsdkintf.c vput(vp);
vp 2814 dev/raidframe/rf_openbsdkintf.c if (!raidread_component_label(dev, vp, clabel)) {
vp 2841 dev/raidframe/rf_openbsdkintf.c ac->vp = vp;
vp 2851 dev/raidframe/rf_openbsdkintf.c VOP_CLOSE(vp, FREAD | FWRITE, NOCRED, 0);
vp 2852 dev/raidframe/rf_openbsdkintf.c vrele(vp);
vp 3301 dev/raidframe/rf_openbsdkintf.c struct vnode *vp;
vp 3311 dev/raidframe/rf_openbsdkintf.c vp = raidPtr->raid_cinfo[row][column].ci_vp;
vp 3312 dev/raidframe/rf_openbsdkintf.c raidread_component_label(dev, vp, &clabel);
vp 3314 dev/raidframe/rf_openbsdkintf.c raidwrite_component_label(dev, vp, &clabel);
vp 3325 dev/raidframe/rf_openbsdkintf.c struct vnode *vp;
vp 3335 dev/raidframe/rf_openbsdkintf.c vp = raidPtr->raid_cinfo[row][column].ci_vp;
vp 3336 dev/raidframe/rf_openbsdkintf.c raidread_component_label(dev, vp, &clabel);
vp 3338 dev/raidframe/rf_openbsdkintf.c raidwrite_component_label(dev, vp, &clabel);
vp 3353 dev/raidframe/rf_openbsdkintf.c if (ac->vp) {
vp 3354 dev/raidframe/rf_openbsdkintf.c VOP_CLOSE(ac->vp, FREAD, NOCRED, 0);
vp 3355 dev/raidframe/rf_openbsdkintf.c vrele(ac->vp);
vp 3356 dev/raidframe/rf_openbsdkintf.c ac->vp = NULL;
vp 475 dev/raidframe/rf_reconstruct.c struct vnode *vp;
vp 573 dev/raidframe/rf_reconstruct.c vp = raidPtr->raid_cinfo[row][col].ci_vp;
vp 575 dev/raidframe/rf_reconstruct.c rf_close_component(raidPtr, vp, ac);
vp 587 dev/raidframe/rf_reconstruct.c proc, &vp);
vp 610 dev/raidframe/rf_reconstruct.c VOP_GETATTR(vp, &va, proc->p_ucred, proc)) != 0) {
vp 615 dev/raidframe/rf_reconstruct.c retcode = VOP_IOCTL(vp, DIOCGPART, (caddr_t) & dpart,
vp 628 dev/raidframe/rf_reconstruct.c raidPtr->raid_cinfo[row][col].ci_vp = vp;
vp 503 dev/vnd.c struct vnode *vp;
vp 509 dev/vnd.c error = VOP_BMAP(vnd->sc_vp, bn / bsize, &vp, &nbn, &nra);
vp 526 dev/vnd.c vnd->sc_vp, vp, bn, nbn, sz);
vp 535 dev/vnd.c if (vp->v_type == VBLK || vp->v_type == VCHR)
vp 536 dev/vnd.c nbp->vb_buf.b_dev = vp->v_rdev;
vp 543 dev/vnd.c nbp->vb_buf.b_vp = vp;
vp 981 dev/vnd.c struct vnode *vp = vnd->sc_vp;
vp 984 dev/vnd.c DNPRINTF(VDB_FOLLOW, "vndclear(%p): vp %p\n", vnd, vp);
vp 987 dev/vnd.c if (vp == NULL)
vp 989 dev/vnd.c (void) vn_close(vp, VNDRW(vnd), vnd->sc_cred, p);
vp 437 isofs/cd9660/cd9660_lookup.c struct vnode *vp = ITOV(ip);
vp 443 isofs/cd9660/cd9660_lookup.c if ((error = bread(vp, lbn, bsize, NOCRED, &bp)) != 0) {
vp 159 isofs/cd9660/cd9660_node.c struct vnode *vp;
vp 165 isofs/cd9660/cd9660_node.c vp = ITOV(ip);
vp 167 isofs/cd9660/cd9660_node.c if (vget(vp, LK_EXCLUSIVE, p))
vp 169 isofs/cd9660/cd9660_node.c return (vp);
vp 238 isofs/cd9660/cd9660_node.c struct vnode *vp = ap->a_vp;
vp 240 isofs/cd9660/cd9660_node.c register struct iso_node *ip = VTOI(vp);
vp 244 isofs/cd9660/cd9660_node.c if (prtactive && vp->v_usecount != 0)
vp 245 isofs/cd9660/cd9660_node.c vprint("cd9660_inactive: pushing active", vp);
vp 249 isofs/cd9660/cd9660_node.c VOP_UNLOCK(vp, 0, p);
vp 255 isofs/cd9660/cd9660_node.c vrecycle(vp, p);
vp 268 isofs/cd9660/cd9660_node.c register struct vnode *vp = ap->a_vp;
vp 269 isofs/cd9660/cd9660_node.c register struct iso_node *ip = VTOI(vp);
vp 272 isofs/cd9660/cd9660_node.c if (prtactive && vp->v_usecount != 0)
vp 273 isofs/cd9660/cd9660_node.c vprint("cd9660_reclaim: pushing active", vp);
vp 283 isofs/cd9660/cd9660_node.c cache_purge(vp);
vp 288 isofs/cd9660/cd9660_node.c FREE(vp->v_data, M_ISOFSNODE);
vp 289 isofs/cd9660/cd9660_node.c vp->v_data = NULL;
vp 101 isofs/cd9660/cd9660_node.h #define VTOI(vp) ((struct iso_node *)(vp)->v_data)
vp 753 isofs/cd9660/cd9660_vfsops.c struct vnode *vp, *nvp;
vp 764 isofs/cd9660/cd9660_vfsops.c if ((error = getnewvnode(VT_ISOFS, mp, cd9660_vnodeop_p, &vp)) != 0) {
vp 772 isofs/cd9660/cd9660_vfsops.c vp->v_data = ip;
vp 773 isofs/cd9660/cd9660_vfsops.c ip->i_vnode = vp;
vp 786 isofs/cd9660/cd9660_vfsops.c vrele(vp);
vp 799 isofs/cd9660/cd9660_vfsops.c vput(vp);
vp 807 isofs/cd9660/cd9660_vfsops.c vput(vp);
vp 817 isofs/cd9660/cd9660_vfsops.c vput(vp);
vp 826 isofs/cd9660/cd9660_vfsops.c vput(vp);
vp 863 isofs/cd9660/cd9660_vfsops.c vput(vp);
vp 876 isofs/cd9660/cd9660_vfsops.c vp->v_type = VNON;
vp 905 isofs/cd9660/cd9660_vfsops.c switch (vp->v_type = IFTOVT(ip->inode.iso_mode)) {
vp 908 isofs/cd9660/cd9660_vfsops.c vp->v_op = cd9660_fifoop_p;
vp 911 isofs/cd9660/cd9660_vfsops.c vput(vp);
vp 923 isofs/cd9660/cd9660_vfsops.c vp->v_op = cd9660_specop_p;
vp 924 isofs/cd9660/cd9660_vfsops.c if ((nvp = checkalias(vp, ip->inode.iso_rdev, mp)) != NULL) {
vp 929 isofs/cd9660/cd9660_vfsops.c nvp->v_data = vp->v_data;
vp 930 isofs/cd9660/cd9660_vfsops.c vp->v_data = NULL;
vp 931 isofs/cd9660/cd9660_vfsops.c vp->v_op = spec_vnodeop_p;
vp 932 isofs/cd9660/cd9660_vfsops.c vrele(vp);
vp 933 isofs/cd9660/cd9660_vfsops.c vgone(vp);
vp 937 isofs/cd9660/cd9660_vfsops.c vp = nvp;
vp 938 isofs/cd9660/cd9660_vfsops.c ip->i_vnode = vp;
vp 948 isofs/cd9660/cd9660_vfsops.c uvm_vnp_setsize(vp, ip->i_size);
vp 953 isofs/cd9660/cd9660_vfsops.c vp->v_flag |= VROOT;
vp 959 isofs/cd9660/cd9660_vfsops.c *vpp = vp;
vp 968 isofs/cd9660/cd9660_vfsops.c cd9660_vptofh(vp, fhp)
vp 969 isofs/cd9660/cd9660_vfsops.c struct vnode *vp;
vp 972 isofs/cd9660/cd9660_vfsops.c register struct iso_node *ip = VTOI(vp);
vp 106 isofs/cd9660/cd9660_vnops.c register struct vnode *vp;
vp 111 isofs/cd9660/cd9660_vnops.c vp = ndp->ni_vp;
vp 112 isofs/cd9660/cd9660_vnops.c ip = VTOI(vp);
vp 115 isofs/cd9660/cd9660_vnops.c || vap->va_type != vp->v_type
vp 137 isofs/cd9660/cd9660_vnops.c vput(vp);
vp 138 isofs/cd9660/cd9660_vnops.c vp->v_type = VNON;
vp 139 isofs/cd9660/cd9660_vnops.c vgone(vp);
vp 153 isofs/cd9660/cd9660_vnops.c struct vnode *vp = ap->a_vp;
vp 161 isofs/cd9660/cd9660_vnops.c switch (vp->v_type) {
vp 227 isofs/cd9660/cd9660_vnops.c struct vnode *vp = ap->a_vp;
vp 229 isofs/cd9660/cd9660_vnops.c register struct iso_node *ip = VTOI(vp);
vp 244 isofs/cd9660/cd9660_vnops.c if (ip->i_size == 0 && vp->v_type == VLNK) {
vp 271 isofs/cd9660/cd9660_vnops.c vap->va_type = vp->v_type;
vp 283 isofs/cd9660/cd9660_vnops.c struct vnode *vp = ap->a_vp;
vp 285 isofs/cd9660/cd9660_vnops.c register struct iso_node *ip = VTOI(vp);
vp 329 isofs/cd9660/cd9660_vnops.c error = breadn(vp, lbn, size, ra->blks,
vp 333 isofs/cd9660/cd9660_vnops.c error = bread(vp, lbn, size, NOCRED, &bp);
vp 795 isofs/cd9660/cd9660_vnops.c struct vnode *vp = ap->a_vp;
vp 797 isofs/cd9660/cd9660_vnops.c return (lockmgr(&VTOI(vp)->i_lock, ap->a_flags, NULL));
vp 808 isofs/cd9660/cd9660_vnops.c struct vnode *vp = ap->a_vp;
vp 810 isofs/cd9660/cd9660_vnops.c return (lockmgr(&VTOI(vp)->i_lock, ap->a_flags | LK_RELEASE, NULL));
vp 823 isofs/cd9660/cd9660_vnops.c struct vnode *vp = bp->b_vp;
vp 828 isofs/cd9660/cd9660_vnops.c ip = VTOI(vp);
vp 829 isofs/cd9660/cd9660_vnops.c if (vp->v_type == VBLK || vp->v_type == VCHR)
vp 832 isofs/cd9660/cd9660_vnops.c error = VOP_BMAP(vp, bp->b_lblkno, NULL, &bp->b_blkno, NULL);
vp 850 isofs/cd9660/cd9660_vnops.c vp = ip->i_devvp;
vp 851 isofs/cd9660/cd9660_vnops.c bp->b_dev = vp->v_rdev;
vp 852 isofs/cd9660/cd9660_vnops.c VOCALL (vp->v_op, VOFFSET(vop_strategy), ap);
vp 94 isofs/udf/udf.h #define VTOU(vp) ((struct unode *)((vp)->v_data))
vp 190 isofs/udf/udf_subr.c struct vnode *vp;
vp 194 isofs/udf/udf_subr.c error = udf_vget(ump->um_mountp, lb - ump->um_start - 3, &vp);
vp 198 isofs/udf/udf_subr.c up = VTOU(vp);
vp 207 isofs/udf/udf_subr.c vput(vp);
vp 456 isofs/udf/udf_vfsops.c struct vnode *vp;
vp 468 isofs/udf/udf_vfsops.c vp = *vpp;
vp 469 isofs/udf/udf_vfsops.c vp->v_flag |= VROOT;
vp 512 isofs/udf/udf_vfsops.c struct vnode *vp;
vp 571 isofs/udf/udf_vfsops.c if ((error = udf_allocv(mp, &vp, p))) {
vp 577 isofs/udf/udf_vfsops.c up->u_vnode = vp;
vp 582 isofs/udf/udf_vfsops.c vp->v_data = up;
vp 594 isofs/udf/udf_vfsops.c vp->v_type = VBAD;
vp 597 isofs/udf/udf_vfsops.c vp->v_type = VDIR;
vp 600 isofs/udf/udf_vfsops.c vp->v_type = VREG;
vp 603 isofs/udf/udf_vfsops.c vp->v_type = VBLK;
vp 606 isofs/udf/udf_vfsops.c vp->v_type = VCHR;
vp 609 isofs/udf/udf_vfsops.c vp->v_type = VFIFO;
vp 612 isofs/udf/udf_vfsops.c vp->v_type = VSOCK;
vp 615 isofs/udf/udf_vfsops.c vp->v_type = VLNK;
vp 618 isofs/udf/udf_vfsops.c vp->v_type = VREG;
vp 622 isofs/udf/udf_vfsops.c *vpp = vp;
vp 654 isofs/udf/udf_vfsops.c udf_vptofh(struct vnode *vp, struct fid *fhp)
vp 659 isofs/udf/udf_vfsops.c up = VTOU(vp);
vp 166 isofs/udf/udf_vnops.c struct vnode *vp;
vp 168 isofs/udf/udf_vnops.c error = getnewvnode(VT_UDF, mp, udf_vnodeop_p, &vp);
vp 174 isofs/udf/udf_vnops.c *vpp = vp;
vp 203 isofs/udf/udf_vnops.c struct vnode *vp;
vp 207 isofs/udf/udf_vnops.c vp = ap->a_vp;
vp 208 isofs/udf/udf_vnops.c up = VTOU(vp);
vp 212 isofs/udf/udf_vnops.c switch (vp->v_type) {
vp 310 isofs/udf/udf_vnops.c struct vnode *vp;
vp 318 isofs/udf/udf_vnops.c vp = ap->a_vp;
vp 320 isofs/udf/udf_vnops.c up = VTOU(vp);
vp 337 isofs/udf/udf_vnops.c if (vp->v_type & VDIR) {
vp 358 isofs/udf/udf_vnops.c vap->va_type = vp->v_type;
vp 417 isofs/udf/udf_vnops.c struct vnode *vp = ap->a_vp;
vp 419 isofs/udf/udf_vnops.c struct unode *up = VTOU(vp);
vp 698 isofs/udf/udf_vnops.c struct vnode *vp;
vp 713 isofs/udf/udf_vnops.c vp = ap->a_vp;
vp 715 isofs/udf/udf_vnops.c up = VTOU(vp);
vp 833 isofs/udf/udf_vnops.c struct vnode *vp;
vp 838 isofs/udf/udf_vnops.c vp = bp->b_vp;
vp 839 isofs/udf/udf_vnops.c up = VTOU(vp);
vp 853 isofs/udf/udf_vnops.c error = VOP_BMAP(vp, bp->b_lblkno, NULL, &bp->b_blkno, NULL);
vp 872 isofs/udf/udf_vnops.c bp->b_dev = vp->v_rdev;
vp 884 isofs/udf/udf_vnops.c struct vnode *vp = ap->a_vp;
vp 886 isofs/udf/udf_vnops.c return (lockmgr(&VTOU(vp)->u_lock, ap->a_flags, NULL));
vp 894 isofs/udf/udf_vnops.c struct vnode *vp = ap->a_vp;
vp 896 isofs/udf/udf_vnops.c return (lockmgr(&VTOU(vp)->u_lock, ap->a_flags | LK_RELEASE, NULL));
vp 911 isofs/udf/udf_vnops.c struct vnode *vp = ap->a_vp;
vp 912 isofs/udf/udf_vnops.c struct unode *up = VTOU(vp);
vp 1120 isofs/udf/udf_vnops.c struct vnode *vp = ap->a_vp;
vp 1126 isofs/udf/udf_vnops.c VOP_UNLOCK(vp, 0, p);
vp 1135 isofs/udf/udf_vnops.c struct vnode *vp;
vp 1138 isofs/udf/udf_vnops.c vp = ap->a_vp;
vp 1139 isofs/udf/udf_vnops.c up = VTOU(vp);
vp 1152 isofs/udf/udf_vnops.c vp->v_data = NULL;
vp 245 kern/exec_elf.c ELFNAME(load_psection)(struct exec_vmcmd_set *vcset, struct vnode *vp,
vp 292 kern/exec_elf.c NEW_VMCMD2(vcset, vmcmd_map_pagedvn, psize, base, vp,
vp 296 kern/exec_elf.c base + psize, vp, offset + psize, *prot, flags);
vp 299 kern/exec_elf.c NEW_VMCMD2(vcset, vmcmd_map_pagedvn, psize, base, vp, offset,
vp 320 kern/exec_elf.c ELFNAME(read_from)(struct proc *p, struct vnode *vp, u_long off, caddr_t buf,
vp 326 kern/exec_elf.c if ((error = vn_rdwr(UIO_READ, vp, buf, size, off, UIO_SYSSPACE,
vp 351 kern/exec_elf.c struct vnode *vp;
vp 368 kern/exec_elf.c vp = nd.ni_vp;
vp 369 kern/exec_elf.c if (vp->v_type != VREG) {
vp 373 kern/exec_elf.c if ((error = VOP_GETATTR(vp, epp->ep_vap, p->p_ucred, p)) != 0)
vp 375 kern/exec_elf.c if (vp->v_mount->mnt_flag & MNT_NOEXEC) {
vp 379 kern/exec_elf.c if ((error = VOP_ACCESS(vp, VREAD, p->p_ucred, p)) != 0)
vp 431 kern/exec_elf.c uobj = &vp->v_uvm.u_obj;
vp 57 kern/exec_subr.c new_vmcmd(evsp, proc, len, addr, vp, offset, prot, flags)
vp 62 kern/exec_subr.c struct vnode *vp;
vp 75 kern/exec_subr.c if ((vcp->ev_vp = vp) != NULL)
vp 76 kern/exec_subr.c vref(vp);
vp 165 kern/kern_acct.c struct vnode *vp;
vp 170 kern/kern_acct.c vp = acctp;
vp 171 kern/kern_acct.c if (vp == NULL)
vp 230 kern/kern_acct.c error = vn_rdwr(UIO_WRITE, vp, (caddr_t)&acct, sizeof (acct),
vp 303 kern/kern_descrip.c struct vnode *vp;
vp 412 kern/kern_descrip.c vp = (struct vnode *)fp->f_data;
vp 434 kern/kern_descrip.c error = (VOP_ADVLOCK(vp, (caddr_t)p, F_SETLK, &fl, flg));
vp 443 kern/kern_descrip.c error = (VOP_ADVLOCK(vp, (caddr_t)p, F_SETLK, &fl, flg));
vp 447 kern/kern_descrip.c error = (VOP_ADVLOCK(vp, (caddr_t)p, F_UNLCK, &fl,
vp 461 kern/kern_descrip.c vp = (struct vnode *)fp->f_data;
vp 482 kern/kern_descrip.c error = VOP_ADVLOCK(vp, (caddr_t)p, F_GETLK, &fl, F_POSIX);
vp 638 kern/kern_descrip.c struct vnode *vp;
vp 656 kern/kern_descrip.c vp = (struct vnode *)fp->f_data;
vp 657 kern/kern_descrip.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 658 kern/kern_descrip.c error = VOP_PATHCONF(vp, SCARG(uap, name), retval);
vp 659 kern/kern_descrip.c VOP_UNLOCK(vp, 0, p);
vp 1018 kern/kern_descrip.c struct vnode *vp;
vp 1038 kern/kern_descrip.c vp = (struct vnode *)fp->f_data;
vp 1039 kern/kern_descrip.c (void) VOP_ADVLOCK(vp, (caddr_t)p, F_UNLCK, &lf, F_POSIX);
vp 1073 kern/kern_descrip.c vp = (struct vnode *)fp->f_data;
vp 1074 kern/kern_descrip.c (void) VOP_ADVLOCK(vp, (caddr_t)fp, F_UNLCK, &lf, F_FLOCK);
vp 1112 kern/kern_descrip.c struct vnode *vp;
vp 1120 kern/kern_descrip.c vp = (struct vnode *)fp->f_data;
vp 1127 kern/kern_descrip.c error = VOP_ADVLOCK(vp, (caddr_t)fp, F_UNLCK, &lf, F_FLOCK);
vp 1140 kern/kern_descrip.c error = VOP_ADVLOCK(vp, (caddr_t)fp, F_SETLK, &lf, F_FLOCK);
vp 1142 kern/kern_descrip.c error = VOP_ADVLOCK(vp, (caddr_t)fp, F_SETLK, &lf, F_FLOCK|F_WAIT);
vp 115 kern/kern_exec.c struct vnode *vp;
vp 125 kern/kern_exec.c epp->ep_vp = vp = ndp->ni_vp;
vp 128 kern/kern_exec.c if (vp->v_type == VDIR) {
vp 132 kern/kern_exec.c if (vp->v_type != VREG) {
vp 138 kern/kern_exec.c if ((error = VOP_GETATTR(vp, epp->ep_vap, p->p_ucred, p)) != 0)
vp 142 kern/kern_exec.c if (vp->v_mount->mnt_flag & MNT_NOEXEC) {
vp 147 kern/kern_exec.c if ((vp->v_mount->mnt_flag & MNT_NOSUID))
vp 151 kern/kern_exec.c if ((error = VOP_ACCESS(vp, VEXEC, p->p_ucred, p)) != 0)
vp 159 kern/kern_exec.c if ((error = VOP_OPEN(vp, FREAD, p->p_ucred, p)) != 0)
vp 163 kern/kern_exec.c VOP_UNLOCK(vp, 0, p);
vp 166 kern/kern_exec.c error = vn_rdwr(UIO_READ, vp, epp->ep_hdr, epp->ep_hdrlen, 0,
vp 216 kern/kern_exec.c vn_close(vp, FREAD, p->p_ucred, p);
vp 226 kern/kern_exec.c vput(vp);
vp 543 kern/kern_exec.c struct vnode *vp;
vp 552 kern/kern_exec.c if ((error = cdevvp(getnulldev(), &vp)) != 0) {
vp 557 kern/kern_exec.c if ((error = VOP_OPEN(vp, flags, p->p_ucred, p)) != 0) {
vp 560 kern/kern_exec.c vrele(vp);
vp 564 kern/kern_exec.c vp->v_writecount++;
vp 568 kern/kern_exec.c fp->f_data = (caddr_t)vp;
vp 68 kern/kern_ktrace.c struct vnode *vp;
vp 76 kern/kern_ktrace.c vp = p->p_tracep;
vp 79 kern/kern_ktrace.c if (vp != NULL)
vp 80 kern/kern_ktrace.c vrele(vp);
vp 291 kern/kern_ktrace.c struct vnode *vp = NULL;
vp 312 kern/kern_ktrace.c vp = nd.ni_vp;
vp 314 kern/kern_ktrace.c VOP_UNLOCK(vp, 0, curp);
vp 315 kern/kern_ktrace.c if (vp->v_type != VREG) {
vp 316 kern/kern_ktrace.c (void) vn_close(vp, FREAD|FWRITE, curp->p_ucred, curp);
vp 326 kern/kern_ktrace.c if (p->p_tracep == vp) {
vp 357 kern/kern_ktrace.c ret |= ktrsetchildren(curp, p, ops, facs, vp);
vp 359 kern/kern_ktrace.c ret |= ktrops(curp, p, ops, facs, vp);
vp 371 kern/kern_ktrace.c ret |= ktrsetchildren(curp, p, ops, facs, vp);
vp 373 kern/kern_ktrace.c ret |= ktrops(curp, p, ops, facs, vp);
vp 378 kern/kern_ktrace.c if (vp != NULL)
vp 379 kern/kern_ktrace.c (void) vn_close(vp, FWRITE, curp->p_ucred, curp);
vp 385 kern/kern_ktrace.c ktrops(struct proc *curp, struct proc *p, int ops, int facs, struct vnode *vp)
vp 391 kern/kern_ktrace.c ktrsettracevnode(p, vp);
vp 416 kern/kern_ktrace.c struct vnode *vp)
vp 423 kern/kern_ktrace.c ret |= ktrops(curp, p, ops, facs, vp);
vp 450 kern/kern_ktrace.c struct vnode *vp = p->p_tracep;
vp 452 kern/kern_ktrace.c if (vp == NULL)
vp 469 kern/kern_ktrace.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 470 kern/kern_ktrace.c error = VOP_WRITE(vp, &auio, IO_UNIT|IO_APPEND, p->p_ucred);
vp 471 kern/kern_ktrace.c VOP_UNLOCK(vp, 0, p);
vp 480 kern/kern_ktrace.c if (p->p_tracep == vp) {
vp 1354 kern/kern_sig.c struct vnode *vp;
vp 1406 kern/kern_sig.c vp = nd.ni_vp;
vp 1407 kern/kern_sig.c if ((error = VOP_GETATTR(vp, &vattr, cred, p)) != 0)
vp 1410 kern/kern_sig.c if (vp->v_type != VREG || vattr.va_nlink != 1 ||
vp 1417 kern/kern_sig.c VOP_SETATTR(vp, &vattr, cred, p);
vp 1431 kern/kern_sig.c error = cpu_coredump(p, vp, cred, &core);
vp 1438 kern/kern_sig.c error = uvm_coredump(p, vp, cred, &core);
vp 1441 kern/kern_sig.c error = vn_rdwr(UIO_WRITE, vp, (caddr_t)&core,
vp 1445 kern/kern_sig.c VOP_UNLOCK(vp, 0, p);
vp 1446 kern/kern_sig.c error1 = vn_close(vp, FWRITE, cred, p);
vp 1353 kern/kern_sysctl.c struct proc *vp;
vp 1384 kern/kern_sysctl.c if ((vp = pfind(pid)) == NULL)
vp 1395 kern/kern_sysctl.c if (P_ZOMBIE(vp) || (vp->p_flag & P_SYSTEM))
vp 1399 kern/kern_sysctl.c if ((vp->p_flag & P_WEXIT))
vp 1403 kern/kern_sysctl.c if ((vp->p_flag & P_INEXEC))
vp 1406 kern/kern_sysctl.c vp->p_vmspace->vm_refcnt++; /* XXX */
vp 1419 kern/kern_sysctl.c if ((error = uvm_io(&vp->p_vmspace->vm_map, &uio, 0)) != 0)
vp 1476 kern/kern_sysctl.c if ((error = uvm_io(&vp->p_vmspace->vm_map, &uio, 0)) != 0)
vp 1498 kern/kern_sysctl.c if ((error = uvm_io(&vp->p_vmspace->vm_map, &uio, 0)) != 0)
vp 1546 kern/kern_sysctl.c uvmspace_free(vp->p_vmspace);
vp 658 kern/subr_prf.c if ((vp != NULL) && (sbuf == tailp)) { \
vp 664 kern/subr_prf.c kputchar(chr, oflags, (struct tty *)vp); \
vp 669 kern/subr_prf.c kprintf(const char *fmt0, int oflags, void *vp, char *sbuf, va_list ap)
vp 690 kern/subr_prf.c if ((oflags & TOBUFONLY) && (vp != NULL))
vp 691 kern/subr_prf.c tailp = *(char **)vp;
vp 1116 kern/subr_prf.c if ((oflags & TOBUFONLY) && (vp != NULL))
vp 1117 kern/subr_prf.c *(char **)vp = sbuf;
vp 1015 kern/tty_pty.c struct vnode *vp;
vp 1020 kern/tty_pty.c vp = ndp->ni_vp;
vp 1021 kern/tty_pty.c if (vp->v_type != VCHR) {
vp 1030 kern/tty_pty.c error = VOP_OPEN(vp, FREAD|FWRITE, cred, p);
vp 1037 kern/tty_pty.c (void)VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 1044 kern/tty_pty.c vp->v_writecount++;
vp 1048 kern/tty_pty.c vput(vp);
vp 399 kern/uipc_usrreq.c struct vnode *vp;
vp 416 kern/uipc_usrreq.c vp = nd.ni_vp;
vp 417 kern/uipc_usrreq.c if (vp != NULL) {
vp 419 kern/uipc_usrreq.c if (nd.ni_dvp == vp)
vp 423 kern/uipc_usrreq.c vrele(vp);
vp 432 kern/uipc_usrreq.c vp = nd.ni_vp;
vp 433 kern/uipc_usrreq.c vp->v_socket = unp->unp_socket;
vp 434 kern/uipc_usrreq.c unp->unp_vnode = vp;
vp 439 kern/uipc_usrreq.c VOP_UNLOCK(vp, 0, p);
vp 447 kern/uipc_usrreq.c struct vnode *vp;
vp 461 kern/uipc_usrreq.c vp = nd.ni_vp;
vp 462 kern/uipc_usrreq.c if (vp->v_type != VSOCK) {
vp 466 kern/uipc_usrreq.c if ((error = VOP_ACCESS(vp, VWRITE, p->p_ucred, p)) != 0)
vp 468 kern/uipc_usrreq.c so2 = vp->v_socket;
vp 501 kern/uipc_usrreq.c vput(vp);
vp 639 kern/uipc_usrreq.c struct vnode *vp = (struct vnode *)fp->f_data;
vp 640 kern/uipc_usrreq.c if ((vp->v_type == VDIR) &&
vp 641 kern/uipc_usrreq.c !vn_isunder(vp, p->p_cwdi->cwdi_rdir, p)) {
vp 222 kern/vfs_bio.c buf_stub(struct vnode *vp, daddr64_t lblkno)
vp 255 kern/vfs_bio.c bgetvp(vp, bp);
vp 377 kern/vfs_bio.c bio_doread(struct vnode *vp, daddr64_t blkno, int size, int async)
vp 381 kern/vfs_bio.c bp = getblk(vp, blkno, size, 0, 0);
vp 406 kern/vfs_bio.c bread(struct vnode *vp, daddr64_t blkno, int size, struct ucred *cred,
vp 412 kern/vfs_bio.c bp = *bpp = bio_doread(vp, blkno, size, 0);
vp 423 kern/vfs_bio.c breadn(struct vnode *vp, daddr64_t blkno, int size, daddr64_t rablks[],
vp 429 kern/vfs_bio.c bp = *bpp = bio_doread(vp, blkno, size, 0);
vp 436 kern/vfs_bio.c if (incore(vp, rablks[i]))
vp 440 kern/vfs_bio.c (void) bio_doread(vp, rablks[i], rasizes[i], B_ASYNC);
vp 470 kern/vfs_bio.c bread_cluster(struct vnode *vp, daddr64_t blkno, int size, struct buf **rbpp)
vp 477 kern/vfs_bio.c *rbpp = bio_doread(vp, blkno, size, 0);
vp 482 kern/vfs_bio.c if (VOP_BMAP(vp, blkno + 1, NULL, &sblkno, &maxra))
vp 498 kern/vfs_bio.c if (incore(vp, blkno + i + 1)) {
vp 506 kern/vfs_bio.c xbpp[i] = buf_stub(vp, blkno + i + 1);
vp 533 kern/vfs_bio.c binshash(xbpp[i], BUFHASH(vp, xbpp[i]->b_lblkno));
vp 544 kern/vfs_bio.c bp->b_vp = vp;
vp 564 kern/vfs_bio.c struct vnode *vp;
vp 567 kern/vfs_bio.c vp = bp->b_vp;
vp 568 kern/vfs_bio.c if (vp != NULL)
vp 569 kern/vfs_bio.c mp = vp->v_type == VBLK? vp->v_specmountpoint : vp->v_mount;
vp 842 kern/vfs_bio.c incore(struct vnode *vp, daddr64_t blkno)
vp 847 kern/vfs_bio.c LIST_FOREACH(bp, BUFHASH(vp, blkno), b_hash) {
vp 848 kern/vfs_bio.c if (bp->b_lblkno == blkno && bp->b_vp == vp &&
vp 865 kern/vfs_bio.c getblk(struct vnode *vp, daddr64_t blkno, int size, int slpflag, int slptimeo)
vp 881 kern/vfs_bio.c bh = BUFHASH(vp, blkno);
vp 883 kern/vfs_bio.c LIST_FOREACH(bp, BUFHASH(vp, blkno), b_hash) {
vp 884 kern/vfs_bio.c if (bp->b_lblkno != blkno || bp->b_vp != vp)
vp 931 kern/vfs_bio.c bgetvp(vp, bp);
vp 83 kern/vfs_cache.c #define NCVHASH(vp) (vp)->v_id & ncvhash
vp 111 kern/vfs_cache.c struct vnode *vp;
vp 165 kern/vfs_cache.c vp = ncp->nc_vp;
vp 166 kern/vfs_cache.c vpid = vp->v_id;
vp 167 kern/vfs_cache.c if (vp == dvp) { /* lookup on "." */
vp 173 kern/vfs_cache.c error = vget(vp, LK_EXCLUSIVE, p);
vp 180 kern/vfs_cache.c vput(vp);
vp 186 kern/vfs_cache.c error = vget(vp, LK_EXCLUSIVE, p);
vp 201 kern/vfs_cache.c if (error || vpid != vp->v_id) {
vp 203 kern/vfs_cache.c vput(vp);
vp 212 kern/vfs_cache.c if (vp == dvp || error ||
vp 229 kern/vfs_cache.c *vpp = vp;
vp 268 kern/vfs_cache.c cache_revlookup(struct vnode *vp, struct vnode **dvpp, char **bpp, char *bufp)
vp 278 kern/vfs_cache.c nvcpp = &ncvhashtbl[NCVHASH(vp)];
vp 281 kern/vfs_cache.c if (ncp->nc_vp == vp &&
vp 282 kern/vfs_cache.c ncp->nc_vpid == vp->v_id &&
vp 285 kern/vfs_cache.c dvp != vp && ncp->nc_dvpid == dvp->v_id) {
vp 335 kern/vfs_cache.c cache_enter(struct vnode *dvp, struct vnode *vp, struct componentname *cnp)
vp 364 kern/vfs_cache.c ncp->nc_vp = vp;
vp 365 kern/vfs_cache.c if (vp)
vp 366 kern/vfs_cache.c ncp->nc_vpid = vp->v_id;
vp 384 kern/vfs_cache.c if (vp && vp != dvp && vp->v_type == VDIR &&
vp 388 kern/vfs_cache.c nvcpp = &ncvhashtbl[NCVHASH(vp)];
vp 412 kern/vfs_cache.c cache_purge(struct vnode *vp)
vp 417 kern/vfs_cache.c vp->v_id = ++nextvnodeid;
vp 426 kern/vfs_cache.c vp->v_id = ++nextvnodeid;
vp 64 kern/vfs_cluster.c struct vnode *vp;
vp 68 kern/vfs_cluster.c vp = bp->b_vp;
vp 77 kern/vfs_cluster.c maxclen = MAXBSIZE / vp->v_mount->mnt_stat.f_iosize - 1;
vp 91 kern/vfs_cluster.c cluster_wbuild(vp, NULL, bp->b_bcount,
vp 97 kern/vfs_cluster.c buflist = cluster_collectbufs(vp, ci, bp);
vp 100 kern/vfs_cluster.c if (VOP_REALLOCBLKS(vp, buflist)) {
vp 108 kern/vfs_cluster.c cluster_wbuild(vp, NULL, bp->b_bcount,
vp 130 kern/vfs_cluster.c (VOP_BMAP(vp, lbn, NULL, &bp->b_blkno, &maxclen) ||
vp 151 kern/vfs_cluster.c cluster_wbuild(vp, bp, bp->b_bcount, ci->ci_cstart,
vp 171 kern/vfs_cluster.c cluster_wbuild(struct vnode *vp, struct buf *last_bp, long size,
vp 177 kern/vfs_cluster.c if (size != vp->v_mount->mnt_stat.f_iosize)
vp 179 kern/vfs_cluster.c size, vp->v_mount->mnt_stat.f_iosize);
vp 182 kern/vfs_cluster.c while ((!incore(vp, start_lbn) || start_lbn == lbn) && len) {
vp 192 kern/vfs_cluster.c bp = getblk(vp, start_lbn, size, 0, 0);
vp 206 kern/vfs_cluster.c bp = getblk(vp, start_lbn, size, 0, 0);
vp 225 kern/vfs_cluster.c cluster_collectbufs(struct vnode *vp, struct cluster_info *ci,
vp 238 kern/vfs_cluster.c (void)bread(vp, lbn, last_bp->b_bcount, NOCRED,
vp 61 kern/vfs_default.c struct vnode *vp, *vq;
vp 69 kern/vfs_default.c vp = ap->a_vp;
vp 71 kern/vfs_default.c if (vp->v_flag & VALIASED) {
vp 76 kern/vfs_default.c if (vp->v_flag & VXLOCK) {
vp 77 kern/vfs_default.c vp->v_flag |= VXWANT;
vp 78 kern/vfs_default.c tsleep(vp, PINOD, "vop_generic_revokeall", 0);
vp 87 kern/vfs_default.c vp->v_flag |= VXLOCK;
vp 88 kern/vfs_default.c while (vp->v_flag & VALIASED) {
vp 89 kern/vfs_default.c for (vq = *vp->v_hashchain; vq; vq = vq->v_specnext) {
vp 90 kern/vfs_default.c if (vq->v_rdev != vp->v_rdev ||
vp 91 kern/vfs_default.c vq->v_type != vp->v_type || vp == vq)
vp 103 kern/vfs_default.c vp->v_flag &= ~VXLOCK;
vp 106 kern/vfs_default.c vgonel(vp, p);
vp 343 kern/vfs_subr.c struct vnode *vp;
vp 370 kern/vfs_subr.c vp = pool_get(&vnode_pool, PR_WAITOK);
vp 371 kern/vfs_subr.c bzero((char *)vp, sizeof *vp);
vp 374 kern/vfs_subr.c for (vp = TAILQ_FIRST(listhd); vp != NULLVP;
vp 375 kern/vfs_subr.c vp = TAILQ_NEXT(vp, v_freelist)) {
vp 376 kern/vfs_subr.c if (VOP_ISLOCKED(vp) == 0)
vp 384 kern/vfs_subr.c if (vp == NULL) {
vp 392 kern/vfs_subr.c if (vp->v_usecount) {
vp 393 kern/vfs_subr.c vprint("free vnode", vp);
vp 398 kern/vfs_subr.c TAILQ_REMOVE(listhd, vp, v_freelist);
vp 399 kern/vfs_subr.c vp->v_bioflag &= ~VBIOONFREELIST;
vp 402 kern/vfs_subr.c if (vp->v_type != VBAD)
vp 403 kern/vfs_subr.c vgonel(vp, p);
vp 405 kern/vfs_subr.c if (vp->v_data) {
vp 406 kern/vfs_subr.c vprint("cleaned vnode", vp);
vp 410 kern/vfs_subr.c if (vp->v_numoutput)
vp 414 kern/vfs_subr.c vp->v_flag = 0;
vp 415 kern/vfs_subr.c vp->v_socket = 0;
vp 417 kern/vfs_subr.c vp->v_type = VNON;
vp 418 kern/vfs_subr.c cache_purge(vp);
vp 419 kern/vfs_subr.c vp->v_tag = tag;
vp 420 kern/vfs_subr.c vp->v_op = vops;
vp 421 kern/vfs_subr.c insmntque(vp, mp);
vp 422 kern/vfs_subr.c *vpp = vp;
vp 423 kern/vfs_subr.c vp->v_usecount = 1;
vp 424 kern/vfs_subr.c vp->v_data = 0;
vp 425 kern/vfs_subr.c simple_lock_init(&vp->v_uvm.u_obj.vmobjlock);
vp 433 kern/vfs_subr.c insmntque(struct vnode *vp, struct mount *mp)
vp 438 kern/vfs_subr.c if (vp->v_mount != NULL)
vp 439 kern/vfs_subr.c LIST_REMOVE(vp, v_mntvnodes);
vp 443 kern/vfs_subr.c if ((vp->v_mount = mp) != NULL)
vp 444 kern/vfs_subr.c LIST_INSERT_HEAD(&mp->mnt_vnodelist, vp, v_mntvnodes);
vp 476 kern/vfs_subr.c struct vnode *vp;
vp 489 kern/vfs_subr.c vp = nvp;
vp 490 kern/vfs_subr.c vp->v_type = type;
vp 491 kern/vfs_subr.c if ((nvp = checkalias(vp, dev, NULL)) != 0) {
vp 492 kern/vfs_subr.c vput(vp);
vp 493 kern/vfs_subr.c vp = nvp;
vp 495 kern/vfs_subr.c *vpp = vp;
vp 511 kern/vfs_subr.c struct vnode *vp;
vp 519 kern/vfs_subr.c for (vp = *vpp; vp; vp = vp->v_specnext) {
vp 520 kern/vfs_subr.c if (nvp_rdev != vp->v_rdev || nvp->v_type != vp->v_type) {
vp 526 kern/vfs_subr.c if (vp->v_usecount == 0) {
vp 527 kern/vfs_subr.c vgonel(vp, p);
vp 530 kern/vfs_subr.c if (vget(vp, LK_EXCLUSIVE, p)) {
vp 539 kern/vfs_subr.c if (vp == NULL || !(vp->v_tag == VT_NON && vp->v_type == VBLK)) {
vp 549 kern/vfs_subr.c if (vp != NULLVP) {
vp 551 kern/vfs_subr.c vp->v_flag |= VALIASED;
vp 552 kern/vfs_subr.c vput(vp);
vp 568 kern/vfs_subr.c VOP_UNLOCK(vp, 0, p);
vp 569 kern/vfs_subr.c vclean(vp, 0, p);
vp 570 kern/vfs_subr.c vp->v_op = nvp->v_op;
vp 571 kern/vfs_subr.c vp->v_tag = nvp->v_tag;
vp 573 kern/vfs_subr.c insmntque(vp, mp);
vp 574 kern/vfs_subr.c return (vp);
vp 587 kern/vfs_subr.c vget(struct vnode *vp, int flags, struct proc *p)
vp 598 kern/vfs_subr.c if (vp->v_flag & VXLOCK) {
vp 603 kern/vfs_subr.c vp->v_flag |= VXWANT;
vp 604 kern/vfs_subr.c ltsleep(vp, PINOD | PNORELOCK, "vget", 0, NULL);
vp 608 kern/vfs_subr.c onfreelist = vp->v_bioflag & VBIOONFREELIST;
vp 609 kern/vfs_subr.c if (vp->v_usecount == 0 && onfreelist) {
vp 611 kern/vfs_subr.c if (vp->v_holdcnt > 0)
vp 612 kern/vfs_subr.c TAILQ_REMOVE(&vnode_hold_list, vp, v_freelist);
vp 614 kern/vfs_subr.c TAILQ_REMOVE(&vnode_free_list, vp, v_freelist);
vp 615 kern/vfs_subr.c vp->v_bioflag &= ~VBIOONFREELIST;
vp 619 kern/vfs_subr.c vp->v_usecount++;
vp 621 kern/vfs_subr.c if ((error = vn_lock(vp, flags, p)) != 0) {
vp 622 kern/vfs_subr.c vp->v_usecount--;
vp 623 kern/vfs_subr.c if (vp->v_usecount == 0 && onfreelist)
vp 624 kern/vfs_subr.c vputonfreelist(vp);
vp 635 kern/vfs_subr.c vref(struct vnode *vp)
vp 638 kern/vfs_subr.c if (vp->v_usecount == 0)
vp 641 kern/vfs_subr.c vp->v_usecount++;
vp 645 kern/vfs_subr.c vputonfreelist(struct vnode *vp)
vp 652 kern/vfs_subr.c if (vp->v_usecount != 0)
vp 655 kern/vfs_subr.c if (vp->v_bioflag & VBIOONFREELIST) {
vp 656 kern/vfs_subr.c vprint("vnode already on free list: ", vp);
vp 661 kern/vfs_subr.c vp->v_bioflag |= VBIOONFREELIST;
vp 663 kern/vfs_subr.c if (vp->v_holdcnt > 0)
vp 668 kern/vfs_subr.c if (vp->v_type == VBAD)
vp 669 kern/vfs_subr.c TAILQ_INSERT_HEAD(lst, vp, v_freelist);
vp 671 kern/vfs_subr.c TAILQ_INSERT_TAIL(lst, vp, v_freelist);
vp 680 kern/vfs_subr.c vput(struct vnode *vp)
vp 685 kern/vfs_subr.c if (vp == NULL)
vp 690 kern/vfs_subr.c if (vp->v_usecount == 0) {
vp 691 kern/vfs_subr.c vprint("vput: bad ref count", vp);
vp 695 kern/vfs_subr.c vp->v_usecount--;
vp 696 kern/vfs_subr.c if (vp->v_usecount > 0) {
vp 697 kern/vfs_subr.c VOP_UNLOCK(vp, 0, p);
vp 702 kern/vfs_subr.c if (vp->v_writecount != 0) {
vp 703 kern/vfs_subr.c vprint("vput: bad writecount", vp);
vp 708 kern/vfs_subr.c VOP_INACTIVE(vp, p);
vp 710 kern/vfs_subr.c if (vp->v_usecount == 0 && !(vp->v_bioflag & VBIOONFREELIST))
vp 711 kern/vfs_subr.c vputonfreelist(vp);
vp 719 kern/vfs_subr.c vrele(struct vnode *vp)
vp 724 kern/vfs_subr.c if (vp == NULL)
vp 728 kern/vfs_subr.c if (vp->v_usecount == 0) {
vp 729 kern/vfs_subr.c vprint("vrele: bad ref count", vp);
vp 733 kern/vfs_subr.c vp->v_usecount--;
vp 734 kern/vfs_subr.c if (vp->v_usecount > 0) {
vp 739 kern/vfs_subr.c if (vp->v_writecount != 0) {
vp 740 kern/vfs_subr.c vprint("vrele: bad writecount", vp);
vp 745 kern/vfs_subr.c if (vn_lock(vp, LK_EXCLUSIVE, p)) {
vp 747 kern/vfs_subr.c vprint("vrele: cannot lock", vp);
vp 752 kern/vfs_subr.c VOP_INACTIVE(vp, p);
vp 754 kern/vfs_subr.c if (vp->v_usecount == 0 && !(vp->v_bioflag & VBIOONFREELIST))
vp 755 kern/vfs_subr.c vputonfreelist(vp);
vp 758 kern/vfs_subr.c void vhold(struct vnode *vp);
vp 764 kern/vfs_subr.c vhold(struct vnode *vp)
vp 770 kern/vfs_subr.c if ((vp->v_bioflag & VBIOONFREELIST) &&
vp 771 kern/vfs_subr.c vp->v_holdcnt == 0 && vp->v_usecount == 0) {
vp 772 kern/vfs_subr.c TAILQ_REMOVE(&vnode_free_list, vp, v_freelist);
vp 773 kern/vfs_subr.c TAILQ_INSERT_TAIL(&vnode_hold_list, vp, v_freelist);
vp 775 kern/vfs_subr.c vp->v_holdcnt++;
vp 794 kern/vfs_subr.c struct vnode *vp, *nvp;
vp 798 kern/vfs_subr.c for (vp = LIST_FIRST(&mp->mnt_vnodelist); vp != NULL; vp = nvp) {
vp 799 kern/vfs_subr.c if (vp->v_mount != mp)
vp 801 kern/vfs_subr.c nvp = LIST_NEXT(vp, v_mntvnodes);
vp 803 kern/vfs_subr.c error = func(vp, arg);
vp 819 kern/vfs_subr.c vflush_vnode(struct vnode *vp, void *arg) {
vp 823 kern/vfs_subr.c if (vp == va->skipvp) {
vp 827 kern/vfs_subr.c if ((va->flags & SKIPSYSTEM) && (vp->v_flag & VSYSTEM)) {
vp 836 kern/vfs_subr.c (vp->v_writecount == 0 || vp->v_type != VREG)) {
vp 844 kern/vfs_subr.c if (vp->v_usecount == 0) {
vp 845 kern/vfs_subr.c vgonel(vp, p);
vp 855 kern/vfs_subr.c if (vp->v_type != VBLK && vp->v_type != VCHR) {
vp 856 kern/vfs_subr.c vgonel(vp, p);
vp 858 kern/vfs_subr.c vclean(vp, 0, p);
vp 859 kern/vfs_subr.c vp->v_op = spec_vnodeop_p;
vp 860 kern/vfs_subr.c insmntque(vp, (struct mount *)0);
vp 867 kern/vfs_subr.c vprint("vflush: busy vnode", vp);
vp 892 kern/vfs_subr.c vclean(struct vnode *vp, int flags, struct proc *p)
vp 902 kern/vfs_subr.c if ((active = vp->v_usecount) != 0)
vp 903 kern/vfs_subr.c vp->v_usecount++;
vp 909 kern/vfs_subr.c if (vp->v_flag & VXLOCK)
vp 911 kern/vfs_subr.c vp->v_flag |= VXLOCK;
vp 919 kern/vfs_subr.c VOP_LOCK(vp, LK_DRAIN, p);
vp 924 kern/vfs_subr.c uvm_vnp_terminate(vp);
vp 929 kern/vfs_subr.c vinvalbuf(vp, V_SAVE, NOCRED, p, 0, 0);
vp 937 kern/vfs_subr.c VOP_CLOSE(vp, FNONBLOCK, NOCRED, p);
vp 938 kern/vfs_subr.c VOP_INACTIVE(vp, p);
vp 944 kern/vfs_subr.c VOP_UNLOCK(vp, 0, p);
vp 950 kern/vfs_subr.c if (VOP_RECLAIM(vp, p))
vp 953 kern/vfs_subr.c vp->v_usecount--;
vp 954 kern/vfs_subr.c if (vp->v_usecount == 0) {
vp 955 kern/vfs_subr.c if (vp->v_holdcnt > 0)
vp 957 kern/vfs_subr.c vputonfreelist(vp);
vp 960 kern/vfs_subr.c cache_purge(vp);
vp 965 kern/vfs_subr.c vp->v_op = dead_vnodeop_p;
vp 966 kern/vfs_subr.c VN_KNOTE(vp, NOTE_REVOKE);
vp 967 kern/vfs_subr.c vp->v_tag = VT_NON;
vp 968 kern/vfs_subr.c vp->v_flag &= ~VXLOCK;
vp 970 kern/vfs_subr.c vp->v_flag &= ~VLOCKSWORK;
vp 972 kern/vfs_subr.c if (vp->v_flag & VXWANT) {
vp 973 kern/vfs_subr.c vp->v_flag &= ~VXWANT;
vp 974 kern/vfs_subr.c wakeup(vp);
vp 982 kern/vfs_subr.c vrecycle(struct vnode *vp, struct proc *p)
vp 984 kern/vfs_subr.c if (vp->v_usecount == 0) {
vp 985 kern/vfs_subr.c vgonel(vp, p);
vp 996 kern/vfs_subr.c vgone(struct vnode *vp)
vp 999 kern/vfs_subr.c vgonel(vp, p);
vp 1006 kern/vfs_subr.c vgonel(struct vnode *vp, struct proc *p)
vp 1017 kern/vfs_subr.c if (vp->v_flag & VXLOCK) {
vp 1018 kern/vfs_subr.c vp->v_flag |= VXWANT;
vp 1019 kern/vfs_subr.c ltsleep(vp, PINOD | PNORELOCK, "vgone", 0, NULL);
vp 1026 kern/vfs_subr.c vclean(vp, DOCLOSE, p);
vp 1030 kern/vfs_subr.c if (vp->v_mount != NULL)
vp 1031 kern/vfs_subr.c insmntque(vp, (struct mount *)0);
vp 1036 kern/vfs_subr.c if ((vp->v_type == VBLK || vp->v_type == VCHR) && vp->v_specinfo != 0) {
vp 1037 kern/vfs_subr.c if (*vp->v_hashchain == vp) {
vp 1038 kern/vfs_subr.c *vp->v_hashchain = vp->v_specnext;
vp 1040 kern/vfs_subr.c for (vq = *vp->v_hashchain; vq; vq = vq->v_specnext) {
vp 1041 kern/vfs_subr.c if (vq->v_specnext != vp)
vp 1043 kern/vfs_subr.c vq->v_specnext = vp->v_specnext;
vp 1049 kern/vfs_subr.c if (vp->v_flag & VALIASED) {
vp 1051 kern/vfs_subr.c for (vq = *vp->v_hashchain; vq; vq = vq->v_specnext) {
vp 1052 kern/vfs_subr.c if (vq->v_rdev != vp->v_rdev ||
vp 1053 kern/vfs_subr.c vq->v_type != vp->v_type)
vp 1063 kern/vfs_subr.c vp->v_flag &= ~VALIASED;
vp 1071 kern/vfs_subr.c mp = vp->v_specmountpoint;
vp 1079 kern/vfs_subr.c FREE(vp->v_specinfo, M_VNODE);
vp 1080 kern/vfs_subr.c vp->v_specinfo = NULL;
vp 1086 kern/vfs_subr.c vp->v_type = VBAD;
vp 1092 kern/vfs_subr.c if (vp->v_usecount == 0 &&
vp 1093 kern/vfs_subr.c (vp->v_bioflag & VBIOONFREELIST)) {
vp 1098 kern/vfs_subr.c if (vp->v_holdcnt > 0)
vp 1101 kern/vfs_subr.c if (TAILQ_FIRST(&vnode_free_list) != vp) {
vp 1102 kern/vfs_subr.c TAILQ_REMOVE(&vnode_free_list, vp, v_freelist);
vp 1103 kern/vfs_subr.c TAILQ_INSERT_HEAD(&vnode_free_list, vp, v_freelist);
vp 1115 kern/vfs_subr.c struct vnode *vp;
vp 1118 kern/vfs_subr.c for (vp = speclisth[SPECHASH(dev)]; vp; vp = vp->v_specnext) {
vp 1119 kern/vfs_subr.c if (dev != vp->v_rdev || type != vp->v_type)
vp 1121 kern/vfs_subr.c *vpp = vp;
vp 1135 kern/vfs_subr.c struct vnode *vp;
vp 1139 kern/vfs_subr.c if (vfinddev(makedev(maj, mn), type, &vp))
vp 1140 kern/vfs_subr.c VOP_REVOKE(vp, REVOKEALL);
vp 1147 kern/vfs_subr.c vcount(struct vnode *vp)
vp 1153 kern/vfs_subr.c if ((vp->v_flag & VALIASED) == 0)
vp 1154 kern/vfs_subr.c return (vp->v_usecount);
vp 1155 kern/vfs_subr.c for (count = 0, vq = *vp->v_hashchain; vq; vq = vnext) {
vp 1157 kern/vfs_subr.c if (vq->v_rdev != vp->v_rdev || vq->v_type != vp->v_type)
vp 1162 kern/vfs_subr.c if (vq->v_usecount == 0 && vq != vp) {
vp 1179 kern/vfs_subr.c vprint(char *label, struct vnode *vp)
vp 1186 kern/vfs_subr.c vp, typename[vp->v_type], vp->v_usecount, vp->v_writecount,
vp 1187 kern/vfs_subr.c vp->v_holdcnt);
vp 1189 kern/vfs_subr.c if (vp->v_flag & VROOT)
vp 1191 kern/vfs_subr.c if (vp->v_flag & VTEXT)
vp 1193 kern/vfs_subr.c if (vp->v_flag & VSYSTEM)
vp 1195 kern/vfs_subr.c if (vp->v_flag & VXLOCK)
vp 1197 kern/vfs_subr.c if (vp->v_flag & VXWANT)
vp 1199 kern/vfs_subr.c if (vp->v_bioflag & VBIOWAIT)
vp 1201 kern/vfs_subr.c if (vp->v_bioflag & VBIOONFREELIST)
vp 1203 kern/vfs_subr.c if (vp->v_bioflag & VBIOONSYNCLIST)
vp 1205 kern/vfs_subr.c if (vp->v_flag & VALIASED)
vp 1209 kern/vfs_subr.c if (vp->v_data == NULL) {
vp 1213 kern/vfs_subr.c VOP_PRINT(vp);
vp 1227 kern/vfs_subr.c struct vnode *vp;
vp 1237 kern/vfs_subr.c LIST_FOREACH(vp, &mp->mnt_vnodelist, v_mntvnodes) {
vp 1238 kern/vfs_subr.c if (VOP_ISLOCKED(vp))
vp 1239 kern/vfs_subr.c vprint((char *)0, vp);
vp 1317 kern/vfs_subr.c struct vnode *vp, *nvp;
vp 1336 kern/vfs_subr.c for (vp = LIST_FIRST(&mp->mnt_vnodelist); vp != NULL;
vp 1337 kern/vfs_subr.c vp = nvp) {
vp 1343 kern/vfs_subr.c if (vp->v_mount != mp) {
vp 1349 kern/vfs_subr.c nvp = LIST_NEXT(vp, v_mntvnodes);
vp 1355 kern/vfs_subr.c if ((error = copyout(&vp,
vp 1358 kern/vfs_subr.c (error = copyout(vp,
vp 1380 kern/vfs_subr.c vfs_mountedon(struct vnode *vp)
vp 1385 kern/vfs_subr.c if (vp->v_specmountpoint != NULL)
vp 1387 kern/vfs_subr.c if (vp->v_flag & VALIASED) {
vp 1388 kern/vfs_subr.c for (vq = *vp->v_hashchain; vq; vq = vq->v_specnext) {
vp 1389 kern/vfs_subr.c if (vq->v_rdev != vp->v_rdev ||
vp 1390 kern/vfs_subr.c vq->v_type != vp->v_type)
vp 1778 kern/vfs_subr.c vwaitforio(struct vnode *vp, int slpflag, char *wmesg, int timeo)
vp 1784 kern/vfs_subr.c while (vp->v_numoutput) {
vp 1785 kern/vfs_subr.c vp->v_bioflag |= VBIOWAIT;
vp 1786 kern/vfs_subr.c error = tsleep(&vp->v_numoutput,
vp 1801 kern/vfs_subr.c vwakeup(struct vnode *vp)
vp 1805 kern/vfs_subr.c if (vp != NULL) {
vp 1806 kern/vfs_subr.c if (vp->v_numoutput-- == 0)
vp 1808 kern/vfs_subr.c if ((vp->v_bioflag & VBIOWAIT) && vp->v_numoutput == 0) {
vp 1809 kern/vfs_subr.c vp->v_bioflag &= ~VBIOWAIT;
vp 1810 kern/vfs_subr.c wakeup(&vp->v_numoutput);
vp 1820 kern/vfs_subr.c vinvalbuf(struct vnode *vp, int flags, struct ucred *cred, struct proc *p,
vp 1828 kern/vfs_subr.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 1834 kern/vfs_subr.c vwaitforio(vp, 0, "vinvalbuf", 0);
vp 1835 kern/vfs_subr.c if (!LIST_EMPTY(&vp->v_dirtyblkhd)) {
vp 1837 kern/vfs_subr.c if ((error = VOP_FSYNC(vp, cred, MNT_WAIT, p)) != 0)
vp 1840 kern/vfs_subr.c if (vp->v_numoutput > 0 ||
vp 1841 kern/vfs_subr.c !LIST_EMPTY(&vp->v_dirtyblkhd))
vp 1849 kern/vfs_subr.c if ((blist = LIST_FIRST(&vp->v_cleanblkhd)) &&
vp 1854 kern/vfs_subr.c (blist = LIST_FIRST(&vp->v_dirtyblkhd)) &&
vp 1892 kern/vfs_subr.c (!LIST_EMPTY(&vp->v_dirtyblkhd) || !LIST_EMPTY(&vp->v_cleanblkhd)))
vp 1899 kern/vfs_subr.c vflushbuf(struct vnode *vp, int sync)
vp 1906 kern/vfs_subr.c for (bp = LIST_FIRST(&vp->v_dirtyblkhd);
vp 1907 kern/vfs_subr.c bp != LIST_END(&vp->v_dirtyblkhd); bp = nbp) {
vp 1920 kern/vfs_subr.c if (bp->b_vp == vp || sync == 0)
vp 1930 kern/vfs_subr.c vwaitforio(vp, 0, "vflushbuf", 0);
vp 1931 kern/vfs_subr.c if (!LIST_EMPTY(&vp->v_dirtyblkhd)) {
vp 1934 kern/vfs_subr.c vprint("vflushbuf: dirty", vp);
vp 1947 kern/vfs_subr.c bgetvp(struct vnode *vp, struct buf *bp)
vp 1954 kern/vfs_subr.c vhold(vp);
vp 1955 kern/vfs_subr.c bp->b_vp = vp;
vp 1956 kern/vfs_subr.c if (vp->v_type == VBLK || vp->v_type == VCHR)
vp 1957 kern/vfs_subr.c bp->b_dev = vp->v_rdev;
vp 1963 kern/vfs_subr.c bufinsvn(bp, &vp->v_cleanblkhd);
vp 1974 kern/vfs_subr.c struct vnode *vp;
vp 1978 kern/vfs_subr.c if ((vp = bp->b_vp) == (struct vnode *) 0)
vp 1985 kern/vfs_subr.c if ((vp->v_bioflag & VBIOONSYNCLIST) &&
vp 1986 kern/vfs_subr.c LIST_FIRST(&vp->v_dirtyblkhd) == NULL) {
vp 1987 kern/vfs_subr.c vp->v_bioflag &= ~VBIOONSYNCLIST;
vp 1988 kern/vfs_subr.c LIST_REMOVE(vp, v_synclist);
vp 1993 kern/vfs_subr.c if (vp->v_holdcnt == 0)
vp 1996 kern/vfs_subr.c vp->v_holdcnt--;
vp 2002 kern/vfs_subr.c if ((vp->v_bioflag & VBIOONFREELIST) &&
vp 2003 kern/vfs_subr.c vp->v_holdcnt == 0 && vp->v_usecount == 0) {
vp 2004 kern/vfs_subr.c TAILQ_REMOVE(&vnode_hold_list, vp, v_freelist);
vp 2005 kern/vfs_subr.c TAILQ_INSERT_TAIL(&vnode_free_list, vp, v_freelist);
vp 2049 kern/vfs_subr.c struct vnode *vp = bp->b_vp;
vp 2064 kern/vfs_subr.c listheadp = &vp->v_cleanblkhd;
vp 2065 kern/vfs_subr.c if ((vp->v_bioflag & VBIOONSYNCLIST) &&
vp 2066 kern/vfs_subr.c LIST_FIRST(&vp->v_dirtyblkhd) == NULL) {
vp 2067 kern/vfs_subr.c vp->v_bioflag &= ~VBIOONSYNCLIST;
vp 2068 kern/vfs_subr.c LIST_REMOVE(vp, v_synclist);
vp 2071 kern/vfs_subr.c listheadp = &vp->v_dirtyblkhd;
vp 2072 kern/vfs_subr.c if ((vp->v_bioflag & VBIOONSYNCLIST) == 0) {
vp 2073 kern/vfs_subr.c switch (vp->v_type) {
vp 2078 kern/vfs_subr.c if (vp->v_specmountpoint != NULL) {
vp 2086 kern/vfs_subr.c vn_syncer_add_to_worklist(vp, delay);
vp 2162 kern/vfs_subr.c vn_isdisk(struct vnode *vp, int *errp)
vp 2164 kern/vfs_subr.c if (vp->v_type != VBLK && vp->v_type != VCHR)
vp 2202 kern/vfs_subr.c vfs_vnode_print(struct vnode *vp, int full, int (*pr)(const char *, ...))
vp 2207 kern/vfs_subr.c vp->v_tag > NENTS(vtags)? "<unk>":vtags[vp->v_tag], vp->v_tag,
vp 2208 kern/vfs_subr.c vp->v_type > NENTS(vtypes)? "<unk>":vtypes[vp->v_type],
vp 2209 kern/vfs_subr.c vp->v_type, vp->v_mount, vp->v_mountedhere);
vp 2212 kern/vfs_subr.c vp->v_data, vp->v_usecount, vp->v_writecount,
vp 2213 kern/vfs_subr.c vp->v_holdcnt, vp->v_numoutput);
vp 2221 kern/vfs_subr.c LIST_FOREACH(bp, &vp->v_cleanblkhd, b_vnbufs) {
vp 2227 kern/vfs_subr.c LIST_FOREACH(bp, &vp->v_dirtyblkhd, b_vnbufs) {
vp 2238 kern/vfs_subr.c struct vnode *vp;
vp 2269 kern/vfs_subr.c LIST_FOREACH(vp, &mp->mnt_vnodelist, v_mntvnodes)
vp 2270 kern/vfs_subr.c if (VOP_ISLOCKED(vp)) {
vp 2271 kern/vfs_subr.c if (!LIST_NEXT(vp, v_mntvnodes))
vp 2272 kern/vfs_subr.c (*pr)(" %p", vp);
vp 2274 kern/vfs_subr.c (*pr)("\n\t%p", vp);
vp 2276 kern/vfs_subr.c (*pr)(", %p", vp);
vp 2283 kern/vfs_subr.c LIST_FOREACH(vp, &mp->mnt_vnodelist, v_mntvnodes)
vp 2284 kern/vfs_subr.c if (!LIST_NEXT(vp, v_mntvnodes))
vp 2285 kern/vfs_subr.c (*pr)(" %p", vp);
vp 2287 kern/vfs_subr.c (*pr)(" %p,\n\t", vp);
vp 2289 kern/vfs_subr.c (*pr)(" %p,", vp);
vp 2290 kern/vfs_subr.c (*pr)("\n", vp);
vp 115 kern/vfs_sync.c vn_syncer_add_to_worklist(struct vnode *vp, int delay)
vp 124 kern/vfs_sync.c if (vp->v_bioflag & VBIOONSYNCLIST)
vp 125 kern/vfs_sync.c LIST_REMOVE(vp, v_synclist);
vp 127 kern/vfs_sync.c vp->v_bioflag |= VBIOONSYNCLIST;
vp 128 kern/vfs_sync.c LIST_INSERT_HEAD(&syncer_workitem_pending[slot], vp, v_synclist);
vp 139 kern/vfs_sync.c struct vnode *vp;
vp 158 kern/vfs_sync.c while ((vp = LIST_FIRST(slp)) != NULL) {
vp 159 kern/vfs_sync.c if (vget(vp, LK_EXCLUSIVE | LK_NOWAIT, p)) {
vp 165 kern/vfs_sync.c vn_syncer_add_to_worklist(vp, 1);
vp 169 kern/vfs_sync.c (void) VOP_FSYNC(vp, p->p_ucred, MNT_LAZY, p);
vp 170 kern/vfs_sync.c vput(vp);
vp 172 kern/vfs_sync.c if (LIST_FIRST(slp) == vp) {
vp 180 kern/vfs_sync.c if (LIST_FIRST(&vp->v_dirtyblkhd) == NULL &&
vp 181 kern/vfs_sync.c vp->v_type != VBLK) {
vp 182 kern/vfs_sync.c vprint("fsync failed", vp);
vp 183 kern/vfs_sync.c if (vp->v_mount != NULL)
vp 185 kern/vfs_sync.c vp->v_mount->mnt_stat.f_mntonname);
vp 195 kern/vfs_sync.c vn_syncer_add_to_worklist(vp, syncdelay);
vp 292 kern/vfs_sync.c struct vnode *vp;
vp 297 kern/vfs_sync.c if ((error = getnewvnode(VT_VFS, mp, sync_vnodeop_p, &vp)) != 0) {
vp 301 kern/vfs_sync.c vp->v_writecount = 1;
vp 302 kern/vfs_sync.c vp->v_type = VNON;
vp 319 kern/vfs_sync.c vn_syncer_add_to_worklist(vp, next);
vp 320 kern/vfs_sync.c mp->mnt_syncer = vp;
vp 370 kern/vfs_sync.c struct vnode *vp = ap->a_vp;
vp 373 kern/vfs_sync.c if (vp->v_usecount == 0) {
vp 374 kern/vfs_sync.c VOP_UNLOCK(vp, 0, ap->a_p);
vp 378 kern/vfs_sync.c vp->v_mount->mnt_syncer = NULL;
vp 382 kern/vfs_sync.c LIST_REMOVE(vp, v_synclist);
vp 383 kern/vfs_sync.c vp->v_bioflag &= ~VBIOONSYNCLIST;
vp 387 kern/vfs_sync.c vp->v_writecount = 0;
vp 388 kern/vfs_sync.c vput(vp);
vp 88 kern/vfs_syscalls.c struct vnode *vp;
vp 116 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 118 kern/vfs_syscalls.c if ((vp->v_flag & VROOT) == 0) {
vp 119 kern/vfs_syscalls.c vput(vp);
vp 122 kern/vfs_syscalls.c mp = vp->v_mount;
vp 130 kern/vfs_syscalls.c vput(vp);
vp 140 kern/vfs_syscalls.c vput(vp);
vp 150 kern/vfs_syscalls.c vput(vp);
vp 158 kern/vfs_syscalls.c vput(vp);
vp 161 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 169 kern/vfs_syscalls.c if ((error = VOP_GETATTR(vp, &va, p->p_ucred, p)) ||
vp 172 kern/vfs_syscalls.c vput(vp);
vp 182 kern/vfs_syscalls.c vput(vp);
vp 186 kern/vfs_syscalls.c if (vp->v_mount->mnt_flag & MNT_NOEXEC)
vp 189 kern/vfs_syscalls.c if ((error = vinvalbuf(vp, V_SAVE, p->p_ucred, p, 0, 0)) != 0) {
vp 190 kern/vfs_syscalls.c vput(vp);
vp 193 kern/vfs_syscalls.c if (vp->v_type != VDIR) {
vp 194 kern/vfs_syscalls.c vput(vp);
vp 212 kern/vfs_syscalls.c vput(vp);
vp 218 kern/vfs_syscalls.c vput(vp);
vp 228 kern/vfs_syscalls.c vput(vp);
vp 232 kern/vfs_syscalls.c if (vp->v_mountedhere != NULL) {
vp 233 kern/vfs_syscalls.c vput(vp);
vp 248 kern/vfs_syscalls.c mp->mnt_vnodecovered = vp;
vp 272 kern/vfs_syscalls.c vrele(vp);
vp 293 kern/vfs_syscalls.c vp->v_mountedhere = mp;
vp 298 kern/vfs_syscalls.c cache_purge(vp);
vp 302 kern/vfs_syscalls.c checkdirs(vp);
vp 303 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 309 kern/vfs_syscalls.c vrele(vp);
vp 314 kern/vfs_syscalls.c vput(vp);
vp 370 kern/vfs_syscalls.c struct vnode *vp;
vp 379 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 380 kern/vfs_syscalls.c mp = vp->v_mount;
vp 388 kern/vfs_syscalls.c vput(vp);
vp 396 kern/vfs_syscalls.c vput(vp);
vp 403 kern/vfs_syscalls.c if ((vp->v_flag & VROOT) == 0) {
vp 404 kern/vfs_syscalls.c vput(vp);
vp 407 kern/vfs_syscalls.c vput(vp);
vp 412 kern/vfs_syscalls.c return (dounmount(mp, SCARG(uap, flags), p, vp));
vp 700 kern/vfs_syscalls.c struct vnode *vp, *tdp;
vp 707 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 708 kern/vfs_syscalls.c VREF(vp);
vp 710 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 711 kern/vfs_syscalls.c if (vp->v_type != VDIR)
vp 714 kern/vfs_syscalls.c error = VOP_ACCESS(vp, VEXEC, p->p_ucred, p);
vp 716 kern/vfs_syscalls.c while (!error && (mp = vp->v_mountedhere) != NULL) {
vp 723 kern/vfs_syscalls.c vput(vp);
vp 724 kern/vfs_syscalls.c vp = tdp;
vp 727 kern/vfs_syscalls.c vput(vp);
vp 730 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 732 kern/vfs_syscalls.c fdp->fd_cdir = vp;
vp 799 kern/vfs_syscalls.c struct vnode *vp;
vp 804 kern/vfs_syscalls.c vp = ndp->ni_vp;
vp 805 kern/vfs_syscalls.c if (vp->v_type != VDIR)
vp 808 kern/vfs_syscalls.c error = VOP_ACCESS(vp, VEXEC, p->p_ucred, p);
vp 810 kern/vfs_syscalls.c vput(vp);
vp 812 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 830 kern/vfs_syscalls.c struct vnode *vp;
vp 866 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 870 kern/vfs_syscalls.c fp->f_data = vp;
vp 882 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 883 kern/vfs_syscalls.c error = VOP_ADVLOCK(vp, (caddr_t)fp, F_SETLK, &lf, type);
vp 890 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 896 kern/vfs_syscalls.c else if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 898 kern/vfs_syscalls.c else if (vp->v_type == VDIR)
vp 900 kern/vfs_syscalls.c else if ((error = vn_writechk(vp)) == 0) {
vp 903 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, fp->f_cred, p);
vp 906 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 913 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 931 kern/vfs_syscalls.c struct vnode *vp;
vp 947 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 949 kern/vfs_syscalls.c fh.fh_fsid = vp->v_mount->mnt_stat.f_fsid;
vp 950 kern/vfs_syscalls.c error = VFS_VPTOFH(vp, &fh.fh_fid);
vp 951 kern/vfs_syscalls.c vput(vp);
vp 973 kern/vfs_syscalls.c struct vnode *vp = NULL;
vp 1008 kern/vfs_syscalls.c if ((error = VFS_FHTOVP(mp, &fh.fh_fid, &vp)) != 0) {
vp 1009 kern/vfs_syscalls.c vp = NULL; /* most likely unnecessary sanity for bad: */
vp 1015 kern/vfs_syscalls.c if (vp->v_type == VSOCK) {
vp 1020 kern/vfs_syscalls.c if ((error = VOP_ACCESS(vp, VREAD, cred, p)) != 0)
vp 1024 kern/vfs_syscalls.c if (vp->v_type == VDIR) {
vp 1028 kern/vfs_syscalls.c if ((error = vn_writechk(vp)) != 0 ||
vp 1029 kern/vfs_syscalls.c (error = VOP_ACCESS(vp, VWRITE, cred, p)) != 0)
vp 1035 kern/vfs_syscalls.c if ((error = VOP_SETATTR(vp, &va, cred, p)) != 0)
vp 1038 kern/vfs_syscalls.c if ((error = VOP_OPEN(vp, flags, cred, p)) != 0)
vp 1041 kern/vfs_syscalls.c vp->v_writecount++;
vp 1048 kern/vfs_syscalls.c fp->f_data = vp;
vp 1060 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 1061 kern/vfs_syscalls.c error = VOP_ADVLOCK(vp, (caddr_t)fp, F_SETLK, &lf, type);
vp 1064 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 1067 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 1078 kern/vfs_syscalls.c if (vp != NULL)
vp 1079 kern/vfs_syscalls.c vput(vp);
vp 1097 kern/vfs_syscalls.c struct vnode *vp;
vp 1110 kern/vfs_syscalls.c if ((error = VFS_FHTOVP(mp, &fh.fh_fid, &vp)))
vp 1112 kern/vfs_syscalls.c error = vn_stat(vp, &sb, p);
vp 1113 kern/vfs_syscalls.c vput(vp);
vp 1131 kern/vfs_syscalls.c struct vnode *vp;
vp 1145 kern/vfs_syscalls.c if ((error = VFS_FHTOVP(mp, &fh.fh_fid, &vp)))
vp 1147 kern/vfs_syscalls.c mp = vp->v_mount;
vp 1149 kern/vfs_syscalls.c vput(vp);
vp 1168 kern/vfs_syscalls.c struct vnode *vp;
vp 1180 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1181 kern/vfs_syscalls.c if (vp != NULL)
vp 1207 kern/vfs_syscalls.c if (nd.ni_dvp == vp)
vp 1211 kern/vfs_syscalls.c if (vp)
vp 1212 kern/vfs_syscalls.c vrele(vp);
vp 1265 kern/vfs_syscalls.c struct vnode *vp;
vp 1273 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1276 kern/vfs_syscalls.c if (vp->v_type == VDIR) {
vp 1293 kern/vfs_syscalls.c error = VOP_LINK(nd.ni_dvp, vp, &nd.ni_cnd);
vp 1295 kern/vfs_syscalls.c vrele(vp);
vp 1350 kern/vfs_syscalls.c struct vnode *vp;
vp 1358 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1363 kern/vfs_syscalls.c if (vp->v_flag & VROOT) {
vp 1365 kern/vfs_syscalls.c if (nd.ni_dvp == vp)
vp 1369 kern/vfs_syscalls.c vput(vp);
vp 1374 kern/vfs_syscalls.c (void)uvm_vnp_uncache(vp);
vp 1397 kern/vfs_syscalls.c struct vnode *vp;
vp 1405 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 1406 kern/vfs_syscalls.c if (vp->v_type == VFIFO)
vp 1408 kern/vfs_syscalls.c if (vp->v_type == VCHR)
vp 1463 kern/vfs_syscalls.c struct vnode *vp;
vp 1477 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1488 kern/vfs_syscalls.c if ((flags & VWRITE) == 0 || (error = vn_writechk(vp)) == 0)
vp 1489 kern/vfs_syscalls.c error = VOP_ACCESS(vp, flags, cred, p);
vp 1491 kern/vfs_syscalls.c vput(vp);
vp 1593 kern/vfs_syscalls.c struct vnode *vp;
vp 1603 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1604 kern/vfs_syscalls.c if (vp->v_type != VLNK)
vp 1616 kern/vfs_syscalls.c error = VOP_READLINK(vp, &auio, p->p_ucred);
vp 1618 kern/vfs_syscalls.c vput(vp);
vp 1634 kern/vfs_syscalls.c struct vnode *vp;
vp 1642 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1643 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 1644 kern/vfs_syscalls.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 1650 kern/vfs_syscalls.c if ((error = VOP_GETATTR(vp, &vattr, p->p_ucred, p)) != 0)
vp 1659 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 1662 kern/vfs_syscalls.c vput(vp);
vp 1678 kern/vfs_syscalls.c struct vnode *vp;
vp 1684 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 1685 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 1686 kern/vfs_syscalls.c if (vp->v_mount && vp->v_mount->mnt_flag & MNT_RDONLY)
vp 1692 kern/vfs_syscalls.c if ((error = VOP_GETATTR(vp, &vattr, p->p_ucred, p))
vp 1702 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 1705 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 1721 kern/vfs_syscalls.c struct vnode *vp;
vp 1732 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1733 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 1734 kern/vfs_syscalls.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 1739 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 1741 kern/vfs_syscalls.c vput(vp);
vp 1757 kern/vfs_syscalls.c struct vnode *vp;
vp 1766 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 1767 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 1768 kern/vfs_syscalls.c if (vp->v_mount && vp->v_mount->mnt_flag & MNT_RDONLY)
vp 1773 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 1775 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 1792 kern/vfs_syscalls.c struct vnode *vp;
vp 1801 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1802 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 1803 kern/vfs_syscalls.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 1808 kern/vfs_syscalls.c error = VOP_GETATTR(vp, &vattr, p->p_ucred, p);
vp 1821 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 1824 kern/vfs_syscalls.c vput(vp);
vp 1840 kern/vfs_syscalls.c struct vnode *vp;
vp 1849 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1850 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 1851 kern/vfs_syscalls.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 1856 kern/vfs_syscalls.c error = VOP_GETATTR(vp, &vattr, p->p_ucred, p);
vp 1869 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 1872 kern/vfs_syscalls.c vput(vp);
vp 1888 kern/vfs_syscalls.c struct vnode *vp;
vp 1896 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 1897 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 1898 kern/vfs_syscalls.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 1903 kern/vfs_syscalls.c error = VOP_GETATTR(vp, &vattr, p->p_ucred, p);
vp 1915 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 1918 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 1934 kern/vfs_syscalls.c struct vnode *vp;
vp 1959 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 1960 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 1961 kern/vfs_syscalls.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 1968 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 1970 kern/vfs_syscalls.c vput(vp);
vp 1985 kern/vfs_syscalls.c struct vnode *vp;
vp 2009 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 2010 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 2011 kern/vfs_syscalls.c if (vp->v_mount && vp->v_mount->mnt_flag & MNT_RDONLY)
vp 2018 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 2020 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 2037 kern/vfs_syscalls.c struct vnode *vp;
vp 2045 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 2046 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 2047 kern/vfs_syscalls.c if (vp->v_type == VDIR)
vp 2049 kern/vfs_syscalls.c else if ((error = vn_writechk(vp)) == 0 &&
vp 2050 kern/vfs_syscalls.c (error = VOP_ACCESS(vp, VWRITE, p->p_ucred, p)) == 0) {
vp 2053 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, p->p_ucred, p);
vp 2055 kern/vfs_syscalls.c vput(vp);
vp 2072 kern/vfs_syscalls.c struct vnode *vp;
vp 2084 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 2085 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 2086 kern/vfs_syscalls.c if (vp->v_type == VDIR)
vp 2088 kern/vfs_syscalls.c else if ((error = vn_writechk(vp)) == 0) {
vp 2091 kern/vfs_syscalls.c error = VOP_SETATTR(vp, &vattr, fp->f_cred, p);
vp 2093 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 2109 kern/vfs_syscalls.c struct vnode *vp;
vp 2115 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 2116 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 2117 kern/vfs_syscalls.c error = VOP_FSYNC(vp, fp->f_cred, MNT_WAIT, p);
vp 2119 kern/vfs_syscalls.c if (error == 0 && vp->v_mount && (vp->v_mount->mnt_flag & MNT_SOFTDEP))
vp 2120 kern/vfs_syscalls.c error = softdep_fsync(vp);
vp 2123 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 2226 kern/vfs_syscalls.c struct vnode *vp;
vp 2235 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 2236 kern/vfs_syscalls.c if (vp != NULL) {
vp 2238 kern/vfs_syscalls.c if (nd.ni_dvp == vp)
vp 2242 kern/vfs_syscalls.c vrele(vp);
vp 2264 kern/vfs_syscalls.c struct vnode *vp;
vp 2272 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 2273 kern/vfs_syscalls.c if (vp->v_type != VDIR) {
vp 2280 kern/vfs_syscalls.c if (nd.ni_dvp == vp) {
vp 2287 kern/vfs_syscalls.c if (vp->v_flag & VROOT)
vp 2294 kern/vfs_syscalls.c if (nd.ni_dvp == vp)
vp 2298 kern/vfs_syscalls.c vput(vp);
vp 2315 kern/vfs_syscalls.c struct vnode *vp;
vp 2330 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 2331 kern/vfs_syscalls.c if (vp->v_type != VDIR) {
vp 2343 kern/vfs_syscalls.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 2345 kern/vfs_syscalls.c error = VOP_READDIR(vp, &auio, fp->f_cred, &eofflag, 0, 0);
vp 2347 kern/vfs_syscalls.c VOP_UNLOCK(vp, 0, p);
vp 2386 kern/vfs_syscalls.c struct vnode *vp;
vp 2394 kern/vfs_syscalls.c vp = nd.ni_vp;
vp 2395 kern/vfs_syscalls.c if ((error = VOP_GETATTR(vp, &vattr, p->p_ucred, p)) != 0)
vp 2400 kern/vfs_syscalls.c if (vp->v_usecount > 1 || (vp->v_flag & (VALIASED)))
vp 2401 kern/vfs_syscalls.c VOP_REVOKE(vp, REVOKEALL);
vp 2403 kern/vfs_syscalls.c vrele(vp);
vp 2416 kern/vfs_syscalls.c struct vnode *vp;
vp 2424 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 2425 kern/vfs_syscalls.c if (vp->v_type == VBAD)
vp 2449 kern/vfs_syscalls.c struct vnode *vp;
vp 2458 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 2459 kern/vfs_syscalls.c if (fp->f_type != DTYPE_VNODE || vp->v_type == VFIFO) {
vp 2487 kern/vfs_syscalls.c struct vnode *vp;
vp 2496 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 2497 kern/vfs_syscalls.c if (fp->f_type != DTYPE_VNODE || vp->v_type == VFIFO) {
vp 2525 kern/vfs_syscalls.c struct vnode *vp;
vp 2534 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 2535 kern/vfs_syscalls.c if (fp->f_type != DTYPE_VNODE || vp->v_type == VFIFO) {
vp 2563 kern/vfs_syscalls.c struct vnode *vp;
vp 2572 kern/vfs_syscalls.c vp = (struct vnode *)fp->f_data;
vp 2573 kern/vfs_syscalls.c if (fp->f_type != DTYPE_VNODE || vp->v_type == VFIFO) {
vp 76 kern/vfs_vnops.c struct vnode *vp;
vp 103 kern/vfs_vnops.c vp = ndp->ni_vp;
vp 111 kern/vfs_vnops.c vp = ndp->ni_vp;
vp 124 kern/vfs_vnops.c vp = ndp->ni_vp;
vp 126 kern/vfs_vnops.c if (vp->v_type == VSOCK) {
vp 130 kern/vfs_vnops.c if (vp->v_type == VLNK) {
vp 136 kern/vfs_vnops.c if ((error = VOP_ACCESS(vp, VREAD, cred, p)) != 0)
vp 140 kern/vfs_vnops.c if (vp->v_type == VDIR) {
vp 144 kern/vfs_vnops.c if ((error = vn_writechk(vp)) != 0 ||
vp 145 kern/vfs_vnops.c (error = VOP_ACCESS(vp, VWRITE, cred, p)) != 0)
vp 149 kern/vfs_vnops.c if ((fmode & O_TRUNC) && vp->v_type == VREG) {
vp 152 kern/vfs_vnops.c if ((error = VOP_SETATTR(vp, &va, cred, p)) != 0)
vp 155 kern/vfs_vnops.c if ((error = VOP_OPEN(vp, fmode, cred, p)) != 0)
vp 158 kern/vfs_vnops.c if (vp->v_flag & VCLONED) {
vp 159 kern/vfs_vnops.c struct cloneinfo *cip = (struct cloneinfo *) vp->v_data;
vp 161 kern/vfs_vnops.c vp->v_flag &= ~VCLONED;
vp 163 kern/vfs_vnops.c vp->v_data = cip->ci_data; /* restore v_data */
vp 164 kern/vfs_vnops.c VOP_UNLOCK(vp, 0, p); /* keep a reference */
vp 165 kern/vfs_vnops.c vp = ndp->ni_vp; /* for the increment below */
vp 171 kern/vfs_vnops.c vp->v_writecount++;
vp 174 kern/vfs_vnops.c vput(vp);
vp 183 kern/vfs_vnops.c vn_writechk(struct vnode *vp)
vp 190 kern/vfs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY) {
vp 191 kern/vfs_vnops.c switch (vp->v_type) {
vp 210 kern/vfs_vnops.c if ((vp->v_flag & VTEXT) && !uvm_vnp_uncache(vp))
vp 220 kern/vfs_vnops.c vn_marktext(struct vnode *vp)
vp 222 kern/vfs_vnops.c vp->v_flag |= VTEXT;
vp 229 kern/vfs_vnops.c vn_close(struct vnode *vp, int flags, struct ucred *cred, struct proc *p)
vp 234 kern/vfs_vnops.c vp->v_writecount--;
vp 235 kern/vfs_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 236 kern/vfs_vnops.c error = VOP_CLOSE(vp, flags, cred, p);
vp 237 kern/vfs_vnops.c vput(vp);
vp 245 kern/vfs_vnops.c vn_rdwr(enum uio_rw rw, struct vnode *vp, caddr_t base, int len, off_t offset,
vp 254 kern/vfs_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 265 kern/vfs_vnops.c error = VOP_READ(vp, &auio, ioflg, cred);
vp 267 kern/vfs_vnops.c error = VOP_WRITE(vp, &auio, ioflg, cred);
vp 275 kern/vfs_vnops.c VOP_UNLOCK(vp, 0, p);
vp 285 kern/vfs_vnops.c struct vnode *vp = (struct vnode *)fp->f_data;
vp 290 kern/vfs_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 293 kern/vfs_vnops.c if (vp->v_type != VDIR)
vp 294 kern/vfs_vnops.c error = VOP_READ(vp, uio,
vp 297 kern/vfs_vnops.c VOP_UNLOCK(vp, 0, p);
vp 307 kern/vfs_vnops.c struct vnode *vp = (struct vnode *)fp->f_data;
vp 312 kern/vfs_vnops.c if (vp->v_type == VREG && (fp->f_flag & O_APPEND))
vp 317 kern/vfs_vnops.c (vp->v_mount && (vp->v_mount->mnt_flag & MNT_SYNCHRONOUS)))
vp 319 kern/vfs_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 322 kern/vfs_vnops.c error = VOP_WRITE(vp, uio, ioflag, cred);
vp 327 kern/vfs_vnops.c VOP_UNLOCK(vp, 0, p);
vp 337 kern/vfs_vnops.c struct vnode *vp = (struct vnode *)fp->f_data;
vp 338 kern/vfs_vnops.c return vn_stat(vp, sb, p);
vp 345 kern/vfs_vnops.c vn_stat(struct vnode *vp, struct stat *sb, struct proc *p)
vp 351 kern/vfs_vnops.c error = VOP_GETATTR(vp, &va, p->p_ucred, p);
vp 360 kern/vfs_vnops.c switch (vp->v_type) {
vp 407 kern/vfs_vnops.c struct vnode *vp = ((struct vnode *)fp->f_data);
vp 411 kern/vfs_vnops.c switch (vp->v_type) {
vp 416 kern/vfs_vnops.c error = VOP_GETATTR(vp, &vattr, p->p_ucred, p);
vp 431 kern/vfs_vnops.c error = VOP_IOCTL(vp, com, data, fp->f_flag, p->p_ucred, p);
vp 435 kern/vfs_vnops.c p->p_session->s_ttyvp = vp;
vp 436 kern/vfs_vnops.c VREF(vp);
vp 456 kern/vfs_vnops.c vn_lock(struct vnode *vp, int flags, struct proc *p)
vp 464 kern/vfs_vnops.c if (vp->v_flag & VXLOCK) {
vp 465 kern/vfs_vnops.c vp->v_flag |= VXWANT;
vp 466 kern/vfs_vnops.c tsleep(vp, PINOD, "vn_lock", 0);
vp 469 kern/vfs_vnops.c error = VOP_LOCK(vp, flags, p);
vp 73 kern/vnode_if.c int VOP_ISLOCKED(struct vnode *vp)
vp 77 kern/vnode_if.c a.a_vp = vp;
vp 78 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_islocked), &a));
vp 188 kern/vnode_if.c int VOP_OPEN(struct vnode *vp, int mode, struct ucred *cred, struct proc *p)
vp 192 kern/vnode_if.c a.a_vp = vp;
vp 196 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_open), &a));
vp 215 kern/vnode_if.c int VOP_CLOSE(struct vnode *vp, int fflag, struct ucred *cred, struct proc *p)
vp 219 kern/vnode_if.c a.a_vp = vp;
vp 221 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 227 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_close), &a));
vp 246 kern/vnode_if.c int VOP_ACCESS(struct vnode *vp, int mode, struct ucred *cred, struct proc *p)
vp 250 kern/vnode_if.c a.a_vp = vp;
vp 252 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 258 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_access), &a));
vp 277 kern/vnode_if.c int VOP_GETATTR(struct vnode *vp, struct vattr *vap, struct ucred *cred,
vp 282 kern/vnode_if.c a.a_vp = vp;
vp 286 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_getattr), &a));
vp 305 kern/vnode_if.c int VOP_SETATTR(struct vnode *vp, struct vattr *vap, struct ucred *cred,
vp 310 kern/vnode_if.c a.a_vp = vp;
vp 312 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 318 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_setattr), &a));
vp 337 kern/vnode_if.c int VOP_READ(struct vnode *vp, struct uio *uio, int ioflag, struct ucred *cred)
vp 341 kern/vnode_if.c a.a_vp = vp;
vp 343 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 349 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_read), &a));
vp 368 kern/vnode_if.c int VOP_WRITE(struct vnode *vp, struct uio *uio, int ioflag,
vp 373 kern/vnode_if.c a.a_vp = vp;
vp 375 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 381 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_write), &a));
vp 400 kern/vnode_if.c int VOP_IOCTL(struct vnode *vp, u_long command, void *data, int fflag,
vp 405 kern/vnode_if.c a.a_vp = vp;
vp 411 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_ioctl), &a));
vp 430 kern/vnode_if.c int VOP_POLL(struct vnode *vp, int events, struct proc *p)
vp 434 kern/vnode_if.c a.a_vp = vp;
vp 437 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_poll), &a));
vp 456 kern/vnode_if.c int VOP_KQFILTER(struct vnode *vp, struct knote *kn)
vp 460 kern/vnode_if.c a.a_vp = vp;
vp 462 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_kqfilter), &a));
vp 481 kern/vnode_if.c int VOP_REVOKE(struct vnode *vp, int flags)
vp 485 kern/vnode_if.c a.a_vp = vp;
vp 487 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_revoke), &a));
vp 506 kern/vnode_if.c int VOP_FSYNC(struct vnode *vp, struct ucred *cred, int waitfor,
vp 511 kern/vnode_if.c a.a_vp = vp;
vp 513 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 519 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_fsync), &a));
vp 539 kern/vnode_if.c int VOP_REMOVE(struct vnode *dvp, struct vnode *vp, struct componentname *cnp)
vp 548 kern/vnode_if.c a.a_vp = vp;
vp 550 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 574 kern/vnode_if.c int VOP_LINK(struct vnode *dvp, struct vnode *vp, struct componentname *cnp)
vp 583 kern/vnode_if.c a.a_vp = vp;
vp 675 kern/vnode_if.c int VOP_RMDIR(struct vnode *dvp, struct vnode *vp, struct componentname *cnp)
vp 684 kern/vnode_if.c a.a_vp = vp;
vp 686 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 742 kern/vnode_if.c int VOP_READDIR(struct vnode *vp, struct uio *uio, struct ucred *cred,
vp 747 kern/vnode_if.c a.a_vp = vp;
vp 749 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 757 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_readdir), &a));
vp 776 kern/vnode_if.c int VOP_READLINK(struct vnode *vp, struct uio *uio, struct ucred *cred)
vp 780 kern/vnode_if.c a.a_vp = vp;
vp 782 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 787 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_readlink), &a));
vp 831 kern/vnode_if.c int VOP_INACTIVE(struct vnode *vp, struct proc *p)
vp 835 kern/vnode_if.c a.a_vp = vp;
vp 837 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 841 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_inactive), &a));
vp 860 kern/vnode_if.c int VOP_RECLAIM(struct vnode *vp, struct proc *p)
vp 864 kern/vnode_if.c a.a_vp = vp;
vp 866 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_reclaim), &a));
vp 885 kern/vnode_if.c int VOP_LOCK(struct vnode *vp, int flags, struct proc *p)
vp 889 kern/vnode_if.c a.a_vp = vp;
vp 892 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_lock), &a));
vp 911 kern/vnode_if.c int VOP_UNLOCK(struct vnode *vp, int flags, struct proc *p)
vp 915 kern/vnode_if.c a.a_vp = vp;
vp 918 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_unlock), &a));
vp 937 kern/vnode_if.c int VOP_BMAP(struct vnode *vp, daddr64_t bn, struct vnode **vpp,
vp 942 kern/vnode_if.c a.a_vp = vp;
vp 944 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 951 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_bmap), &a));
vp 970 kern/vnode_if.c int VOP_PRINT(struct vnode *vp)
vp 974 kern/vnode_if.c a.a_vp = vp;
vp 975 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_print), &a));
vp 994 kern/vnode_if.c int VOP_PATHCONF(struct vnode *vp, int name, register_t *retval)
vp 998 kern/vnode_if.c a.a_vp = vp;
vp 1000 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 1005 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_pathconf), &a));
vp 1024 kern/vnode_if.c int VOP_ADVLOCK(struct vnode *vp, void *id, int op, struct flock *fl, int flags)
vp 1028 kern/vnode_if.c a.a_vp = vp;
vp 1033 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_advlock), &a));
vp 1052 kern/vnode_if.c int VOP_REALLOCBLKS(struct vnode *vp, struct cluster_save *buflist)
vp 1056 kern/vnode_if.c a.a_vp = vp;
vp 1058 kern/vnode_if.c if ((vp->v_flag & VLOCKSWORK) && !VOP_ISLOCKED(vp))
vp 1062 kern/vnode_if.c return (VCALL(vp, VOFFSET(vop_reallocblks), &a));
vp 266 lib/libsa/bootp.c struct cmu_vend *vp;
vp 272 lib/libsa/bootp.c vp = (struct cmu_vend *)cp;
vp 274 lib/libsa/bootp.c if (vp->v_smask.s_addr != 0)
vp 275 lib/libsa/bootp.c smask = vp->v_smask.s_addr;
vp 276 lib/libsa/bootp.c if (vp->v_dgate.s_addr != 0)
vp 277 lib/libsa/bootp.c gateip = vp->v_dgate;
vp 239 miscfs/deadfs/dead_vnops.c struct vnode *vp = ap->a_vp;
vp 241 miscfs/deadfs/dead_vnops.c if (ap->a_flags & LK_DRAIN || !chkvnlock(vp))
vp 244 miscfs/deadfs/dead_vnops.c return (VCALL(vp, VOFFSET(vop_lock), ap));
vp 297 miscfs/deadfs/dead_vnops.c chkvnlock(struct vnode *vp)
vp 301 miscfs/deadfs/dead_vnops.c while (vp->v_flag & VXLOCK) {
vp 302 miscfs/deadfs/dead_vnops.c vp->v_flag |= VXWANT;
vp 303 miscfs/deadfs/dead_vnops.c tsleep(vp, PINOD, "chkvnlock", 0);
vp 148 miscfs/fifofs/fifo_vnops.c struct vnode *vp = ap->a_vp;
vp 154 miscfs/fifofs/fifo_vnops.c if ((fip = vp->v_fifoinfo) == NULL) {
vp 156 miscfs/fifofs/fifo_vnops.c vp->v_fifoinfo = fip;
vp 159 miscfs/fifofs/fifo_vnops.c vp->v_fifoinfo = NULL;
vp 166 miscfs/fifofs/fifo_vnops.c vp->v_fifoinfo = NULL;
vp 174 miscfs/fifofs/fifo_vnops.c vp->v_fifoinfo = NULL;
vp 203 miscfs/fifofs/fifo_vnops.c VOP_UNLOCK(vp, 0, p);
vp 206 miscfs/fifofs/fifo_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 211 miscfs/fifofs/fifo_vnops.c VOP_UNLOCK(vp, 0, p);
vp 214 miscfs/fifofs/fifo_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 221 miscfs/fifofs/fifo_vnops.c VOP_CLOSE(vp, ap->a_mode, ap->a_cred, ap->a_p);
vp 377 miscfs/fifofs/fifo_vnops.c struct vnode *vp = ap->a_vp;
vp 378 miscfs/fifofs/fifo_vnops.c struct fifoinfo *fip = vp->v_fifoinfo;
vp 396 miscfs/fifofs/fifo_vnops.c vp->v_fifoinfo = NULL;
vp 405 miscfs/fifofs/fifo_vnops.c struct vnode *vp = ap->a_vp;
vp 406 miscfs/fifofs/fifo_vnops.c struct fifoinfo *fip = vp->v_fifoinfo;
vp 414 miscfs/fifofs/fifo_vnops.c vp->v_fifoinfo = NULL;
vp 437 miscfs/fifofs/fifo_vnops.c fifo_printinfo(struct vnode *vp)
vp 439 miscfs/fifofs/fifo_vnops.c struct fifoinfo *fip = vp->v_fifoinfo;
vp 65 miscfs/portal/portal.h #define VTOPORTAL(vp) ((struct portalnode *)(vp)->v_data)
vp 200 miscfs/portal/portal_vfsops.c struct vnode *vp;
vp 206 miscfs/portal/portal_vfsops.c vp = VFSTOPORTAL(mp)->pm_root;
vp 207 miscfs/portal/portal_vfsops.c VREF(vp);
vp 208 miscfs/portal/portal_vfsops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 209 miscfs/portal/portal_vfsops.c *vpp = vp;
vp 276 miscfs/portal/portal_vnops.c struct vnode *vp = ap->a_vp;
vp 295 miscfs/portal/portal_vnops.c if (vp->v_flag & VROOT)
vp 306 miscfs/portal/portal_vnops.c pt = VTOPORTAL(vp);
vp 307 miscfs/portal/portal_vnops.c fmp = VFSTOPORTAL(vp->v_mount);
vp 509 miscfs/portal/portal_vnops.c struct vnode *vp = ap->a_vp;
vp 516 miscfs/portal/portal_vnops.c vap->va_fsid = vp->v_mount->mnt_stat.f_fsid.val[0];
vp 528 miscfs/portal/portal_vnops.c if (vp->v_flag & VROOT) {
vp 541 miscfs/portal/portal_vnops.c vap->va_fileid = VTOPORTAL(vp)->pt_fileid;
vp 101 miscfs/procfs/procfs.h #define VTOPFS(vp) ((struct pfsnode *)(vp)->v_data)
vp 94 miscfs/procfs/procfs_subr.c struct vnode *vp;
vp 105 miscfs/procfs/procfs_subr.c vp = PFSTOV(pfs);
vp 108 miscfs/procfs/procfs_subr.c vp->v_mount == mp) {
vp 109 miscfs/procfs/procfs_subr.c if (vget(vp, 0, p))
vp 111 miscfs/procfs/procfs_subr.c *vpp = vp;
vp 118 miscfs/procfs/procfs_subr.c vp = *vpp;
vp 121 miscfs/procfs/procfs_subr.c vp->v_data = pfs;
vp 125 miscfs/procfs/procfs_subr.c pfs->pfs_vnode = vp;
vp 132 miscfs/procfs/procfs_subr.c vp->v_type = VDIR;
vp 133 miscfs/procfs/procfs_subr.c vp->v_flag = VROOT;
vp 139 miscfs/procfs/procfs_subr.c vp->v_type = VLNK;
vp 144 miscfs/procfs/procfs_subr.c vp->v_type = VDIR;
vp 152 miscfs/procfs/procfs_subr.c vp->v_type = VREG;
vp 159 miscfs/procfs/procfs_subr.c vp->v_type = VREG;
vp 167 miscfs/procfs/procfs_subr.c vp->v_type = VREG;
vp 176 miscfs/procfs/procfs_subr.c uvm_vnp_setsize(vp, 0);
vp 184 miscfs/procfs/procfs_subr.c procfs_freevp(struct vnode *vp)
vp 186 miscfs/procfs/procfs_subr.c struct pfsnode *pfs = VTOPFS(vp);
vp 189 miscfs/procfs/procfs_subr.c FREE(vp->v_data, M_TEMP);
vp 190 miscfs/procfs/procfs_subr.c vp->v_data = 0;
vp 198 miscfs/procfs/procfs_subr.c struct vnode *vp = ap->a_vp;
vp 201 miscfs/procfs/procfs_subr.c struct pfsnode *pfs = VTOPFS(vp);
vp 331 miscfs/procfs/procfs_vnops.c struct vnode *vp = ap->a_vp;
vp 332 miscfs/procfs/procfs_vnops.c struct pfsnode *pfs = VTOPFS(vp);
vp 334 miscfs/procfs/procfs_vnops.c if (pfind(pfs->pfs_pid) == NULL && !(vp->v_flag & VXLOCK))
vp 335 miscfs/procfs/procfs_vnops.c vgone(vp);
vp 850 miscfs/procfs/procfs_vnops.c struct vnode *vp;
vp 854 miscfs/procfs/procfs_vnops.c vp = ap->a_vp;
vp 855 miscfs/procfs/procfs_vnops.c pfs = VTOPFS(vp);
vp 884 miscfs/procfs/procfs_vnops.c (*pt->pt_valid)(p, vp->v_mount) == 0)
vp 946 miscfs/procfs/procfs_vnops.c if (VFSTOPROC(vp->v_mount)->pmnt_flags &
vp 957 miscfs/procfs/procfs_vnops.c if (VFSTOPROC(vp->v_mount)->pmnt_flags &
vp 31 miscfs/specfs/spec_subr.c struct vnode *cvp, *vp = ap->a_vp;
vp 35 miscfs/specfs/spec_subr.c for (i = 1; i < sizeof(vp->v_specbitmap) * NBBY; i++)
vp 36 miscfs/specfs/spec_subr.c if (isclr(vp->v_specbitmap, i)) {
vp 37 miscfs/specfs/spec_subr.c setbit(vp->v_specbitmap, i);
vp 41 miscfs/specfs/spec_subr.c if (i == sizeof(vp->v_specbitmap) * NBBY)
vp 45 miscfs/specfs/spec_subr.c major(vp->v_rdev), minor(vp->v_rdev), curproc->p_pid);
vp 47 miscfs/specfs/spec_subr.c error = cdevvp(makedev(major(vp->v_rdev), i), &cvp);
vp 51 miscfs/specfs/spec_subr.c VOP_UNLOCK(vp, 0, ap->a_p);
vp 53 miscfs/specfs/spec_subr.c error = cdevsw[major(vp->v_rdev)].d_open(cvp->v_rdev, ap->a_mode,
vp 56 miscfs/specfs/spec_subr.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, ap->a_p);
vp 59 miscfs/specfs/spec_subr.c clrbit(vp->v_specbitmap, i);
vp 64 miscfs/specfs/spec_subr.c cip->ci_data = vp->v_data;
vp 67 miscfs/specfs/spec_subr.c cvp->v_specparent = vp;
vp 68 miscfs/specfs/spec_subr.c vp->v_flag |= VCLONED;
vp 69 miscfs/specfs/spec_subr.c vp->v_data = cip;
vp 80 miscfs/specfs/spec_subr.c struct vnode *pvp, *vp = ap->a_vp;
vp 83 miscfs/specfs/spec_subr.c error = cdevsw[major(vp->v_rdev)].d_close(vp->v_rdev, ap->a_fflag,
vp 88 miscfs/specfs/spec_subr.c pvp = vp->v_specparent; /* get parent device */
vp 89 miscfs/specfs/spec_subr.c clrbit(pvp->v_specbitmap, minor(vp->v_rdev));
vp 92 miscfs/specfs/spec_subr.c " pid %u\n", minor(vp->v_rdev), major(vp->v_rdev), curproc->p_pid);
vp 130 miscfs/specfs/spec_vnops.c struct vnode *vp = ap->a_vp;
vp 133 miscfs/specfs/spec_vnops.c dev_t dev = (dev_t)vp->v_rdev;
vp 140 miscfs/specfs/spec_vnops.c if (vp->v_mount && (vp->v_mount->mnt_flag & MNT_NODEV))
vp 143 miscfs/specfs/spec_vnops.c switch (vp->v_type) {
vp 172 miscfs/specfs/spec_vnops.c vp->v_flag |= VISTTY;
vp 175 miscfs/specfs/spec_vnops.c VOP_UNLOCK(vp, 0, p);
vp 177 miscfs/specfs/spec_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 194 miscfs/specfs/spec_vnops.c if ((error = vfs_mountedon(vp)) != 0)
vp 217 miscfs/specfs/spec_vnops.c struct vnode *vp = ap->a_vp;
vp 237 miscfs/specfs/spec_vnops.c switch (vp->v_type) {
vp 240 miscfs/specfs/spec_vnops.c VOP_UNLOCK(vp, 0, p);
vp 241 miscfs/specfs/spec_vnops.c error = (*cdevsw[major(vp->v_rdev)].d_read)
vp 242 miscfs/specfs/spec_vnops.c (vp->v_rdev, uio, ap->a_ioflag);
vp 243 miscfs/specfs/spec_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 250 miscfs/specfs/spec_vnops.c if ((majordev = major(vp->v_rdev)) < nblkdev &&
vp 252 miscfs/specfs/spec_vnops.c (*ioctl)(vp->v_rdev, DIOCGPART, (caddr_t)&dpart, FREAD, p) == 0) {
vp 266 miscfs/specfs/spec_vnops.c if (vp->v_lastr + bscale == bn) {
vp 268 miscfs/specfs/spec_vnops.c error = breadn(vp, bn, bsize, &nextbn, &bsize,
vp 271 miscfs/specfs/spec_vnops.c error = bread(vp, bn, bsize, NOCRED, &bp);
vp 272 miscfs/specfs/spec_vnops.c vp->v_lastr = bn;
vp 306 miscfs/specfs/spec_vnops.c struct vnode *vp = ap->a_vp;
vp 324 miscfs/specfs/spec_vnops.c switch (vp->v_type) {
vp 327 miscfs/specfs/spec_vnops.c VOP_UNLOCK(vp, 0, p);
vp 328 miscfs/specfs/spec_vnops.c error = (*cdevsw[major(vp->v_rdev)].d_write)
vp 329 miscfs/specfs/spec_vnops.c (vp->v_rdev, uio, ap->a_ioflag);
vp 330 miscfs/specfs/spec_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 339 miscfs/specfs/spec_vnops.c if ((majordev = major(vp->v_rdev)) < nblkdev &&
vp 341 miscfs/specfs/spec_vnops.c (*ioctl)(vp->v_rdev, DIOCGPART, (caddr_t)&dpart, FREAD, p) == 0) {
vp 355 miscfs/specfs/spec_vnops.c error = bread(vp, bn, bsize, NOCRED, &bp);
vp 442 miscfs/specfs/spec_vnops.c struct vnode *vp = ap->a_vp;
vp 447 miscfs/specfs/spec_vnops.c if (vp->v_type == VCHR)
vp 454 miscfs/specfs/spec_vnops.c for (bp = LIST_FIRST(&vp->v_dirtyblkhd);
vp 455 miscfs/specfs/spec_vnops.c bp != LIST_END(&vp->v_dirtyblkhd); bp = nbp) {
vp 468 miscfs/specfs/spec_vnops.c vwaitforio (vp, 0, "spec_fsync", 0);
vp 471 miscfs/specfs/spec_vnops.c if (!LIST_EMPTY(&vp->v_dirtyblkhd)) {
vp 473 miscfs/specfs/spec_vnops.c vprint("spec_fsync: dirty", vp);
vp 522 miscfs/specfs/spec_vnops.c struct vnode *vp = ap->a_vp;
vp 523 miscfs/specfs/spec_vnops.c dev_t dev = vp->v_rdev;
vp 527 miscfs/specfs/spec_vnops.c switch (vp->v_type) {
vp 539 miscfs/specfs/spec_vnops.c if (vcount(vp) == 2 && ap->a_p &&
vp 540 miscfs/specfs/spec_vnops.c vp == ap->a_p->p_session->s_ttyvp) {
vp 541 miscfs/specfs/spec_vnops.c vrele(vp);
vp 549 miscfs/specfs/spec_vnops.c if (vcount(vp) > 1 && (vp->v_flag & VXLOCK) == 0)
vp 565 miscfs/specfs/spec_vnops.c if (!(vp->v_flag & VXLOCK))
vp 566 miscfs/specfs/spec_vnops.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, ap->a_p);
vp 567 miscfs/specfs/spec_vnops.c error = vinvalbuf(vp, V_SAVE, ap->a_cred, ap->a_p, 0, 0);
vp 568 miscfs/specfs/spec_vnops.c if (!(vp->v_flag & VXLOCK))
vp 569 miscfs/specfs/spec_vnops.c VOP_UNLOCK(vp, 0, ap->a_p);
vp 581 miscfs/specfs/spec_vnops.c if (vcount(vp) > 1 && (vp->v_flag & VXLOCK) == 0)
vp 648 miscfs/specfs/spec_vnops.c struct vnode *vp = ap->a_vp;
vp 650 miscfs/specfs/spec_vnops.c return (lf_advlock(&vp->v_speclockf, (off_t)0, ap->a_id,
vp 225 msdosfs/denode.h #define VTODE(vp) ((struct denode *)(vp)->v_data)
vp 106 msdosfs/msdosfs_denode.c struct vnode *vp = DETOV(dep);
vp 108 msdosfs/msdosfs_denode.c if (!vget(vp, LK_EXCLUSIVE, p))
vp 591 msdosfs/msdosfs_denode.c struct vnode *vp = ap->a_vp;
vp 592 msdosfs/msdosfs_denode.c struct denode *dep = VTODE(vp);
vp 596 msdosfs/msdosfs_denode.c if (prtactive && vp->v_usecount != 0)
vp 597 msdosfs/msdosfs_denode.c vprint("msdosfs_reclaim(): pushing active", vp);
vp 612 msdosfs/msdosfs_denode.c cache_purge(vp);
vp 621 msdosfs/msdosfs_denode.c vp->v_data = NULL;
vp 630 msdosfs/msdosfs_denode.c struct vnode *vp = ap->a_vp;
vp 631 msdosfs/msdosfs_denode.c struct denode *dep = VTODE(vp);
vp 637 msdosfs/msdosfs_denode.c if (prtactive && vp->v_usecount != 0)
vp 638 msdosfs/msdosfs_denode.c vprint("msdosfs_inactive(): pushing active", vp);
vp 660 msdosfs/msdosfs_denode.c dep, dep->de_refcnt, vp->v_mount->mnt_flag, MNT_RDONLY);
vp 662 msdosfs/msdosfs_denode.c if (dep->de_refcnt <= 0 && (vp->v_mount->mnt_flag & MNT_RDONLY) == 0) {
vp 669 msdosfs/msdosfs_denode.c VOP_UNLOCK(vp, 0, p);
vp 675 msdosfs/msdosfs_denode.c printf("msdosfs_inactive(): v_usecount %d, de_Name[0] %x\n", vp->v_usecount,
vp 679 msdosfs/msdosfs_denode.c vrecycle(vp, p);
vp 589 msdosfs/msdosfs_vfsops.c struct vnode *vp;
vp 600 msdosfs/msdosfs_vfsops.c vp = pmp->pm_devvp;
vp 602 msdosfs/msdosfs_vfsops.c vprint("msdosfs_umount(): just before calling VOP_CLOSE()\n", vp);
vp 604 msdosfs/msdosfs_vfsops.c error = VOP_CLOSE(vp,
vp 606 msdosfs/msdosfs_vfsops.c vrele(vp);
vp 670 msdosfs/msdosfs_vfsops.c msdosfs_sync_vnode(struct vnode *vp, void *arg)
vp 676 msdosfs/msdosfs_vfsops.c dep = VTODE(vp);
vp 677 msdosfs/msdosfs_vfsops.c if (vp->v_type == VNON ||
vp 679 msdosfs/msdosfs_vfsops.c && LIST_EMPTY(&vp->v_dirtyblkhd)) ||
vp 684 msdosfs/msdosfs_vfsops.c if (vget(vp, LK_EXCLUSIVE | LK_NOWAIT, msa->p))
vp 687 msdosfs/msdosfs_vfsops.c if ((error = VOP_FSYNC(vp, msa->cred, msa->waitfor, msa->p)) != 0)
vp 689 msdosfs/msdosfs_vfsops.c VOP_UNLOCK(vp, 0, msa->p);
vp 690 msdosfs/msdosfs_vfsops.c vrele(vp);
vp 762 msdosfs/msdosfs_vfsops.c msdosfs_vptofh(vp, fhp)
vp 763 msdosfs/msdosfs_vfsops.c struct vnode *vp;
vp 769 msdosfs/msdosfs_vfsops.c dep = VTODE(vp);
vp 202 msdosfs/msdosfs_vnops.c struct vnode *vp = ap->a_vp;
vp 203 msdosfs/msdosfs_vnops.c struct denode *dep = VTODE(vp);
vp 206 msdosfs/msdosfs_vnops.c if (vp->v_usecount > 1 && !VOP_ISLOCKED(vp)) {
vp 439 msdosfs/msdosfs_vnops.c struct vnode *vp = ap->a_vp;
vp 440 msdosfs/msdosfs_vnops.c struct denode *dep = VTODE(vp);
vp 488 msdosfs/msdosfs_vnops.c error = breadn(vp, de_cn2bn(pmp, lbn),
vp 492 msdosfs/msdosfs_vnops.c error = bread(vp, de_cn2bn(pmp, lbn),
vp 504 msdosfs/msdosfs_vnops.c if (!isadir && !(vp->v_mount->mnt_flag & MNT_NOATIME))
vp 528 msdosfs/msdosfs_vnops.c struct vnode *vp = ap->a_vp;
vp 530 msdosfs/msdosfs_vnops.c struct denode *dep = VTODE(vp);
vp 536 msdosfs/msdosfs_vnops.c vp, uio, ioflag, cred);
vp 541 msdosfs/msdosfs_vnops.c switch (vp->v_type) {
vp 545 msdosfs/msdosfs_vnops.c thisvp = vp;
vp 654 msdosfs/msdosfs_vnops.c uvm_vnp_setsize(vp, dep->de_FileSize);
vp 656 msdosfs/msdosfs_vnops.c uvm_vnp_uncache(vp);
vp 741 msdosfs/msdosfs_vnops.c struct vnode *vp = ap->a_vp;
vp 743 msdosfs/msdosfs_vnops.c vflushbuf(vp, ap->a_waitfor == MNT_WAIT);
vp 744 msdosfs/msdosfs_vnops.c return (deupdat(VTODE(vp), ap->a_waitfor == MNT_WAIT));
vp 1299 msdosfs/msdosfs_vnops.c register struct vnode *vp = ap->a_vp;
vp 1305 msdosfs/msdosfs_vnops.c ip = VTODE(vp);
vp 1312 msdosfs/msdosfs_vnops.c vput(vp);
vp 1349 msdosfs/msdosfs_vnops.c cache_purge(vp);
vp 1353 msdosfs/msdosfs_vnops.c vput(vp);
vp 1665 msdosfs/msdosfs_vnops.c struct vnode *vp = ap->a_vp;
vp 1667 msdosfs/msdosfs_vnops.c return (lockmgr(&VTODE(vp)->de_lock, ap->a_flags, NULL));
vp 1675 msdosfs/msdosfs_vnops.c struct vnode *vp = ap->a_vp;
vp 1677 msdosfs/msdosfs_vnops.c return (lockmgr(&VTODE(vp)->de_lock, ap->a_flags | LK_RELEASE, NULL));
vp 1724 msdosfs/msdosfs_vnops.c struct vnode *vp;
vp 1756 msdosfs/msdosfs_vnops.c vp = dep->de_devvp;
vp 1757 msdosfs/msdosfs_vnops.c bp->b_dev = vp->v_rdev;
vp 1758 msdosfs/msdosfs_vnops.c VOCALL(vp->v_op, VOFFSET(vop_strategy), ap);
vp 68 nfs/nfs_bio.c nfs_bioread(vp, uio, ioflag, cred)
vp 69 nfs/nfs_bio.c struct vnode *vp;
vp 74 nfs/nfs_bio.c struct nfsnode *np = VTONFS(vp);
vp 79 nfs/nfs_bio.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 95 nfs/nfs_bio.c (void)nfs_fsinfo(nmp, vp, cred, p);
vp 115 nfs/nfs_bio.c error = VOP_GETATTR(vp, &vattr, cred, p);
vp 120 nfs/nfs_bio.c error = VOP_GETATTR(vp, &vattr, cred, p);
vp 124 nfs/nfs_bio.c error = nfs_vinvalbuf(vp, V_SAVE, cred, p, 1);
vp 140 nfs/nfs_bio.c if ((vp->v_flag & VROOT) && vp->v_type == VLNK) {
vp 141 nfs/nfs_bio.c return (nfs_readlinkrpc(vp, uio, cred));
vp 144 nfs/nfs_bio.c switch (vp->v_type) {
vp 159 nfs/nfs_bio.c if (!incore(vp, rabn)) {
vp 160 nfs/nfs_bio.c rabp = nfs_getcacheblk(vp, rabn, biosize, p);
vp 181 nfs/nfs_bio.c if ((bp = incore(vp, bn)) &&
vp 187 nfs/nfs_bio.c bp = nfs_getcacheblk(vp, bn, biosize, p);
vp 208 nfs/nfs_bio.c bp = nfs_getcacheblk(vp, bn, biosize, p);
vp 230 nfs/nfs_bio.c bp = nfs_getcacheblk(vp, 0, NFS_MAXPATHLEN, p);
vp 246 nfs/nfs_bio.c printf(" nfsbioread: type %x unexpected\n",vp->v_type);
vp 255 nfs/nfs_bio.c switch (vp->v_type) {
vp 262 nfs/nfs_bio.c printf(" nfsbioread: type %x unexpected\n",vp->v_type);
vp 281 nfs/nfs_bio.c struct vnode *vp = ap->a_vp;
vp 282 nfs/nfs_bio.c struct nfsnode *np = VTONFS(vp);
vp 287 nfs/nfs_bio.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 297 nfs/nfs_bio.c if (vp->v_type != VREG)
vp 304 nfs/nfs_bio.c (void)nfs_fsinfo(nmp, vp, cred, p);
vp 308 nfs/nfs_bio.c error = nfs_vinvalbuf(vp, V_SAVE, cred, p, 1);
vp 314 nfs/nfs_bio.c error = VOP_GETATTR(vp, &vattr, cred, p);
vp 353 nfs/nfs_bio.c uvm_vnp_uncache(vp);
vp 361 nfs/nfs_bio.c bp = nfs_getcacheblk(vp, bn, biosize, p);
vp 367 nfs/nfs_bio.c uvm_vnp_setsize(vp, (u_long)np->n_size);
vp 414 nfs/nfs_bio.c if (NFS_ISV3(vp)) {
vp 418 nfs/nfs_bio.c nfs_del_tobecommitted_range(vp, bp);
vp 420 nfs/nfs_bio.c nfs_del_committed_range(vp, bp);
vp 443 nfs/nfs_bio.c VN_KNOTE(vp, NOTE_WRITE | (extended ? NOTE_EXTEND : 0) |
vp 457 nfs/nfs_bio.c nfs_getcacheblk(vp, bn, size, p)
vp 458 nfs/nfs_bio.c struct vnode *vp;
vp 464 nfs/nfs_bio.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 467 nfs/nfs_bio.c bp = getblk(vp, bn, size, PCATCH, 0);
vp 471 nfs/nfs_bio.c bp = getblk(vp, bn, size, 0, 2 * hz);
vp 474 nfs/nfs_bio.c bp = getblk(vp, bn, size, 0, 0);
vp 483 nfs/nfs_bio.c nfs_vinvalbuf(vp, flags, cred, p, intrflg)
vp 484 nfs/nfs_bio.c struct vnode *vp;
vp 490 nfs/nfs_bio.c struct nfsnode *np = VTONFS(vp);
vp 491 nfs/nfs_bio.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 518 nfs/nfs_bio.c error = vinvalbuf(vp, flags, cred, p, slpflag, 0);
vp 528 nfs/nfs_bio.c error = vinvalbuf(vp, flags, cred, p, 0, slptimeo);
vp 594 nfs/nfs_bio.c struct vnode *vp;
vp 601 nfs/nfs_bio.c vp = bp->b_vp;
vp 602 nfs/nfs_bio.c np = VTONFS(vp);
vp 603 nfs/nfs_bio.c nmp = VFSTONFS(vp->v_mount);
vp 624 nfs/nfs_bio.c error = nfs_readrpc(vp, uiop);
vp 629 nfs/nfs_bio.c error = nfs_writerpc(vp, uiop, &iomode, &must_commit);
vp 639 nfs/nfs_bio.c switch (vp->v_type) {
vp 643 nfs/nfs_bio.c error = nfs_readrpc(vp, uiop);
vp 665 nfs/nfs_bio.c if (p && (vp->v_flag & VTEXT) &&
vp 674 nfs/nfs_bio.c error = nfs_readlinkrpc(vp, uiop, curproc->p_ucred);
vp 677 nfs/nfs_bio.c printf("nfs_doio: type %x unexpected\n", vp->v_type);
vp 697 nfs/nfs_bio.c error = nfs_writerpc(vp, uiop, &iomode, &must_commit);
vp 702 nfs/nfs_bio.c nfs_add_tobecommitted_range(vp, bp);
vp 705 nfs/nfs_bio.c nfs_del_committed_range(vp, bp);
vp 741 nfs/nfs_bio.c nfs_clearcommit(vp->v_mount);
vp 71 nfs/nfs_kq.c struct vnode *vp;
vp 113 nfs/nfs_kq.c struct nfsnode *np = VTONFS(ke->vp);
vp 117 nfs/nfs_kq.c VOP_PRINT(ke->vp);
vp 120 nfs/nfs_kq.c if (nfs_getattrcache(ke->vp, &attr) != ENOENT)
vp 133 nfs/nfs_kq.c error = VOP_GETATTR(ke->vp, &attr, p->p_ucred, p);
vp 136 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_DELETE);
vp 144 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_WRITE
vp 149 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_WRITE);
vp 155 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_ATTRIB);
vp 160 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_LINK);
vp 190 nfs/nfs_kq.c struct vnode *vp = (struct vnode *)kn->kn_hook;
vp 193 nfs/nfs_kq.c SLIST_REMOVE(&vp->v_selectinfo.si_note, kn, knote, kn_selnext);
vp 198 nfs/nfs_kq.c if (ke->vp == vp) {
vp 223 nfs/nfs_kq.c struct vnode *vp = (struct vnode *)kn->kn_hook;
vp 224 nfs/nfs_kq.c struct nfsnode *np = VTONFS(vp);
vp 263 nfs/nfs_kq.c struct vnode *vp;
vp 270 nfs/nfs_kq.c vp = ap->a_vp;
vp 275 nfs/nfs_kq.c VOP_PRINT(vp);
vp 289 nfs/nfs_kq.c kn->kn_hook = vp;
vp 301 nfs/nfs_kq.c (void) VOP_GETATTR(vp, &attr, p->p_ucred, p);
vp 314 nfs/nfs_kq.c if (ke->vp == vp)
vp 324 nfs/nfs_kq.c ke->vp = vp;
vp 336 nfs/nfs_kq.c SLIST_INSERT_HEAD(&vp->v_selectinfo.si_note, kn, kn_selnext);
vp 99 nfs/nfs_node.c struct vnode *vp;
vp 110 nfs/nfs_node.c vp = NFSTOV(np);
vp 111 nfs/nfs_node.c if (vget(vp, LK_EXCLUSIVE, p))
vp 124 nfs/nfs_node.c vp = nvp;
vp 127 nfs/nfs_node.c vp->v_data = np;
vp 128 nfs/nfs_node.c np->n_vnode = vp;
vp 140 nfs/nfs_node.c if (vp->v_type == VNON)
vp 141 nfs/nfs_node.c vp->v_type = VDIR;
vp 142 nfs/nfs_node.c vp->v_flag |= VROOT;
vp 203 nfs/nfs_node.c struct vnode *vp = ap->a_vp;
vp 204 nfs/nfs_node.c struct nfsnode *np = VTONFS(vp);
vp 208 nfs/nfs_node.c if (prtactive && vp->v_usecount != 0)
vp 209 nfs/nfs_node.c vprint("nfs_reclaim: pushing active", vp);
vp 220 nfs/nfs_node.c if (vp->v_type == VDIR) {
vp 236 nfs/nfs_node.c cache_purge(vp);
vp 237 nfs/nfs_node.c pool_put(&nfs_node_pool, vp->v_data);
vp 238 nfs/nfs_node.c vp->v_data = NULL;
vp 106 nfs/nfs_serv.c struct vnode *vp;
vp 122 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam, &rdonly,
vp 131 nfs/nfs_serv.c nfsrv_access(vp, VREAD, cred, rdonly, procp, 0))
vp 133 nfs/nfs_serv.c if (vp->v_type == VDIR)
vp 139 nfs/nfs_serv.c nfsrv_access(vp, VWRITE, cred, rdonly, procp, 0))
vp 141 nfs/nfs_serv.c if (vp->v_type == VDIR)
vp 146 nfs/nfs_serv.c nfsrv_access(vp, VEXEC, cred, rdonly, procp, 0))
vp 148 nfs/nfs_serv.c getret = VOP_GETATTR(vp, &va, cred, procp);
vp 149 nfs/nfs_serv.c vput(vp);
vp 173 nfs/nfs_serv.c struct vnode *vp;
vp 186 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam, &rdonly,
vp 192 nfs/nfs_serv.c error = VOP_GETATTR(vp, &va, cred, procp);
vp 193 nfs/nfs_serv.c vput(vp);
vp 219 nfs/nfs_serv.c struct vnode *vp;
vp 278 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam, &rdonly,
vp 286 nfs/nfs_serv.c error = preat_ret = VOP_GETATTR(vp, &preat, cred, procp);
vp 292 nfs/nfs_serv.c vput(vp);
vp 304 nfs/nfs_serv.c if (rdonly || (vp->v_mount->mnt_flag & MNT_RDONLY)) {
vp 309 nfs/nfs_serv.c if (vp->v_type == VDIR) {
vp 312 nfs/nfs_serv.c } else if ((error = nfsrv_access(vp, VWRITE, cred, rdonly,
vp 316 nfs/nfs_serv.c error = VOP_SETATTR(vp, &va, cred, procp);
vp 317 nfs/nfs_serv.c postat_ret = VOP_GETATTR(vp, &va, cred, procp);
vp 321 nfs/nfs_serv.c vput(vp);
vp 349 nfs/nfs_serv.c struct vnode *vp, *dirp;
vp 384 nfs/nfs_serv.c vp = nd.ni_vp;
vp 386 nfs/nfs_serv.c fhp->fh_fsid = vp->v_mount->mnt_stat.f_fsid;
vp 387 nfs/nfs_serv.c error = VFS_VPTOFH(vp, &fhp->fh_fid);
vp 389 nfs/nfs_serv.c error = VOP_GETATTR(vp, &va, cred, procp);
vp 390 nfs/nfs_serv.c vput(vp);
vp 431 nfs/nfs_serv.c struct vnode *vp;
vp 469 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam,
vp 477 nfs/nfs_serv.c if (vp->v_type != VLNK) {
vp 484 nfs/nfs_serv.c error = VOP_READLINK(vp, uiop, cred);
vp 486 nfs/nfs_serv.c getret = VOP_GETATTR(vp, &attr, cred, procp);
vp 487 nfs/nfs_serv.c vput(vp);
vp 534 nfs/nfs_serv.c struct vnode *vp;
vp 552 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam,
vp 559 nfs/nfs_serv.c if (vp->v_type != VREG) {
vp 563 nfs/nfs_serv.c error = (vp->v_type == VDIR) ? EISDIR : EACCES;
vp 566 nfs/nfs_serv.c if ((error = nfsrv_access(vp, VREAD, cred, rdonly, procp, 1)) != 0)
vp 567 nfs/nfs_serv.c error = nfsrv_access(vp, VEXEC, cred, rdonly, procp, 1);
vp 569 nfs/nfs_serv.c getret = VOP_GETATTR(vp, &va, cred, procp);
vp 573 nfs/nfs_serv.c vput(vp);
vp 641 nfs/nfs_serv.c error = VOP_READ(vp, uiop, IO_NODELOCKED, cred);
vp 644 nfs/nfs_serv.c if (error || (getret = VOP_GETATTR(vp, &va, cred, procp)) != 0){
vp 648 nfs/nfs_serv.c vput(vp);
vp 655 nfs/nfs_serv.c vput(vp);
vp 702 nfs/nfs_serv.c struct vnode *vp;
vp 764 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam,
vp 772 nfs/nfs_serv.c forat_ret = VOP_GETATTR(vp, &forat, cred, procp);
vp 773 nfs/nfs_serv.c if (vp->v_type != VREG) {
vp 777 nfs/nfs_serv.c error = (vp->v_type == VDIR) ? EISDIR : EACCES;
vp 780 nfs/nfs_serv.c error = nfsrv_access(vp, VWRITE, cred, rdonly, procp, 1);
vp 783 nfs/nfs_serv.c vput(vp);
vp 822 nfs/nfs_serv.c error = VOP_WRITE(vp, uiop, ioflags, cred);
vp 826 nfs/nfs_serv.c aftat_ret = VOP_GETATTR(vp, &va, cred, procp);
vp 827 nfs/nfs_serv.c vput(vp);
vp 886 nfs/nfs_serv.c struct vnode *vp;
vp 1033 nfs/nfs_serv.c error = nfsrv_fhtovp(&nfsd->nd_fh, 1, &vp, cred, slp,
vp 1037 nfs/nfs_serv.c forat_ret = VOP_GETATTR(vp, &forat, cred, procp);
vp 1038 nfs/nfs_serv.c if (vp->v_type != VREG) {
vp 1042 nfs/nfs_serv.c error = (vp->v_type == VDIR) ? EISDIR : EACCES;
vp 1045 nfs/nfs_serv.c vp = NULL;
vp 1047 nfs/nfs_serv.c error = nfsrv_access(vp, VWRITE, cred, rdonly, procp, 1);
vp 1083 nfs/nfs_serv.c error = VOP_WRITE(vp, uiop, ioflags, cred);
vp 1089 nfs/nfs_serv.c if (vp) {
vp 1090 nfs/nfs_serv.c aftat_ret = VOP_GETATTR(vp, &va, cred, procp);
vp 1091 nfs/nfs_serv.c vput(vp);
vp 1251 nfs/nfs_serv.c struct vnode *vp = NULL, *dirp = NULL;
vp 1394 nfs/nfs_serv.c vp = nd.ni_vp;
vp 1398 nfs/nfs_serv.c vp = nd.ni_vp;
vp 1399 nfs/nfs_serv.c if (nd.ni_dvp == vp)
vp 1405 nfs/nfs_serv.c error = nfsrv_access(vp, VWRITE, cred,
vp 1411 nfs/nfs_serv.c error = VOP_SETATTR(vp, &va, cred,
vp 1415 nfs/nfs_serv.c vput(vp);
vp 1420 nfs/nfs_serv.c fhp->fh_fsid = vp->v_mount->mnt_stat.f_fsid;
vp 1421 nfs/nfs_serv.c error = VFS_VPTOFH(vp, &fhp->fh_fid);
vp 1423 nfs/nfs_serv.c error = VOP_GETATTR(vp, &va, cred, procp);
vp 1424 nfs/nfs_serv.c vput(vp);
vp 1487 nfs/nfs_serv.c struct vnode *vp, *dirp = (struct vnode *)0;
vp 1576 nfs/nfs_serv.c vp = nd.ni_vp;
vp 1579 nfs/nfs_serv.c fhp->fh_fsid = vp->v_mount->mnt_stat.f_fsid;
vp 1580 nfs/nfs_serv.c error = VFS_VPTOFH(vp, &fhp->fh_fid);
vp 1582 nfs/nfs_serv.c error = VOP_GETATTR(vp, &va, cred, procp);
vp 1583 nfs/nfs_serv.c vput(vp);
vp 1633 nfs/nfs_serv.c struct vnode *vp, *dirp;
vp 1640 nfs/nfs_serv.c vp = (struct vnode *)0;
vp 1658 nfs/nfs_serv.c vp = nd.ni_vp;
vp 1659 nfs/nfs_serv.c if (vp->v_type == VDIR &&
vp 1665 nfs/nfs_serv.c if (vp->v_flag & VROOT) {
vp 1669 nfs/nfs_serv.c if (vp->v_flag & VTEXT)
vp 1670 nfs/nfs_serv.c uvm_vnp_uncache(vp);
vp 1676 nfs/nfs_serv.c if (nd.ni_dvp == vp)
vp 1680 nfs/nfs_serv.c vput(vp);
vp 1897 nfs/nfs_serv.c struct vnode *vp, *xp, *dirp = (struct vnode *)0;
vp 1908 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, FALSE, &vp, cred, slp, nam,
vp 1916 nfs/nfs_serv.c if (vp->v_type == VDIR && (error = suser_ucred(cred)) != 0)
vp 1940 nfs/nfs_serv.c if (vp->v_mount != xp->v_mount)
vp 1944 nfs/nfs_serv.c error = VOP_LINK(nd.ni_dvp, vp, &nd.ni_cnd);
vp 1956 nfs/nfs_serv.c getret = VOP_GETATTR(vp, &at, cred, procp);
vp 1961 nfs/nfs_serv.c vrele(vp);
vp 2137 nfs/nfs_serv.c struct vnode *vp, *dirp = (struct vnode *)0;
vp 2174 nfs/nfs_serv.c vp = nd.ni_vp;
vp 2175 nfs/nfs_serv.c if (vp != NULL) {
vp 2177 nfs/nfs_serv.c if (nd.ni_dvp == vp)
vp 2181 nfs/nfs_serv.c vrele(vp);
vp 2187 nfs/nfs_serv.c vp = nd.ni_vp;
vp 2189 nfs/nfs_serv.c fhp->fh_fsid = vp->v_mount->mnt_stat.f_fsid;
vp 2190 nfs/nfs_serv.c error = VFS_VPTOFH(vp, &fhp->fh_fid);
vp 2192 nfs/nfs_serv.c error = VOP_GETATTR(vp, &va, cred, procp);
vp 2193 nfs/nfs_serv.c vput(vp);
vp 2247 nfs/nfs_serv.c struct vnode *vp, *dirp = (struct vnode *)0;
vp 2278 nfs/nfs_serv.c vp = nd.ni_vp;
vp 2279 nfs/nfs_serv.c if (vp->v_type != VDIR) {
vp 2286 nfs/nfs_serv.c if (nd.ni_dvp == vp) {
vp 2293 nfs/nfs_serv.c if (vp->v_flag & VROOT)
vp 2304 nfs/nfs_serv.c vput(vp);
vp 2376 nfs/nfs_serv.c struct vnode *vp;
vp 2409 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam,
vp 2417 nfs/nfs_serv.c error = getret = VOP_GETATTR(vp, &at, cred, procp);
vp 2427 nfs/nfs_serv.c error = nfsrv_access(vp, VEXEC, cred, rdonly, procp, 0);
vp 2429 nfs/nfs_serv.c vput(vp);
vp 2434 nfs/nfs_serv.c VOP_UNLOCK(vp, 0, procp);
vp 2453 nfs/nfs_serv.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, procp);
vp 2454 nfs/nfs_serv.c error = VOP_READDIR(vp, &io, cred, &eofflag, &ncookies, &cookies);
vp 2460 nfs/nfs_serv.c getret = VOP_GETATTR(vp, &at, cred, procp);
vp 2465 nfs/nfs_serv.c VOP_UNLOCK(vp, 0, procp);
vp 2467 nfs/nfs_serv.c vrele(vp);
vp 2483 nfs/nfs_serv.c vrele(vp);
vp 2598 nfs/nfs_serv.c vrele(vp);
vp 2638 nfs/nfs_serv.c struct vnode *vp, *nvp;
vp 2668 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam,
vp 2675 nfs/nfs_serv.c error = getret = VOP_GETATTR(vp, &at, cred, procp);
vp 2684 nfs/nfs_serv.c error = nfsrv_access(vp, VEXEC, cred, rdonly, procp, 0);
vp 2687 nfs/nfs_serv.c vput(vp);
vp 2692 nfs/nfs_serv.c VOP_UNLOCK(vp, 0, procp);
vp 2712 nfs/nfs_serv.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, procp);
vp 2713 nfs/nfs_serv.c error = VOP_READDIR(vp, &io, cred, &eofflag, &ncookies, &cookies);
vp 2716 nfs/nfs_serv.c getret = VOP_GETATTR(vp, &at, cred, procp);
vp 2718 nfs/nfs_serv.c VOP_UNLOCK(vp, 0, procp);
vp 2725 nfs/nfs_serv.c vrele(vp);
vp 2741 nfs/nfs_serv.c vrele(vp);
vp 2808 nfs/nfs_serv.c if (VFS_VGET(vp->v_mount, dp->d_fileno, &nvp))
vp 2916 nfs/nfs_serv.c vrele(vp);
vp 2951 nfs/nfs_serv.c struct vnode *vp;
vp 2973 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam,
vp 2980 nfs/nfs_serv.c for_ret = VOP_GETATTR(vp, &bfor, cred, procp);
vp 2981 nfs/nfs_serv.c error = VOP_FSYNC(vp, cred, MNT_WAIT, procp);
vp 2982 nfs/nfs_serv.c aft_ret = VOP_GETATTR(vp, &aft, cred, procp);
vp 2983 nfs/nfs_serv.c vput(vp);
vp 3018 nfs/nfs_serv.c struct vnode *vp;
vp 3027 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam,
vp 3035 nfs/nfs_serv.c error = VFS_STATFS(vp->v_mount, sf, procp);
vp 3036 nfs/nfs_serv.c getret = VOP_GETATTR(vp, &at, cred, procp);
vp 3037 nfs/nfs_serv.c vput(vp);
vp 3092 nfs/nfs_serv.c struct vnode *vp;
vp 3100 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam,
vp 3107 nfs/nfs_serv.c getret = VOP_GETATTR(vp, &at, cred, procp);
vp 3108 nfs/nfs_serv.c vput(vp);
vp 3161 nfs/nfs_serv.c struct vnode *vp;
vp 3169 nfs/nfs_serv.c error = nfsrv_fhtovp(fhp, 1, &vp, cred, slp, nam,
vp 3176 nfs/nfs_serv.c error = VOP_PATHCONF(vp, _PC_LINK_MAX, &linkmax);
vp 3178 nfs/nfs_serv.c error = VOP_PATHCONF(vp, _PC_NAME_MAX, &namemax);
vp 3180 nfs/nfs_serv.c error = VOP_PATHCONF(vp, _PC_CHOWN_RESTRICTED, &chownres);
vp 3182 nfs/nfs_serv.c error = VOP_PATHCONF(vp, _PC_NO_TRUNC, ¬runc);
vp 3183 nfs/nfs_serv.c getret = VOP_GETATTR(vp, &at, cred, procp);
vp 3184 nfs/nfs_serv.c vput(vp);
vp 3264 nfs/nfs_serv.c nfsrv_access(vp, flags, cred, rdonly, p, override)
vp 3265 nfs/nfs_serv.c struct vnode *vp;
vp 3282 nfs/nfs_serv.c if (rdonly || (vp->v_mount->mnt_flag & MNT_RDONLY)) {
vp 3283 nfs/nfs_serv.c switch (vp->v_type) {
vp 3297 nfs/nfs_serv.c if ((vp->v_flag & VTEXT) && !uvm_vnp_uncache(vp))
vp 3300 nfs/nfs_serv.c error = VOP_ACCESS(vp, flags, cred, p);
vp 3306 nfs/nfs_serv.c VOP_GETATTR(vp, &vattr, cred, p) == 0 &&
vp 823 nfs/nfs_socket.c nfs_request(vp, mrest, procnum, procp, cred, mrp, mdp, dposp)
vp 824 nfs/nfs_socket.c struct vnode *vp;
vp 851 nfs/nfs_socket.c nmp = VFSTONFS(vp->v_mount);
vp 854 nfs/nfs_socket.c rep->r_vp = vp;
vp 1057 nfs/nfs_socket.c cache_purge(vp);
vp 539 nfs/nfs_subs.c nfsm_reqh(vp, procid, hsiz, bposp)
vp 540 nfs/nfs_subs.c struct vnode *vp;
vp 1112 nfs/nfs_subs.c struct vnode *vp = *vpp;
vp 1126 nfs/nfs_subs.c int v3 = NFS_ISV3(vp);
vp 1162 nfs/nfs_subs.c np = VTONFS(vp);
vp 1163 nfs/nfs_subs.c if (vp->v_type != vtyp) {
vp 1164 nfs/nfs_subs.c vp->v_type = vtyp;
vp 1165 nfs/nfs_subs.c if (vp->v_type == VFIFO) {
vp 1170 nfs/nfs_subs.c vp->v_op = fifo_nfsv2nodeop_p;
vp 1173 nfs/nfs_subs.c if (vp->v_type == VCHR || vp->v_type == VBLK) {
vp 1174 nfs/nfs_subs.c vp->v_op = spec_nfsv2nodeop_p;
vp 1175 nfs/nfs_subs.c nvp = checkalias(vp, (dev_t)rdev, vp->v_mount);
vp 1183 nfs/nfs_subs.c nvp->v_data = vp->v_data;
vp 1184 nfs/nfs_subs.c vp->v_data = NULL;
vp 1185 nfs/nfs_subs.c vp->v_op = spec_vnodeop_p;
vp 1186 nfs/nfs_subs.c vrele(vp);
vp 1187 nfs/nfs_subs.c vgone(vp);
vp 1192 nfs/nfs_subs.c *vpp = vp = nvp;
vp 1202 nfs/nfs_subs.c vap->va_fsid = vp->v_mount->mnt_stat.f_fsid.val[0];
vp 1243 nfs/nfs_subs.c uvm_vnp_setsize(vp, np->n_size);
vp 1264 nfs/nfs_subs.c struct vnode *vp = np->n_vnode;
vp 1265 nfs/nfs_subs.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 1269 nfs/nfs_subs.c if (vp->v_type == VDIR) {
vp 1292 nfs/nfs_subs.c nfs_getattrcache(vp, vaper)
vp 1293 nfs/nfs_subs.c struct vnode *vp;
vp 1296 nfs/nfs_subs.c struct nfsnode *np = VTONFS(vp);
vp 1315 nfs/nfs_subs.c uvm_vnp_setsize(vp, np->n_size);
vp 1721 nfs/nfs_subs.c struct vnode *vp, *nvp;
vp 1727 nfs/nfs_subs.c for (vp = LIST_FIRST(&mp->mnt_vnodelist); vp != NULL; vp = nvp) {
vp 1728 nfs/nfs_subs.c if (vp->v_mount != mp) /* Paranoia */
vp 1730 nfs/nfs_subs.c nvp = LIST_NEXT(vp, v_mntvnodes);
vp 1731 nfs/nfs_subs.c for (bp = LIST_FIRST(&vp->v_dirtyblkhd); bp != NULL; bp = nbp) {
vp 1742 nfs/nfs_subs.c nfs_merge_commit_ranges(vp)
vp 1743 nfs/nfs_subs.c struct vnode *vp;
vp 1745 nfs/nfs_subs.c struct nfsnode *np = VTONFS(vp);
vp 1763 nfs/nfs_subs.c nfs_in_committed_range(vp, bp)
vp 1764 nfs/nfs_subs.c struct vnode *vp;
vp 1767 nfs/nfs_subs.c struct nfsnode *np = VTONFS(vp);
vp 1779 nfs/nfs_subs.c nfs_in_tobecommitted_range(vp, bp)
vp 1780 nfs/nfs_subs.c struct vnode *vp;
vp 1783 nfs/nfs_subs.c struct nfsnode *np = VTONFS(vp);
vp 1795 nfs/nfs_subs.c nfs_add_committed_range(vp, bp)
vp 1796 nfs/nfs_subs.c struct vnode *vp;
vp 1799 nfs/nfs_subs.c struct nfsnode *np = VTONFS(vp);
vp 1818 nfs/nfs_subs.c nfs_del_committed_range(vp, bp)
vp 1819 nfs/nfs_subs.c struct vnode *vp;
vp 1822 nfs/nfs_subs.c struct nfsnode *np = VTONFS(vp);
vp 1851 nfs/nfs_subs.c nfs_add_tobecommitted_range(vp, bp)
vp 1852 nfs/nfs_subs.c struct vnode *vp;
vp 1855 nfs/nfs_subs.c struct nfsnode *np = VTONFS(vp);
vp 1874 nfs/nfs_subs.c nfs_del_tobecommitted_range(vp, bp)
vp 1875 nfs/nfs_subs.c struct vnode *vp;
vp 1878 nfs/nfs_subs.c struct nfsnode *np = VTONFS(vp);
vp 699 nfs/nfs_syscalls.c struct vnode *vp;
vp 736 nfs/nfs_syscalls.c vp = bp->b_vp;
vp 738 nfs/nfs_syscalls.c for (nbp = LIST_FIRST(&vp->v_dirtyblkhd); nbp != NULL;
vp 110 nfs/nfs_vfsops.c struct vnode *vp;
vp 126 nfs/nfs_vfsops.c vp = NFSTOV(np);
vp 130 nfs/nfs_vfsops.c (void)nfs_fsinfo(nmp, vp, cred, p);
vp 132 nfs/nfs_vfsops.c nfsm_reqhead(vp, NFSPROC_FSSTAT, NFSX_FH(v3));
vp 133 nfs/nfs_vfsops.c nfsm_fhtom(vp, v3);
vp 134 nfs/nfs_vfsops.c nfsm_request(vp, NFSPROC_FSSTAT, p, cred);
vp 136 nfs/nfs_vfsops.c nfsm_postop_attr(vp, retattr);
vp 173 nfs/nfs_vfsops.c vrele(vp);
vp 182 nfs/nfs_vfsops.c nfs_fsinfo(nmp, vp, cred, p)
vp 184 nfs/nfs_vfsops.c struct vnode *vp;
vp 197 nfs/nfs_vfsops.c nfsm_reqhead(vp, NFSPROC_FSINFO, NFSX_FH(1));
vp 198 nfs/nfs_vfsops.c nfsm_fhtom(vp, 1);
vp 199 nfs/nfs_vfsops.c nfsm_request(vp, NFSPROC_FSINFO, p, cred);
vp 200 nfs/nfs_vfsops.c nfsm_postop_attr(vp, retattr);
vp 252 nfs/nfs_vfsops.c struct vnode *vp;
vp 324 nfs/nfs_vfsops.c nfs_root(mp, &vp);
vp 331 nfs/nfs_vfsops.c vp->v_type = VREG;
vp 332 nfs/nfs_vfsops.c vp->v_flag = 0;
vp 340 nfs/nfs_vfsops.c swdevt[0].sw_vp = vp;
vp 345 nfs/nfs_vfsops.c error = VOP_GETATTR(vp, &attr, procp->p_ucred, procp);
vp 786 nfs/nfs_vfsops.c struct vnode *vp;
vp 799 nfs/nfs_vfsops.c for (vp = LIST_FIRST(&mp->mnt_vnodelist); vp != NULL;
vp 800 nfs/nfs_vfsops.c vp = LIST_NEXT(vp, v_mntvnodes)) {
vp 805 nfs/nfs_vfsops.c if (vp->v_mount != mp)
vp 807 nfs/nfs_vfsops.c if (VOP_ISLOCKED(vp) || LIST_FIRST(&vp->v_dirtyblkhd) == NULL)
vp 809 nfs/nfs_vfsops.c if (vget(vp, LK_EXCLUSIVE, p))
vp 811 nfs/nfs_vfsops.c error = VOP_FSYNC(vp, cred, waitfor, p);
vp 814 nfs/nfs_vfsops.c vput(vp);
vp 907 nfs/nfs_vfsops.c nfs_vptofh(vp, fhp)
vp 908 nfs/nfs_vfsops.c struct vnode *vp;
vp 193 nfs/nfs_vnops.c nfs_null(vp, cred, procp)
vp 194 nfs/nfs_vnops.c struct vnode *vp;
vp 202 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_NULL, 0);
vp 203 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_NULL, procp, cred);
vp 219 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 227 nfs/nfs_vnops.c int v3 = NFS_ISV3(vp);
vp 234 nfs/nfs_vnops.c if ((ap->a_mode & VWRITE) && (vp->v_mount->mnt_flag & MNT_RDONLY)) {
vp 235 nfs/nfs_vnops.c switch (vp->v_type) {
vp 254 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_ACCESS, NFSX_FH(v3) + NFSX_UNSIGNED);
vp 255 nfs/nfs_vnops.c nfsm_fhtom(vp, v3);
vp 261 nfs/nfs_vnops.c if (vp->v_type == VDIR) {
vp 274 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_ACCESS, ap->a_p, ap->a_cred);
vp 275 nfs/nfs_vnops.c nfsm_postop_attr(vp, attrflag);
vp 306 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 307 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 311 nfs/nfs_vnops.c if (vp->v_type != VREG && vp->v_type != VDIR && vp->v_type != VLNK) {
vp 313 nfs/nfs_vnops.c printf("open eacces vtyp=%d\n",vp->v_type);
vp 339 nfs/nfs_vnops.c if ((error = nfs_vinvalbuf(vp, V_SAVE, ap->a_cred,
vp 342 nfs/nfs_vnops.c uvm_vnp_uncache(vp);
vp 344 nfs/nfs_vnops.c if (vp->v_type == VDIR)
vp 346 nfs/nfs_vnops.c error = VOP_GETATTR(vp, &vattr, ap->a_cred, ap->a_p);
vp 351 nfs/nfs_vnops.c error = VOP_GETATTR(vp, &vattr, ap->a_cred, ap->a_p);
vp 355 nfs/nfs_vnops.c if (vp->v_type == VDIR)
vp 357 nfs/nfs_vnops.c if ((error = nfs_vinvalbuf(vp, V_SAVE,
vp 360 nfs/nfs_vnops.c uvm_vnp_uncache(vp);
vp 401 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 402 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 405 nfs/nfs_vnops.c if (vp->v_type == VREG) {
vp 407 nfs/nfs_vnops.c if (NFS_ISV3(vp)) {
vp 408 nfs/nfs_vnops.c error = nfs_flush(vp, ap->a_cred, MNT_WAIT, ap->a_p, 0);
vp 411 nfs/nfs_vnops.c error = nfs_vinvalbuf(vp, V_SAVE, ap->a_cred, ap->a_p, 1);
vp 430 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 431 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 438 nfs/nfs_vnops.c int v3 = NFS_ISV3(vp);
vp 448 nfs/nfs_vnops.c if (nfs_getattrcache(vp, ap->a_vap) == 0)
vp 451 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_GETATTR, NFSX_FH(v3));
vp 452 nfs/nfs_vnops.c nfsm_fhtom(vp, v3);
vp 453 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_GETATTR, ap->a_p, ap->a_cred);
vp 455 nfs/nfs_vnops.c nfsm_loadattr(vp, ap->a_vap);
vp 468 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 469 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 486 nfs/nfs_vnops.c (vp->v_mount->mnt_flag & MNT_RDONLY))
vp 489 nfs/nfs_vnops.c switch (vp->v_type) {
vp 509 nfs/nfs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 512 nfs/nfs_vnops.c error = nfs_vinvalbuf(vp, 0,
vp 515 nfs/nfs_vnops.c error = nfs_vinvalbuf(vp, V_SAVE,
vp 521 nfs/nfs_vnops.c uvm_vnp_setsize(vp, np->n_size);
vp 525 nfs/nfs_vnops.c vp->v_type == VREG &&
vp 526 nfs/nfs_vnops.c (error = nfs_vinvalbuf(vp, V_SAVE, ap->a_cred,
vp 529 nfs/nfs_vnops.c error = nfs_setattrrpc(vp, vap, ap->a_cred, ap->a_p);
vp 532 nfs/nfs_vnops.c uvm_vnp_setsize(vp, np->n_size);
vp 535 nfs/nfs_vnops.c VN_KNOTE(vp, NOTE_ATTRIB); /* XXX setattrrpc? */
vp 544 nfs/nfs_vnops.c nfs_setattrrpc(vp, vap, cred, procp)
vp 545 nfs/nfs_vnops.c struct vnode *vp;
vp 557 nfs/nfs_vnops.c int v3 = NFS_ISV3(vp);
vp 560 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_SETATTR, NFSX_FH(v3) + NFSX_SATTR(v3));
vp 561 nfs/nfs_vnops.c nfsm_fhtom(vp, v3);
vp 571 nfs/nfs_vnops.c sp->sa_mode = vtonfsv2_mode(vp->v_type, vap->va_mode);
vp 584 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_SETATTR, procp, cred);
vp 586 nfs/nfs_vnops.c nfsm_wcc_data(vp, wccflag);
vp 588 nfs/nfs_vnops.c nfsm_loadattr(vp, (struct vattr *)0);
vp 860 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 862 nfs/nfs_vnops.c if (vp->v_type != VREG)
vp 864 nfs/nfs_vnops.c return (nfs_bioread(vp, ap->a_uio, ap->a_ioflag, ap->a_cred));
vp 875 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 877 nfs/nfs_vnops.c if (vp->v_type != VLNK)
vp 879 nfs/nfs_vnops.c return (nfs_bioread(vp, ap->a_uio, 0, ap->a_cred));
vp 887 nfs/nfs_vnops.c nfs_readlinkrpc(vp, uiop, cred)
vp 888 nfs/nfs_vnops.c struct vnode *vp;
vp 898 nfs/nfs_vnops.c int v3 = NFS_ISV3(vp);
vp 901 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_READLINK, NFSX_FH(v3));
vp 902 nfs/nfs_vnops.c nfsm_fhtom(vp, v3);
vp 903 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_READLINK, uiop->uio_procp, cred);
vp 905 nfs/nfs_vnops.c nfsm_postop_attr(vp, attrflag);
vp 919 nfs/nfs_vnops.c nfs_readrpc(vp, uiop)
vp 920 nfs/nfs_vnops.c struct vnode *vp;
vp 930 nfs/nfs_vnops.c int v3 = NFS_ISV3(vp);
vp 935 nfs/nfs_vnops.c nmp = VFSTONFS(vp->v_mount);
vp 942 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_READ, NFSX_FH(v3) + NFSX_UNSIGNED * 3);
vp 943 nfs/nfs_vnops.c nfsm_fhtom(vp, v3);
vp 953 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_READ, uiop->uio_procp,
vp 954 nfs/nfs_vnops.c VTONFS(vp)->n_rcred);
vp 956 nfs/nfs_vnops.c nfsm_postop_attr(vp, attrflag);
vp 964 nfs/nfs_vnops.c nfsm_loadattr(vp, (struct vattr *)0);
vp 983 nfs/nfs_vnops.c nfs_writerpc(vp, uiop, iomode, must_commit)
vp 984 nfs/nfs_vnops.c struct vnode *vp;
vp 993 nfs/nfs_vnops.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 995 nfs/nfs_vnops.c int v3 = NFS_ISV3(vp), committed = NFSV3WRITE_FILESYNC;
vp 1008 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_WRITE,
vp 1010 nfs/nfs_vnops.c nfsm_fhtom(vp, v3);
vp 1032 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_WRITE, uiop->uio_procp,
vp 1033 nfs/nfs_vnops.c VTONFS(vp)->n_wcred);
vp 1036 nfs/nfs_vnops.c nfsm_wcc_data(vp, wccflag);
vp 1075 nfs/nfs_vnops.c nfsm_loadattr(vp, (struct vattr *)0);
vp 1077 nfs/nfs_vnops.c VTONFS(vp)->n_mtime = VTONFS(vp)->n_vattr.va_mtime.tv_sec;
vp 1318 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 1321 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 1328 nfs/nfs_vnops.c if (vp->v_usecount < 1)
vp 1331 nfs/nfs_vnops.c if (vp->v_type == VDIR)
vp 1333 nfs/nfs_vnops.c else if (vp->v_usecount == 1 || (np->n_sillyrename &&
vp 1334 nfs/nfs_vnops.c VOP_GETATTR(vp, &vattr, cnp->cn_cred, cnp->cn_proc) == 0 &&
vp 1343 nfs/nfs_vnops.c cache_purge(vp);
vp 1348 nfs/nfs_vnops.c error = nfs_vinvalbuf(vp, 0, cnp->cn_cred, cnp->cn_proc, 1);
vp 1362 nfs/nfs_vnops.c error = nfs_sillyrename(dvp, vp, cnp);
vp 1366 nfs/nfs_vnops.c vrele(vp);
vp 1368 nfs/nfs_vnops.c VN_KNOTE(vp, NOTE_DELETE);
vp 1554 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 1565 nfs/nfs_vnops.c if (dvp->v_mount != vp->v_mount) {
vp 1567 nfs/nfs_vnops.c if (vp == dvp)
vp 1579 nfs/nfs_vnops.c VOP_FSYNC(vp, cnp->cn_cred, MNT_WAIT, cnp->cn_proc);
vp 1581 nfs/nfs_vnops.c v3 = NFS_ISV3(vp);
vp 1583 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_LINK,
vp 1585 nfs/nfs_vnops.c nfsm_fhtom(vp, v3);
vp 1588 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_LINK, cnp->cn_proc, cnp->cn_cred);
vp 1590 nfs/nfs_vnops.c nfsm_postop_attr(vp, attrflag);
vp 1597 nfs/nfs_vnops.c VTONFS(vp)->n_attrstamp = 0;
vp 1601 nfs/nfs_vnops.c VN_KNOTE(vp, NOTE_LINK);
vp 1761 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 1772 nfs/nfs_vnops.c if (dvp == vp) {
vp 1793 nfs/nfs_vnops.c VN_KNOTE(vp, NOTE_DELETE);
vp 1796 nfs/nfs_vnops.c cache_purge(vp);
vp 1797 nfs/nfs_vnops.c vrele(vp);
vp 1836 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 1837 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 1844 nfs/nfs_vnops.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 1852 nfs/nfs_vnops.c if (vp->v_type != VDIR)
vp 1859 nfs/nfs_vnops.c if (VOP_GETATTR(vp, &vattr, ap->a_cred, uio->uio_procp) == 0 &&
vp 1885 nfs/nfs_vnops.c (void)nfs_fsinfo(nmp, vp, cred, p);
vp 1906 nfs/nfs_vnops.c error = nfs_readdirplusrpc(vp, &readdir_uio, cred,
vp 1912 nfs/nfs_vnops.c error = nfs_readdirrpc(vp, &readdir_uio, cred, &eof);
vp 1982 nfs/nfs_vnops.c nfs_readdirrpc(struct vnode *vp,
vp 1996 nfs/nfs_vnops.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 1997 nfs/nfs_vnops.c struct nfsnode *dnp = VTONFS(vp);
vp 2001 nfs/nfs_vnops.c int v3 = NFS_ISV3(vp);
vp 2018 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_READDIR, NFSX_FH(v3) +
vp 2020 nfs/nfs_vnops.c nfsm_fhtom(vp, v3);
vp 2038 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_READDIR, uiop->uio_procp, cred);
vp 2040 nfs/nfs_vnops.c nfsm_postop_attr(vp, attrflag);
vp 2168 nfs/nfs_vnops.c nfs_readdirplusrpc(struct vnode *vp, struct uio *uiop, struct ucred *cred,
vp 2183 nfs/nfs_vnops.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 2184 nfs/nfs_vnops.c struct nfsnode *dnp = VTONFS(vp), *np;
vp 2195 nfs/nfs_vnops.c ndp->ni_dvp = vp;
vp 2207 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_READDIRPLUS,
vp 2209 nfs/nfs_vnops.c nfsm_fhtom(vp, 1);
vp 2223 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_READDIRPLUS, uiop->uio_procp, cred);
vp 2224 nfs/nfs_vnops.c nfsm_postop_attr(vp, attrflag);
vp 2302 nfs/nfs_vnops.c VREF(vp);
vp 2303 nfs/nfs_vnops.c newvp = vp;
vp 2306 nfs/nfs_vnops.c error = nfs_nget(vp->v_mount, fhp,
vp 2394 nfs/nfs_vnops.c nfs_sillyrename(dvp, vp, cnp)
vp 2395 nfs/nfs_vnops.c struct vnode *dvp, *vp;
vp 2403 nfs/nfs_vnops.c np = VTONFS(vp);
vp 2410 nfs/nfs_vnops.c if (vp->v_type == VDIR) {
vp 2529 nfs/nfs_vnops.c nfs_commit(vp, offset, cnt, procp)
vp 2530 nfs/nfs_vnops.c struct vnode *vp;
vp 2538 nfs/nfs_vnops.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 2546 nfs/nfs_vnops.c nfsm_reqhead(vp, NFSPROC_COMMIT, NFSX_FH(1));
vp 2547 nfs/nfs_vnops.c nfsm_fhtom(vp, 1);
vp 2552 nfs/nfs_vnops.c nfsm_request(vp, NFSPROC_COMMIT, procp, VTONFS(vp)->n_wcred);
vp 2553 nfs/nfs_vnops.c nfsm_wcc_data(vp, wccflag);
vp 2581 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 2584 nfs/nfs_vnops.c *ap->a_vpp = vp;
vp 2586 nfs/nfs_vnops.c *ap->a_bnp = ap->a_bn * btodb(vp->v_mount->mnt_stat.f_iosize);
vp 2640 nfs/nfs_vnops.c nfs_flush(vp, cred, waitfor, p, commit)
vp 2641 nfs/nfs_vnops.c struct vnode *vp;
vp 2647 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 2651 nfs/nfs_vnops.c struct nfsmount *nmp = VFSTONFS(vp->v_mount);
vp 2673 nfs/nfs_vnops.c if (NFS_ISV3(vp) && commit) {
vp 2675 nfs/nfs_vnops.c for (bp = LIST_FIRST(&vp->v_dirtyblkhd); bp != NULL; bp = nbp) {
vp 2706 nfs/nfs_vnops.c retv = nfs_commit(vp, off, (int)(endoff - off), p);
vp 2708 nfs/nfs_vnops.c nfs_clearcommit(vp->v_mount);
vp 2722 nfs/nfs_vnops.c vp->v_numoutput++;
vp 2737 nfs/nfs_vnops.c for (bp = LIST_FIRST(&vp->v_dirtyblkhd); bp != NULL; bp = nbp) {
vp 2777 nfs/nfs_vnops.c error = vwaitforio(vp, slpflag, "nfs_fsync", slptimeo);
vp 2789 nfs/nfs_vnops.c if (LIST_FIRST(&vp->v_dirtyblkhd) && commit) {
vp 2791 nfs/nfs_vnops.c vprint("nfs_fsync: dirty", vp);
vp 2843 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 2844 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 2849 nfs/nfs_vnops.c if (vp->v_type == VFIFO)
vp 2850 nfs/nfs_vnops.c fifo_printinfo(vp);
vp 2882 nfs/nfs_vnops.c struct vnode *vp;
vp 2888 nfs/nfs_vnops.c vp = bp->b_vp;
vp 2889 nfs/nfs_vnops.c np = VTONFS(vp);
vp 2921 nfs/nfs_vnops.c if (!nfs_in_committed_range(vp, bp)) {
vp 2928 nfs/nfs_vnops.c if (nfs_in_tobecommitted_range(vp, bp)) {
vp 2940 nfs/nfs_vnops.c nfs_merge_commit_ranges(vp);
vp 2942 nfs/nfs_vnops.c nfs_add_committed_range(vp, bp);
vp 2986 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 2994 nfs/nfs_vnops.c if ((ap->a_mode & VWRITE) && (vp->v_mount->mnt_flag & MNT_RDONLY)) {
vp 2995 nfs/nfs_vnops.c switch (vp->v_type) {
vp 3005 nfs/nfs_vnops.c error = VOP_GETATTR(vp, &va, ap->a_cred, ap->a_p);
vp 3072 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 3073 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 3078 nfs/nfs_vnops.c if (vp->v_usecount == 1 &&
vp 3079 nfs/nfs_vnops.c (vp->v_mount->mnt_flag & MNT_RDONLY) == 0) {
vp 3085 nfs/nfs_vnops.c (void)VOP_SETATTR(vp, &vattr, ap->a_cred, ap->a_p);
vp 3140 nfs/nfs_vnops.c struct vnode *vp = ap->a_vp;
vp 3141 nfs/nfs_vnops.c struct nfsnode *np = VTONFS(vp);
vp 3153 nfs/nfs_vnops.c if (vp->v_usecount == 1 &&
vp 3154 nfs/nfs_vnops.c (vp->v_mount->mnt_flag & MNT_RDONLY) == 0) {
vp 3160 nfs/nfs_vnops.c (void)VOP_SETATTR(vp, &vattr, ap->a_cred, ap->a_p);
vp 113 nfs/nfsmount.h int nfs_vptofh(struct vnode *vp, struct fid *fhp);
vp 155 nfs/nfsnode.h #define VTONFS(vp) ((struct nfsnode *)(vp)->v_data)
vp 325 ntfs/ntfs_subr.c struct vnode *vp;
vp 327 ntfs/ntfs_subr.c vp = ntmp->ntm_sysvn[NTFS_MFTINO];
vp 328 ntfs/ntfs_subr.c error = ntfs_readattr(ntmp, VTONT(vp), NTFS_A_DATA, NULL,
vp 900 ntfs/ntfs_subr.c struct vnode * vp,
vp 909 ntfs/ntfs_subr.c struct fnode *fp = VTOF(vp);
vp 1033 ntfs/ntfs_subr.c VREF(vp);
vp 1034 ntfs/ntfs_subr.c *vpp = vp;
vp 2089 ntfs/ntfs_subr.c struct vnode *vp;
vp 2107 ntfs/ntfs_subr.c if ((error = VFS_VGET(mp, NTFS_UPCASEINO, &vp)))
vp 2109 ntfs/ntfs_subr.c error = ntfs_readattr(ntmp, VTONT(vp), NTFS_A_DATA, NULL,
vp 2112 ntfs/ntfs_subr.c vput(vp);
vp 518 ntfs/ntfs_vfsops.c struct vnode *vp;
vp 635 ntfs/ntfs_vfsops.c error = VFS_VGET(mp, NTFS_ATTRDEFINO, &vp );
vp 641 ntfs/ntfs_vfsops.c error = ntfs_readattr(ntmp, VTONT(vp),
vp 660 ntfs/ntfs_vfsops.c error = ntfs_readattr(ntmp, VTONT(vp),
vp 674 ntfs/ntfs_vfsops.c vput(vp);
vp 842 ntfs/ntfs_vfsops.c struct vnode *vp;
vp 848 ntfs/ntfs_vfsops.c vp = ntmp->ntm_sysvn[NTFS_BITMAPINO];
vp 850 ntfs/ntfs_vfsops.c bmsize = VTOF(vp)->f_size;
vp 854 ntfs/ntfs_vfsops.c error = ntfs_readattr(ntmp, VTONT(vp), NTFS_A_DATA, NULL,
vp 965 ntfs/ntfs_vfsops.c struct vnode *vp,
vp 972 ntfs/ntfs_vfsops.c ddprintf(("ntfs_fhtovp(): %s: %p\n", vp->v_mount->mnt_stat.f_mntonname,
vp 973 ntfs/ntfs_vfsops.c vp));
vp 975 ntfs/ntfs_vfsops.c fn = VTOF(vp);
vp 976 ntfs/ntfs_vfsops.c ntp = VTONT(vp);
vp 1002 ntfs/ntfs_vfsops.c struct vnode *vp;
vp 1095 ntfs/ntfs_vfsops.c error = getnewvnode(VT_NTFS, ntmp->ntm_mountp, ntfs_vnodeop_p, &vp);
vp 1105 ntfs/ntfs_vfsops.c dprintf(("ntfs_vget: vnode: %p for ntnode: %d\n", vp,ino));
vp 1110 ntfs/ntfs_vfsops.c fp->f_vp = vp;
vp 1111 ntfs/ntfs_vfsops.c vp->v_data = fp;
vp 1112 ntfs/ntfs_vfsops.c vp->v_type = f_type;
vp 1115 ntfs/ntfs_vfsops.c vp->v_flag |= VROOT;
vp 1118 ntfs/ntfs_vfsops.c error = VN_LOCK(vp, lkflags, p);
vp 1120 ntfs/ntfs_vfsops.c vput(vp);
vp 1126 ntfs/ntfs_vfsops.c genfs_node_init(vp, &ntfs_genfsops);
vp 1128 ntfs/ntfs_vfsops.c *vpp = vp;
vp 151 ntfs/ntfs_vnops.c struct vnode *vp = ap->a_vp;
vp 152 ntfs/ntfs_vnops.c struct fnode *fp = VTOF(vp);
vp 198 ntfs/ntfs_vnops.c struct vnode *vp = ap->a_vp;
vp 199 ntfs/ntfs_vnops.c struct fnode *fp = VTOF(vp);
vp 224 ntfs/ntfs_vnops.c vap->va_type = vp->v_type;
vp 237 ntfs/ntfs_vnops.c struct vnode *vp = ap->a_vp;
vp 242 ntfs/ntfs_vnops.c struct ntnode *ip = VTONT(vp);
vp 245 ntfs/ntfs_vnops.c dprintf(("ntfs_inactive: vnode: %p, ntnode: %d\n", vp, ip->i_number));
vp 248 ntfs/ntfs_vnops.c if (ntfs_prtactive && vp->v_usecount != 0)
vp 249 ntfs/ntfs_vnops.c vprint("ntfs_inactive: pushing active", vp);
vp 252 ntfs/ntfs_vnops.c VOP__UNLOCK(vp, 0, p);
vp 267 ntfs/ntfs_vnops.c struct vnode *vp = ap->a_vp;
vp 268 ntfs/ntfs_vnops.c struct fnode *fp = VTOF(vp);
vp 275 ntfs/ntfs_vnops.c dprintf(("ntfs_reclaim: vnode: %p, ntnode: %d\n", vp, ip->i_number));
vp 278 ntfs/ntfs_vnops.c if (ntfs_prtactive && vp->v_usecount != 0)
vp 279 ntfs/ntfs_vnops.c vprint("ntfs_reclaim: pushing active", vp);
vp 290 ntfs/ntfs_vnops.c cache_purge(vp);
vp 298 ntfs/ntfs_vnops.c vp->v_data = NULL;
vp 324 ntfs/ntfs_vnops.c struct vnode *vp = bp->b_vp;
vp 325 ntfs/ntfs_vnops.c struct fnode *fp = VTOF(vp);
vp 400 ntfs/ntfs_vnops.c struct vnode *vp = ap->a_vp;
vp 401 ntfs/ntfs_vnops.c struct fnode *fp = VTOF(vp);
vp 435 ntfs/ntfs_vnops.c struct vnode *vp = ap->a_vp;
vp 436 ntfs/ntfs_vnops.c struct ntnode *ip = VTONT(vp);
vp 450 ntfs/ntfs_vnops.c switch ((int)vp->v_type) {
vp 454 ntfs/ntfs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 510 ntfs/ntfs_vnops.c struct vnode *vp = ap->a_vp;
vp 511 ntfs/ntfs_vnops.c struct ntnode *ip = VTONT(vp);
vp 534 ntfs/ntfs_vnops.c struct vnode *vp = ap->a_vp;
vp 535 ntfs/ntfs_vnops.c struct ntnode *ip = VTONT(vp);
vp 547 ntfs/ntfs_vnops.c struct vnode *vp = ap->a_vp;
vp 548 ntfs/ntfs_vnops.c struct fnode *fp = VTOF(vp);
vp 189 sys/exec.h u_long len, u_long addr, struct vnode *vp, u_long offset,
vp 191 sys/exec.h #define NEW_VMCMD(evsp,proc,len,addr,vp,offset,prot) \
vp 192 sys/exec.h new_vmcmd(evsp,proc,len,addr,vp,offset,prot, 0);
vp 193 sys/exec.h #define NEW_VMCMD2(evsp,proc,len,addr,vp,offset,prot,flags) \
vp 194 sys/exec.h new_vmcmd(evsp,proc,len,addr,vp,offset,prot,flags)
vp 196 sys/exec.h #define NEW_VMCMD(evsp,proc,len,addr,vp,offset,prot) \
vp 197 sys/exec.h NEW_VMCMD2(evsp,proc,len,addr,vp,offset,prot,0)
vp 198 sys/exec.h #define NEW_VMCMD2(evsp,proc,len,addr,vp,offset,prot,flags) do { \
vp 206 sys/exec.h if ((vcp->ev_vp = (vp)) != NULLVP) \
vp 207 sys/exec.h VREF(vp); \
vp 488 sys/mount.h int (*vfs_vptofh)(struct vnode *vp, struct fid *fhp);
vp 232 sys/vnode.h #define VREF(vp) vref(vp) /* increase reference */
vp 234 sys/vnode.h #define VN_KNOTE(vp, b) \
vp 235 sys/vnode.h KNOTE(&vp->v_selectinfo.si_note, (b))
vp 65 ufs/ext2fs/ext2fs_balloc.c struct vnode *vp = ITOV(ip);
vp 85 ufs/ext2fs/ext2fs_balloc.c error = bread(vp, bn, fs->e2fs_bsize, NOCRED, &bp);
vp 106 ufs/ext2fs/ext2fs_balloc.c bp = getblk(vp, bn, fs->e2fs_bsize, 0, 0);
vp 117 ufs/ext2fs/ext2fs_balloc.c if ((error = ufs_getlbns(vp, bn, indirs, &num)) != 0)
vp 138 ufs/ext2fs/ext2fs_balloc.c bp = getblk(vp, indirs[1].in_lbn, fs->e2fs_bsize, 0, 0);
vp 156 ufs/ext2fs/ext2fs_balloc.c error = bread(vp,
vp 180 ufs/ext2fs/ext2fs_balloc.c nbp = getblk(vp, indirs[i].in_lbn, fs->e2fs_bsize, 0, 0);
vp 228 ufs/ext2fs/ext2fs_balloc.c nbp = getblk(vp, lbn, fs->e2fs_bsize, 0, 0);
vp 237 ufs/ext2fs/ext2fs_balloc.c error = bread(vp, lbn, (int)fs->e2fs_bsize, NOCRED, &nbp);
vp 243 ufs/ext2fs/ext2fs_balloc.c nbp = getblk(vp, lbn, fs->e2fs_bsize, 0, 0);
vp 264 ufs/ext2fs/ext2fs_balloc.c r = bread(vp, indirs[unwindidx].in_lbn,
vp 279 ufs/ext2fs/ext2fs_balloc.c bp = getblk(vp, indirs[i].in_lbn, (int)fs->e2fs_bsize,
vp 99 ufs/ext2fs/ext2fs_bmap.c ext2fs_bmaparray(struct vnode *vp, int32_t bn, daddr64_t *bnp,
vp 112 ufs/ext2fs/ext2fs_bmap.c ip = VTOI(vp);
vp 113 ufs/ext2fs/ext2fs_bmap.c mp = vp->v_mount;
vp 134 ufs/ext2fs/ext2fs_bmap.c if ((error = ufs_getlbns(vp, bn, xap, nump)) != 0)
vp 154 ufs/ext2fs/ext2fs_bmap.c devvp = VFSTOUFS(vp->v_mount)->um_devvp;
vp 170 ufs/ext2fs/ext2fs_bmap.c if ((daddr == 0 && !incore(vp, metalbn)) || metalbn == bn)
vp 180 ufs/ext2fs/ext2fs_bmap.c bp = getblk(vp, metalbn, mp->mnt_stat.f_iosize, 0, 0);
vp 152 ufs/ext2fs/ext2fs_extern.h #define IS_EXT2_VNODE(vp) (vp->v_tag == VT_EXT2FS)
vp 111 ufs/ext2fs/ext2fs_inode.c struct vnode *vp = ap->a_vp;
vp 112 ufs/ext2fs/ext2fs_inode.c struct inode *ip = VTOI(vp);
vp 119 ufs/ext2fs/ext2fs_inode.c if (prtactive && vp->v_usecount != 0)
vp 120 ufs/ext2fs/ext2fs_inode.c vprint("ext2fs_inactive: pushing active", vp);
vp 128 ufs/ext2fs/ext2fs_inode.c if (ip->i_e2fs_nlink == 0 && (vp->v_mount->mnt_flag & MNT_RDONLY) == 0) {
vp 141 ufs/ext2fs/ext2fs_inode.c VOP_UNLOCK(vp, 0, p);
vp 147 ufs/ext2fs/ext2fs_inode.c vrecycle(vp, p);
vp 439 ufs/ext2fs/ext2fs_inode.c struct vnode *vp;
vp 465 ufs/ext2fs/ext2fs_inode.c vp = ITOV(ip);
vp 466 ufs/ext2fs/ext2fs_inode.c bp = getblk(vp, lbn, (int)fs->e2fs_bsize, 0, 0);
vp 133 ufs/ext2fs/ext2fs_lookup.c struct vnode *vp = ap->a_vp;
vp 134 ufs/ext2fs/ext2fs_lookup.c struct m_ext2fs *fs = VTOI(vp)->i_e2fs;
vp 146 ufs/ext2fs/ext2fs_lookup.c if (vp->v_type != VDIR)
vp 1015 ufs/ext2fs/ext2fs_lookup.c struct vnode *vp;
vp 1020 ufs/ext2fs/ext2fs_lookup.c vp = ITOV(target);
vp 1031 ufs/ext2fs/ext2fs_lookup.c if (vp->v_type != VDIR) {
vp 1035 ufs/ext2fs/ext2fs_lookup.c error = vn_rdwr(UIO_READ, vp, (caddr_t)&dirbuf,
vp 1055 ufs/ext2fs/ext2fs_lookup.c vput(vp);
vp 1056 ufs/ext2fs/ext2fs_lookup.c error = VFS_VGET(vp->v_mount, ino, &vp);
vp 1058 ufs/ext2fs/ext2fs_lookup.c vp = NULL;
vp 1068 ufs/ext2fs/ext2fs_lookup.c if (vp != NULL)
vp 1069 ufs/ext2fs/ext2fs_lookup.c vput(vp);
vp 68 ufs/ext2fs/ext2fs_readwrite.c struct vnode *vp;
vp 78 ufs/ext2fs/ext2fs_readwrite.c vp = ap->a_vp;
vp 79 ufs/ext2fs/ext2fs_readwrite.c ip = VTOI(vp);
vp 86 ufs/ext2fs/ext2fs_readwrite.c if (vp->v_type == VLNK) {
vp 87 ufs/ext2fs/ext2fs_readwrite.c if ((int)ext2fs_size(ip) < vp->v_mount->mnt_maxsymlinklen ||
vp 88 ufs/ext2fs/ext2fs_readwrite.c (vp->v_mount->mnt_maxsymlinklen == 0 &&
vp 91 ufs/ext2fs/ext2fs_readwrite.c } else if (vp->v_type != VREG && vp->v_type != VDIR)
vp 92 ufs/ext2fs/ext2fs_readwrite.c panic("%s: type %d", "ext2fs_read", vp->v_type);
vp 115 ufs/ext2fs/ext2fs_readwrite.c error = bread(vp, lbn, size, NOCRED, &bp);
vp 118 ufs/ext2fs/ext2fs_readwrite.c error = breadn(vp, lbn,
vp 121 ufs/ext2fs/ext2fs_readwrite.c error = bread(vp, lbn, size, NOCRED, &bp);
vp 147 ufs/ext2fs/ext2fs_readwrite.c if (!(vp->v_mount->mnt_flag & MNT_NOATIME)) {
vp 160 ufs/ext2fs/ext2fs_readwrite.c struct vnode *vp;
vp 172 ufs/ext2fs/ext2fs_readwrite.c vp = ap->a_vp;
vp 173 ufs/ext2fs/ext2fs_readwrite.c ip = VTOI(vp);
vp 187 ufs/ext2fs/ext2fs_readwrite.c switch (vp->v_type) {
vp 215 ufs/ext2fs/ext2fs_readwrite.c if (vp->v_type == VREG && p &&
vp 245 ufs/ext2fs/ext2fs_readwrite.c uvm_vnp_setsize(vp, ip->i_e2fs_size);
vp 247 ufs/ext2fs/ext2fs_readwrite.c uvm_vnp_uncache(vp);
vp 82 ufs/ext2fs/ext2fs_subr.c struct vnode *vp;
vp 88 ufs/ext2fs/ext2fs_subr.c vp = ITOV(ip);
vp 93 ufs/ext2fs/ext2fs_subr.c if ((error = bread(vp, lbn, fs->e2fs_bsize, NOCRED, &bp)) != 0) {
vp 110 ufs/ext2fs/ext2fs_subr.c struct vnode *vp;
vp 118 ufs/ext2fs/ext2fs_subr.c if (VOP_BMAP(ep->b_vp, (daddr_t)0, &vp, (daddr_t)0, NULL))
vp 120 ufs/ext2fs/ext2fs_subr.c if (vp != ip->i_devvp)
vp 126 ufs/ext2fs/ext2fs_subr.c vprint("Disk overlap", vp);
vp 143 ufs/ext2fs/ext2fs_subr.c struct vnode *vp, *nvp;
vp 146 ufs/ext2fs/ext2fs_subr.c vp = *vpp;
vp 147 ufs/ext2fs/ext2fs_subr.c ip = VTOI(vp);
vp 148 ufs/ext2fs/ext2fs_subr.c vp->v_type = IFTOVT(ip->i_e2fs_mode);
vp 150 ufs/ext2fs/ext2fs_subr.c switch(vp->v_type) {
vp 153 ufs/ext2fs/ext2fs_subr.c vp->v_op = specops;
vp 155 ufs/ext2fs/ext2fs_subr.c nvp = checkalias(vp, fs2h32(ip->i_e2din->e2di_rdev), mp);
vp 162 ufs/ext2fs/ext2fs_subr.c nvp->v_data = vp->v_data;
vp 163 ufs/ext2fs/ext2fs_subr.c vp->v_data = NULL;
vp 164 ufs/ext2fs/ext2fs_subr.c vp->v_op = spec_vnodeop_p;
vp 166 ufs/ext2fs/ext2fs_subr.c vp->v_flag &= ~VLOCKSWORK;
vp 168 ufs/ext2fs/ext2fs_subr.c vrele(vp);
vp 169 ufs/ext2fs/ext2fs_subr.c vgone(vp);
vp 171 ufs/ext2fs/ext2fs_subr.c vp = nvp;
vp 172 ufs/ext2fs/ext2fs_subr.c ip->i_vnode = vp;
vp 179 ufs/ext2fs/ext2fs_subr.c vp->v_op = fifoops;
vp 191 ufs/ext2fs/ext2fs_subr.c vp->v_flag |= VROOT;
vp 198 ufs/ext2fs/ext2fs_subr.c *vpp = vp;
vp 330 ufs/ext2fs/ext2fs_vfsops.c ext2fs_reload_vnode(struct vnode *vp, void *args)
vp 341 ufs/ext2fs/ext2fs_vfsops.c if (vp->v_usecount == 0) {
vp 342 ufs/ext2fs/ext2fs_vfsops.c vgonel(vp, era->p);
vp 349 ufs/ext2fs/ext2fs_vfsops.c if (vget(vp, LK_EXCLUSIVE, era->p))
vp 352 ufs/ext2fs/ext2fs_vfsops.c if (vinvalbuf(vp, 0, era->cred, era->p, 0, 0))
vp 357 ufs/ext2fs/ext2fs_vfsops.c ip = VTOI(vp);
vp 362 ufs/ext2fs/ext2fs_vfsops.c vput(vp);
vp 369 ufs/ext2fs/ext2fs_vfsops.c vput(vp);
vp 715 ufs/ext2fs/ext2fs_vfsops.c int ext2fs_sync_vnode(struct vnode *vp, void *);
vp 725 ufs/ext2fs/ext2fs_vfsops.c ext2fs_sync_vnode(struct vnode *vp, void *args)
vp 731 ufs/ext2fs/ext2fs_vfsops.c ip = VTOI(vp);
vp 732 ufs/ext2fs/ext2fs_vfsops.c if (vp->v_type == VNON ||
vp 734 ufs/ext2fs/ext2fs_vfsops.c LIST_EMPTY(&vp->v_dirtyblkhd)) ||
vp 739 ufs/ext2fs/ext2fs_vfsops.c if (vget(vp, LK_EXCLUSIVE | LK_NOWAIT, esa->p))
vp 742 ufs/ext2fs/ext2fs_vfsops.c if ((error = VOP_FSYNC(vp, esa->cred, esa->waitfor, esa->p)) != 0)
vp 744 ufs/ext2fs/ext2fs_vfsops.c vput(vp);
vp 815 ufs/ext2fs/ext2fs_vfsops.c struct vnode *vp;
vp 827 ufs/ext2fs/ext2fs_vfsops.c if ((error = getnewvnode(VT_EXT2FS, mp, ext2fs_vnodeop_p, &vp)) != 0) {
vp 835 ufs/ext2fs/ext2fs_vfsops.c vp->v_data = ip;
vp 836 ufs/ext2fs/ext2fs_vfsops.c ip->i_vnode = vp;
vp 853 ufs/ext2fs/ext2fs_vfsops.c vrele(vp);
vp 871 ufs/ext2fs/ext2fs_vfsops.c vput(vp);
vp 905 ufs/ext2fs/ext2fs_vfsops.c error = ext2fs_vinit(mp, ext2fs_specop_p, EXT2FS_FIFOOPS, &vp);
vp 907 ufs/ext2fs/ext2fs_vfsops.c vput(vp);
vp 924 ufs/ext2fs/ext2fs_vfsops.c if ((vp->v_mount->mnt_flag & MNT_RDONLY) == 0)
vp 928 ufs/ext2fs/ext2fs_vfsops.c *vpp = vp;
vp 977 ufs/ext2fs/ext2fs_vfsops.c ext2fs_vptofh(struct vnode *vp, struct fid *fhp)
vp 982 ufs/ext2fs/ext2fs_vfsops.c ip = VTOI(vp);
vp 150 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp = ap->a_vp;
vp 151 ufs/ext2fs/ext2fs_vnops.c struct inode *ip = VTOI(vp);
vp 167 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp = ap->a_vp;
vp 168 ufs/ext2fs/ext2fs_vnops.c struct inode *ip = VTOI(vp);
vp 200 ufs/ext2fs/ext2fs_vnops.c if (vp->v_type == VBLK)
vp 202 ufs/ext2fs/ext2fs_vnops.c else if (vp->v_type == VCHR)
vp 205 ufs/ext2fs/ext2fs_vnops.c vap->va_blocksize = vp->v_mount->mnt_stat.f_iosize;
vp 207 ufs/ext2fs/ext2fs_vnops.c vap->va_type = vp->v_type;
vp 220 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp = ap->a_vp;
vp 221 ufs/ext2fs/ext2fs_vnops.c struct inode *ip = VTOI(vp);
vp 236 ufs/ext2fs/ext2fs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 269 ufs/ext2fs/ext2fs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 271 ufs/ext2fs/ext2fs_vnops.c error = ext2fs_chown(vp, vap->va_uid, vap->va_gid, cred, p);
vp 281 ufs/ext2fs/ext2fs_vnops.c switch (vp->v_type) {
vp 286 ufs/ext2fs/ext2fs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 296 ufs/ext2fs/ext2fs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 301 ufs/ext2fs/ext2fs_vnops.c (error = VOP_ACCESS(vp, VWRITE, cred, p))))
vp 304 ufs/ext2fs/ext2fs_vnops.c if (!(vp->v_mount->mnt_flag & MNT_NOATIME))
vp 314 ufs/ext2fs/ext2fs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 316 ufs/ext2fs/ext2fs_vnops.c error = ext2fs_chmod(vp, (int)vap->va_mode, cred, p);
vp 326 ufs/ext2fs/ext2fs_vnops.c ext2fs_chmod(struct vnode *vp, mode_t mode, struct ucred *cred, struct proc *p)
vp 328 ufs/ext2fs/ext2fs_vnops.c struct inode *ip = VTOI(vp);
vp 334 ufs/ext2fs/ext2fs_vnops.c if (vp->v_type != VDIR && (mode & S_ISTXT))
vp 342 ufs/ext2fs/ext2fs_vnops.c if ((vp->v_flag & VTEXT) && (ip->i_e2fs_mode & S_ISTXT) == 0)
vp 343 ufs/ext2fs/ext2fs_vnops.c (void) uvm_vnp_uncache(vp);
vp 352 ufs/ext2fs/ext2fs_vnops.c ext2fs_chown(struct vnode *vp, uid_t uid, gid_t gid, struct ucred *cred, struct proc *p)
vp 354 ufs/ext2fs/ext2fs_vnops.c struct inode *ip = VTOI(vp);
vp 391 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp = ap->a_vp;
vp 395 ufs/ext2fs/ext2fs_vnops.c ip = VTOI(vp);
vp 396 ufs/ext2fs/ext2fs_vnops.c if (vp->v_type == VDIR ||
vp 408 ufs/ext2fs/ext2fs_vnops.c if (dvp == vp)
vp 409 ufs/ext2fs/ext2fs_vnops.c vrele(vp);
vp 411 ufs/ext2fs/ext2fs_vnops.c vput(vp);
vp 424 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp = ap->a_vp;
vp 434 ufs/ext2fs/ext2fs_vnops.c if (vp->v_type == VDIR) {
vp 439 ufs/ext2fs/ext2fs_vnops.c if (dvp->v_mount != vp->v_mount) {
vp 444 ufs/ext2fs/ext2fs_vnops.c if (dvp != vp && (error = vn_lock(vp, LK_EXCLUSIVE, p))) {
vp 448 ufs/ext2fs/ext2fs_vnops.c ip = VTOI(vp);
vp 470 ufs/ext2fs/ext2fs_vnops.c if (dvp != vp)
vp 471 ufs/ext2fs/ext2fs_vnops.c VOP_UNLOCK(vp, 0, p);
vp 1001 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp = ap->a_vp;
vp 1007 ufs/ext2fs/ext2fs_vnops.c ip = VTOI(vp);
vp 1014 ufs/ext2fs/ext2fs_vnops.c vput(vp);
vp 1065 ufs/ext2fs/ext2fs_vnops.c vput(vp);
vp 1076 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp, **vpp = ap->a_vpp;
vp 1084 ufs/ext2fs/ext2fs_vnops.c vp = *vpp;
vp 1086 ufs/ext2fs/ext2fs_vnops.c if (len < vp->v_mount->mnt_maxsymlinklen) {
vp 1087 ufs/ext2fs/ext2fs_vnops.c ip = VTOI(vp);
vp 1094 ufs/ext2fs/ext2fs_vnops.c error = vn_rdwr(UIO_WRITE, vp, ap->a_target, len, (off_t)0,
vp 1098 ufs/ext2fs/ext2fs_vnops.c vput(vp);
vp 1109 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp = ap->a_vp;
vp 1110 ufs/ext2fs/ext2fs_vnops.c struct inode *ip = VTOI(vp);
vp 1114 ufs/ext2fs/ext2fs_vnops.c if (isize < vp->v_mount->mnt_maxsymlinklen ||
vp 1115 ufs/ext2fs/ext2fs_vnops.c (vp->v_mount->mnt_maxsymlinklen == 0 && ip->i_e2fs_nblock == 0)) {
vp 1119 ufs/ext2fs/ext2fs_vnops.c return (VOP_READ(vp, ap->a_uio, 0, ap->a_cred));
vp 1209 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp = ap->a_vp;
vp 1211 ufs/ext2fs/ext2fs_vnops.c vflushbuf(vp, ap->a_waitfor == MNT_WAIT);
vp 1223 ufs/ext2fs/ext2fs_vnops.c struct vnode *vp = ap->a_vp;
vp 1228 ufs/ext2fs/ext2fs_vnops.c if (prtactive && vp->v_usecount != 0)
vp 1229 ufs/ext2fs/ext2fs_vnops.c vprint("ext2fs_reclaim: pushing active", vp);
vp 1235 ufs/ext2fs/ext2fs_vnops.c ip = VTOI(vp);
vp 1241 ufs/ext2fs/ext2fs_vnops.c cache_purge(vp);
vp 1250 ufs/ext2fs/ext2fs_vnops.c vp->v_data = NULL;
vp 372 ufs/ffs/ffs_alloc.c struct vnode *vp;
vp 380 ufs/ffs/ffs_alloc.c vp = ap->a_vp;
vp 381 ufs/ffs/ffs_alloc.c ip = VTOI(vp);
vp 414 ufs/ffs/ffs_alloc.c if (ufs_getlbns(vp, start_lbn, start_ap, &start_lvl) ||
vp 415 ufs/ffs/ffs_alloc.c ufs_getlbns(vp, end_lbn, end_ap, &end_lvl))
vp 425 ufs/ffs/ffs_alloc.c if (bread(vp, idp->in_lbn, (int)fs->fs_bsize, NOCRED, &sbp)) {
vp 448 ufs/ffs/ffs_alloc.c if (bread(vp, idp->in_lbn, (int)fs->fs_bsize, NOCRED, &ebp))
vp 487 ufs/ffs/ffs_alloc.c if (DOINGSOFTDEP(vp)) {
vp 539 ufs/ffs/ffs_alloc.c if (!DOINGSOFTDEP(vp))
vp 575 ufs/ffs/ffs_alloc.c struct vnode *vp;
vp 585 ufs/ffs/ffs_alloc.c vp = ap->a_vp;
vp 586 ufs/ffs/ffs_alloc.c ip = VTOI(vp);
vp 624 ufs/ffs/ffs_alloc.c if (ufs_getlbns(vp, start_lbn, start_ap, &start_lvl) ||
vp 625 ufs/ffs/ffs_alloc.c ufs_getlbns(vp, end_lbn, end_ap, &end_lvl))
vp 636 ufs/ffs/ffs_alloc.c if (bread(vp, idp->in_lbn, (int)fs->fs_bsize, NOCRED, &sbp)) {
vp 655 ufs/ffs/ffs_alloc.c if (bread(vp, idp->in_lbn, (int)fs->fs_bsize, NOCRED, &ebp))
vp 703 ufs/ffs/ffs_alloc.c if (DOINGSOFTDEP(vp)) {
vp 756 ufs/ffs/ffs_alloc.c if (!DOINGSOFTDEP(vp))
vp 80 ufs/ffs/ffs_balloc.c struct vnode *vp;
vp 88 ufs/ffs/ffs_balloc.c vp = ITOV(ip);
vp 114 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp))
vp 120 ufs/ffs/ffs_balloc.c uvm_vnp_setsize(vp, ip->i_ffs1_size);
vp 145 ufs/ffs/ffs_balloc.c error = bread(vp, lbn, fs->fs_bsize, NOCRED,
vp 167 ufs/ffs/ffs_balloc.c error = bread(vp, lbn, fs->fs_bsize,
vp 187 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp))
vp 208 ufs/ffs/ffs_balloc.c *bpp = getblk(vp, lbn, fs->fs_bsize, 0, 0);
vp 215 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp))
vp 228 ufs/ffs/ffs_balloc.c if ((error = ufs_getlbns(vp, lbn, indirs, &num)) != 0)
vp 251 ufs/ffs/ffs_balloc.c bp = getblk(vp, indirs[1].in_lbn, fs->fs_bsize, 0, 0);
vp 255 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp)) {
vp 276 ufs/ffs/ffs_balloc.c error = bread(vp,
vp 301 ufs/ffs/ffs_balloc.c nbp = getblk(vp, indirs[i].in_lbn, fs->fs_bsize, 0, 0);
vp 305 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp)) {
vp 346 ufs/ffs/ffs_balloc.c nbp = getblk(vp, lbn, fs->fs_bsize, 0, 0);
vp 352 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp))
vp 370 ufs/ffs/ffs_balloc.c error = bread(vp, lbn, (int)fs->fs_bsize, NOCRED, &nbp);
vp 376 ufs/ffs/ffs_balloc.c nbp = getblk(vp, lbn, fs->fs_bsize, 0, 0);
vp 400 ufs/ffs/ffs_balloc.c VOP_FSYNC(vp, p->p_ucred, MNT_WAIT, p);
vp 410 ufs/ffs/ffs_balloc.c r = bread(vp, indirs[unwindidx].in_lbn,
vp 431 ufs/ffs/ffs_balloc.c VOP_FSYNC(vp, p->p_ucred, MNT_WAIT, p);
vp 447 ufs/ffs/ffs_balloc.c struct vnode *vp;
vp 450 ufs/ffs/ffs_balloc.c vp = ITOV(ip);
vp 483 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp))
vp 489 ufs/ffs/ffs_balloc.c uvm_vnp_setsize(vp, ip->i_ffs2_size);
vp 516 ufs/ffs/ffs_balloc.c error = bread(vp, lbn, fs->fs_bsize, NOCRED,
vp 540 ufs/ffs/ffs_balloc.c error = bread(vp, lbn, fs->fs_bsize,
vp 562 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp))
vp 583 ufs/ffs/ffs_balloc.c bp = getblk(vp, lbn, fs->fs_bsize, 0, 0);
vp 592 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp))
vp 607 ufs/ffs/ffs_balloc.c error = ufs_getlbns(vp, lbn, indirs, &num);
vp 633 ufs/ffs/ffs_balloc.c bp = getblk(vp, indirs[1].in_lbn, fs->fs_bsize, 0, 0);
vp 637 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp)) {
vp 661 ufs/ffs/ffs_balloc.c error = bread(vp, indirs[i].in_lbn, (int) fs->fs_bsize,
vp 693 ufs/ffs/ffs_balloc.c nbp = getblk(vp, indirs[i].in_lbn, fs->fs_bsize, 0, 0);
vp 697 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp)) {
vp 745 ufs/ffs/ffs_balloc.c nbp = getblk(vp, lbn, fs->fs_bsize, 0, 0);
vp 752 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp))
vp 777 ufs/ffs/ffs_balloc.c error = bread(vp, lbn, (int)fs->fs_bsize, NOCRED, &nbp);
vp 783 ufs/ffs/ffs_balloc.c nbp = getblk(vp, lbn, fs->fs_bsize, 0, 0);
vp 810 ufs/ffs/ffs_balloc.c VOP_FSYNC(vp, p->p_ucred, MNT_WAIT, p);
vp 822 ufs/ffs/ffs_balloc.c bp = getblk(vp, indirs[i].in_lbn, (int) fs->fs_bsize,
vp 842 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp) && unwindidx == 0) {
vp 854 ufs/ffs/ffs_balloc.c if (DOINGSOFTDEP(vp))
vp 857 ufs/ffs/ffs_balloc.c r = bread(vp, indirs[unwindidx].in_lbn,
vp 868 ufs/ffs/ffs_balloc.c bp = getblk(vp, indirs[i].in_lbn, (int)fs->fs_bsize, 0,
vp 889 ufs/ffs/ffs_balloc.c VOP_FSYNC(vp, p->p_ucred, MNT_WAIT, p);
vp 71 ufs/ffs/ffs_inode.c struct vnode *vp;
vp 77 ufs/ffs/ffs_inode.c vp = ITOV(ip);
vp 78 ufs/ffs/ffs_inode.c if (vp->v_mount->mnt_flag & MNT_RDONLY) {
vp 84 ufs/ffs/ffs_inode.c if ((vp->v_mount->mnt_flag & MNT_NOATIME) &&
vp 131 ufs/ffs/ffs_inode.c if (DOINGSOFTDEP(vp))
vp 145 ufs/ffs/ffs_inode.c if (waitfor && !DOINGASYNC(vp)) {
vp 474 ufs/ffs/ffs_inode.c struct vnode *vp;
vp 505 ufs/ffs/ffs_inode.c vp = ITOV(ip);
vp 506 ufs/ffs/ffs_inode.c bp = getblk(vp, lbn, (int)fs->fs_bsize, 0, 0);
vp 536 ufs/ffs/ffs_inode.c if (!DOINGASYNC(vp)) {
vp 739 ufs/ffs/ffs_softdep.c struct vnode *vp;
vp 753 ufs/ffs/ffs_softdep.c vp = ufs_ihashlookup(VFSTOUFS(dirrem->dm_mnt)->um_dev,
vp 755 ufs/ffs/ffs_softdep.c if (vp == NULL || !VOP_ISLOCKED(vp))
vp 1923 ufs/ffs/ffs_softdep.c struct vnode *vp;
vp 2012 ufs/ffs/ffs_softdep.c vp = ITOV(ip);
vp 2014 ufs/ffs/ffs_softdep.c drain_output(vp, 1);
vp 2015 ufs/ffs/ffs_softdep.c while ((bp = LIST_FIRST(&vp->v_dirtyblkhd))) {
vp 3141 ufs/ffs/ffs_softdep.c struct vnode *vp;
vp 3146 ufs/ffs/ffs_softdep.c if ((error = VFS_VGET(dirrem->dm_mnt, dirrem->dm_oldinum, &vp)) != 0) {
vp 3150 ufs/ffs/ffs_softdep.c ip = VTOI(vp);
vp 3169 ufs/ffs/ffs_softdep.c vput(vp);
vp 3195 ufs/ffs/ffs_softdep.c vput(vp);
vp 3213 ufs/ffs/ffs_softdep.c vput(vp);
vp 3220 ufs/ffs/ffs_softdep.c UFS_UPDATE(VTOI(vp), 0);
vp 3221 ufs/ffs/ffs_softdep.c vput(vp);
vp 3243 ufs/ffs/ffs_softdep.c struct vnode vp;
vp 3261 ufs/ffs/ffs_softdep.c tip.i_vnode = &vp;
vp 3262 ufs/ffs/ffs_softdep.c vp.v_data = &tip;
vp 4546 ufs/ffs/ffs_softdep.c softdep_fsync(vp)
vp 4547 ufs/ffs/ffs_softdep.c struct vnode *vp; /* the "in_core" copy of the inode */
vp 4564 ufs/ffs/ffs_softdep.c ip = VTOI(vp);
vp 4615 ufs/ffs/ffs_softdep.c if (vp->v_flag & VXLOCK)
vp 4625 ufs/ffs/ffs_softdep.c VOP_UNLOCK(vp, 0, p);
vp 4627 ufs/ffs/ffs_softdep.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p);
vp 4679 ufs/ffs/ffs_softdep.c softdep_fsync_mountdev(vp, waitfor)
vp 4680 ufs/ffs/ffs_softdep.c struct vnode *vp;
vp 4686 ufs/ffs/ffs_softdep.c if (!vn_isdisk(vp, NULL))
vp 4689 ufs/ffs/ffs_softdep.c for (bp = LIST_FIRST(&vp->v_dirtyblkhd); bp; bp = nbp) {
vp 4719 ufs/ffs/ffs_softdep.c nbp = LIST_FIRST(&vp->v_dirtyblkhd);
vp 4722 ufs/ffs/ffs_softdep.c drain_output(vp, 1);
vp 4741 ufs/ffs/ffs_softdep.c struct vnode *vp = ap->a_vp;
vp 4753 ufs/ffs/ffs_softdep.c if (!vn_isdisk(vp, NULL)) {
vp 4754 ufs/ffs/ffs_softdep.c if (!DOINGSOFTDEP(vp))
vp 4757 ufs/ffs/ffs_softdep.c if (vp->v_specmountpoint == NULL ||
vp 4758 ufs/ffs/ffs_softdep.c (vp->v_specmountpoint->mnt_flag & MNT_SOFTDEP) == 0)
vp 4764 ufs/ffs/ffs_softdep.c if ((error = flush_inodedep_deps(VTOI(vp)->i_fs, VTOI(vp)->i_number))) {
vp 4791 ufs/ffs/ffs_softdep.c drain_output(vp, 1);
vp 4792 ufs/ffs/ffs_softdep.c bp = LIST_FIRST(&vp->v_dirtyblkhd);
vp 4880 ufs/ffs/ffs_softdep.c flush_pagedep_deps(vp, pagedep->pd_mnt,
vp 4966 ufs/ffs/ffs_softdep.c drain_output(vp, 1);
vp 4967 ufs/ffs/ffs_softdep.c if (LIST_FIRST(&vp->v_dirtyblkhd) == NULL) {
vp 4980 ufs/ffs/ffs_softdep.c if (vn_isdisk(vp, NULL) &&
vp 4981 ufs/ffs/ffs_softdep.c vp->v_specmountpoint && !VOP_ISLOCKED(vp) &&
vp 4982 ufs/ffs/ffs_softdep.c (error = VFS_SYNC(vp->v_specmountpoint, MNT_WAIT, ap->a_cred,
vp 5094 ufs/ffs/ffs_softdep.c struct vnode *vp;
vp 5137 ufs/ffs/ffs_softdep.c if ((error = VFS_VGET(mp, inum, &vp)) != 0)
vp 5139 ufs/ffs/ffs_softdep.c if ((error=VOP_FSYNC(vp, p->p_ucred, MNT_NOWAIT, p)) ||
vp 5140 ufs/ffs/ffs_softdep.c (error=VOP_FSYNC(vp, p->p_ucred, MNT_NOWAIT, p))) {
vp 5141 ufs/ffs/ffs_softdep.c vput(vp);
vp 5144 ufs/ffs/ffs_softdep.c drain_output(vp, 0);
vp 5152 ufs/ffs/ffs_softdep.c bp = incore(vp, 0);
vp 5169 ufs/ffs/ffs_softdep.c vput(vp);
vp 5248 ufs/ffs/ffs_softdep.c softdep_slowdown(vp)
vp 5249 ufs/ffs/ffs_softdep.c struct vnode *vp;
vp 5379 ufs/ffs/ffs_softdep.c struct vnode *vp;
vp 5398 ufs/ffs/ffs_softdep.c if ((error = VFS_VGET(mp, ino, &vp)) != 0) {
vp 5405 ufs/ffs/ffs_softdep.c if ((error = VOP_FSYNC(vp, p->p_ucred, MNT_NOWAIT, p)))
vp 5407 ufs/ffs/ffs_softdep.c drain_output(vp, 0);
vp 5408 ufs/ffs/ffs_softdep.c vput(vp);
vp 5430 ufs/ffs/ffs_softdep.c struct vnode *vp;
vp 5479 ufs/ffs/ffs_softdep.c if ((error = VFS_VGET(mp, ino, &vp)) != 0) {
vp 5487 ufs/ffs/ffs_softdep.c if ((error = VOP_FSYNC(vp, p->p_ucred, MNT_WAIT, p)))
vp 5490 ufs/ffs/ffs_softdep.c if ((error = VOP_FSYNC(vp, p->p_ucred, MNT_NOWAIT, p)))
vp 5492 ufs/ffs/ffs_softdep.c drain_output(vp, 0);
vp 5494 ufs/ffs/ffs_softdep.c vput(vp);
vp 5628 ufs/ffs/ffs_softdep.c drain_output(vp, islocked)
vp 5629 ufs/ffs/ffs_softdep.c struct vnode *vp;
vp 5639 ufs/ffs/ffs_softdep.c while (vp->v_numoutput) {
vp 5640 ufs/ffs/ffs_softdep.c vp->v_bioflag |= VBIOWAIT;
vp 5642 ufs/ffs/ffs_softdep.c tsleep((caddr_t)&vp->v_numoutput, PRIBIO + 1, "drain_output", 0);
vp 165 ufs/ffs/ffs_softdep_stub.c softdep_fsync_mountdev(struct vnode *vp, int waitfor)
vp 186 ufs/ffs/ffs_softdep_stub.c softdep_slowdown(struct vnode *vp)
vp 60 ufs/ffs/ffs_subr.c struct vnode *vp;
vp 65 ufs/ffs/ffs_subr.c vp = ITOV(ip);
vp 71 ufs/ffs/ffs_subr.c if ((error = bread(vp, lbn, fs->fs_bsize, NOCRED, &bp)) != 0) {
vp 127 ufs/ffs/ffs_subr.c struct vnode *vp;
vp 136 ufs/ffs/ffs_subr.c if (VOP_BMAP(ep->b_vp, (daddr_t)0, &vp, (daddr_t)0, NULL))
vp 138 ufs/ffs/ffs_subr.c if (vp != ip->i_devvp)
vp 144 ufs/ffs/ffs_subr.c vprint("Disk overlap", vp);
vp 465 ufs/ffs/ffs_vfsops.c ffs_reload_vnode(struct vnode *vp, void *args)
vp 475 ufs/ffs/ffs_vfsops.c if (vp->v_usecount == 0) {
vp 476 ufs/ffs/ffs_vfsops.c vgonel(vp, fra->p);
vp 483 ufs/ffs/ffs_vfsops.c if (vget(vp, LK_EXCLUSIVE, fra->p))
vp 486 ufs/ffs/ffs_vfsops.c if (vinvalbuf(vp, 0, fra->cred, fra->p, 0, 0))
vp 492 ufs/ffs/ffs_vfsops.c ip = VTOI(vp);
vp 499 ufs/ffs/ffs_vfsops.c vput(vp);
vp 507 ufs/ffs/ffs_vfsops.c vput(vp);
vp 1134 ufs/ffs/ffs_vfsops.c ffs_sync_vnode(struct vnode *vp, void *arg) {
vp 1139 ufs/ffs/ffs_vfsops.c ip = VTOI(vp);
vp 1140 ufs/ffs/ffs_vfsops.c if (vp->v_type == VNON ||
vp 1143 ufs/ffs/ffs_vfsops.c LIST_EMPTY(&vp->v_dirtyblkhd)) ) {
vp 1147 ufs/ffs/ffs_vfsops.c if (vget(vp, LK_EXCLUSIVE | LK_NOWAIT, fsa->p))
vp 1150 ufs/ffs/ffs_vfsops.c if ((error = VOP_FSYNC(vp, fsa->cred, fsa->waitfor, fsa->p)))
vp 1152 ufs/ffs/ffs_vfsops.c VOP_UNLOCK(vp, 0, fsa->p);
vp 1153 ufs/ffs/ffs_vfsops.c vrele(vp);
vp 1244 ufs/ffs/ffs_vfsops.c struct vnode *vp;
vp 1255 ufs/ffs/ffs_vfsops.c if ((error = getnewvnode(VT_UFS, mp, ffs_vnodeop_p, &vp)) != 0) {
vp 1260 ufs/ffs/ffs_vfsops.c vp->v_flag |= VLOCKSWORK;
vp 1268 ufs/ffs/ffs_vfsops.c vp->v_data = ip;
vp 1269 ufs/ffs/ffs_vfsops.c ip->i_vnode = vp;
vp 1288 ufs/ffs/ffs_vfsops.c vrele(vp);
vp 1307 ufs/ffs/ffs_vfsops.c vput(vp);
vp 1328 ufs/ffs/ffs_vfsops.c if (DOINGSOFTDEP(vp))
vp 1337 ufs/ffs/ffs_vfsops.c error = ufs_vinit(mp, ffs_specop_p, FFS_FIFOOPS, &vp);
vp 1339 ufs/ffs/ffs_vfsops.c vput(vp);
vp 1352 ufs/ffs/ffs_vfsops.c if ((vp->v_mount->mnt_flag & MNT_RDONLY) == 0)
vp 1365 ufs/ffs/ffs_vfsops.c *vpp = vp;
vp 1397 ufs/ffs/ffs_vfsops.c ffs_vptofh(struct vnode *vp, struct fid *fhp)
vp 1402 ufs/ffs/ffs_vfsops.c ip = VTOI(vp);
vp 172 ufs/ffs/ffs_vnops.c struct vnode *vp;
vp 183 ufs/ffs/ffs_vnops.c vp = ap->a_vp;
vp 184 ufs/ffs/ffs_vnops.c ip = VTOI(vp);
vp 192 ufs/ffs/ffs_vnops.c if (vp->v_type == VLNK) {
vp 193 ufs/ffs/ffs_vnops.c if ((int)DIP(ip, size) < vp->v_mount->mnt_maxsymlinklen ||
vp 194 ufs/ffs/ffs_vnops.c (vp->v_mount->mnt_maxsymlinklen == 0 &&
vp 197 ufs/ffs/ffs_vnops.c } else if (vp->v_type != VREG && vp->v_type != VDIR)
vp 198 ufs/ffs/ffs_vnops.c panic("ffs_read: type %d", vp->v_type);
vp 221 ufs/ffs/ffs_vnops.c error = bread(vp, lbn, size, NOCRED, &bp);
vp 223 ufs/ffs/ffs_vnops.c error = bread_cluster(vp, lbn, size, &bp);
vp 225 ufs/ffs/ffs_vnops.c error = bread(vp, lbn, size, NOCRED, &bp);
vp 263 ufs/ffs/ffs_vnops.c struct vnode *vp;
vp 276 ufs/ffs/ffs_vnops.c vp = ap->a_vp;
vp 277 ufs/ffs/ffs_vnops.c ip = VTOI(vp);
vp 291 ufs/ffs/ffs_vnops.c switch (vp->v_type) {
vp 317 ufs/ffs/ffs_vnops.c if (vp->v_type == VREG && p && !(ioflag & IO_NOLIMIT) &&
vp 344 ufs/ffs/ffs_vnops.c uvm_vnp_setsize(vp, DIP(ip, size));
vp 347 ufs/ffs/ffs_vnops.c (void)uvm_vnp_uncache(vp);
vp 381 ufs/ffs/ffs_vnops.c VN_KNOTE(vp, NOTE_WRITE | (extended ? NOTE_EXTEND : 0));
vp 403 ufs/ffs/ffs_vnops.c struct vnode *vp = ap->a_vp;
vp 407 ufs/ffs/ffs_vnops.c if (vp->v_type == VBLK &&
vp 408 ufs/ffs/ffs_vnops.c vp->v_specmountpoint != NULL &&
vp 409 ufs/ffs/ffs_vnops.c (vp->v_specmountpoint->mnt_flag & MNT_SOFTDEP))
vp 410 ufs/ffs/ffs_vnops.c softdep_fsync_mountdev(vp, ap->a_waitfor);
vp 421 ufs/ffs/ffs_vnops.c for (bp = LIST_FIRST(&vp->v_dirtyblkhd); bp;
vp 424 ufs/ffs/ffs_vnops.c for (bp = LIST_FIRST(&vp->v_dirtyblkhd); bp; bp = nbp) {
vp 465 ufs/ffs/ffs_vnops.c nbp = LIST_FIRST(&vp->v_dirtyblkhd);
vp 472 ufs/ffs/ffs_vnops.c vwaitforio(vp, 0, "ffs_fsync", 0);
vp 482 ufs/ffs/ffs_vnops.c if (!LIST_EMPTY(&vp->v_dirtyblkhd)) {
vp 496 ufs/ffs/ffs_vnops.c if (vp->v_type != VBLK)
vp 497 ufs/ffs/ffs_vnops.c vprint("ffs_fsync: dirty", vp);
vp 502 ufs/ffs/ffs_vnops.c return (UFS_UPDATE(VTOI(vp), ap->a_waitfor == MNT_WAIT));
vp 512 ufs/ffs/ffs_vnops.c struct vnode *vp = ap->a_vp;
vp 513 ufs/ffs/ffs_vnops.c struct inode *ip = VTOI(vp);
vp 516 ufs/ffs/ffs_vnops.c if ((error = ufs_reclaim(vp, ap->a_p)) != 0)
vp 530 ufs/ffs/ffs_vnops.c vp->v_data = NULL;
vp 246 ufs/mfs/mfs_vfsops.c struct vnode *vp = VFSTOUFS(mp)->um_devvp;
vp 247 ufs/mfs/mfs_vfsops.c struct mfsnode *mfsp = VTOMFS(vp);
vp 272 ufs/mfs/mfs_vfsops.c sleepreturn = tsleep((caddr_t)vp, mfs_pri, "mfsidl", 0);
vp 141 ufs/mfs/mfs_vnops.c struct vnode *vp;
vp 145 ufs/mfs/mfs_vnops.c if (!vfinddev(bp->b_dev, VBLK, &vp) || vp->v_usecount == 0)
vp 148 ufs/mfs/mfs_vnops.c mfsp = VTOMFS(vp);
vp 166 ufs/mfs/mfs_vnops.c wakeup((caddr_t)vp);
vp 221 ufs/mfs/mfs_vnops.c struct vnode *vp = ap->a_vp;
vp 222 ufs/mfs/mfs_vnops.c struct mfsnode *mfsp = VTOMFS(vp);
vp 239 ufs/mfs/mfs_vnops.c if ((error = vinvalbuf(vp, V_SAVE, ap->a_cred, ap->a_p, 0, 0)) != 0)
vp 246 ufs/mfs/mfs_vnops.c if (vp->v_usecount > 1)
vp 247 ufs/mfs/mfs_vnops.c printf("mfs_close: ref count %d > 1\n", vp->v_usecount);
vp 250 ufs/mfs/mfs_vnops.c if (vp->v_usecount > 1 || mfsp->mfs_buflist)
vp 257 ufs/mfs/mfs_vnops.c wakeup((caddr_t)vp);
vp 287 ufs/mfs/mfs_vnops.c struct vnode *vp = ap->a_vp;
vp 289 ufs/mfs/mfs_vnops.c free(vp->v_data, M_MFSNODE);
vp 290 ufs/mfs/mfs_vnops.c vp->v_data = NULL;
vp 51 ufs/mfs/mfsnode.h #define VTOMFS(vp) ((struct mfsnode *)(vp)->v_data)
vp 299 ufs/ufs/inode.h #define VTOI(vp) ((struct inode *)(vp)->v_data)
vp 342 ufs/ufs/inode.h #define DOINGSOFTDEP(vp) ((vp)->v_mount->mnt_flag & MNT_SOFTDEP)
vp 344 ufs/ufs/inode.h #define DOINGSOFTDEP(vp) (0)
vp 346 ufs/ufs/inode.h #define DOINGASYNC(vp) ((vp)->v_mount->mnt_flag & MNT_ASYNC)
vp 92 ufs/ufs/ufs_bmap.c ufs_bmaparray(struct vnode *vp, daddr_t bn, daddr64_t *bnp, struct indir *ap,
vp 105 ufs/ufs/ufs_bmap.c ip = VTOI(vp);
vp 106 ufs/ufs/ufs_bmap.c mp = vp->v_mount;
vp 127 ufs/ufs/ufs_bmap.c if ((error = ufs_getlbns(vp, bn, xap, nump)) != 0)
vp 147 ufs/ufs/ufs_bmap.c devvp = VFSTOUFS(vp->v_mount)->um_devvp;
vp 156 ufs/ufs/ufs_bmap.c if ((daddr == 0 && !incore(vp, metalbn)) || metalbn == bn)
vp 166 ufs/ufs/ufs_bmap.c bp = getblk(vp, metalbn, mp->mnt_stat.f_iosize, 0, 0);
vp 228 ufs/ufs/ufs_bmap.c ufs_getlbns(struct vnode *vp, daddr_t bn, struct indir *ap, int *nump)
vp 235 ufs/ufs/ufs_bmap.c ump = VFSTOUFS(vp->v_mount);
vp 58 ufs/ufs/ufs_dirhash.c #define OFSFMT(vp) ((vp)->v_mount->mnt_maxsymlinklen <= 0)
vp 112 ufs/ufs/ufs_dirhash.c struct vnode *vp;
vp 138 ufs/ufs/ufs_dirhash.c vp = ip->i_vnode;
vp 206 ufs/ufs/ufs_dirhash.c bmask = VFSTOUFS(vp->v_mount)->um_mountp->mnt_stat.f_iosize - 1;
vp 317 ufs/ufs/ufs_dirhash.c struct vnode *vp;
vp 364 ufs/ufs/ufs_dirhash.c vp = ip->i_vnode;
vp 365 ufs/ufs/ufs_dirhash.c bmask = VFSTOUFS(vp->v_mount)->um_mountp->mnt_stat.f_iosize - 1;
vp 93 ufs/ufs/ufs_ihash.c struct vnode *vp;
vp 98 ufs/ufs/ufs_ihash.c vp = ITOV(ip);
vp 100 ufs/ufs/ufs_ihash.c if (vget(vp, LK_EXCLUSIVE, p))
vp 102 ufs/ufs/ufs_ihash.c return (vp);
vp 65 ufs/ufs/ufs_inode.c struct vnode *vp = ap->a_vp;
vp 66 ufs/ufs/ufs_inode.c struct inode *ip = VTOI(vp);
vp 73 ufs/ufs/ufs_inode.c if (prtactive && vp->v_usecount != 0)
vp 74 ufs/ufs/ufs_inode.c vprint("ffs_inactive: pushing active", vp);
vp 83 ufs/ufs/ufs_inode.c if (DIP(ip, nlink) <= 0 && (vp->v_mount->mnt_flag & MNT_RDONLY) == 0) {
vp 106 ufs/ufs/ufs_inode.c if (DOINGSOFTDEP(vp))
vp 116 ufs/ufs/ufs_inode.c VOP_UNLOCK(vp, 0, p);
vp 123 ufs/ufs/ufs_inode.c vrecycle(vp, p);
vp 132 ufs/ufs/ufs_inode.c ufs_reclaim(struct vnode *vp, struct proc *p)
vp 138 ufs/ufs/ufs_inode.c if (prtactive && vp->v_usecount != 0)
vp 139 ufs/ufs/ufs_inode.c vprint("ufs_reclaim: pushing active", vp);
vp 145 ufs/ufs/ufs_inode.c ip = VTOI(vp);
vp 150 ufs/ufs/ufs_inode.c cache_purge(vp);
vp 69 ufs/ufs/ufs_lookup.c #define FSFMT(vp) ((vp)->v_mount->mnt_maxsymlinklen <= 0)
vp 1126 ufs/ufs/ufs_lookup.c struct vnode *vp;
vp 1130 ufs/ufs/ufs_lookup.c vp = ITOV(target);
vp 1141 ufs/ufs/ufs_lookup.c if (vp->v_type != VDIR) {
vp 1145 ufs/ufs/ufs_lookup.c error = vn_rdwr(UIO_READ, vp, (caddr_t)&dirbuf,
vp 1151 ufs/ufs/ufs_lookup.c if (vp->v_mount->mnt_maxsymlinklen > 0)
vp 1170 ufs/ufs/ufs_lookup.c vput(vp);
vp 1171 ufs/ufs/ufs_lookup.c error = VFS_VGET(vp->v_mount, dirbuf.dotdot_ino, &vp);
vp 1173 ufs/ufs/ufs_lookup.c vp = NULL;
vp 1181 ufs/ufs/ufs_lookup.c if (vp != NULL)
vp 1182 ufs/ufs/ufs_lookup.c vput(vp);
vp 151 ufs/ufs/ufs_quota.c struct vnode *vp = ITOV(ip);
vp 161 ufs/ufs/ufs_quota.c dqget(vp, DIP(ip, uid), ump, USRQUOTA, &ip->i_dquot[USRQUOTA])) &&
vp 170 ufs/ufs/ufs_quota.c dqget(vp, DIP(ip, gid), ump, GRPQUOTA, &ip->i_dquot[GRPQUOTA])) &&
vp 437 ufs/ufs/ufs_quota.c struct vnode *vp = ITOV(ip);
vp 439 ufs/ufs/ufs_quota.c if (!VOP_ISLOCKED(vp))
vp 459 ufs/ufs/ufs_quota.c quotaon_vnode(struct vnode *vp, void *arg)
vp 464 ufs/ufs/ufs_quota.c if (vp->v_type == VNON || vp->v_writecount == 0)
vp 467 ufs/ufs/ufs_quota.c if (vget(vp, LK_EXCLUSIVE, p)) {
vp 471 ufs/ufs/ufs_quota.c error = getinoquota(VTOI(vp));
vp 472 ufs/ufs/ufs_quota.c vput(vp);
vp 484 ufs/ufs/ufs_quota.c struct vnode *vp, **vpp;
vp 498 ufs/ufs/ufs_quota.c vp = nd.ni_vp;
vp 499 ufs/ufs/ufs_quota.c VOP_UNLOCK(vp, 0, p);
vp 500 ufs/ufs/ufs_quota.c if (vp->v_type != VREG) {
vp 501 ufs/ufs/ufs_quota.c (void) vn_close(vp, FREAD|FWRITE, p->p_ucred, p);
vp 504 ufs/ufs/ufs_quota.c if (*vpp != vp)
vp 508 ufs/ufs/ufs_quota.c vp->v_flag |= VSYSTEM;
vp 509 ufs/ufs/ufs_quota.c *vpp = vp;
vp 544 ufs/ufs/ufs_quota.c quotaoff_vnode(struct vnode *vp, void *arg)
vp 550 ufs/ufs/ufs_quota.c if (vp->v_type == VNON)
vp 554 ufs/ufs/ufs_quota.c if (vget(vp, LK_EXCLUSIVE, qa->p))
vp 556 ufs/ufs/ufs_quota.c ip = VTOI(vp);
vp 559 ufs/ufs/ufs_quota.c dqrele(vp, dq);
vp 560 ufs/ufs/ufs_quota.c vput(vp);
vp 719 ufs/ufs/ufs_quota.c qsync_vnode(struct vnode *vp, void *arg)
vp 725 ufs/ufs/ufs_quota.c if (vp->v_type == VNON)
vp 728 ufs/ufs/ufs_quota.c if (vget(vp, LK_EXCLUSIVE | LK_NOWAIT, p))
vp 732 ufs/ufs/ufs_quota.c dq = VTOI(vp)->i_dquot[i];
vp 734 ufs/ufs/ufs_quota.c dqsync(vp, dq);
vp 736 ufs/ufs/ufs_quota.c vput(vp);
vp 796 ufs/ufs/ufs_quota.c dqget(struct vnode *vp, u_long id, struct ufsmount *ump, int type,
vp 856 ufs/ufs/ufs_quota.c if (vp != dqvp)
vp 878 ufs/ufs/ufs_quota.c if (vp != dqvp)
vp 889 ufs/ufs/ufs_quota.c dqrele(vp, dq);
vp 914 ufs/ufs/ufs_quota.c dqrele(struct vnode *vp, struct dquot *dq)
vp 924 ufs/ufs/ufs_quota.c (void) dqsync(vp, dq);
vp 934 ufs/ufs/ufs_quota.c dqsync(struct vnode *vp, struct dquot *dq)
vp 949 ufs/ufs/ufs_quota.c if (vp != dqvp)
vp 955 ufs/ufs/ufs_quota.c if (vp != dqvp)
vp 976 ufs/ufs/ufs_quota.c if (vp != dqvp)
vp 984 ufs/ufs/ufs_quota.c struct vnode *vp = ITOV(ip);
vp 988 ufs/ufs/ufs_quota.c dqrele(vp, ip->i_dquot[i]);
vp 203 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 204 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 206 ufs/ufs/ufs_vnops.c if (vp->v_usecount > 1) {
vp 219 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 220 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 229 ufs/ufs/ufs_vnops.c switch (vp->v_type) {
vp 234 ufs/ufs/ufs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 264 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 265 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 291 ufs/ufs/ufs_vnops.c if (vp->v_type == VBLK)
vp 293 ufs/ufs/ufs_vnops.c else if (vp->v_type == VCHR)
vp 296 ufs/ufs/ufs_vnops.c vap->va_blocksize = vp->v_mount->mnt_stat.f_iosize;
vp 298 ufs/ufs/ufs_vnops.c vap->va_type = vp->v_type;
vp 311 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 312 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 329 ufs/ufs/ufs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 356 ufs/ufs/ufs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 358 ufs/ufs/ufs_vnops.c error = ufs_chown(vp, vap->va_uid, vap->va_gid, cred, p);
vp 369 ufs/ufs/ufs_vnops.c switch (vp->v_type) {
vp 374 ufs/ufs/ufs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 386 ufs/ufs/ufs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 391 ufs/ufs/ufs_vnops.c (error = VOP_ACCESS(vp, VWRITE, cred, p))))
vp 403 ufs/ufs/ufs_vnops.c if (vp->v_mount->mnt_flag & MNT_RDONLY)
vp 405 ufs/ufs/ufs_vnops.c error = ufs_chmod(vp, (int)vap->va_mode, cred, p);
vp 407 ufs/ufs/ufs_vnops.c VN_KNOTE(vp, hint);
vp 416 ufs/ufs/ufs_vnops.c ufs_chmod(struct vnode *vp, int mode, struct ucred *cred, struct proc *p)
vp 418 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 425 ufs/ufs/ufs_vnops.c if (vp->v_type != VDIR && (mode & S_ISTXT))
vp 433 ufs/ufs/ufs_vnops.c if ((vp->v_flag & VTEXT) && (DIP(ip, mode) & S_ISTXT) == 0)
vp 434 ufs/ufs/ufs_vnops.c (void) uvm_vnp_uncache(vp);
vp 443 ufs/ufs/ufs_vnops.c ufs_chown(struct vnode *vp, uid_t uid, gid_t gid, struct ucred *cred,
vp 446 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 554 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 558 ufs/ufs/ufs_vnops.c ip = VTOI(vp);
vp 559 ufs/ufs/ufs_vnops.c if (vp->v_type == VDIR || (DIP(ip, flags) & (IMMUTABLE | APPEND)) ||
vp 565 ufs/ufs/ufs_vnops.c VN_KNOTE(vp, NOTE_DELETE);
vp 568 ufs/ufs/ufs_vnops.c if (dvp == vp)
vp 569 ufs/ufs/ufs_vnops.c vrele(vp);
vp 571 ufs/ufs/ufs_vnops.c vput(vp);
vp 584 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 595 ufs/ufs/ufs_vnops.c if (vp->v_type == VDIR) {
vp 600 ufs/ufs/ufs_vnops.c if (dvp->v_mount != vp->v_mount) {
vp 605 ufs/ufs/ufs_vnops.c if (dvp != vp && (error = vn_lock(vp, LK_EXCLUSIVE, p))) {
vp 609 ufs/ufs/ufs_vnops.c ip = VTOI(vp);
vp 623 ufs/ufs/ufs_vnops.c if (DOINGSOFTDEP(vp))
vp 625 ufs/ufs/ufs_vnops.c if ((error = UFS_UPDATE(ip, !DOINGSOFTDEP(vp))) == 0) {
vp 627 ufs/ufs/ufs_vnops.c error = ufs_direnter(dvp, vp, &newdir, cnp, NULL);
vp 633 ufs/ufs/ufs_vnops.c if (DOINGSOFTDEP(vp))
vp 637 ufs/ufs/ufs_vnops.c VN_KNOTE(vp, NOTE_LINK);
vp 640 ufs/ufs/ufs_vnops.c if (dvp != vp)
vp 641 ufs/ufs/ufs_vnops.c VOP_UNLOCK(vp, 0, p);
vp 1242 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 1248 ufs/ufs/ufs_vnops.c ip = VTOI(vp);
vp 1253 ufs/ufs/ufs_vnops.c if (dp == ip || vp->v_mountedhere != 0) {
vp 1258 ufs/ufs/ufs_vnops.c vput(vp);
vp 1289 ufs/ufs/ufs_vnops.c if (DOINGSOFTDEP(vp)) {
vp 1296 ufs/ufs/ufs_vnops.c if (DOINGSOFTDEP(vp)) {
vp 1312 ufs/ufs/ufs_vnops.c if (!DOINGSOFTDEP(vp)) {
vp 1319 ufs/ufs/ufs_vnops.c ioflag = DOINGASYNC(vp) ? 0 : IO_SYNC;
vp 1322 ufs/ufs/ufs_vnops.c cache_purge(vp);
vp 1330 ufs/ufs/ufs_vnops.c VN_KNOTE(vp, NOTE_DELETE);
vp 1332 ufs/ufs/ufs_vnops.c vput(vp);
vp 1343 ufs/ufs/ufs_vnops.c struct vnode *vp, **vpp = ap->a_vpp;
vp 1352 ufs/ufs/ufs_vnops.c vp = *vpp;
vp 1354 ufs/ufs/ufs_vnops.c if (len < vp->v_mount->mnt_maxsymlinklen) {
vp 1355 ufs/ufs/ufs_vnops.c ip = VTOI(vp);
vp 1360 ufs/ufs/ufs_vnops.c error = vn_rdwr(UIO_WRITE, vp, ap->a_target, len, (off_t)0,
vp 1363 ufs/ufs/ufs_vnops.c vput(vp);
vp 1490 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 1491 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 1495 ufs/ufs/ufs_vnops.c if (isize < vp->v_mount->mnt_maxsymlinklen ||
vp 1496 ufs/ufs/ufs_vnops.c (vp->v_mount->mnt_maxsymlinklen == 0 && DIP(ip, blocks) == 0)) {
vp 1500 ufs/ufs/ufs_vnops.c return (VOP_READ(vp, ap->a_uio, 0, ap->a_cred));
vp 1510 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 1512 ufs/ufs/ufs_vnops.c return (lockmgr(&VTOI(vp)->i_lock, ap->a_flags, NULL));
vp 1522 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 1524 ufs/ufs/ufs_vnops.c return (lockmgr(&VTOI(vp)->i_lock, ap->a_flags | LK_RELEASE, NULL));
vp 1547 ufs/ufs/ufs_vnops.c struct vnode *vp = bp->b_vp;
vp 1552 ufs/ufs/ufs_vnops.c ip = VTOI(vp);
vp 1553 ufs/ufs/ufs_vnops.c if (vp->v_type == VBLK || vp->v_type == VCHR)
vp 1556 ufs/ufs/ufs_vnops.c error = VOP_BMAP(vp, bp->b_lblkno, NULL, &bp->b_blkno,
vp 1575 ufs/ufs/ufs_vnops.c vp = ip->i_devvp;
vp 1576 ufs/ufs/ufs_vnops.c bp->b_dev = vp->v_rdev;
vp 1577 ufs/ufs/ufs_vnops.c VOCALL (vp->v_op, VOFFSET(vop_strategy), ap);
vp 1590 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 1591 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 1601 ufs/ufs/ufs_vnops.c if (vp->v_type == VFIFO)
vp 1602 ufs/ufs/ufs_vnops.c fifo_printinfo(vp);
vp 1651 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 1652 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 1706 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 1707 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 1774 ufs/ufs/ufs_vnops.c struct vnode *vp, *nvp;
vp 1777 ufs/ufs/ufs_vnops.c vp = *vpp;
vp 1778 ufs/ufs/ufs_vnops.c ip = VTOI(vp);
vp 1779 ufs/ufs/ufs_vnops.c switch(vp->v_type = IFTOVT(DIP(ip, mode))) {
vp 1782 ufs/ufs/ufs_vnops.c vp->v_op = specops;
vp 1783 ufs/ufs/ufs_vnops.c if ((nvp = checkalias(vp, DIP(ip, rdev), mntp)) != NULL) {
vp 1789 ufs/ufs/ufs_vnops.c nvp->v_data = vp->v_data;
vp 1790 ufs/ufs/ufs_vnops.c vp->v_data = NULL;
vp 1791 ufs/ufs/ufs_vnops.c vp->v_op = spec_vnodeop_p;
vp 1793 ufs/ufs/ufs_vnops.c vp->v_flag &= ~VLOCKSWORK;
vp 1795 ufs/ufs/ufs_vnops.c vrele(vp);
vp 1796 ufs/ufs/ufs_vnops.c vgone(vp);
vp 1800 ufs/ufs/ufs_vnops.c vp = nvp;
vp 1801 ufs/ufs/ufs_vnops.c ip->i_vnode = vp;
vp 1806 ufs/ufs/ufs_vnops.c vp->v_op = fifoops;
vp 1820 ufs/ufs/ufs_vnops.c vp->v_flag |= VROOT;
vp 1827 ufs/ufs/ufs_vnops.c *vpp = vp;
vp 1929 ufs/ufs/ufs_vnops.c struct vnode *vp = ap->a_vp;
vp 1946 ufs/ufs/ufs_vnops.c kn->kn_hook = (caddr_t)vp;
vp 1948 ufs/ufs/ufs_vnops.c SLIST_INSERT_HEAD(&vp->v_selectinfo.si_note, kn, kn_selnext);
vp 1956 ufs/ufs/ufs_vnops.c struct vnode *vp = (struct vnode *)kn->kn_hook;
vp 1958 ufs/ufs/ufs_vnops.c SLIST_REMOVE(&vp->v_selectinfo.si_note, kn, knote, kn_selnext);
vp 1965 ufs/ufs/ufs_vnops.c struct vnode *vp = (struct vnode *)kn->kn_hook;
vp 1966 ufs/ufs/ufs_vnops.c struct inode *ip = VTOI(vp);
vp 402 uvm/uvm_mmap.c struct vnode *vp;
vp 481 uvm/uvm_mmap.c vp = (struct vnode *)fp->f_data; /* convert to vnode */
vp 483 uvm/uvm_mmap.c if (vp->v_type != VREG && vp->v_type != VCHR &&
vp 484 uvm/uvm_mmap.c vp->v_type != VBLK) {
vp 489 uvm/uvm_mmap.c if (vp->v_type == VREG && (pos + size) < pos) {
vp 495 uvm/uvm_mmap.c if (vp->v_type == VCHR && iszerodev(vp->v_rdev)) {
vp 511 uvm/uvm_mmap.c "%s (pid %d comm %s)\n", vp->v_type == VCHR ?
vp 515 uvm/uvm_mmap.c if (vp->v_type == VCHR)
vp 526 uvm/uvm_mmap.c if (vp->v_type == VCHR && (flags & MAP_PRIVATE) != 0) {
vp 554 uvm/uvm_mmap.c VOP_GETATTR(vp, &va, p->p_ucred, p)))
vp 575 uvm/uvm_mmap.c handle = (caddr_t)vp;
vp 1101 uvm/uvm_mmap.c struct vnode *vp;
vp 1151 uvm/uvm_mmap.c vp = (struct vnode *) handle; /* get vnode */
vp 1152 uvm/uvm_mmap.c if (vp->v_type != VCHR) {
vp 1153 uvm/uvm_mmap.c uobj = uvn_attach((void *) vp, (flags & MAP_SHARED) ?
vp 1183 uvm/uvm_mmap.c uvm_vnp_uncache(vp);
vp 1188 uvm/uvm_mmap.c VREF(vp);
vp 1191 uvm/uvm_mmap.c uobj = udv_attach((void *) &vp->v_rdev,
vp 1201 uvm/uvm_mmap.c uobj = udv_attach((void *) &vp->v_rdev,
vp 1209 uvm/uvm_mmap.c return((vp->v_type == VREG) ? ENOMEM : EINVAL);
vp 512 uvm/uvm_swap.c swaplist_find(vp, remove)
vp 513 uvm/uvm_swap.c struct vnode *vp;
vp 527 uvm/uvm_swap.c if (sdp->swd_vp == vp) {
vp 624 uvm/uvm_swap.c struct vnode *vp;
vp 726 uvm/uvm_swap.c vp = rootvp; /* miniroot */
vp 727 uvm/uvm_swap.c if (vget(vp, LK_EXCLUSIVE, p)) {
vp 751 uvm/uvm_swap.c vp = nd.ni_vp;
vp 759 uvm/uvm_swap.c if (vp->v_type != VBLK) {
vp 763 uvm/uvm_swap.c dumpdev = vp->v_rdev;
vp 775 uvm/uvm_swap.c if ((sdp = swaplist_find(vp, 1)) == NULL) {
vp 797 uvm/uvm_swap.c if ((sdp = swaplist_find(vp, 0)) != NULL) {
vp 806 uvm/uvm_swap.c sdp->swd_vp = vp;
vp 807 uvm/uvm_swap.c sdp->swd_dev = (vp->v_type == VBLK) ? vp->v_rdev : NODEV;
vp 812 uvm/uvm_swap.c if (vp->v_type == VREG) {
vp 833 uvm/uvm_swap.c (void) swaplist_find(vp, 1); /* kill fake entry */
vp 836 uvm/uvm_swap.c if (vp->v_type == VREG) {
vp 847 uvm/uvm_swap.c if ((sdp = swaplist_find(vp, 0)) == NULL) {
vp 876 uvm/uvm_swap.c vput(vp);
vp 901 uvm/uvm_swap.c struct vnode *vp;
vp 917 uvm/uvm_swap.c vp = sdp->swd_vp;
vp 927 uvm/uvm_swap.c if (vp != rootvp) {
vp 928 uvm/uvm_swap.c if ((error = VOP_OPEN(vp, FREAD|FWRITE, p->p_ucred, p)))
vp 944 uvm/uvm_swap.c switch (vp->v_type) {
vp 954 uvm/uvm_swap.c if ((error = VOP_GETATTR(vp, &va, p->p_ucred, p)))
vp 958 uvm/uvm_swap.c VFS_STATFS(vp->v_mount, &vp->v_mount->mnt_stat, p)) != 0)
vp 961 uvm/uvm_swap.c sdp->swd_bsize = vp->v_mount->mnt_stat.f_iosize;
vp 967 uvm/uvm_swap.c if (vp->v_op == nfsv2_vnodeop_p)
vp 993 uvm/uvm_swap.c if (vp->v_type == VBLK) {
vp 1038 uvm/uvm_swap.c if (vp == rootvp) {
vp 1062 uvm/uvm_swap.c vref(vp);
vp 1084 uvm/uvm_swap.c if (vp != rootvp)
vp 1085 uvm/uvm_swap.c (void)VOP_CLOSE(vp, FREAD|FWRITE, p->p_ucred, p);
vp 1286 uvm/uvm_swap.c struct vnode *vp;
vp 1326 uvm/uvm_swap.c &vp, &nbn, &nra);
vp 1366 uvm/uvm_swap.c sdp->swd_vp, vp, (u_long)byteoff, nbn);
vp 1426 uvm/uvm_swap.c bgetvp(vp, &nbp->vb_buf);
vp 2076 uvm/uvm_swap.c struct vnode *vp;
vp 2089 uvm/uvm_swap.c if (bdevvp(swap_dev, &vp)) {
vp 2100 uvm/uvm_swap.c sdp->swd_vp = vp;
vp 2108 uvm/uvm_swap.c swaplist_find(vp, 1);
vp 2116 uvm/uvm_swap.c VOP_UNLOCK(vp, 0, curproc);
vp 180 uvm/uvm_unix.c uvm_coredump(p, vp, cred, chdr)
vp 182 uvm/uvm_unix.c struct vnode *vp;
vp 247 uvm/uvm_unix.c error = vn_rdwr(UIO_WRITE, vp,
vp 259 uvm/uvm_unix.c error = vn_rdwr(UIO_WRITE, vp,
vp 160 uvm/uvm_vnode.c struct vnode *vp = arg;
vp 161 uvm/uvm_vnode.c struct uvm_vnode *uvn = &vp->v_uvm;
vp 190 uvm/uvm_vnode.c if (vp->v_type == VBLK && bdevsw[major(vp->v_rdev)].d_type != D_DISK) {
vp 206 uvm/uvm_vnode.c VREF(vp);
vp 239 uvm/uvm_vnode.c if (vp->v_type == VBLK) {
vp 248 uvm/uvm_vnode.c result = (*bdevsw[major(vp->v_rdev)].d_ioctl)(vp->v_rdev,
vp 256 uvm/uvm_vnode.c result = VOP_GETATTR(vp, &vattr, curproc->p_ucred, curproc);
vp 278 uvm/uvm_vnode.c if (vp->v_type == VBLK)
vp 305 uvm/uvm_vnode.c VREF(vp);
vp 364 uvm/uvm_vnode.c struct vnode *vp;
vp 383 uvm/uvm_vnode.c vp = (struct vnode *) uobj;
vp 389 uvm/uvm_vnode.c vp->v_flag &= ~VTEXT;
vp 400 uvm/uvm_vnode.c vrele(vp); /* drop vnode reference */
vp 478 uvm/uvm_vnode.c vrele(vp);
vp 513 uvm/uvm_vnode.c uvm_vnp_terminate(vp)
vp 514 uvm/uvm_vnode.c struct vnode *vp;
vp 516 uvm/uvm_vnode.c struct uvm_vnode *uvn = &vp->v_uvm;
vp 524 uvm/uvm_vnode.c UVMHIST_LOG(maphist, " vp=%p, ref=%ld, flag=0x%lx", vp,
vp 1769 uvm/uvm_vnode.c uvm_vnp_uncache(vp)
vp 1770 uvm/uvm_vnode.c struct vnode *vp;
vp 1772 uvm/uvm_vnode.c struct uvm_vnode *uvn = &vp->v_uvm;
vp 1801 uvm/uvm_vnode.c VREF(vp); /* seems ok, even with VOP_LOCK */
vp 1811 uvm/uvm_vnode.c if (!VOP_ISLOCKED(vp)) {
vp 1821 uvm/uvm_vnode.c if (vp->v_op == nfsv2_vnodeop_p ||
vp 1822 uvm/uvm_vnode.c vp->v_op == spec_nfsv2nodeop_p) {
vp 1826 uvm/uvm_vnode.c if (vp->v_op == fifo_nfsv2nodeop_p) {
vp 1846 uvm/uvm_vnode.c VOP_UNLOCK(vp, 0, curproc);
vp 1848 uvm/uvm_vnode.c vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, curproc);
vp 1876 uvm/uvm_vnode.c uvm_vnp_setsize(vp, newsize)
vp 1877 uvm/uvm_vnode.c struct vnode *vp;
vp 1880 uvm/uvm_vnode.c struct uvm_vnode *uvn = &vp->v_uvm;
vp 1923 uvm/uvm_vnode.c struct vnode *vp;
vp 1939 uvm/uvm_vnode.c vp = (struct vnode *) uvn;
vp 1940 uvm/uvm_vnode.c if (mp && vp->v_mount != mp)
vp 1974 uvm/uvm_vnode.c VREF(vp);
vp 72 xfs/xfs_fs.h #define NNPFS_FROM_VNODE(vp) VFS_TO_NNPFS((vp)->v_mount)
vp 301 xfs/xfs_locl.h #define xfs_pushdirty(vp, cred, p)
vp 370 xfs/xfs_locl.h #define xfs_vfs_object_create(vp,proc,ucred) vfs_object_create(vp,proc,ucred,TRUE)
vp 372 xfs/xfs_locl.h #define xfs_vfs_object_create(vp,proc,ucred) vfs_object_create(vp,proc,ucred)
vp 376 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) uvm_vnp_setsize(vp, sz)
vp 378 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) vnode_pager_setsize(vp, sz)
vp 380 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) ubc_setsize(vp, sz)
vp 382 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz)
vp 393 xfs/xfs_locl.h void vgonel (struct vnode *vp, d_thread_t *p);
vp 186 xfs/xfs_message.c struct vnode *vp;
vp 195 xfs/xfs_message.c error = xfs_fhlookup (p, fh, &vp);
vp 214 xfs/xfs_message.c vp = ndp->ni_vp;
vp 219 xfs/xfs_message.c xfs_vfs_unlock(vp, p);
vp 223 xfs/xfs_message.c DATA_FROM_XNODE(t) = vp;
vp 285 xfs/xfs_message.c struct vnode *vp = XNODE_TO_VNODE(t);
vp 288 xfs/xfs_message.c if (vp->v_usecount >= 0 && vp->xfs_writecount >= 1)
vp 293 xfs/xfs_message.c vm_object_t obj = vp->v_object;
vp 307 xfs/xfs_message.c if (vp->v_usecount > 0 && vp->v_type != VDIR) {
vp 309 xfs/xfs_message.c if (vget(vp, 0, p))
vp 312 xfs/xfs_message.c if (UBCISVALID(vp) && !ubc_isinuse(vp, 1)) {
vp 313 xfs/xfs_message.c ubc_setsize(vp, 0);
vp 314 xfs/xfs_message.c vrele(vp);
vp 316 xfs/xfs_message.c vrele(vp);
vp 334 xfs/xfs_message.c if (vp->v_type == VDIR)
vp 335 xfs/xfs_message.c xfs_dnlc_purge(vp);
vp 336 xfs/xfs_message.c if (vp->v_usecount == 0) {
vp 339 xfs/xfs_message.c vrecycle(vp, p);
vp 386 xfs/xfs_message.c gc_vnode (struct vnode *vp,
vp 390 xfs/xfs_message.c if (vp->v_usecount <= 0) {
vp 393 xfs/xfs_message.c if (vp->v_usecount < 0) {
vp 394 xfs/xfs_message.c vprint("vrele: bad ref count", vp);
vp 400 xfs/xfs_message.c vgone(vp, VX_NOSLEEP, NULL);
vp 415 xfs/xfs_message.c gc_vnode (struct vnode *vp,
vp 419 xfs/xfs_message.c mtx_lock(&vp->v_interlock);
vp 421 xfs/xfs_message.c simple_lock(&vp->v_interlock);
vp 425 xfs/xfs_message.c if (vp->v_usecount <= 0) {
vp 429 xfs/xfs_message.c obj = vp->v_object;
vp 438 xfs/xfs_message.c mtx_unlock(&vp->v_interlock);
vp 440 xfs/xfs_message.c simple_unlock (&vp->v_interlock);
vp 447 xfs/xfs_message.c if (vp->v_usecount < 0 || vp->v_writecount != 0) {
vp 448 xfs/xfs_message.c vprint("vrele: bad ref count", vp);
vp 456 xfs/xfs_message.c vgonel (vp, p);
vp 459 xfs/xfs_message.c mtx_unlock(&vp->v_interlock);
vp 461 xfs/xfs_message.c simple_unlock(&vp->v_interlock);
vp 463 xfs/xfs_message.c vgone (vp);
vp 468 xfs/xfs_message.c mtx_unlock(&vp->v_interlock);
vp 470 xfs/xfs_message.c simple_unlock(&vp->v_interlock);
vp 408 xfs/xfs_node-bsd.c struct vnode *dvp, *vp;
vp 423 xfs/xfs_node-bsd.c tbl.dvp = tbl.vp = NULL;
vp 434 xfs/xfs_node-bsd.c tbl_enter (size_t len, const char *name, struct vnode *dvp, struct vnode *vp)
vp 439 xfs/xfs_node-bsd.c tbl.vp = vp;
vp 441 xfs/xfs_node-bsd.c tbl.vpid = vp->v_id;
vp 458 xfs/xfs_node-bsd.c && tbl.vpid == tbl.vp->v_id) {
vp 460 xfs/xfs_node-bsd.c *res = tbl.vp;
vp 473 xfs/xfs_node-bsd.c struct vnode *vp)
vp 478 xfs/xfs_node-bsd.c (unsigned long)vp));
vp 485 xfs/xfs_node-bsd.c (unsigned long)vp,
vp 526 xfs/xfs_node-bsd.c cache_enter(dvp, vp, cnp);
vp 529 xfs/xfs_node-bsd.c if (vp != NULL)
vp 530 xfs/xfs_node-bsd.c tbl_enter (cnp->cn_namelen, cnp->cn_nameptr, dvp, vp);
vp 586 xfs/xfs_node-bsd.c struct vnode *vp)
vp 593 xfs/xfs_node-bsd.c (unsigned long)vp));
vp 596 xfs/xfs_node-bsd.c return xfs_dnlc_enter (dvp, &cn, vp);
vp 746 xfs/xfs_node-bsd.c xfs_dnlc_purge (struct vnode *vp)
vp 750 xfs/xfs_node-bsd.c if (tbl.dvp == vp || tbl.vp == vp)
vp 753 xfs/xfs_node-bsd.c cache_purge(vp);
vp 107 xfs/xfs_node.h #define DATA_FROM_VNODE(vp) DATA_FROM_XNODE(VNODE_TO_XNODE(vp))
vp 112 xfs/xfs_node.h #define VNODE_TO_XNODE(vp) ((struct xfs_node *) (vp)->v_data)
vp 115 xfs/xfs_node.h #define xfs_do_vget(vp, lockflag, proc) vget((vp))
vp 117 xfs/xfs_node.h #define xfs_do_vget(vp, lockflag, proc) vget((vp), (lockflag))
vp 119 xfs/xfs_node.h #define xfs_do_vget(vp, lockflag, proc) vget((vp), (lockflag), (proc))
vp 314 xfs/xfs_syscalls-common.c struct vnode *vp;
vp 338 xfs/xfs_syscalls-common.c vp = ndp->ni_vp;
vp 340 xfs/xfs_syscalls-common.c *res = vp;
vp 352 xfs/xfs_syscalls-common.c struct vnode *vp)
vp 359 xfs/xfs_syscalls-common.c fh.fh_fsid = vp->v_mount->mnt_stat.f_fsid;
vp 361 xfs/xfs_syscalls-common.c VFS_VPTOFH(vp, &fh.fh_fid, error);
vp 363 xfs/xfs_syscalls-common.c error = VFS_VPTOFH(vp, &fh.fh_fid);
vp 382 xfs/xfs_syscalls-common.c struct vnode *vp)
vp 392 xfs/xfs_syscalls-common.c xfs_vop_getattr(vp, &vattr, xfs_thread_to_cred(p), p, error);
vp 394 xfs/xfs_syscalls-common.c xfs_vop_getattr(vp, &vattr, xfs_proc_to_cred(p), p, error);
vp 399 xfs/xfs_syscalls-common.c mnt = vp->v_mount;
vp 428 xfs/xfs_syscalls-common.c struct vnode *vp)
vp 434 xfs/xfs_syscalls-common.c if (vp == NULL)
vp 447 xfs/xfs_syscalls-common.c error = getfh_compat (p, vice_ioctl, vp);
vp 449 xfs/xfs_syscalls-common.c error = trad_fhget (p, vice_ioctl, vp);
vp 452 xfs/xfs_syscalls-common.c vrele(vp);
vp 463 xfs/xfs_syscalls-common.c struct vnode *vp,
vp 470 xfs/xfs_syscalls-common.c if (vp != NULL) {
vp 471 xfs/xfs_syscalls-common.c vrele (vp);
vp 493 xfs/xfs_syscalls-common.c struct vnode *vp)
vp 506 xfs/xfs_syscalls-common.c if (vp != NULL) {
vp 509 xfs/xfs_syscalls-common.c if (vp->v_tag != VT_XFS) {
vp 511 xfs/xfs_syscalls-common.c vrele(vp);
vp 516 xfs/xfs_syscalls-common.c xn = VNODE_TO_XNODE(vp);
vp 519 xfs/xfs_syscalls-common.c vrele(vp);
vp 632 xfs/xfs_syscalls-common.c struct vnode *vp = NULL;
vp 653 xfs/xfs_syscalls-common.c error = lookup_node (pathptr, SCARG(arg, a_followSymlinks), &vp,
vp 661 xfs/xfs_syscalls-common.c return fhget_call (proc, &vice_ioctl, vp);
vp 663 xfs/xfs_syscalls-common.c return fhopen_call (proc, &vice_ioctl, vp,
vp 666 xfs/xfs_syscalls-common.c if (vp != NULL)
vp 667 xfs/xfs_syscalls-common.c vrele (vp);
vp 671 xfs/xfs_syscalls-common.c return remote_pioctl (proc, arg, &vice_ioctl, vp);
vp 143 xfs/xfs_vfsops-bsd.c struct vnode *vp;
vp 164 xfs/xfs_vfsops-bsd.c error = xfs_getnewvnode(xfs[0].mp, &vp, &handle);
vp 168 xfs/xfs_vfsops-bsd.c xfs_do_vget(vp, 0, curproc);
vp 172 xfs/xfs_vfsops-bsd.c vp = XNODE_TO_VNODE(xn);
vp 175 xfs/xfs_vfsops-bsd.c if (vp->v_usecount <= 0)
vp 176 xfs/xfs_vfsops-bsd.c xfs_do_vget(vp, 0, curproc);
vp 178 xfs/xfs_vfsops-bsd.c VREF(vp);
vp 182 xfs/xfs_vfsops-bsd.c *vpp = vp;
vp 270 xfs/xfs_vfsops-bsd.c xfs_vptofh(struct vnode * vp,
vp 280 xfs/xfs_vfsops-bsd.c xn = VNODE_TO_XNODE(vp);
vp 425 xfs/xfs_vfsops-bsd.c struct vnode *vp;
vp 443 xfs/xfs_vfsops-bsd.c error = xfs_fhlookup (proc, &fh, &vp);
vp 448 xfs/xfs_vfsops-bsd.c switch (vp->v_type) {
vp 462 xfs/xfs_vfsops-bsd.c switch (vp->v_type) {
vp 473 xfs/xfs_vfsops-bsd.c error = vn_writechk (vp);
vp 483 xfs/xfs_vfsops-bsd.c error = VOP_ACCESS(vp, mode, cred, proc);
vp 488 xfs/xfs_vfsops-bsd.c error = VOP_OPEN(vp, flags, cred, proc);
vp 497 xfs/xfs_vfsops-bsd.c vp->v_writecount++;
vp 500 xfs/xfs_vfsops-bsd.c if (vp->v_type == VREG) {
vp 502 xfs/xfs_vfsops-bsd.c error = xfs_vfs_object_create(vp, proc, proc->td_proc->p_ucred);
vp 504 xfs/xfs_vfsops-bsd.c error = xfs_vfs_object_create(vp, proc, proc->p_ucred);
vp 514 xfs/xfs_vfsops-bsd.c fp->f_data = (caddr_t)vp;
vp 515 xfs/xfs_vfsops-bsd.c xfs_vfs_unlock(vp, proc);
vp 526 xfs/xfs_vfsops-bsd.c vput(vp);
vp 97 xfs/xfs_vfsops-bsd.h xfs_vptofh(struct vnode * vp,
vp 405 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 408 xfs/xfs_vnodeops-bsd.c int error = xfs_remove_common(dvp, vp, cnp->cn_nameptr,
vp 411 xfs/xfs_vnodeops-bsd.c int error = xfs_remove_common(dvp, vp, cnp->cn_nameptr,
vp 418 xfs/xfs_vnodeops-bsd.c if (dvp == vp)
vp 419 xfs/xfs_vnodeops-bsd.c vrele(vp);
vp 421 xfs/xfs_vnodeops-bsd.c vput(vp);
vp 427 xfs/xfs_vnodeops-bsd.c if (UBCINFOEXISTS(vp)) {
vp 428 xfs/xfs_vnodeops-bsd.c ubc_setsize(vp, 0);
vp 429 xfs/xfs_vnodeops-bsd.c ubc_release(vp);
vp 430 xfs/xfs_vnodeops-bsd.c ubc_uncache(vp);
vp 526 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 538 xfs/xfs_vnodeops-bsd.c if (dvp == vp)
vp 539 xfs/xfs_vnodeops-bsd.c vrele(vp);
vp 541 xfs/xfs_vnodeops-bsd.c vput(vp);
vp 625 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 642 xfs/xfs_vnodeops-bsd.c if (vp->v_type == VDIR) {
vp 649 xfs/xfs_vnodeops-bsd.c if (dvp->v_mount != vp->v_mount) {
vp 659 xfs/xfs_vnodeops-bsd.c if (dvp != vp && (error = xfs_vfs_writelock(vp, p))) {
vp 669 xfs/xfs_vnodeops-bsd.c vp,
vp 680 xfs/xfs_vnodeops-bsd.c if (dvp != vp)
vp 681 xfs/xfs_vnodeops-bsd.c xfs_vfs_unlock(vp, p);
vp 760 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 763 xfs/xfs_vnodeops-bsd.c ret = xfs_reclaim_common(vp);
vp 764 xfs/xfs_vnodeops-bsd.c vp->v_data = NULL;
vp 779 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 780 xfs/xfs_vnodeops-bsd.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 786 xfs/xfs_vnodeops-bsd.c (unsigned long)vp, flags));
vp 801 xfs/xfs_vnodeops-bsd.c ret = lockmgr(l, flags, &vp->v_interlock, ap->a_td);
vp 803 xfs/xfs_vnodeops-bsd.c ret = lockmgr(l, flags, &vp->v_interlock, ap->a_p);
vp 810 xfs/xfs_vnodeops-bsd.c ret = debuglockmgr(l, flags, &vp->v_interlock, ap->a_td,
vp 813 xfs/xfs_vnodeops-bsd.c ret = debuglockmgr(l, flags, &vp->v_interlock, ap->a_p,
vp 831 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 832 xfs/xfs_vnodeops-bsd.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 842 xfs/xfs_vnodeops-bsd.c (unsigned long)vp, flags,
vp 855 xfs/xfs_vnodeops-bsd.c ret = lockmgr (l, flags | LK_RELEASE, &vp->v_interlock, ap->a_td);
vp 857 xfs/xfs_vnodeops-bsd.c ret = lockmgr (l, flags | LK_RELEASE, &vp->v_interlock, ap->a_p);
vp 864 xfs/xfs_vnodeops-bsd.c ret = debuglockmgr (l, flags | LK_RELEASE, &vp->v_interlock, ap->a_td,
vp 867 xfs/xfs_vnodeops-bsd.c ret = debuglockmgr (l, flags | LK_RELEASE, &vp->v_interlock, ap->a_p,
vp 880 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 881 xfs/xfs_vnodeops-bsd.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 885 xfs/xfs_vnodeops-bsd.c (unsigned long)vp));
vp 907 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 908 xfs/xfs_vnodeops-bsd.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 911 xfs/xfs_vnodeops-bsd.c (unsigned long)vp, xn->vnlocks));
vp 913 xfs/xfs_vnodeops-bsd.c while (vp->v_flag & VXLOCK) {
vp 914 xfs/xfs_vnodeops-bsd.c vp->v_flag |= VXWANT;
vp 915 xfs/xfs_vnodeops-bsd.c (void) tsleep((caddr_t)vp, PINOD, "xfs_vnlock", 0);
vp 917 xfs/xfs_vnodeops-bsd.c if (vp->v_tag == VT_NON)
vp 928 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 929 xfs/xfs_vnodeops-bsd.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 931 xfs/xfs_vnodeops-bsd.c (unsigned long)vp, xn->vnlocks));
vp 948 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 949 xfs/xfs_vnodeops-bsd.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 952 xfs/xfs_vnodeops-bsd.c (unsigned long)vp, xn->vnlocks));
vp 1096 xfs/xfs_vnodeops-bsd.c struct vnode *vp = ap->a_vp;
vp 1097 xfs/xfs_vnodeops-bsd.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 1114 xfs/xfs_vnodeops-bsd.c if (vp->v_type == VREG && ap->a_flags & PGO_CLEANIT) {
vp 1150 xfs/xfs_vnodeops-bsd.c xfs_realvp(struct vnode * vp,
vp 1160 xfs/xfs_vnodeops-bsd.c xfs_cntl(struct vnode * vp,
vp 1387 xfs/xfs_vnodeops-bsd.c xfs_pushdirty(struct vnode *vp, struct ucred *cred, d_thread_t *p)
vp 1390 xfs/xfs_vnodeops-bsd.c VOP_PUTPAGES(vp, 0, 0, PGO_ALLPAGES|PGO_SYNCIO|PGO_CLEANIT);
vp 1392 xfs/xfs_vnodeops-bsd.c ubc_pushdirty(vp);
vp 53 xfs/xfs_vnodeops-common.c struct vnode *vp = XNODE_TO_VNODE(xn);
vp 60 xfs/xfs_vnodeops-common.c if (UBCISVALID(vp) && !ubc_isinuse(vp, 1)) {
vp 62 xfs/xfs_vnodeops-common.c ubc_setsize(vp, 0);
vp 71 xfs/xfs_vnodeops-common.c xfs_open_valid(struct vnode *vp, struct ucred *cred, d_thread_t *p,
vp 74 xfs/xfs_vnodeops-common.c struct xfs *xfsp = NNPFS_FROM_VNODE(vp);
vp 75 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 108 xfs/xfs_vnodeops-common.c xfs_attr_valid(struct vnode *vp, struct ucred *cred, d_thread_t *p,
vp 111 xfs/xfs_vnodeops-common.c struct xfs *xfsp = NNPFS_FROM_VNODE(vp);
vp 112 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 137 xfs/xfs_vnodeops-common.c xfs_data_valid(struct vnode *vp, struct ucred *cred, d_thread_t *p,
vp 140 xfs/xfs_vnodeops-common.c struct xfs *xfsp = NNPFS_FROM_VNODE(vp);
vp 141 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 182 xfs/xfs_vnodeops-common.c xfs_open_common(struct vnode *vp,
vp 187 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 193 xfs/xfs_vnodeops-common.c ret = xfs_open_valid(vp, cred, p, NNPFS_OPEN_NW);
vp 195 xfs/xfs_vnodeops-common.c ret = xfs_open_valid(vp, cred, p, NNPFS_OPEN_NR);
vp 241 xfs/xfs_vnodeops-common.c xfs_fsync_common(struct vnode *vp, struct ucred *cred,
vp 244 xfs/xfs_vnodeops-common.c struct xfs *xfsp = NNPFS_FROM_VNODE(vp);
vp 245 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 248 xfs/xfs_vnodeops-common.c NNPFSDEB(XDEBVNOPS, ("xfs_fsync: %lx\n", (unsigned long)vp));
vp 260 xfs/xfs_vnodeops-common.c xfs_pushdirty(vp, cred, proc);
vp 278 xfs/xfs_vnodeops-common.c xfs_close_common(struct vnode *vp, int fflag,
vp 281 xfs/xfs_vnodeops-common.c struct xfs *xfsp = NNPFS_FROM_VNODE(vp);
vp 282 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 289 xfs/xfs_vnodeops-common.c if (vp->v_type == VREG)
vp 290 xfs/xfs_vnodeops-common.c xfs_pushdirty(vp, cred, proc);
vp 315 xfs/xfs_vnodeops-common.c xfs_read_common(struct vnode *vp, struct uio *uio, int ioflag,
vp 323 xfs/xfs_vnodeops-common.c xfs_update_read_cred(VNODE_TO_XNODE(vp), cred);
vp 326 xfs/xfs_vnodeops-common.c error = xfs_data_valid(vp, cred, xfs_uio_to_thread(uio), NNPFS_DATA_R,
vp 329 xfs/xfs_vnodeops-common.c error = xfs_data_valid(vp, cred, xfs_uio_to_proc(uio), NNPFS_DATA_R,
vp 340 xfs/xfs_vnodeops-common.c struct vnode *t = DATA_FROM_VNODE(vp);
vp 362 xfs/xfs_vnodeops-common.c xfs_write_common(struct vnode *vp, struct uio *uiop, int ioflag,
vp 365 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 373 xfs/xfs_vnodeops-common.c error = xfs_data_valid(vp, cred, xfs_uio_to_thread(uiop), NNPFS_DATA_W,
vp 374 xfs/xfs_vnodeops-common.c VNODE_TO_XNODE(vp)->attr.va_size);
vp 376 xfs/xfs_vnodeops-common.c error = xfs_data_valid(vp, cred, xfs_uio_to_proc(uiop), NNPFS_DATA_W,
vp 377 xfs/xfs_vnodeops-common.c VNODE_TO_XNODE(vp)->attr.va_size);
vp 388 xfs/xfs_vnodeops-common.c VNODE_TO_XNODE(vp)->flags |= NNPFS_DATA_DIRTY;
vp 393 xfs/xfs_vnodeops-common.c VNODE_TO_XNODE(vp)->flags |= NNPFS_DATA_DIRTY;
vp 401 xfs/xfs_vnodeops-common.c xfs_set_vp_size(vp, sub_attr.va_size);
vp 415 xfs/xfs_vnodeops-common.c xfs_getattr_common(struct vnode *vp, struct vattr *vap,
vp 420 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 424 xfs/xfs_vnodeops-common.c error = xfs_attr_valid(vp, cred, p, NNPFS_ATTR_R);
vp 431 xfs/xfs_vnodeops-common.c xfs_setattr_common(struct vnode *vp, struct vattr *vap,
vp 434 xfs/xfs_vnodeops-common.c struct xfs *xfsp = NNPFS_FROM_VNODE(vp);
vp 435 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 454 xfs/xfs_vnodeops-common.c VNODE_TO_XNODE(vp)->flags |= NNPFS_ATTR_DIRTY;
vp 471 xfs/xfs_vnodeops-common.c if (vp->v_type == VREG) {
vp 478 xfs/xfs_vnodeops-common.c if (UBCINFOEXISTS(vp))
vp 479 xfs/xfs_vnodeops-common.c ubc_setsize(vp, msg.attr.xa_size);
vp 516 xfs/xfs_vnodeops-common.c xfs_access_common(struct vnode *vp, int mode, struct ucred *cred,
vp 524 xfs/xfs_vnodeops-common.c error = xfs_attr_valid(vp, cred, p, NNPFS_ATTR_R);
vp 526 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 691 xfs/xfs_vnodeops-common.c struct vnode *vp,
vp 716 xfs/xfs_vnodeops-common.c xfs_dnlc_purge (vp);
vp 801 xfs/xfs_vnodeops-common.c struct vnode *vp,
vp 825 xfs/xfs_vnodeops-common.c xfs_dnlc_purge (vp);
vp 833 xfs/xfs_vnodeops-common.c xfs_readdir_common(struct vnode *vp,
vp 846 xfs/xfs_vnodeops-common.c error = xfs_data_valid(vp, cred, xfs_uio_to_thread(uiop), NNPFS_DATA_R,
vp 849 xfs/xfs_vnodeops-common.c error = xfs_data_valid(vp, cred, xfs_uio_to_proc(uiop), NNPFS_DATA_R,
vp 853 xfs/xfs_vnodeops-common.c struct vnode *t = DATA_FROM_VNODE(vp);
vp 884 xfs/xfs_vnodeops-common.c struct vnode *vp,
vp 891 xfs/xfs_vnodeops-common.c struct xfs_node *xn2 = VNODE_TO_XNODE(vp);
vp 964 xfs/xfs_vnodeops-common.c xfs_readlink_common(struct vnode *vp, struct uio *uiop, struct ucred *cred)
vp 971 xfs/xfs_vnodeops-common.c error = xfs_data_valid(vp, cred, xfs_uio_to_thread(uiop), NNPFS_DATA_R,
vp 974 xfs/xfs_vnodeops-common.c error = xfs_data_valid(vp, cred, xfs_uio_to_proc(uiop), NNPFS_DATA_R,
vp 978 xfs/xfs_vnodeops-common.c struct vnode *t = DATA_FROM_VNODE(vp);
vp 994 xfs/xfs_vnodeops-common.c xfs_inactive_common(struct vnode *vp, d_thread_t *p)
vp 997 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 1000 xfs/xfs_vnodeops-common.c (unsigned long)vp));
vp 1014 xfs/xfs_vnodeops-common.c if (vp->v_type == VREG)
vp 1015 xfs/xfs_vnodeops-common.c xfs_pushdirty(vp, xn->wr_cred, p);
vp 1017 xfs/xfs_vnodeops-common.c error = xfs_fsync_common(vp, xn->wr_cred, /* XXX */ 0, p);
vp 1028 xfs/xfs_vnodeops-common.c xfs_vfs_unlock(vp, p);
vp 1030 xfs/xfs_vnodeops-common.c vrecycle(vp, p);
vp 1033 xfs/xfs_vnodeops-common.c (unsigned long)vp, vp?vp->v_usecount:0));
vp 1037 xfs/xfs_vnodeops-common.c xfs_vfs_unlock(vp, p);
vp 1048 xfs/xfs_vnodeops-common.c xfs_reclaim_common(struct vnode *vp)
vp 1051 xfs/xfs_vnodeops-common.c struct xfs *xfsp = NNPFS_FROM_VNODE(vp);
vp 1052 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 1055 xfs/xfs_vnodeops-common.c (unsigned long)vp));
vp 1074 xfs/xfs_vnodeops-common.c xfs_dnlc_purge(vp);
vp 1135 xfs/xfs_vnodeops-common.c xfs_printnode_common (struct vnode *vp)
vp 1137 xfs/xfs_vnodeops-common.c struct xfs_node *xn = VNODE_TO_XNODE(vp);
vp 47 xfs/xfs_vnodeops.h #define xfs_vfs_readlock(vp, proc) VREF((vp))
vp 48 xfs/xfs_vnodeops.h #define xfs_vfs_writelock(vp, proc) VREF((vp))
vp 49 xfs/xfs_vnodeops.h #define xfs_vfs_unlock(vp, proc) vrele((vp))
vp 53 xfs/xfs_vnodeops.h #define xfs_vfs_vn_lock(vp, flags, proc) (0)
vp 57 xfs/xfs_vnodeops.h #define xfs_vfs_readlock(vp, proc) vn_lock((vp), LK_SHARED | LK_RETRY)
vp 58 xfs/xfs_vnodeops.h #define xfs_vfs_writelock(vp, proc) vn_lock((vp), LK_EXCLUSIVE | LK_RETRY)
vp 59 xfs/xfs_vnodeops.h #define xfs_vfs_unlock(vp, proc) VOP_UNLOCK((vp), 0)
vp 60 xfs/xfs_vnodeops.h #define xfs_vfs_vn_lock(vp, flags, proc) vn_lock((vp), (flags))
vp 64 xfs/xfs_vnodeops.h #define xfs_vfs_readlock(vp, proc) vn_lock((vp), LK_SHARED | LK_RETRY, (proc))
vp 65 xfs/xfs_vnodeops.h #define xfs_vfs_writelock(vp, proc) vn_lock((vp), LK_EXCLUSIVE | LK_RETRY, (proc))
vp 66 xfs/xfs_vnodeops.h #define xfs_vfs_unlock(vp, proc) VOP_UNLOCK((vp), 0, (proc))
vp 67 xfs/xfs_vnodeops.h #define xfs_vfs_vn_lock(vp, flags, proc) vn_lock((vp), (flags), (proc))
vp 71 xfs/xfs_vnodeops.h #define xfs_vfs_readlock(vp, proc) VOP_LOCK((vp))
vp 72 xfs/xfs_vnodeops.h #define xfs_vfs_writelock(vp, proc) VOP_LOCK((vp))
vp 73 xfs/xfs_vnodeops.h #define xfs_vfs_unlock(vp, proc) VOP_UNLOCK((vp))
vp 77 xfs/xfs_vnodeops.h #define xfs_vfs_vn_lock(vp, flags, proc) (0)
vp 86 xfs/xfs_vnodeops.h xfs_open_valid(struct vnode *vp, struct ucred *cred, d_thread_t *p,
vp 90 xfs/xfs_vnodeops.h xfs_attr_valid(struct vnode *vp, struct ucred *cred, d_thread_t *p,
vp 94 xfs/xfs_vnodeops.h xfs_fetch_rights(struct vnode *vp, struct ucred *cred, d_thread_t *p);
vp 97 xfs/xfs_vnodeops.h xfs_data_valid(struct vnode *vp, struct ucred *cred, d_thread_t *p,
vp 101 xfs/xfs_vnodeops.h xfs_open_common(struct vnode *vp,
vp 107 xfs/xfs_vnodeops.h xfs_fsync_common(struct vnode *vp, struct ucred *cred,
vp 111 xfs/xfs_vnodeops.h xfs_close_common(struct vnode *vp, int fflag,
vp 115 xfs/xfs_vnodeops.h xfs_read_common(struct vnode *vp, struct uio *uio, int ioflag,
vp 119 xfs/xfs_vnodeops.h xfs_write_common(struct vnode *vp, struct uio *uiop, int ioflag,
vp 123 xfs/xfs_vnodeops.h xfs_getattr_common(struct vnode *vp, struct vattr *vap,
vp 127 xfs/xfs_vnodeops.h xfs_setattr_common(struct vnode *vp, struct vattr *vap,
vp 131 xfs/xfs_vnodeops.h xfs_access_common(struct vnode *vp, int mode,
vp 148 xfs/xfs_vnodeops.h struct vnode *vp,
vp 172 xfs/xfs_vnodeops.h struct vnode *vp,
vp 178 xfs/xfs_vnodeops.h xfs_readdir_common(struct vnode *vp,
vp 186 xfs/xfs_vnodeops.h struct vnode *vp,
vp 199 xfs/xfs_vnodeops.h xfs_readlink_common(struct vnode *vp, struct uio *uiop, struct ucred *cred);
vp 202 xfs/xfs_vnodeops.h xfs_inactive_common(struct vnode *vp, d_thread_t *p);
vp 205 xfs/xfs_vnodeops.h xfs_reclaim_common(struct vnode *vp);
vp 214 xfs/xfs_vnodeops.h xfs_printnode_common (struct vnode *vp);