sz 385 arch/i386/i386/machdep.c int sz; sz 423 arch/i386/i386/machdep.c sz = (int)allocsys((caddr_t)0); sz 424 arch/i386/i386/machdep.c if ((v = (caddr_t)uvm_km_zalloc(kernel_map, round_page(sz))) == 0) sz 426 arch/i386/i386/machdep.c if (allocsys(v) - v != sz) sz 701 arch/i386/include/bus.h #define i386_space_copy(a1, a2, sz, cnt) do { \ sz 703 arch/i386/include/bus.h a1 += ((cnt) - 1) * (sz); a2 += ((cnt) - 1) * (sz); \ sz 704 arch/i386/include/bus.h switch (sz) { \ sz 710 arch/i386/include/bus.h switch (sz) { \ sz 71 compat/common/compat_util.c size_t sz, len; sz 79 compat/common/compat_util.c sz = MAXPATHLEN - (ptr - buf); sz 85 compat/common/compat_util.c error = copystr(path, ptr, sz, &len); sz 87 compat/common/compat_util.c error = copyinstr(path, ptr, sz, &len); sz 153 compat/common/compat_util.c sz = &ptr[len] - buf; sz 154 compat/common/compat_util.c *pbuf = stackgap_alloc(sgp, sz + 1); sz 159 compat/common/compat_util.c if ((error = copyout(buf, *pbuf, sz)) != 0) { sz 212 compat/common/compat_util.c stackgap_alloc(sgp, sz) sz 214 compat/common/compat_util.c size_t sz; sz 219 compat/common/compat_util.c sz = ALIGN(sz); sz 220 compat/common/compat_util.c nsgp = *sgp + sz; sz 65 compat/common/kern_info_09.c size_t sz; sz 68 compat/common/kern_info_09.c sz = SCARG(uap,len); sz 69 compat/common/kern_info_09.c return (kern_sysctl(&name, 1, SCARG(uap, domainname), &sz, 0, 0, p)); sz 94 compat/common/kern_info_43.c size_t sz; sz 97 compat/common/kern_info_43.c sz = SCARG(uap, len); sz 98 compat/common/kern_info_43.c return (kern_sysctl(&name, 1, SCARG(uap, hostname), &sz, 0, 0, p)); sz 186 compat/hpux/hpux_compat.c size_t sz = sizeof(*SCARG(&w4, status)); sz 194 compat/hpux/hpux_compat.c SCARG(&w4, status) = stackgap_alloc(&sg, sz); sz 118 compat/svr4/svr4_misc.c size_t sz = sizeof(*SCARG(&w4, status)); sz 125 compat/svr4/svr4_misc.c SCARG(&w4, status) = stackgap_alloc(&sg, sz); sz 140 compat/svr4/svr4_misc.c return copyin(SCARG(&w4, status), &retval[1], sz); sz 98 dev/acpi/dsdt.c #define acpi_os_malloc(sz) _acpi_os_malloc(sz, __FUNCTION__, __LINE__) sz 662 dev/ic/gem.c gem_ringsize(int sz) sz 664 dev/ic/gem.c switch (sz) { sz 684 dev/ic/gem.c printf("gem: invalid Receive Descriptor ring size %d\n", sz); sz 299 dev/ic/hmereg.h #define HME_XD_ENCODE_RSIZE(sz) \ sz 300 dev/ic/hmereg.h (((sz) << HME_XD_RXLENSHIFT) & HME_XD_RXLENMSK) sz 305 dev/ic/hmereg.h #define HME_XD_ENCODE_TSIZE(sz) \ sz 306 dev/ic/hmereg.h (((sz) << 0) & HME_XD_TXLENMSK) sz 107 dev/ic/if_wi.c #define BPFATTACH(if_bpf,if,dlt,sz) sz 118 dev/ic/if_wireg.h #define CSR_READ_RAW_2(sc, ba, dst, sz) \ sz 121 dev/ic/if_wireg.h (sc->sc_pci? ba * 2: ba), (dst), (sz)) sz 122 dev/ic/if_wireg.h #define CSR_WRITE_RAW_2(sc, ba, dst, sz) \ sz 125 dev/ic/if_wireg.h (sc->sc_pci? ba * 2: ba), (dst), (sz)) sz 520 dev/ic/midway.c STATIC INLINE int en_sz2b(sz) sz 522 dev/ic/midway.c int sz; sz 525 dev/ic/midway.c switch (sz) { sz 548 dev/ic/midway.c int result, needalign, sz; sz 568 dev/ic/midway.c sz = min(len, sizeof(u_int32_t) - needalign); sz 569 dev/ic/midway.c len -= sz; sz 570 dev/ic/midway.c data += sz; sz 578 dev/ic/midway.c sz = min(len, sc->bestburstlen - needalign); sz 579 dev/ic/midway.c len -= sz; sz 584 dev/ic/midway.c sz = len / sc->bestburstlen; sz 585 dev/ic/midway.c sz = sz * sc->bestburstlen; sz 586 dev/ic/midway.c len -= sz; sz 668 dev/ic/midway.c int sz; sz 762 dev/ic/midway.c sz = sc->en_obmemsz - (MID_BUFOFF - MID_RAMOFF); sz 765 dev/ic/midway.c sz = sz - (ptr - sav); sz 766 dev/ic/midway.c if (EN_TXSZ*1024 * EN_NTX > sz) { sz 774 dev/ic/midway.c sz -= (EN_TXSZ * 1024); sz 787 dev/ic/midway.c sz = sz - (ptr - sav); sz 788 dev/ic/midway.c sc->en_nrx = sz / (EN_RXSZ * 1024); sz 808 dev/ic/midway.c sz -= (EN_RXSZ * 1024); sz 92 dev/ic/mk48txx.c int sz; sz 110 dev/ic/mk48txx.c sz = ALIGN(sizeof(struct todr_chip_handle)) + sizeof(struct mk48txx); sz 111 dev/ic/mk48txx.c handle = malloc(sz, M_DEVBUF, M_NOWAIT); sz 347 dev/isa/fd.c int sz; sz 365 dev/isa/fd.c sz = howmany(bp->b_bcount, DEV_BSIZE); sz 367 dev/isa/fd.c if (bp->b_blkno + sz > fd->sc_type->size * bf) { sz 368 dev/isa/fd.c sz = fd->sc_type->size * bf - bp->b_blkno; sz 369 dev/isa/fd.c if (sz == 0) sz 372 dev/isa/fd.c if (sz < 0) { sz 378 dev/isa/fd.c bp->b_bcount = sz << DEV_BSHIFT; sz 385 dev/isa/fd.c bp->b_blkno, bp->b_bcount, fd->sc_blkno, bp->b_cylinder, sz); sz 122 dev/isa/isadma.c int i, sz; sz 129 dev/isa/isadma.c sz = (i & 4) ? 1 << 17 : 1 << 16; sz 130 dev/isa/isadma.c if ((bus_dmamap_create(sc->sc_dmat, sz, 1, sz, sz, sz 99 dev/isa/mcd.c int sz; sz 582 dev/isa/mcd.c sc->mbx.sz = sc->blksize; sz 1234 dev/isa/mcd.c bp->b_data + mbx->skip, mbx->sz); sz 1237 dev/isa/mcd.c mbx->skip += mbx->sz; sz 56 dev/pci/emuxkireg.h #define EMU_MKSUBREG(sz, idx, reg) (((sz) << 24) | ((idx) << 16) | (reg)) sz 886 dev/pci/if_cas.c cas_ringsize(int sz) sz 888 dev/pci/if_cas.c switch (sz) { sz 908 dev/pci/if_cas.c printf("cas: invalid Receive Descriptor ring size %d\n", sz); sz 914 dev/pci/if_cas.c cas_cringsize(int sz) sz 919 dev/pci/if_cas.c if (sz == (128 << i)) sz 922 dev/pci/if_cas.c printf("cas: invalid completion ring size %d\n", sz); sz 394 dev/pci/neo.c int ofs, sz, i; sz 411 dev/pci/neo.c sz = nf->coefficientSizes[num]; sz 415 dev/pci/neo.c for (i = 0; i < sz; i++) sz 419 dev/pci/neo.c sz--; sz 420 dev/pci/neo.c nm_wr(sc, addr + 4, sc->cbuf + sz, 4); sz 88 dev/raidframe/rf_debugMem.c rf_unrecord_malloc(void *p, int sz) sz 93 dev/raidframe/rf_debugMem.c size = rf_memory_hash_remove(p, sz); sz 180 dev/raidframe/rf_debugMem.c rf_memory_hash_remove(void *addr, int sz) sz 198 dev/raidframe/rf_debugMem.c if (sz > 0 && p->size != sz) { sz 205 dev/raidframe/rf_debugMem.c (unsigned long) addr, sz, p->size, p->line, p->filen); sz 548 dev/sbus/cgsix.c bus_size_t sz; sz 573 dev/sbus/cgsix.c sz = mo->mo_size ? mo->mo_size : sc->sc_sunfb.sf_fbsize; sz 574 dev/sbus/cgsix.c if (u < sz) { sz 107 dev/sbus/sbusvar.h #define sbus_bus_map(t, slot, offset, sz, flags, unused, hp) \ sz 108 dev/sbus/sbusvar.h bus_space_map(t, BUS_ADDR(slot, offset), sz, flags, hp) sz 1109 dev/softraid.c size_t sz = SR_META_SIZE * 512; sz 1115 dev/softraid.c m = malloc(sz , M_DEVBUF, M_WAITOK); sz 1116 dev/softraid.c bzero(m, sz); sz 1123 dev/softraid.c b.b_bcount = sz; sz 1124 dev/softraid.c b.b_bufsize = sz; sz 1125 dev/softraid.c b.b_resid = sz; sz 1159 dev/softraid.c bcopy(m, sm, sz); sz 1791 dev/softraid.c size_t sz = SR_META_SIZE * 512; sz 1797 dev/softraid.c m = malloc(sz , M_DEVBUF, M_WAITOK); sz 1798 dev/softraid.c bzero(m, sz); sz 1805 dev/softraid.c b.b_bcount = sz; sz 1806 dev/softraid.c b.b_bufsize = sz; sz 1807 dev/softraid.c b.b_resid = sz; sz 1874 dev/softraid.c size_t sz = SR_META_SIZE * 512; sz 1889 dev/softraid.c sz) { sz 1964 dev/softraid.c b.b_bcount = sz; sz 1965 dev/softraid.c b.b_bufsize = sz; sz 1966 dev/softraid.c b.b_resid = sz; sz 2019 dev/softraid.c size_t sz = SR_META_SIZE * 512; sz 2024 dev/softraid.c bp = geteblk(sz); sz 2095 dev/softraid.c bp->b_bcount = sz; sz 2096 dev/softraid.c bp->b_bufsize = sz; sz 2097 dev/softraid.c bp->b_resid = sz; sz 2113 dev/softraid.c mle->sml_metadata = malloc(sz, M_DEVBUF, sz 2115 dev/softraid.c bzero(mle->sml_metadata, sz); sz 2116 dev/softraid.c bcopy(sm, mle->sml_metadata, sz); sz 1094 dev/usb/usbdi.c usb_match_device(const struct usb_devno *tbl, u_int nentries, u_int sz, sz 1102 dev/usb/usbdi.c tbl = (const struct usb_devno *)((const char *)tbl + sz); sz 205 dev/usb/usbdi.h u_int nentries, u_int sz, u_int16_t vendor, u_int16_t product); sz 388 dev/vnd.c int sz, flags, error, s; sz 431 dev/vnd.c sz = howmany(bp->b_bcount, DEV_BSIZE); sz 502 dev/vnd.c for (resid = bp->b_resid; resid; resid -= sz) { sz 519 dev/vnd.c sz = bsize - off; sz 521 dev/vnd.c sz = (1 + nra) * bsize; sz 522 dev/vnd.c if (resid < sz) sz 523 dev/vnd.c sz = resid; sz 526 dev/vnd.c vnd->sc_vp, vp, bn, nbn, sz); sz 532 dev/vnd.c nbp->vb_buf.b_bcount = sz; sz 567 dev/vnd.c bp->b_resid -= (resid - sz); sz 585 dev/vnd.c bn += sz; sz 586 dev/vnd.c addr += sz; sz 770 kern/kern_exec.c vsize_t sz; sz 772 kern/kern_exec.c sz = (vaddr_t)e->e_esigcode - (vaddr_t)e->e_sigcode; sz 789 kern/kern_exec.c e->e_sigobject = uao_create(sz, 0); sz 793 kern/kern_exec.c if ((r = uvm_map(kernel_map, &va, round_page(sz), e->e_sigobject, sz 799 kern/kern_exec.c memcpy((void *)va, e->e_sigcode, sz); sz 800 kern/kern_exec.c uvm_unmap(kernel_map, va, va + round_page(sz)); sz 806 kern/kern_exec.c if (uvm_map(&p->p_vmspace->vm_map, &p->p_sigcode, round_page(sz), sz 635 kern/kern_malloc.c malloc_roundup(size_t sz) sz 637 kern/kern_malloc.c if (sz > MAXALLOCSAVE) sz 638 kern/kern_malloc.c return round_page(sz); sz 640 kern/kern_malloc.c return (1 << BUCKETINDX(sz)); sz 607 kern/subr_disk.c daddr64_t sz = howmany(bp->b_bcount, DEV_BSIZE); sz 614 kern/subr_disk.c if (bp->b_blkno + sz > blockpersec(DL_GETPSIZE(p), lp)) { sz 615 kern/subr_disk.c sz = blockpersec(DL_GETPSIZE(p), lp) - bp->b_blkno; sz 616 kern/subr_disk.c if (sz == 0) { sz 621 kern/subr_disk.c if (sz < 0) sz 626 kern/subr_disk.c bp->b_bcount = sz << DEV_BSHIFT; sz 166 kern/subr_extent.c size_t sz = storagesize; sz 200 kern/subr_extent.c sz -= ALIGN(sizeof(struct extent_fixed)); sz 209 kern/subr_extent.c while (sz >= ALIGN(sizeof(struct extent_region))) { sz 212 kern/subr_extent.c sz -= ALIGN(sizeof(struct extent_region)); sz 868 kern/sys_generic.c size_t sz; sz 880 kern/sys_generic.c sz = sizeof(struct pollfd) * nfds; sz 883 kern/sys_generic.c if (sz > sizeof(pfds)) sz 884 kern/sys_generic.c pl = (struct pollfd *) malloc(sz, M_TEMP, M_WAITOK); sz 886 kern/sys_generic.c if ((error = copyin(SCARG(uap, fds), pl, sz)) != 0) sz 942 kern/sys_generic.c error = copyout(pl, SCARG(uap, fds), sz); sz 948 kern/sys_generic.c error = copyout(pl, SCARG(uap, fds), sz); sz 55 lib/libsa/exec.c ssize_t sz; sz 69 lib/libsa/exec.c sz = read(io, (char *)&x, sizeof(x)); sz 70 lib/libsa/exec.c if (sz != sizeof(x) || N_BADMAG(x)) { sz 85 lib/libsa/exec.c sz = x.a_text; sz 89 lib/libsa/exec.c sz -= sizeof x; sz 91 lib/libsa/exec.c if (read(io, (char *)addr, sz) != sz) sz 93 lib/libsa/exec.c addr += sz; sz 132 lib/libsa/exec.c sz = i - sizeof(int); sz 134 lib/libsa/exec.c if (read(io, addr, sz) != sz) sz 136 lib/libsa/exec.c addr += sz; sz 140 lib/libsa/exec.c printf("+%d]", sz); sz 84 lib/libsa/loadfile_elf.c size_t sz; sz 90 lib/libsa/loadfile_elf.c sz = elf->e_phnum * sizeof(Elf_Phdr); sz 91 lib/libsa/loadfile_elf.c phdr = ALLOC(sz); sz 95 lib/libsa/loadfile_elf.c FREE(phdr, sz); sz 98 lib/libsa/loadfile_elf.c if (read(fd, phdr, sz) != sz) { sz 100 lib/libsa/loadfile_elf.c FREE(phdr, sz); sz 125 lib/libsa/loadfile_elf.c FREE(phdr, sz); sz 131 lib/libsa/loadfile_elf.c FREE(phdr, sz); sz 160 lib/libsa/loadfile_elf.c FREE(phdr, sz); sz 174 lib/libsa/loadfile_elf.c sz = elf->e_shnum * sizeof(Elf_Shdr); sz 175 lib/libsa/loadfile_elf.c shp = ALLOC(sz); sz 177 lib/libsa/loadfile_elf.c if (read(fd, shp, sz) != sz) { sz 179 lib/libsa/loadfile_elf.c FREE(shp, sz); sz 184 lib/libsa/loadfile_elf.c maxp += roundup(sz, sizeof(long)); sz 191 lib/libsa/loadfile_elf.c off = roundup((sizeof(Elf_Ehdr) + sz), sizeof(long)); sz 206 lib/libsa/loadfile_elf.c FREE(shp, sz); sz 212 lib/libsa/loadfile_elf.c FREE(shp, sz); sz 224 lib/libsa/loadfile_elf.c BCOPY(shp, shpp, sz); sz 229 lib/libsa/loadfile_elf.c FREE(shp, sz); sz 138 net/if_atmsubr.c int s, error = 0, sz, len; sz 215 net/if_atmsubr.c sz = sizeof(atmdst); sz 217 net/if_atmsubr.c if (atm_flags & ATM_PH_LLCSNAP) sz += 8; /* sizeof snap == 8 */ sz 218 net/if_atmsubr.c M_PREPEND(m, sz, M_DONTWAIT); sz 653 ntfs/ntfs_subr.c u_int32_t sz, i; sz 675 ntfs/ntfs_subr.c sz = run[off++]; sz 678 ntfs/ntfs_subr.c for (i = 0; i < (sz & 0xF); i++) sz 681 ntfs/ntfs_subr.c sz >>= 4; sz 682 ntfs/ntfs_subr.c if (run[off + sz - 1] & 0x80) { sz 683 ntfs/ntfs_subr.c tmp = ((u_int64_t) - 1) << (sz << 3); sz 684 ntfs/ntfs_subr.c for (i = 0; i < sz; i++) sz 688 ntfs/ntfs_subr.c for (i = 0; i < sz; i++) sz 1445 ntfs/ntfs_subr.c u_int64_t sz, bn; sz 1456 ntfs/ntfs_subr.c sz = vap->va_datalen; sz 1459 ntfs/ntfs_subr.c (u_int32_t) sz, (u_int32_t) bn)); sz 1462 ntfs/ntfs_subr.c *size = sz; sz 1934 ntfs/ntfs_subr.c u_int8_t sz; sz 1941 ntfs/ntfs_subr.c sz = run[(*off)++]; sz 1942 ntfs/ntfs_subr.c if (0 == sz) { sz 1947 ntfs/ntfs_subr.c if ((sz & 0xF) > 8 || (*off) + (sz & 0xF) > len) { sz 1950 ntfs/ntfs_subr.c sz, len, *off); sz 1953 ntfs/ntfs_subr.c for (i = 0; i < (sz & 0xF); i++) sz 1956 ntfs/ntfs_subr.c sz >>= 4; sz 1957 ntfs/ntfs_subr.c if ((sz & 0xF) > 8 || (*off) + (sz & 0xF) > len) { sz 1960 ntfs/ntfs_subr.c sz, len, *off); sz 1963 ntfs/ntfs_subr.c for (i = 0; i < (sz & 0xF); i++) sz 601 ntfs/ntfs_vnops.c int sz; sz 623 ntfs/ntfs_vnops.c sz = (*ntmp->ntm_wput)(fname, remains, sz 625 ntfs/ntfs_vnops.c fname += sz; sz 626 ntfs/ntfs_vnops.c remains -= sz; sz 238 uvm/uvm_amap.c amap_alloc(vaddr_t sz, vaddr_t padsz, int waitf) sz 244 uvm/uvm_amap.c AMAP_B2SLOT(slots, sz); /* load slots */ sz 254 uvm/uvm_amap.c UVMHIST_LOG(maphist,"<- done, amap = %p, sz=%lu", amap, sz, 0, 0); sz 66 uvm/uvm_io.c vsize_t chunksz, togo, sz; sz 129 uvm/uvm_io.c sz = chunksz - pageoffset; sz 130 uvm/uvm_io.c if (sz > togo) sz 131 uvm/uvm_io.c sz = togo; sz 132 uvm/uvm_io.c error = uiomove((caddr_t) (kva + pageoffset), sz, uio); sz 133 uvm/uvm_io.c togo -= sz; sz 1291 uvm/uvm_swap.c int s, off, nra, error, sz, resid; sz 1314 uvm/uvm_swap.c for (resid = bp->b_resid; resid; resid -= sz) { sz 1360 uvm/uvm_swap.c sz = (1 + nra) * sdp->swd_bsize - off; sz 1361 uvm/uvm_swap.c if (sz > resid) sz 1362 uvm/uvm_swap.c sz = resid; sz 1375 uvm/uvm_swap.c nbp->vb_buf.b_bcount = sz; sz 1376 uvm/uvm_swap.c nbp->vb_buf.b_bufsize = sz; sz 1393 uvm/uvm_swap.c nbp->vb_buf.b_dirtyend = sz; sz 1398 uvm/uvm_swap.c min(sz, sz 1403 uvm/uvm_swap.c nbp->vb_buf.b_validend = sz; sz 1408 uvm/uvm_swap.c min(sz, sz 1436 uvm/uvm_swap.c byteoff += sz; sz 1437 uvm/uvm_swap.c addr += sz; sz 94 xfs/xfs_common-bsd.c xfs_devtoname_r (dev_t dev, char *buf, size_t sz) sz 102 xfs/xfs_common-bsd.c xfs_devtoname_r (dev_t dev, char *buf, size_t sz) sz 105 xfs/xfs_common-bsd.c snprintf (buf, sz, "%u/%u", major(dev), minor(dev)); sz 72 xfs/xfs_common.h xfs_devtoname_r (dev_t dev, char *buf, size_t sz); sz 376 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) uvm_vnp_setsize(vp, sz) sz 378 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) vnode_pager_setsize(vp, sz) sz 380 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) ubc_setsize(vp, sz) sz 382 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) sz 302 xfs/xfs_vnodeops-common.c size_t sz = 0; sz 306 xfs/xfs_vnodeops-common.c sz += uio->uio_iov[i].iov_len; sz 307 xfs/xfs_vnodeops-common.c if (sz != uio->uio_resid)