sz 385 arch/i386/i386/machdep.c int sz;
sz 423 arch/i386/i386/machdep.c sz = (int)allocsys((caddr_t)0);
sz 424 arch/i386/i386/machdep.c if ((v = (caddr_t)uvm_km_zalloc(kernel_map, round_page(sz))) == 0)
sz 426 arch/i386/i386/machdep.c if (allocsys(v) - v != sz)
sz 701 arch/i386/include/bus.h #define i386_space_copy(a1, a2, sz, cnt) do { \
sz 703 arch/i386/include/bus.h a1 += ((cnt) - 1) * (sz); a2 += ((cnt) - 1) * (sz); \
sz 704 arch/i386/include/bus.h switch (sz) { \
sz 710 arch/i386/include/bus.h switch (sz) { \
sz 71 compat/common/compat_util.c size_t sz, len;
sz 79 compat/common/compat_util.c sz = MAXPATHLEN - (ptr - buf);
sz 85 compat/common/compat_util.c error = copystr(path, ptr, sz, &len);
sz 87 compat/common/compat_util.c error = copyinstr(path, ptr, sz, &len);
sz 153 compat/common/compat_util.c sz = &ptr[len] - buf;
sz 154 compat/common/compat_util.c *pbuf = stackgap_alloc(sgp, sz + 1);
sz 159 compat/common/compat_util.c if ((error = copyout(buf, *pbuf, sz)) != 0) {
sz 212 compat/common/compat_util.c stackgap_alloc(sgp, sz)
sz 214 compat/common/compat_util.c size_t sz;
sz 219 compat/common/compat_util.c sz = ALIGN(sz);
sz 220 compat/common/compat_util.c nsgp = *sgp + sz;
sz 65 compat/common/kern_info_09.c size_t sz;
sz 68 compat/common/kern_info_09.c sz = SCARG(uap,len);
sz 69 compat/common/kern_info_09.c return (kern_sysctl(&name, 1, SCARG(uap, domainname), &sz, 0, 0, p));
sz 94 compat/common/kern_info_43.c size_t sz;
sz 97 compat/common/kern_info_43.c sz = SCARG(uap, len);
sz 98 compat/common/kern_info_43.c return (kern_sysctl(&name, 1, SCARG(uap, hostname), &sz, 0, 0, p));
sz 186 compat/hpux/hpux_compat.c size_t sz = sizeof(*SCARG(&w4, status));
sz 194 compat/hpux/hpux_compat.c SCARG(&w4, status) = stackgap_alloc(&sg, sz);
sz 118 compat/svr4/svr4_misc.c size_t sz = sizeof(*SCARG(&w4, status));
sz 125 compat/svr4/svr4_misc.c SCARG(&w4, status) = stackgap_alloc(&sg, sz);
sz 140 compat/svr4/svr4_misc.c return copyin(SCARG(&w4, status), &retval[1], sz);
sz 98 dev/acpi/dsdt.c #define acpi_os_malloc(sz) _acpi_os_malloc(sz, __FUNCTION__, __LINE__)
sz 662 dev/ic/gem.c gem_ringsize(int sz)
sz 664 dev/ic/gem.c switch (sz) {
sz 684 dev/ic/gem.c printf("gem: invalid Receive Descriptor ring size %d\n", sz);
sz 299 dev/ic/hmereg.h #define HME_XD_ENCODE_RSIZE(sz) \
sz 300 dev/ic/hmereg.h (((sz) << HME_XD_RXLENSHIFT) & HME_XD_RXLENMSK)
sz 305 dev/ic/hmereg.h #define HME_XD_ENCODE_TSIZE(sz) \
sz 306 dev/ic/hmereg.h (((sz) << 0) & HME_XD_TXLENMSK)
sz 107 dev/ic/if_wi.c #define BPFATTACH(if_bpf,if,dlt,sz)
sz 118 dev/ic/if_wireg.h #define CSR_READ_RAW_2(sc, ba, dst, sz) \
sz 121 dev/ic/if_wireg.h (sc->sc_pci? ba * 2: ba), (dst), (sz))
sz 122 dev/ic/if_wireg.h #define CSR_WRITE_RAW_2(sc, ba, dst, sz) \
sz 125 dev/ic/if_wireg.h (sc->sc_pci? ba * 2: ba), (dst), (sz))
sz 520 dev/ic/midway.c STATIC INLINE int en_sz2b(sz)
sz 522 dev/ic/midway.c int sz;
sz 525 dev/ic/midway.c switch (sz) {
sz 548 dev/ic/midway.c int result, needalign, sz;
sz 568 dev/ic/midway.c sz = min(len, sizeof(u_int32_t) - needalign);
sz 569 dev/ic/midway.c len -= sz;
sz 570 dev/ic/midway.c data += sz;
sz 578 dev/ic/midway.c sz = min(len, sc->bestburstlen - needalign);
sz 579 dev/ic/midway.c len -= sz;
sz 584 dev/ic/midway.c sz = len / sc->bestburstlen;
sz 585 dev/ic/midway.c sz = sz * sc->bestburstlen;
sz 586 dev/ic/midway.c len -= sz;
sz 668 dev/ic/midway.c int sz;
sz 762 dev/ic/midway.c sz = sc->en_obmemsz - (MID_BUFOFF - MID_RAMOFF);
sz 765 dev/ic/midway.c sz = sz - (ptr - sav);
sz 766 dev/ic/midway.c if (EN_TXSZ*1024 * EN_NTX > sz) {
sz 774 dev/ic/midway.c sz -= (EN_TXSZ * 1024);
sz 787 dev/ic/midway.c sz = sz - (ptr - sav);
sz 788 dev/ic/midway.c sc->en_nrx = sz / (EN_RXSZ * 1024);
sz 808 dev/ic/midway.c sz -= (EN_RXSZ * 1024);
sz 92 dev/ic/mk48txx.c int sz;
sz 110 dev/ic/mk48txx.c sz = ALIGN(sizeof(struct todr_chip_handle)) + sizeof(struct mk48txx);
sz 111 dev/ic/mk48txx.c handle = malloc(sz, M_DEVBUF, M_NOWAIT);
sz 347 dev/isa/fd.c int sz;
sz 365 dev/isa/fd.c sz = howmany(bp->b_bcount, DEV_BSIZE);
sz 367 dev/isa/fd.c if (bp->b_blkno + sz > fd->sc_type->size * bf) {
sz 368 dev/isa/fd.c sz = fd->sc_type->size * bf - bp->b_blkno;
sz 369 dev/isa/fd.c if (sz == 0)
sz 372 dev/isa/fd.c if (sz < 0) {
sz 378 dev/isa/fd.c bp->b_bcount = sz << DEV_BSHIFT;
sz 385 dev/isa/fd.c bp->b_blkno, bp->b_bcount, fd->sc_blkno, bp->b_cylinder, sz);
sz 122 dev/isa/isadma.c int i, sz;
sz 129 dev/isa/isadma.c sz = (i & 4) ? 1 << 17 : 1 << 16;
sz 130 dev/isa/isadma.c if ((bus_dmamap_create(sc->sc_dmat, sz, 1, sz, sz,
sz 99 dev/isa/mcd.c int sz;
sz 582 dev/isa/mcd.c sc->mbx.sz = sc->blksize;
sz 1234 dev/isa/mcd.c bp->b_data + mbx->skip, mbx->sz);
sz 1237 dev/isa/mcd.c mbx->skip += mbx->sz;
sz 56 dev/pci/emuxkireg.h #define EMU_MKSUBREG(sz, idx, reg) (((sz) << 24) | ((idx) << 16) | (reg))
sz 886 dev/pci/if_cas.c cas_ringsize(int sz)
sz 888 dev/pci/if_cas.c switch (sz) {
sz 908 dev/pci/if_cas.c printf("cas: invalid Receive Descriptor ring size %d\n", sz);
sz 914 dev/pci/if_cas.c cas_cringsize(int sz)
sz 919 dev/pci/if_cas.c if (sz == (128 << i))
sz 922 dev/pci/if_cas.c printf("cas: invalid completion ring size %d\n", sz);
sz 394 dev/pci/neo.c int ofs, sz, i;
sz 411 dev/pci/neo.c sz = nf->coefficientSizes[num];
sz 415 dev/pci/neo.c for (i = 0; i < sz; i++)
sz 419 dev/pci/neo.c sz--;
sz 420 dev/pci/neo.c nm_wr(sc, addr + 4, sc->cbuf + sz, 4);
sz 88 dev/raidframe/rf_debugMem.c rf_unrecord_malloc(void *p, int sz)
sz 93 dev/raidframe/rf_debugMem.c size = rf_memory_hash_remove(p, sz);
sz 180 dev/raidframe/rf_debugMem.c rf_memory_hash_remove(void *addr, int sz)
sz 198 dev/raidframe/rf_debugMem.c if (sz > 0 && p->size != sz) {
sz 205 dev/raidframe/rf_debugMem.c (unsigned long) addr, sz, p->size, p->line, p->filen);
sz 548 dev/sbus/cgsix.c bus_size_t sz;
sz 573 dev/sbus/cgsix.c sz = mo->mo_size ? mo->mo_size : sc->sc_sunfb.sf_fbsize;
sz 574 dev/sbus/cgsix.c if (u < sz) {
sz 107 dev/sbus/sbusvar.h #define sbus_bus_map(t, slot, offset, sz, flags, unused, hp) \
sz 108 dev/sbus/sbusvar.h bus_space_map(t, BUS_ADDR(slot, offset), sz, flags, hp)
sz 1109 dev/softraid.c size_t sz = SR_META_SIZE * 512;
sz 1115 dev/softraid.c m = malloc(sz , M_DEVBUF, M_WAITOK);
sz 1116 dev/softraid.c bzero(m, sz);
sz 1123 dev/softraid.c b.b_bcount = sz;
sz 1124 dev/softraid.c b.b_bufsize = sz;
sz 1125 dev/softraid.c b.b_resid = sz;
sz 1159 dev/softraid.c bcopy(m, sm, sz);
sz 1791 dev/softraid.c size_t sz = SR_META_SIZE * 512;
sz 1797 dev/softraid.c m = malloc(sz , M_DEVBUF, M_WAITOK);
sz 1798 dev/softraid.c bzero(m, sz);
sz 1805 dev/softraid.c b.b_bcount = sz;
sz 1806 dev/softraid.c b.b_bufsize = sz;
sz 1807 dev/softraid.c b.b_resid = sz;
sz 1874 dev/softraid.c size_t sz = SR_META_SIZE * 512;
sz 1889 dev/softraid.c sz) {
sz 1964 dev/softraid.c b.b_bcount = sz;
sz 1965 dev/softraid.c b.b_bufsize = sz;
sz 1966 dev/softraid.c b.b_resid = sz;
sz 2019 dev/softraid.c size_t sz = SR_META_SIZE * 512;
sz 2024 dev/softraid.c bp = geteblk(sz);
sz 2095 dev/softraid.c bp->b_bcount = sz;
sz 2096 dev/softraid.c bp->b_bufsize = sz;
sz 2097 dev/softraid.c bp->b_resid = sz;
sz 2113 dev/softraid.c mle->sml_metadata = malloc(sz, M_DEVBUF,
sz 2115 dev/softraid.c bzero(mle->sml_metadata, sz);
sz 2116 dev/softraid.c bcopy(sm, mle->sml_metadata, sz);
sz 1094 dev/usb/usbdi.c usb_match_device(const struct usb_devno *tbl, u_int nentries, u_int sz,
sz 1102 dev/usb/usbdi.c tbl = (const struct usb_devno *)((const char *)tbl + sz);
sz 205 dev/usb/usbdi.h u_int nentries, u_int sz, u_int16_t vendor, u_int16_t product);
sz 388 dev/vnd.c int sz, flags, error, s;
sz 431 dev/vnd.c sz = howmany(bp->b_bcount, DEV_BSIZE);
sz 502 dev/vnd.c for (resid = bp->b_resid; resid; resid -= sz) {
sz 519 dev/vnd.c sz = bsize - off;
sz 521 dev/vnd.c sz = (1 + nra) * bsize;
sz 522 dev/vnd.c if (resid < sz)
sz 523 dev/vnd.c sz = resid;
sz 526 dev/vnd.c vnd->sc_vp, vp, bn, nbn, sz);
sz 532 dev/vnd.c nbp->vb_buf.b_bcount = sz;
sz 567 dev/vnd.c bp->b_resid -= (resid - sz);
sz 585 dev/vnd.c bn += sz;
sz 586 dev/vnd.c addr += sz;
sz 770 kern/kern_exec.c vsize_t sz;
sz 772 kern/kern_exec.c sz = (vaddr_t)e->e_esigcode - (vaddr_t)e->e_sigcode;
sz 789 kern/kern_exec.c e->e_sigobject = uao_create(sz, 0);
sz 793 kern/kern_exec.c if ((r = uvm_map(kernel_map, &va, round_page(sz), e->e_sigobject,
sz 799 kern/kern_exec.c memcpy((void *)va, e->e_sigcode, sz);
sz 800 kern/kern_exec.c uvm_unmap(kernel_map, va, va + round_page(sz));
sz 806 kern/kern_exec.c if (uvm_map(&p->p_vmspace->vm_map, &p->p_sigcode, round_page(sz),
sz 635 kern/kern_malloc.c malloc_roundup(size_t sz)
sz 637 kern/kern_malloc.c if (sz > MAXALLOCSAVE)
sz 638 kern/kern_malloc.c return round_page(sz);
sz 640 kern/kern_malloc.c return (1 << BUCKETINDX(sz));
sz 607 kern/subr_disk.c daddr64_t sz = howmany(bp->b_bcount, DEV_BSIZE);
sz 614 kern/subr_disk.c if (bp->b_blkno + sz > blockpersec(DL_GETPSIZE(p), lp)) {
sz 615 kern/subr_disk.c sz = blockpersec(DL_GETPSIZE(p), lp) - bp->b_blkno;
sz 616 kern/subr_disk.c if (sz == 0) {
sz 621 kern/subr_disk.c if (sz < 0)
sz 626 kern/subr_disk.c bp->b_bcount = sz << DEV_BSHIFT;
sz 166 kern/subr_extent.c size_t sz = storagesize;
sz 200 kern/subr_extent.c sz -= ALIGN(sizeof(struct extent_fixed));
sz 209 kern/subr_extent.c while (sz >= ALIGN(sizeof(struct extent_region))) {
sz 212 kern/subr_extent.c sz -= ALIGN(sizeof(struct extent_region));
sz 868 kern/sys_generic.c size_t sz;
sz 880 kern/sys_generic.c sz = sizeof(struct pollfd) * nfds;
sz 883 kern/sys_generic.c if (sz > sizeof(pfds))
sz 884 kern/sys_generic.c pl = (struct pollfd *) malloc(sz, M_TEMP, M_WAITOK);
sz 886 kern/sys_generic.c if ((error = copyin(SCARG(uap, fds), pl, sz)) != 0)
sz 942 kern/sys_generic.c error = copyout(pl, SCARG(uap, fds), sz);
sz 948 kern/sys_generic.c error = copyout(pl, SCARG(uap, fds), sz);
sz 55 lib/libsa/exec.c ssize_t sz;
sz 69 lib/libsa/exec.c sz = read(io, (char *)&x, sizeof(x));
sz 70 lib/libsa/exec.c if (sz != sizeof(x) || N_BADMAG(x)) {
sz 85 lib/libsa/exec.c sz = x.a_text;
sz 89 lib/libsa/exec.c sz -= sizeof x;
sz 91 lib/libsa/exec.c if (read(io, (char *)addr, sz) != sz)
sz 93 lib/libsa/exec.c addr += sz;
sz 132 lib/libsa/exec.c sz = i - sizeof(int);
sz 134 lib/libsa/exec.c if (read(io, addr, sz) != sz)
sz 136 lib/libsa/exec.c addr += sz;
sz 140 lib/libsa/exec.c printf("+%d]", sz);
sz 84 lib/libsa/loadfile_elf.c size_t sz;
sz 90 lib/libsa/loadfile_elf.c sz = elf->e_phnum * sizeof(Elf_Phdr);
sz 91 lib/libsa/loadfile_elf.c phdr = ALLOC(sz);
sz 95 lib/libsa/loadfile_elf.c FREE(phdr, sz);
sz 98 lib/libsa/loadfile_elf.c if (read(fd, phdr, sz) != sz) {
sz 100 lib/libsa/loadfile_elf.c FREE(phdr, sz);
sz 125 lib/libsa/loadfile_elf.c FREE(phdr, sz);
sz 131 lib/libsa/loadfile_elf.c FREE(phdr, sz);
sz 160 lib/libsa/loadfile_elf.c FREE(phdr, sz);
sz 174 lib/libsa/loadfile_elf.c sz = elf->e_shnum * sizeof(Elf_Shdr);
sz 175 lib/libsa/loadfile_elf.c shp = ALLOC(sz);
sz 177 lib/libsa/loadfile_elf.c if (read(fd, shp, sz) != sz) {
sz 179 lib/libsa/loadfile_elf.c FREE(shp, sz);
sz 184 lib/libsa/loadfile_elf.c maxp += roundup(sz, sizeof(long));
sz 191 lib/libsa/loadfile_elf.c off = roundup((sizeof(Elf_Ehdr) + sz), sizeof(long));
sz 206 lib/libsa/loadfile_elf.c FREE(shp, sz);
sz 212 lib/libsa/loadfile_elf.c FREE(shp, sz);
sz 224 lib/libsa/loadfile_elf.c BCOPY(shp, shpp, sz);
sz 229 lib/libsa/loadfile_elf.c FREE(shp, sz);
sz 138 net/if_atmsubr.c int s, error = 0, sz, len;
sz 215 net/if_atmsubr.c sz = sizeof(atmdst);
sz 217 net/if_atmsubr.c if (atm_flags & ATM_PH_LLCSNAP) sz += 8; /* sizeof snap == 8 */
sz 218 net/if_atmsubr.c M_PREPEND(m, sz, M_DONTWAIT);
sz 653 ntfs/ntfs_subr.c u_int32_t sz, i;
sz 675 ntfs/ntfs_subr.c sz = run[off++];
sz 678 ntfs/ntfs_subr.c for (i = 0; i < (sz & 0xF); i++)
sz 681 ntfs/ntfs_subr.c sz >>= 4;
sz 682 ntfs/ntfs_subr.c if (run[off + sz - 1] & 0x80) {
sz 683 ntfs/ntfs_subr.c tmp = ((u_int64_t) - 1) << (sz << 3);
sz 684 ntfs/ntfs_subr.c for (i = 0; i < sz; i++)
sz 688 ntfs/ntfs_subr.c for (i = 0; i < sz; i++)
sz 1445 ntfs/ntfs_subr.c u_int64_t sz, bn;
sz 1456 ntfs/ntfs_subr.c sz = vap->va_datalen;
sz 1459 ntfs/ntfs_subr.c (u_int32_t) sz, (u_int32_t) bn));
sz 1462 ntfs/ntfs_subr.c *size = sz;
sz 1934 ntfs/ntfs_subr.c u_int8_t sz;
sz 1941 ntfs/ntfs_subr.c sz = run[(*off)++];
sz 1942 ntfs/ntfs_subr.c if (0 == sz) {
sz 1947 ntfs/ntfs_subr.c if ((sz & 0xF) > 8 || (*off) + (sz & 0xF) > len) {
sz 1950 ntfs/ntfs_subr.c sz, len, *off);
sz 1953 ntfs/ntfs_subr.c for (i = 0; i < (sz & 0xF); i++)
sz 1956 ntfs/ntfs_subr.c sz >>= 4;
sz 1957 ntfs/ntfs_subr.c if ((sz & 0xF) > 8 || (*off) + (sz & 0xF) > len) {
sz 1960 ntfs/ntfs_subr.c sz, len, *off);
sz 1963 ntfs/ntfs_subr.c for (i = 0; i < (sz & 0xF); i++)
sz 601 ntfs/ntfs_vnops.c int sz;
sz 623 ntfs/ntfs_vnops.c sz = (*ntmp->ntm_wput)(fname, remains,
sz 625 ntfs/ntfs_vnops.c fname += sz;
sz 626 ntfs/ntfs_vnops.c remains -= sz;
sz 238 uvm/uvm_amap.c amap_alloc(vaddr_t sz, vaddr_t padsz, int waitf)
sz 244 uvm/uvm_amap.c AMAP_B2SLOT(slots, sz); /* load slots */
sz 254 uvm/uvm_amap.c UVMHIST_LOG(maphist,"<- done, amap = %p, sz=%lu", amap, sz, 0, 0);
sz 66 uvm/uvm_io.c vsize_t chunksz, togo, sz;
sz 129 uvm/uvm_io.c sz = chunksz - pageoffset;
sz 130 uvm/uvm_io.c if (sz > togo)
sz 131 uvm/uvm_io.c sz = togo;
sz 132 uvm/uvm_io.c error = uiomove((caddr_t) (kva + pageoffset), sz, uio);
sz 133 uvm/uvm_io.c togo -= sz;
sz 1291 uvm/uvm_swap.c int s, off, nra, error, sz, resid;
sz 1314 uvm/uvm_swap.c for (resid = bp->b_resid; resid; resid -= sz) {
sz 1360 uvm/uvm_swap.c sz = (1 + nra) * sdp->swd_bsize - off;
sz 1361 uvm/uvm_swap.c if (sz > resid)
sz 1362 uvm/uvm_swap.c sz = resid;
sz 1375 uvm/uvm_swap.c nbp->vb_buf.b_bcount = sz;
sz 1376 uvm/uvm_swap.c nbp->vb_buf.b_bufsize = sz;
sz 1393 uvm/uvm_swap.c nbp->vb_buf.b_dirtyend = sz;
sz 1398 uvm/uvm_swap.c min(sz,
sz 1403 uvm/uvm_swap.c nbp->vb_buf.b_validend = sz;
sz 1408 uvm/uvm_swap.c min(sz,
sz 1436 uvm/uvm_swap.c byteoff += sz;
sz 1437 uvm/uvm_swap.c addr += sz;
sz 94 xfs/xfs_common-bsd.c xfs_devtoname_r (dev_t dev, char *buf, size_t sz)
sz 102 xfs/xfs_common-bsd.c xfs_devtoname_r (dev_t dev, char *buf, size_t sz)
sz 105 xfs/xfs_common-bsd.c snprintf (buf, sz, "%u/%u", major(dev), minor(dev));
sz 72 xfs/xfs_common.h xfs_devtoname_r (dev_t dev, char *buf, size_t sz);
sz 376 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) uvm_vnp_setsize(vp, sz)
sz 378 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) vnode_pager_setsize(vp, sz)
sz 380 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz) ubc_setsize(vp, sz)
sz 382 xfs/xfs_locl.h #define xfs_set_vp_size(vp, sz)
sz 302 xfs/xfs_vnodeops-common.c size_t sz = 0;
sz 306 xfs/xfs_vnodeops-common.c sz += uio->uio_iov[i].iov_len;
sz 307 xfs/xfs_vnodeops-common.c if (sz != uio->uio_resid)