curr_md 389 arch/i386/i386/i686_mem.c struct mem_range_desc *first_md, *last_md, *curr_md; curr_md 398 arch/i386/i386/i686_mem.c for (curr_md = first_md; curr_md <= last_md; curr_md++) { curr_md 399 arch/i386/i386/i686_mem.c if ((curr_md->mr_flags & MDF_ATTRMASK) == MDF_UNKNOWN) curr_md 404 arch/i386/i386/i686_mem.c for (curr_md = first_md; curr_md <= last_md; curr_md++) { curr_md 405 arch/i386/i386/i686_mem.c curr_md->mr_flags = mrcopyflags(curr_md->mr_flags & ~MDF_FIRMWARE, mrd->mr_flags); curr_md 406 arch/i386/i386/i686_mem.c bcopy(mrd->mr_owner, curr_md->mr_owner, sizeof(mrd->mr_owner)); curr_md 422 arch/i386/i386/i686_mem.c struct mem_range_desc *curr_md, *free_md; curr_md 433 arch/i386/i386/i686_mem.c curr_md = sc->mr_desc + i; curr_md 435 arch/i386/i386/i686_mem.c for (; i < sc->mr_ndesc; i++, curr_md++) { curr_md 436 arch/i386/i386/i686_mem.c if (curr_md->mr_flags & MDF_ACTIVE) { curr_md 438 arch/i386/i386/i686_mem.c if ((curr_md->mr_base == mrd->mr_base) && curr_md 439 arch/i386/i386/i686_mem.c (curr_md->mr_len == mrd->mr_len)) { curr_md 441 arch/i386/i386/i686_mem.c if (curr_md->mr_flags & MDF_BUSY) curr_md 445 arch/i386/i386/i686_mem.c ((curr_md->mr_flags & MDF_ATTRMASK) curr_md 449 arch/i386/i386/i686_mem.c free_md = curr_md; curr_md 453 arch/i386/i386/i686_mem.c if (mroverlap(curr_md, mrd)) { curr_md 455 arch/i386/i386/i686_mem.c if (i686_mtrrconflict(curr_md->mr_flags, curr_md 460 arch/i386/i386/i686_mem.c free_md = curr_md;