ci 235 arch/i386/i386/amd64errata.c amd64_errata_testmsr(struct cpu_info *ci, errata_t *e) ci 239 arch/i386/i386/amd64errata.c (void)ci; ci 250 arch/i386/i386/amd64errata.c amd64_errata_setmsr(struct cpu_info *ci, errata_t *e) ci 254 arch/i386/i386/amd64errata.c (void)ci; ci 274 arch/i386/i386/amd64errata.c amd64_errata(struct cpu_info *ci) ci 320 arch/i386/i386/amd64errata.c ci->ci_dev.dv_xname, e->e_num); ci 327 arch/i386/i386/amd64errata.c rc = (e->e_act == NULL) ? 1 : (*e->e_act)(ci, e); ci 340 arch/i386/i386/amd64errata.c ci->ci_dev.dv_xname, e->e_num, ci 351 arch/i386/i386/amd64errata.c printf("%s: AMD %s", ci->ci_dev.dv_xname, ci 369 arch/i386/i386/amd64errata.c printf("%s: AMD %s", ci->ci_dev.dv_xname, ci 187 arch/i386/i386/cpu.c struct cpu_info *ci = (struct cpu_info *)self; ci 202 arch/i386/i386/cpu.c ci = &cpu_info_primary; ci 203 arch/i386/i386/cpu.c bcopy(self, &ci->ci_dev, sizeof *self); ci 213 arch/i386/i386/cpu.c cpu_info[cpunum] = ci; ci 216 arch/i386/i386/cpu.c ci->ci_self = ci; ci 217 arch/i386/i386/cpu.c ci->ci_apicid = caa->cpu_number; ci 219 arch/i386/i386/cpu.c ci->ci_cpuid = ci->ci_apicid; ci 221 arch/i386/i386/cpu.c ci->ci_cpuid = 0; /* False for APs, so what, they're not used */ ci 223 arch/i386/i386/cpu.c ci->ci_signature = caa->cpu_signature; ci 224 arch/i386/i386/cpu.c ci->ci_feature_flags = caa->feature_flags; ci 225 arch/i386/i386/cpu.c ci->ci_func = caa->cpu_func; ci 239 arch/i386/i386/cpu.c ci->ci_dev.dv_xname); ci 242 arch/i386/i386/cpu.c pcb = ci->ci_idle_pcb = (struct pcb *)kstack; ci 254 arch/i386/i386/cpu.c cpu_default_ldt(ci); /* Use the `global' ldt until one alloc'd */ ci 266 arch/i386/i386/cpu.c ci->ci_flags |= CPUF_PRESENT | CPUF_SP | CPUF_PRIMARY; ci 267 arch/i386/i386/cpu.c identifycpu(ci); ci 268 arch/i386/i386/cpu.c cpu_init(ci); ci 273 arch/i386/i386/cpu.c ci->ci_flags |= CPUF_PRESENT | CPUF_BSP | CPUF_PRIMARY; ci 274 arch/i386/i386/cpu.c identifycpu(ci); ci 275 arch/i386/i386/cpu.c cpu_init(ci); ci 282 arch/i386/i386/cpu.c lapic_calibrate_timer(ci); ci 294 arch/i386/i386/cpu.c gdt_alloc_cpu(ci); ci 295 arch/i386/i386/cpu.c cpu_alloc_ldt(ci); ci 296 arch/i386/i386/cpu.c ci->ci_flags |= CPUF_PRESENT | CPUF_AP; ci 297 arch/i386/i386/cpu.c identifycpu(ci); ci 298 arch/i386/i386/cpu.c ci->ci_next = cpu_info_list->ci_next; ci 299 arch/i386/i386/cpu.c cpu_info_list->ci_next = ci; ci 308 arch/i386/i386/cpu.c if (ci->ci_apicid < IOAPIC_ID_MAX) ci 309 arch/i386/i386/cpu.c ioapic_id_map &= ~(1 << ci->ci_apicid); ci 313 arch/i386/i386/cpu.c ci->ci_dev.dv_xname, kstack, USPACE); ci 315 arch/i386/i386/cpu.c ci->ci_dev.dv_xname, pcb, pcb->pcb_esp); ci 328 arch/i386/i386/cpu.c cpu_init(struct cpu_info *ci) ci 331 arch/i386/i386/cpu.c if (ci->cpu_setup != NULL) ci 332 arch/i386/i386/cpu.c (*ci->cpu_setup)(ci); ci 343 arch/i386/i386/cpu.c ci->ci_flags |= CPUF_RUNNING; ci 363 arch/i386/i386/cpu.c struct cpu_info *ci; ci 367 arch/i386/i386/cpu.c ci = cpu_info[i]; ci 368 arch/i386/i386/cpu.c if (ci == NULL) ci 370 arch/i386/i386/cpu.c if (ci->ci_idle_pcb == NULL) ci 372 arch/i386/i386/cpu.c if ((ci->ci_flags & CPUF_PRESENT) == 0) ci 374 arch/i386/i386/cpu.c if (ci->ci_flags & (CPUF_BSP|CPUF_SP|CPUF_PRIMARY)) ci 376 arch/i386/i386/cpu.c cpu_boot_secondary(ci); ci 383 arch/i386/i386/cpu.c struct cpu_info *ci; ci 387 arch/i386/i386/cpu.c ci = cpu_info[i]; ci 388 arch/i386/i386/cpu.c if (ci == NULL) ci 390 arch/i386/i386/cpu.c if (ci->ci_idle_pcb == NULL) ci 392 arch/i386/i386/cpu.c if ((ci->ci_flags & CPUF_PRESENT) == 0) ci 394 arch/i386/i386/cpu.c i386_init_pcb_tss_ldt(ci); ci 399 arch/i386/i386/cpu.c cpu_boot_secondary(struct cpu_info *ci) ci 407 arch/i386/i386/cpu.c printf("%s: starting", ci->ci_dev.dv_xname); ci 412 arch/i386/i386/cpu.c pcb = ci->ci_idle_pcb; ci 417 arch/i386/i386/cpu.c CPU_STARTUP(ci); ci 422 arch/i386/i386/cpu.c for (i = 100000; (!(ci->ci_flags & CPUF_RUNNING)) && i > 0; i--) { ci 425 arch/i386/i386/cpu.c if (!(ci->ci_flags & CPUF_RUNNING)) { ci 426 arch/i386/i386/cpu.c printf("%s failed to become ready\n", ci->ci_dev.dv_xname); ci 432 arch/i386/i386/cpu.c CPU_START_CLEANUP(ci); ci 443 arch/i386/i386/cpu.c struct cpu_info *ci = (struct cpu_info *)v; ci 450 arch/i386/i386/cpu.c gdt_init_cpu(ci); ci 451 arch/i386/i386/cpu.c cpu_init_ldt(ci); ci 452 arch/i386/i386/cpu.c npxinit(ci); ci 456 arch/i386/i386/cpu.c cpu_init(ci); ci 463 arch/i386/i386/cpu.c ci->ci_dev.dv_xname, ci->ci_cpuid); ci 464 arch/i386/i386/cpu.c microuptime(&ci->ci_schedstate.spc_runtime); ci 514 arch/i386/i386/cpu.c cpu_set_tss_gates(struct cpu_info *ci) ci 518 arch/i386/i386/cpu.c ci->ci_doubleflt_stack = (char *)uvm_km_alloc(kernel_map, USPACE); ci 519 arch/i386/i386/cpu.c cpu_init_tss(&ci->ci_doubleflt_tss, ci->ci_doubleflt_stack, ci 521 arch/i386/i386/cpu.c setsegment(&sd, &ci->ci_doubleflt_tss, sizeof(struct i386tss) - 1, ci 523 arch/i386/i386/cpu.c ci->ci_gdt[GTRAPTSS_SEL].sd = sd; ci 535 arch/i386/i386/cpu.c ci->ci_ddbipi_stack = (char *)uvm_km_alloc(kernel_map, USPACE); ci 536 arch/i386/i386/cpu.c cpu_init_tss(&ci->ci_ddbipi_tss, ci->ci_ddbipi_stack, ci 539 arch/i386/i386/cpu.c setsegment(&sd, &ci->ci_ddbipi_tss, sizeof(struct i386tss) - 1, ci 541 arch/i386/i386/cpu.c ci->ci_gdt[GIPITSS_SEL].sd = sd; ci 551 arch/i386/i386/cpu.c mp_cpu_start(struct cpu_info *ci) ci 582 arch/i386/i386/cpu.c if (ci->ci_flags & CPUF_AP) { ci 583 arch/i386/i386/cpu.c if ((error = i386_ipi_init(ci->ci_apicid)) != 0) ci 590 arch/i386/i386/cpu.c ci->ci_apicid, LAPIC_DLMODE_STARTUP)) != 0) ci 595 arch/i386/i386/cpu.c ci->ci_apicid, LAPIC_DLMODE_STARTUP)) != 0) ci 605 arch/i386/i386/cpu.c mp_cpu_start_cleanup(struct cpu_info *ci) ci 138 arch/i386/i386/db_mp.c i386_ipi_db(struct cpu_info *ci) ci 110 arch/i386/i386/gdt.c struct cpu_info *ci; ci 115 arch/i386/i386/gdt.c CPU_INFO_FOREACH(cii, ci) ci 116 arch/i386/i386/gdt.c if (ci->ci_gdt != NULL && ci->ci_gdt != gdt) ci 117 arch/i386/i386/gdt.c ci->ci_gdt[sel].sd = *sd; ci 129 arch/i386/i386/gdt.c struct cpu_info *ci = &cpu_info_primary; ci 147 arch/i386/i386/gdt.c ci->ci_gdt = gdt; ci 148 arch/i386/i386/gdt.c setsegment(&ci->ci_gdt[GCPU_SEL].sd, ci, sizeof(struct cpu_info)-1, ci 151 arch/i386/i386/gdt.c gdt_init_cpu(ci); ci 159 arch/i386/i386/gdt.c gdt_alloc_cpu(struct cpu_info *ci) ci 164 arch/i386/i386/gdt.c ci->ci_gdt = (union descriptor *)uvm_km_valloc(kernel_map, max_len); ci 165 arch/i386/i386/gdt.c uvm_map_pageable(kernel_map, (vaddr_t)ci->ci_gdt, ci 166 arch/i386/i386/gdt.c (vaddr_t)ci->ci_gdt + min_len, FALSE, FALSE); ci 167 arch/i386/i386/gdt.c bzero(ci->ci_gdt, min_len); ci 168 arch/i386/i386/gdt.c bcopy(gdt, ci->ci_gdt, gdt_size * sizeof(union descriptor)); ci 169 arch/i386/i386/gdt.c setsegment(&ci->ci_gdt[GCPU_SEL].sd, ci, sizeof(struct cpu_info)-1, ci 180 arch/i386/i386/gdt.c gdt_init_cpu(struct cpu_info *ci) ci 184 arch/i386/i386/gdt.c setregion(®ion, ci->ci_gdt, ci 197 arch/i386/i386/gdt.c struct cpu_info *ci; ci 205 arch/i386/i386/gdt.c CPU_INFO_FOREACH(cii, ci) { ci 206 arch/i386/i386/gdt.c for (va = (vaddr_t)(ci->ci_gdt) + old_len; ci 207 arch/i386/i386/gdt.c va < (vaddr_t)(ci->ci_gdt) + new_len; ci 97 arch/i386/i386/ipifuncs.c i386_ipi_halt(struct cpu_info *ci) ci 101 arch/i386/i386/ipifuncs.c printf("%s: shutting down\n", ci->ci_dev.dv_xname); ci 109 arch/i386/i386/ipifuncs.c i386_ipi_flush_fpu(struct cpu_info *ci) ci 111 arch/i386/i386/ipifuncs.c npxsave_cpu(ci, 0); ci 115 arch/i386/i386/ipifuncs.c i386_ipi_synch_fpu(struct cpu_info *ci) ci 117 arch/i386/i386/ipifuncs.c npxsave_cpu(ci, 1); ci 128 arch/i386/i386/ipifuncs.c i386_send_ipi(struct cpu_info *ci, int ipimask) ci 132 arch/i386/i386/ipifuncs.c i386_atomic_setbits_l(&ci->ci_ipis, ipimask); ci 135 arch/i386/i386/ipifuncs.c if (!(ci->ci_flags & CPUF_RUNNING)) ci 138 arch/i386/i386/ipifuncs.c ret = i386_ipi(LAPIC_IPI_VECTOR, ci->ci_cpuid, LAPIC_DLMODE_FIXED); ci 141 arch/i386/i386/ipifuncs.c ipimask, curcpu()->ci_dev.dv_xname, ci->ci_dev.dv_xname); ci 148 arch/i386/i386/ipifuncs.c i386_fast_ipi(struct cpu_info *ci, int ipi) ci 150 arch/i386/i386/ipifuncs.c if (!(ci->ci_flags & CPUF_RUNNING)) ci 153 arch/i386/i386/ipifuncs.c return (i386_ipi(ipi, ci->ci_cpuid, LAPIC_DLMODE_FIXED)); ci 167 arch/i386/i386/ipifuncs.c struct cpu_info *ci, *self = curcpu(); ci 171 arch/i386/i386/ipifuncs.c CPU_INFO_FOREACH(cii, ci) { ci 172 arch/i386/i386/ipifuncs.c if (ci == self) ci 174 arch/i386/i386/ipifuncs.c if ((ci->ci_flags & CPUF_RUNNING) == 0) ci 176 arch/i386/i386/ipifuncs.c i386_atomic_setbits_l(&ci->ci_ipis, ipimask); ci 189 arch/i386/i386/ipifuncs.c struct cpu_info *ci = curcpu(); ci 193 arch/i386/i386/ipifuncs.c pending = i386_atomic_testset_ul(&ci->ci_ipis, 0); ci 198 arch/i386/i386/ipifuncs.c (*ipifunc[bit])(ci); ci 131 arch/i386/i386/lapic.c struct cpu_info *ci = curcpu(); ci 137 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "prelint", 0, 0, ci 139 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "prelint", 1, 0, ci 147 arch/i386/i386/lapic.c || mpi->cpu_id == ci->ci_apicid)) { ci 162 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "timer", 0, 0, ci 164 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "pcint", 0, 0, ci 166 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "lint", 0, 0, ci 168 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "lint", 1, 0, ci 170 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "err", 0, 0, ci 224 arch/i386/i386/lapic.c struct cpu_info *ci = curcpu(); ci 227 arch/i386/i386/lapic.c if (CPU_IS_PRIMARY(ci)) { ci 228 arch/i386/i386/lapic.c ci->ci_tscbase = rdtsc(); ci 267 arch/i386/i386/lapic.c lapic_calibrate_timer(struct cpu_info *ci) ci 276 arch/i386/i386/lapic.c printf("%s: calibrating local timer\n", ci->ci_dev.dv_xname); ci 351 arch/i386/i386/lapic.c printf("%s: apic clock running at %s\n", ci->ci_dev.dv_xname, tbuf); ci 432 arch/i386/i386/lapic.c i386_ipi_microset(struct cpu_info *ci) ci 434 arch/i386/i386/lapic.c ci->ci_tscbase = rdtsc(); ci 445 arch/i386/i386/lapic.c struct cpu_info *ci = curcpu(); ci 451 arch/i386/i386/lapic.c tmp = rdtsc() - ci->ci_tscbase; ci 502 arch/i386/i386/machdep.c i386_init_pcb_tss_ldt(struct cpu_info *ci) ci 505 arch/i386/i386/machdep.c struct pcb *pcb = ci->ci_idle_pcb; ci 514 arch/i386/i386/machdep.c pcb->pcb_ldt = ci->ci_ldt; ci 516 arch/i386/i386/machdep.c ci->ci_idle_tss_sel = tss_alloc(pcb); ci 1028 arch/i386/i386/machdep.c winchip_cpu_setup(struct cpu_info *ci) ci 1032 arch/i386/i386/machdep.c switch ((ci->ci_signature >> 4) & 15) { /* model */ ci 1034 arch/i386/i386/machdep.c ci->ci_feature_flags &= ~CPUID_TSC; ci 1037 arch/i386/i386/machdep.c printf("%s: TSC disabled\n", ci->ci_dev.dv_xname); ci 1045 arch/i386/i386/machdep.c cyrix3_setperf_setup(struct cpu_info *ci) ci 1049 arch/i386/i386/machdep.c est_init(ci->ci_dev.dv_xname, CPUVENDOR_VIA); ci 1052 arch/i386/i386/machdep.c ci->ci_dev.dv_xname); ci 1058 arch/i386/i386/machdep.c cyrix3_cpu_setup(struct cpu_info *ci) ci 1061 arch/i386/i386/machdep.c int model = (ci->ci_signature >> 4) & 15; ci 1062 arch/i386/i386/machdep.c int step = ci->ci_signature & 15; ci 1073 arch/i386/i386/machdep.c cyrix3_get_bus_clock(ci); ci 1118 arch/i386/i386/machdep.c printf("%s:", ci->ci_dev.dv_xname); ci 1192 arch/i386/i386/machdep.c cyrix6x86_cpu_setup(struct cpu_info *ci) ci 1196 arch/i386/i386/machdep.c switch ((ci->ci_signature >> 4) & 15) { /* model */ ci 1213 arch/i386/i386/machdep.c ci->ci_dev.dv_xname); ci 1219 arch/i386/i386/machdep.c printf("%s: TSC disabled\n", ci->ci_dev.dv_xname); ci 1225 arch/i386/i386/machdep.c natsem6x86_cpu_setup(struct cpu_info *ci) ci 1229 arch/i386/i386/machdep.c int model = (ci->ci_signature >> 4) & 15; ci 1235 arch/i386/i386/machdep.c printf("%s: TSC disabled\n", ci->ci_dev.dv_xname); ci 1242 arch/i386/i386/machdep.c intel586_cpu_setup(struct cpu_info *ci) ci 1248 arch/i386/i386/machdep.c ci->ci_dev.dv_xname); ci 1255 arch/i386/i386/machdep.c amd_family5_setperf_setup(struct cpu_info *ci) ci 1262 arch/i386/i386/machdep.c amd_family5_setup(struct cpu_info *ci) ci 1264 arch/i386/i386/machdep.c int model = (ci->ci_signature >> 4) & 15; ci 1292 arch/i386/i386/machdep.c amd_family6_setperf_setup(struct cpu_info *ci) ci 1294 arch/i386/i386/machdep.c int family = (ci->ci_signature >> 8) & 15; ci 1308 arch/i386/i386/machdep.c amd_family6_setup(struct cpu_info *ci) ci 1338 arch/i386/i386/machdep.c struct cpu_info *ci = (struct cpu_info *) args; ci 1347 arch/i386/i386/machdep.c ci->ci_sensor.value = max - MSR_THERM_STATUS_TEMP(msr); ci 1349 arch/i386/i386/machdep.c ci->ci_sensor.value *= 1000000; ci 1351 arch/i386/i386/machdep.c ci->ci_sensor.value += 273150000; ci 1352 arch/i386/i386/machdep.c ci->ci_sensor.flags &= ~SENSOR_FINVALID; ci 1354 arch/i386/i386/machdep.c ci->ci_sensor.value = 0; ci 1355 arch/i386/i386/machdep.c ci->ci_sensor.flags |= SENSOR_FINVALID; ci 1360 arch/i386/i386/machdep.c intel686_cpusensors_setup(struct cpu_info *ci) ci 1373 arch/i386/i386/machdep.c strlcpy(ci->ci_sensordev.xname, ci->ci_dev.dv_xname, ci 1374 arch/i386/i386/machdep.c sizeof(ci->ci_sensordev.xname)); ci 1375 arch/i386/i386/machdep.c ci->ci_sensor.type = SENSOR_TEMP; ci 1376 arch/i386/i386/machdep.c sensor_task_register(ci, intelcore_update_sensor, 5); ci 1377 arch/i386/i386/machdep.c sensor_attach(&ci->ci_sensordev, &ci->ci_sensor); ci 1378 arch/i386/i386/machdep.c sensordev_install(&ci->ci_sensordev); ci 1384 arch/i386/i386/machdep.c intel686_setperf_setup(struct cpu_info *ci) ci 1386 arch/i386/i386/machdep.c int family = (ci->ci_signature >> 8) & 15; ci 1387 arch/i386/i386/machdep.c int step = ci->ci_signature & 15; ci 1391 arch/i386/i386/machdep.c est_init(ci->ci_dev.dv_xname, CPUVENDOR_INTEL); ci 1394 arch/i386/i386/machdep.c ci->ci_dev.dv_xname); ci 1402 arch/i386/i386/machdep.c intel686_common_cpu_setup(struct cpu_info *ci) ci 1427 arch/i386/i386/machdep.c intel686_cpu_setup(struct cpu_info *ci) ci 1429 arch/i386/i386/machdep.c int model = (ci->ci_signature >> 4) & 15; ci 1430 arch/i386/i386/machdep.c int step = ci->ci_signature & 15; ci 1434 arch/i386/i386/machdep.c p3_get_bus_clock(ci); ci 1437 arch/i386/i386/machdep.c intel686_common_cpu_setup(ci); ci 1444 arch/i386/i386/machdep.c ci->ci_feature_flags &= ~CPUID_SEP; ci 1449 arch/i386/i386/machdep.c if ((model == 7) && (ci->ci_feature_flags & CPUID_SER)) { ci 1455 arch/i386/i386/machdep.c ci->ci_dev.dv_xname); ci 1456 arch/i386/i386/machdep.c ci->ci_feature_flags &= ~CPUID_SER; ci 1457 arch/i386/i386/machdep.c ci->ci_level = 2; ci 1467 arch/i386/i386/machdep.c intel686_p4_cpu_setup(struct cpu_info *ci) ci 1470 arch/i386/i386/machdep.c p4_get_bus_clock(ci); ci 1473 arch/i386/i386/machdep.c intel686_common_cpu_setup(ci); ci 1476 arch/i386/i386/machdep.c p4_model = (ci->ci_signature >> 4) & 15; ci 1482 arch/i386/i386/machdep.c tm86_cpu_setup(struct cpu_info *ci) ci 1560 arch/i386/i386/machdep.c identifycpu(struct cpu_info *ci) ci 1567 arch/i386/i386/machdep.c char *cpu_device = ci->ci_dev.dv_xname; ci 1582 arch/i386/i386/machdep.c ci->cpu_setup = i386_nocpuid_cpus[cpu].cpu_setup; ci 1587 arch/i386/i386/machdep.c modif = (ci->ci_signature >> 12) & 3; ci 1588 arch/i386/i386/machdep.c family = (ci->ci_signature >> 8) & 15; ci 1589 arch/i386/i386/machdep.c model = (ci->ci_signature >> 4) & 15; ci 1590 arch/i386/i386/machdep.c step = ci->ci_signature & 15; ci 1623 arch/i386/i386/machdep.c ci->cpu_setup = NULL; ci 1666 arch/i386/i386/machdep.c ci->cpu_setup = cpup->cpu_family[i].cpu_setup; ci 1706 arch/i386/i386/machdep.c if ((ci->ci_flags & CPUF_PRIMARY) == 0) { ci 1725 arch/i386/i386/machdep.c if (ci->ci_feature_flags && (ci->ci_feature_flags & CPUID_TSC)) { ci 1733 arch/i386/i386/machdep.c if ((ci->ci_flags & CPUF_PRIMARY) == 0) { ci 1740 arch/i386/i386/machdep.c if ((ci->ci_flags & CPUF_PRIMARY) == 0) { ci 1746 arch/i386/i386/machdep.c if ((ci->ci_flags & CPUF_PRIMARY) == 0) { ci 1749 arch/i386/i386/machdep.c if (ci->ci_feature_flags) { ci 1756 arch/i386/i386/machdep.c if (ci->ci_feature_flags & ci 1779 arch/i386/i386/machdep.c if (ci->cpu_setup != NULL) ci 1780 arch/i386/i386/machdep.c (ci->cpu_setup)(ci); ci 1829 arch/i386/i386/machdep.c ci->cpu_class = class; ci 1872 arch/i386/i386/machdep.c amd64_errata(ci); ci 1898 arch/i386/i386/machdep.c cyrix3_get_bus_clock(struct cpu_info *ci) ci 1922 arch/i386/i386/machdep.c p4_get_bus_clock(struct cpu_info *ci) ci 1927 arch/i386/i386/machdep.c model = (ci->ci_signature >> 4) & 15; ci 1941 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, model, bus); ci 1962 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, model, bus); ci 1969 arch/i386/i386/machdep.c p3_get_bus_clock(struct cpu_info *ci) ci 1974 arch/i386/i386/machdep.c model = (ci->ci_signature >> 4) & 15; ci 1991 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, bus); ci 2020 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, bus); ci 2046 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, bus); ci 2052 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, model); ci 2800 arch/i386/i386/machdep.c cpu_default_ldt(struct cpu_info *ci) ci 2802 arch/i386/i386/machdep.c ci->ci_ldt = ldt; ci 2803 arch/i386/i386/machdep.c ci->ci_ldt_len = sizeof(ldt); ci 2807 arch/i386/i386/machdep.c cpu_alloc_ldt(struct cpu_info *ci) ci 2814 arch/i386/i386/machdep.c ci->ci_ldt = cpu_ldt; ci 2815 arch/i386/i386/machdep.c ci->ci_ldt_len = len; ci 2819 arch/i386/i386/machdep.c cpu_init_ldt(struct cpu_info *ci) ci 2821 arch/i386/i386/machdep.c setsegment(&ci->ci_gdt[GLDT_SEL].sd, ci->ci_ldt, ci->ci_ldt_len - 1, ci 3223 arch/i386/i386/machdep.c need_resched(struct cpu_info *ci) ci 3227 arch/i386/i386/machdep.c ci->ci_want_resched = 1; ci 3233 arch/i386/i386/machdep.c if ((p = ci->ci_curproc) != NULL) ci 48 arch/i386/i386/mp_setperf.c struct cpu_info *ci; ci 57 arch/i386/i386/mp_setperf.c CPU_INFO_FOREACH(cii, ci) { ci 58 arch/i386/i386/mp_setperf.c if (ci->ci_setperf_state != CI_SETPERF_INTRANSIT) { ci 59 arch/i386/i386/mp_setperf.c ci->ci_setperf_state = ci 61 arch/i386/i386/mp_setperf.c i386_send_ipi(ci, I386_IPI_SETPERF); ci 68 arch/i386/i386/mp_setperf.c CPU_INFO_FOREACH(cii, ci) { ci 69 arch/i386/i386/mp_setperf.c if ((notready = (ci->ci_setperf_state ci 86 arch/i386/i386/mp_setperf.c CPU_INFO_FOREACH(cii, ci) { ci 87 arch/i386/i386/mp_setperf.c if ((notready = (ci->ci_setperf_state ci 104 arch/i386/i386/mp_setperf.c i386_setperf_ipi(struct cpu_info *ci) ci 107 arch/i386/i386/mp_setperf.c if (ci->ci_setperf_state == CI_SETPERF_SHOULDSTOP) ci 108 arch/i386/i386/mp_setperf.c ci->ci_setperf_state = CI_SETPERF_INTRANSIT; ci 115 arch/i386/i386/mp_setperf.c ci->ci_setperf_state = CI_SETPERF_DONE; ci 119 arch/i386/i386/mp_setperf.c ci->ci_setperf_state = CI_SETPERF_READY; ci 126 arch/i386/i386/mp_setperf.c struct cpu_info *ci; ci 135 arch/i386/i386/mp_setperf.c CPU_INFO_FOREACH(cii, ci) { ci 136 arch/i386/i386/mp_setperf.c ci->ci_setperf_state = CI_SETPERF_READY; ci 2940 arch/i386/i386/pmap.c struct cpu_info *ci, *self = curcpu(); ci 2945 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) { ci 2946 arch/i386/i386/pmap.c if (ci == self || !pmap_is_active(pm, ci->ci_cpuid) || ci 2947 arch/i386/i386/pmap.c !(ci->ci_flags & CPUF_RUNNING)) ci 2949 arch/i386/i386/pmap.c mask |= 1 << ci->ci_cpuid; ci 2961 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) { ci 2962 arch/i386/i386/pmap.c if ((mask & 1 << ci->ci_cpuid) == 0) ci 2964 arch/i386/i386/pmap.c if (i386_fast_ipi(ci, LAPIC_IPI_INVLPG) != 0) ci 2977 arch/i386/i386/pmap.c struct cpu_info *ci, *self = curcpu(); ci 2983 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) { ci 2984 arch/i386/i386/pmap.c if (ci == self || !pmap_is_active(pm, ci->ci_cpuid) || ci 2985 arch/i386/i386/pmap.c !(ci->ci_flags & CPUF_RUNNING)) ci 2987 arch/i386/i386/pmap.c mask |= 1 << ci->ci_cpuid; ci 3000 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) { ci 3001 arch/i386/i386/pmap.c if ((mask & 1 << ci->ci_cpuid) == 0) ci 3003 arch/i386/i386/pmap.c if (i386_fast_ipi(ci, LAPIC_IPI_INVLRANGE) != 0) ci 3017 arch/i386/i386/pmap.c struct cpu_info *ci, *self = curcpu(); ci 3022 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) { ci 3023 arch/i386/i386/pmap.c if (ci == self || !(ci->ci_flags & CPUF_RUNNING)) ci 3025 arch/i386/i386/pmap.c mask |= 1 << ci->ci_cpuid; ci 3037 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) { ci 3038 arch/i386/i386/pmap.c if ((mask & 1 << ci->ci_cpuid) == 0) ci 3040 arch/i386/i386/pmap.c if (i386_fast_ipi(ci, LAPIC_IPI_INVLTLB) != 0) ci 385 arch/i386/i386/powernow-k7.c struct cpu_info *ci; ci 392 arch/i386/i386/powernow-k7.c ci = curcpu(); ci 410 arch/i386/i386/powernow-k7.c if (ci->ci_signature == AMD_ERRATA_A0_CPUSIG) ci 426 arch/i386/i386/powernow-k7.c if (!k7pnow_states(cstate, ci->ci_signature, maxfid, startvid)) ci 436 arch/i386/i386/powernow-k7.c ci->ci_dev.dv_xname, techname, cpuspeed); ci 435 arch/i386/i386/powernow-k8.c struct cpu_info * ci; ci 439 arch/i386/i386/powernow-k8.c ci = curcpu(); ci 482 arch/i386/i386/powernow-k8.c if (!k8pnow_states(cstate, ci->ci_signature, maxfid, maxvid)) ci 487 arch/i386/i386/powernow-k8.c ci->ci_dev.dv_xname, techname, cpuspeed); ci 177 arch/i386/include/cpu.h #define CPU_INFO_FOREACH(cii, ci) for (cii = 0, ci = cpu_info_list; \ ci 178 arch/i386/include/cpu.h ci != NULL; ci = ci->ci_next) ci 180 arch/i386/include/cpu.h #define CPU_INFO_UNIT(ci) ((ci)->ci_dev.dv_unit) ci 195 arch/i386/include/cpu.h struct cpu_info *ci; ci 199 arch/i386/include/cpu.h "=r" (ci) : "m" ci 201 arch/i386/include/cpu.h return ci; ci 205 arch/i386/include/cpu.h #define CPU_IS_PRIMARY(ci) ((ci)->ci_flags & CPUF_PRIMARY) ci 220 arch/i386/include/cpu.h #define CPU_IS_PRIMARY(ci) 1 ci 157 arch/i386/isa/npx.c npxdna_notset(struct cpu_info *ci) ci 350 arch/i386/isa/npx.c npxinit(struct cpu_info *ci) ci 357 arch/i386/isa/npx.c ci->ci_dev.dv_xname); ci 422 arch/i386/isa/npx.c struct cpu_info *ci = curcpu(); ci 423 arch/i386/isa/npx.c struct proc *p = ci->ci_fpcurproc; ci 430 arch/i386/isa/npx.c IPRINTF(("%s: fp intr\n", ci->ci_dev.dv_xname)); ci 446 arch/i386/isa/npx.c if (ci->ci_fpsaving) ci 582 arch/i386/isa/npx.c npxdna_xmm(struct cpu_info *ci) ci 587 arch/i386/isa/npx.c if (ci->ci_fpsaving) { ci 595 arch/i386/isa/npx.c p = ci->ci_curproc; ci 600 arch/i386/isa/npx.c IPRINTF(("%s: dna for %lx%s\n", ci->ci_dev.dv_xname, (u_long)p, ci 611 arch/i386/isa/npx.c if (ci->ci_fpcurproc != NULL) { ci 612 arch/i386/isa/npx.c IPRINTF(("%s: fp save %lx\n", ci->ci_dev.dv_xname, ci 613 arch/i386/isa/npx.c (u_long)ci->ci_fpcurproc)); ci 614 arch/i386/isa/npx.c npxsave_cpu(ci, 1); ci 617 arch/i386/isa/npx.c IPRINTF(("%s: fp init\n", ci->ci_dev.dv_xname)); ci 624 arch/i386/isa/npx.c IPRINTF(("%s: done saving\n", ci->ci_dev.dv_xname)); ci 625 arch/i386/isa/npx.c KDASSERT(ci->ci_fpcurproc == NULL); ci 635 arch/i386/isa/npx.c ci->ci_fpcurproc = p; ci 636 arch/i386/isa/npx.c p->p_addr->u_pcb.pcb_fpcpu = ci; ci 660 arch/i386/isa/npx.c npxdna_s87(struct cpu_info *ci) ci 667 arch/i386/isa/npx.c if (ci->ci_fpsaving) { ci 674 arch/i386/isa/npx.c p = ci->ci_curproc; ci 679 arch/i386/isa/npx.c IPRINTF(("%s: dna for %lx%s\n", ci->ci_dev.dv_xname, (u_long)p, ci 687 arch/i386/isa/npx.c if (ci->ci_fpcurproc != NULL) { ci 688 arch/i386/isa/npx.c IPRINTF(("%s: fp save %lx\n", ci->ci_dev.dv_xname, ci 689 arch/i386/isa/npx.c (u_long)ci->ci_fpcurproc)); ci 690 arch/i386/isa/npx.c npxsave_cpu(ci, 1); ci 693 arch/i386/isa/npx.c IPRINTF(("%s: fp init\n", ci->ci_dev.dv_xname)); ci 700 arch/i386/isa/npx.c IPRINTF(("%s: done saving\n", ci->ci_dev.dv_xname)); ci 701 arch/i386/isa/npx.c KDASSERT(ci->ci_fpcurproc == NULL); ci 711 arch/i386/isa/npx.c ci->ci_fpcurproc = p; ci 712 arch/i386/isa/npx.c p->p_addr->u_pcb.pcb_fpcpu = ci; ci 748 arch/i386/isa/npx.c npxsave_cpu(struct cpu_info *ci, int save) ci 753 arch/i386/isa/npx.c KDASSERT(ci == curcpu()); ci 755 arch/i386/isa/npx.c p = ci->ci_fpcurproc; ci 759 arch/i386/isa/npx.c IPRINTF(("%s: fp cpu %s %lx\n", ci->ci_dev.dv_xname, ci 764 arch/i386/isa/npx.c if (ci->ci_fpsaving != 0) ci 781 arch/i386/isa/npx.c ci->ci_fpsaving = 1; ci 783 arch/i386/isa/npx.c ci->ci_fpsaving = 0; ci 797 arch/i386/isa/npx.c ci->ci_fpcurproc = NULL; ci 807 arch/i386/isa/npx.c struct cpu_info *ci = curcpu(); ci 816 arch/i386/isa/npx.c IPRINTF(("%s: fp proc %s %lx\n", ci->ci_dev.dv_xname, ci 820 arch/i386/isa/npx.c if (oci == ci) { ci 822 arch/i386/isa/npx.c npxsave_cpu(ci, save); ci 829 arch/i386/isa/npx.c IPRINTF(("%s: fp ipi to %s %s %lx\n", ci->ci_dev.dv_xname, ci 843 arch/i386/isa/npx.c ci->ci_dev.dv_xname, oci->ci_dev.dv_xname); ci 848 arch/i386/isa/npx.c KASSERT(ci->ci_fpcurproc == p); ci 849 arch/i386/isa/npx.c npxsave_cpu(ci, save); ci 180 arch/i386/pci/ichpcib.c struct cpu_info *ci; ci 186 arch/i386/pci/ichpcib.c ci = curcpu(); ci 187 arch/i386/pci/ichpcib.c family = (ci->ci_signature >> 8) & 15; ci 188 arch/i386/pci/ichpcib.c model = (ci->ci_signature >> 4) & 15; ci 189 arch/i386/pci/ichpcib.c stepping = ci->ci_signature & 15; ci 269 dev/ccd.c struct ccdcinfo *ci = NULL; ci 307 dev/ccd.c ci = &cs->sc_cinfo[ix]; ci 308 dev/ccd.c ci->ci_vp = vp; ci 315 dev/ccd.c MAXPATHLEN, &ci->ci_pathlen); ci 323 dev/ccd.c ci->ci_path = malloc(ci->ci_pathlen, M_DEVBUF, M_WAITOK); ci 324 dev/ccd.c bcopy(tmppath, ci->ci_path, ci->ci_pathlen); ci 332 dev/ccd.c cs->sc_xname, ci->ci_path, error)); ci 333 dev/ccd.c free(ci->ci_path, M_DEVBUF); ci 337 dev/ccd.c ci->ci_dev = va.va_rdev; ci 347 dev/ccd.c cs->sc_xname, ci->ci_path, error)); ci 348 dev/ccd.c free(ci->ci_path, M_DEVBUF); ci 361 dev/ccd.c cs->sc_xname, ci->ci_path)); ci 362 dev/ccd.c free(ci->ci_path, M_DEVBUF); ci 376 dev/ccd.c ("%s: %s: size == 0\n", cs->sc_xname, ci->ci_path)); ci 377 dev/ccd.c free(ci->ci_path, M_DEVBUF); ci 384 dev/ccd.c ci->ci_size = size; ci 399 dev/ccd.c free(ci->ci_path, M_DEVBUF); ci 414 dev/ccd.c free(ci->ci_path, M_DEVBUF); ci 422 dev/ccd.c free(ci->ci_path, M_DEVBUF); ci 434 dev/ccd.c for (ci = cs->sc_cinfo; ci 435 dev/ccd.c ci < &cs->sc_cinfo[cs->sc_nccdisks]; ci++) ci 436 dev/ccd.c ci->ci_size = minsize; ci 468 dev/ccd.c struct ccdcinfo *ci, *smallci; ci 523 dev/ccd.c for (ci = cs->sc_cinfo; ci 524 dev/ccd.c ci < &cs->sc_cinfo[cs->sc_nccdisks]; ci++) ci 525 dev/ccd.c if (ci->ci_size > size && ci 527 dev/ccd.c ci->ci_size < smallci->ci_size)) ci 528 dev/ccd.c smallci = ci; ci 549 dev/ccd.c for (ci = cs->sc_cinfo; ci 550 dev/ccd.c ci < &cs->sc_cinfo[cs->sc_nccunits]; ci++) ci 551 dev/ccd.c if (ci->ci_size >= smallci->ci_size) ci 552 dev/ccd.c ii->ii_index[ix++] = ci - cs->sc_cinfo; ci 766 dev/ccd.c struct ccdcinfo *ci, *ci2 = NULL; ci 788 dev/ccd.c for (ccdisk = 0, ci = &cs->sc_cinfo[ccdisk]; ci 789 dev/ccd.c cbn >= sblk + ci->ci_size; ci 790 dev/ccd.c ccdisk++, ci = &cs->sc_cinfo[ccdisk]) ci 791 dev/ccd.c sblk += ci->ci_size; ci 819 dev/ccd.c ci->ci_flags & CCIF_FAILED)) ci 823 dev/ccd.c ci = &cs->sc_cinfo[ccdisk]; ci 825 dev/ccd.c ccdisk, cbn, ci, ci2)); ci 830 dev/ccd.c cnt = dbtob(ci->ci_size - cbn); ci 845 dev/ccd.c nbp->b_dev = ci->ci_dev; /* XXX */ ci 847 dev/ccd.c nbp->b_vp = ci->ci_vp; ci 879 dev/ccd.c ci->ci_dev, ci-cs->sc_cinfo, cbp, bp->b_blkno, ci 107 dev/i2o/iopsp.c struct i2o_param_hba_ctlr_info ci; ci 123 dev/i2o/iopsp.c printf("iopsp_match: bustype = %d\n", param.ci.bustype); ci 126 dev/i2o/iopsp.c return (param.ci.bustype == I2O_HBA_BUS_SCSI || ci 127 dev/i2o/iopsp.c param.ci.bustype == I2O_HBA_BUS_FCA); ci 144 dev/i2o/iopsp.c struct i2o_param_hba_ctlr_info ci; ci 171 dev/i2o/iopsp.c fcal = (param.p.ci.bustype == I2O_HBA_BUS_FCA); /* XXX */ ci 1888 dev/ic/aac.c u_int32_t pi, ci; ci 1898 dev/ic/aac.c ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX]; ci 1905 dev/ic/aac.c if ((pi + 1) == ci) { ci 1941 dev/ic/aac.c u_int32_t pi, ci; ci 1948 dev/ic/aac.c ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX]; ci 1951 dev/ic/aac.c if (ci == pi) { ci 1961 dev/ic/aac.c if (ci == pi + 1) ci 1965 dev/ic/aac.c if (ci >= aac_qinfo[queue].size) ci 1966 dev/ic/aac.c ci = 0; ci 1969 dev/ic/aac.c *fib_size = (sc->aac_qentries[queue] + ci)->aq_fib_size; ci 1980 dev/ic/aac.c fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr / ci 1997 dev/ic/aac.c fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr; ci 2018 dev/ic/aac.c sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX] = ci + 1; ci 2035 dev/ic/aac.c u_int32_t pi, ci; ci 2047 dev/ic/aac.c ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX]; ci 2054 dev/ic/aac.c if ((pi + 1) == ci) { ci 443 dev/ic/pdq.c pdq_command_info_t * const ci = &pdq->pdq_command_info; ci 452 dev/ic/pdq.c if (ci->ci_command_active || ci->ci_pending_commands == 0) ci 459 dev/ic/pdq.c for (mask = 1 << ((int) op); (mask & ci->ci_pending_commands) == 0; mask >>= 1) ci 479 dev/ic/pdq.c dbp->pdqdb_command_requests[ci->ci_request_producer].txd_seg_len = cmdlen; ci 480 dev/ic/pdq.c PDQ_ADVANCE(ci->ci_request_producer, 1, PDQ_RING_MASK(dbp->pdqdb_command_requests)); ci 486 dev/ic/pdq.c dbp->pdqdb_command_responses[ci->ci_response_producer].rxd_seg_len_hi = cmdlen / 16; ci 487 dev/ic/pdq.c PDQ_ADVANCE(ci->ci_response_producer, 1, PDQ_RING_MASK(dbp->pdqdb_command_responses)); ci 494 dev/ic/pdq.c PDQ_OS_MEMZERO(ci->ci_bufstart, cmdlen); ci 495 dev/ic/pdq.c *(pdq_cmd_code_t *) ci->ci_bufstart = op; ci 496 dev/ic/pdq.c ci->ci_pending_commands &= ~mask; ci 503 dev/ic/pdq.c pdq_cmd_filter_set_t *filter_set = (pdq_cmd_filter_set_t *) ci->ci_bufstart; ci 521 dev/ic/pdq.c pdq_cmd_addr_filter_set_t *addr_filter_set = (pdq_cmd_addr_filter_set_t *) ci->ci_bufstart; ci 543 dev/ic/pdq.c ci->ci_command_active++; ci 544 dev/ic/pdq.c PDQ_CSR_WRITE(csrs, csr_cmd_response_producer, ci->ci_response_producer | (ci->ci_response_completion << 8)); ci 545 dev/ic/pdq.c PDQ_CSR_WRITE(csrs, csr_cmd_request_producer, ci->ci_request_producer | (ci->ci_request_completion << 8)); ci 553 dev/ic/pdq.c pdq_command_info_t * const ci = &pdq->pdq_command_info; ci 564 dev/ic/pdq.c if (cbp->pdqcb_command_response == ci->ci_response_completion) ci 567 dev/ic/pdq.c PDQ_ASSERT(cbp->pdqcb_command_request != ci->ci_request_completion); ci 569 dev/ic/pdq.c rspgen = (const pdq_response_generic_t *) ci->ci_bufstart; ci 580 dev/ic/pdq.c PDQ_ADVANCE(ci->ci_request_completion, 1, PDQ_RING_MASK(dbp->pdqdb_command_requests)); ci 581 dev/ic/pdq.c PDQ_ADVANCE(ci->ci_response_completion, 1, PDQ_RING_MASK(dbp->pdqdb_command_responses)); ci 582 dev/ic/pdq.c ci->ci_command_active = 0; ci 584 dev/ic/pdq.c if (ci->ci_pending_commands != 0) { ci 588 dev/ic/pdq.c ci->ci_response_producer | (ci->ci_response_completion << 8)); ci 590 dev/ic/pdq.c ci->ci_request_producer | (ci->ci_request_completion << 8)); ci 437 dev/usb/uvisor.c uvisor_init(struct uvisor_softc *sc, struct uvisor_connection_info *ci, ci 474 dev/usb/uvisor.c err = usbd_do_request_flags(sc->sc_udev, &req, ci, ci 300 isofs/cd9660/cd9660_vnops.c struct cluster_info *ci = &ip->i_ci; ci 314 isofs/cd9660/cd9660_vnops.c if (ci->ci_lastr + 1 == lbn) { ci 334 isofs/cd9660/cd9660_vnops.c ci->ci_lastr = lbn; ci 229 kern/kern_clock.c struct cpu_info *ci = curcpu(); ci 254 kern/kern_clock.c if (--ci->ci_schedstate.spc_rrticks <= 0) ci 255 kern/kern_clock.c roundrobin(ci); ci 261 kern/kern_clock.c if (CPU_IS_PRIMARY(ci) == 0) ci 481 kern/kern_clock.c struct cpu_info *ci = curcpu(); ci 482 kern/kern_clock.c struct schedstate_percpu *spc = &ci->ci_schedstate; ci 438 kern/kern_sysctl.c struct cpu_info *ci; ci 443 kern/kern_sysctl.c CPU_INFO_FOREACH(cii, ci) { ci 445 kern/kern_sysctl.c cp_time[i] += ci->ci_schedstate.spc_cp_time[i]; ci 1874 kern/kern_sysctl.c struct cpu_info *ci; ci 1882 kern/kern_sysctl.c CPU_INFO_FOREACH(cii, ci) { ci 1890 kern/kern_sysctl.c &ci->ci_schedstate.spc_cp_time, ci 1891 kern/kern_sysctl.c sizeof(ci->ci_schedstate.spc_cp_time))); ci 95 kern/sched_bsd.c roundrobin(struct cpu_info *ci) ci 97 kern/sched_bsd.c struct schedstate_percpu *spc = &ci->ci_schedstate; ci 480 kern/sched_bsd.c struct cpu_info *ci; ci 504 kern/sched_bsd.c ci = (p->p_cpu != NULL) ? p->p_cpu : curcpu(); ci 505 kern/sched_bsd.c if (pri < ci->ci_schedstate.spc_curpriority) ci 506 kern/sched_bsd.c need_resched(ci); ci 62 kern/vfs_cluster.c cluster_write(struct buf *bp, struct cluster_info *ci, u_quad_t filesize) ci 73 kern/vfs_cluster.c ci->ci_lasta = ci->ci_clen = ci->ci_cstart = ci->ci_lastw = 0; ci 75 kern/vfs_cluster.c if (ci->ci_clen == 0 || lbn != ci->ci_lastw + 1 || ci 76 kern/vfs_cluster.c (bp->b_blkno != ci->ci_lasta + btodb(bp->b_bcount))) { ci 78 kern/vfs_cluster.c if (ci->ci_clen != 0) { ci 88 kern/vfs_cluster.c cursize = ci->ci_lastw - ci->ci_cstart + 1; ci 90 kern/vfs_cluster.c lbn != ci->ci_lastw + 1 || ci->ci_clen <= cursize) { ci 92 kern/vfs_cluster.c ci->ci_cstart, cursize, lbn); ci 97 kern/vfs_cluster.c buflist = cluster_collectbufs(vp, ci, bp); ci 109 kern/vfs_cluster.c ci->ci_cstart, cursize, lbn); ci 118 kern/vfs_cluster.c ci->ci_lastw = lbn; ci 119 kern/vfs_cluster.c ci->ci_lasta = bp->b_blkno; ci 133 kern/vfs_cluster.c ci->ci_clen = 0; ci 134 kern/vfs_cluster.c ci->ci_lasta = bp->b_blkno; ci 135 kern/vfs_cluster.c ci->ci_cstart = lbn + 1; ci 136 kern/vfs_cluster.c ci->ci_lastw = lbn; ci 139 kern/vfs_cluster.c ci->ci_clen = maxclen; ci 141 kern/vfs_cluster.c ci->ci_cstart = lbn + 1; ci 144 kern/vfs_cluster.c ci->ci_cstart = lbn; ci 147 kern/vfs_cluster.c } else if (lbn == ci->ci_cstart + ci->ci_clen) { ci 151 kern/vfs_cluster.c cluster_wbuild(vp, bp, bp->b_bcount, ci->ci_cstart, ci 152 kern/vfs_cluster.c ci->ci_clen + 1, lbn); ci 153 kern/vfs_cluster.c ci->ci_clen = 0; ci 154 kern/vfs_cluster.c ci->ci_cstart = lbn + 1; ci 161 kern/vfs_cluster.c ci->ci_lastw = lbn; ci 162 kern/vfs_cluster.c ci->ci_lasta = bp->b_blkno; ci 225 kern/vfs_cluster.c cluster_collectbufs(struct vnode *vp, struct cluster_info *ci, ci 232 kern/vfs_cluster.c len = ci->ci_lastw - ci->ci_cstart + 1; ci 237 kern/vfs_cluster.c for (lbn = ci->ci_cstart, i = 0; i < len; lbn++, i++)