ci 235 arch/i386/i386/amd64errata.c amd64_errata_testmsr(struct cpu_info *ci, errata_t *e)
ci 239 arch/i386/i386/amd64errata.c (void)ci;
ci 250 arch/i386/i386/amd64errata.c amd64_errata_setmsr(struct cpu_info *ci, errata_t *e)
ci 254 arch/i386/i386/amd64errata.c (void)ci;
ci 274 arch/i386/i386/amd64errata.c amd64_errata(struct cpu_info *ci)
ci 320 arch/i386/i386/amd64errata.c ci->ci_dev.dv_xname, e->e_num);
ci 327 arch/i386/i386/amd64errata.c rc = (e->e_act == NULL) ? 1 : (*e->e_act)(ci, e);
ci 340 arch/i386/i386/amd64errata.c ci->ci_dev.dv_xname, e->e_num,
ci 351 arch/i386/i386/amd64errata.c printf("%s: AMD %s", ci->ci_dev.dv_xname,
ci 369 arch/i386/i386/amd64errata.c printf("%s: AMD %s", ci->ci_dev.dv_xname,
ci 187 arch/i386/i386/cpu.c struct cpu_info *ci = (struct cpu_info *)self;
ci 202 arch/i386/i386/cpu.c ci = &cpu_info_primary;
ci 203 arch/i386/i386/cpu.c bcopy(self, &ci->ci_dev, sizeof *self);
ci 213 arch/i386/i386/cpu.c cpu_info[cpunum] = ci;
ci 216 arch/i386/i386/cpu.c ci->ci_self = ci;
ci 217 arch/i386/i386/cpu.c ci->ci_apicid = caa->cpu_number;
ci 219 arch/i386/i386/cpu.c ci->ci_cpuid = ci->ci_apicid;
ci 221 arch/i386/i386/cpu.c ci->ci_cpuid = 0; /* False for APs, so what, they're not used */
ci 223 arch/i386/i386/cpu.c ci->ci_signature = caa->cpu_signature;
ci 224 arch/i386/i386/cpu.c ci->ci_feature_flags = caa->feature_flags;
ci 225 arch/i386/i386/cpu.c ci->ci_func = caa->cpu_func;
ci 239 arch/i386/i386/cpu.c ci->ci_dev.dv_xname);
ci 242 arch/i386/i386/cpu.c pcb = ci->ci_idle_pcb = (struct pcb *)kstack;
ci 254 arch/i386/i386/cpu.c cpu_default_ldt(ci); /* Use the `global' ldt until one alloc'd */
ci 266 arch/i386/i386/cpu.c ci->ci_flags |= CPUF_PRESENT | CPUF_SP | CPUF_PRIMARY;
ci 267 arch/i386/i386/cpu.c identifycpu(ci);
ci 268 arch/i386/i386/cpu.c cpu_init(ci);
ci 273 arch/i386/i386/cpu.c ci->ci_flags |= CPUF_PRESENT | CPUF_BSP | CPUF_PRIMARY;
ci 274 arch/i386/i386/cpu.c identifycpu(ci);
ci 275 arch/i386/i386/cpu.c cpu_init(ci);
ci 282 arch/i386/i386/cpu.c lapic_calibrate_timer(ci);
ci 294 arch/i386/i386/cpu.c gdt_alloc_cpu(ci);
ci 295 arch/i386/i386/cpu.c cpu_alloc_ldt(ci);
ci 296 arch/i386/i386/cpu.c ci->ci_flags |= CPUF_PRESENT | CPUF_AP;
ci 297 arch/i386/i386/cpu.c identifycpu(ci);
ci 298 arch/i386/i386/cpu.c ci->ci_next = cpu_info_list->ci_next;
ci 299 arch/i386/i386/cpu.c cpu_info_list->ci_next = ci;
ci 308 arch/i386/i386/cpu.c if (ci->ci_apicid < IOAPIC_ID_MAX)
ci 309 arch/i386/i386/cpu.c ioapic_id_map &= ~(1 << ci->ci_apicid);
ci 313 arch/i386/i386/cpu.c ci->ci_dev.dv_xname, kstack, USPACE);
ci 315 arch/i386/i386/cpu.c ci->ci_dev.dv_xname, pcb, pcb->pcb_esp);
ci 328 arch/i386/i386/cpu.c cpu_init(struct cpu_info *ci)
ci 331 arch/i386/i386/cpu.c if (ci->cpu_setup != NULL)
ci 332 arch/i386/i386/cpu.c (*ci->cpu_setup)(ci);
ci 343 arch/i386/i386/cpu.c ci->ci_flags |= CPUF_RUNNING;
ci 363 arch/i386/i386/cpu.c struct cpu_info *ci;
ci 367 arch/i386/i386/cpu.c ci = cpu_info[i];
ci 368 arch/i386/i386/cpu.c if (ci == NULL)
ci 370 arch/i386/i386/cpu.c if (ci->ci_idle_pcb == NULL)
ci 372 arch/i386/i386/cpu.c if ((ci->ci_flags & CPUF_PRESENT) == 0)
ci 374 arch/i386/i386/cpu.c if (ci->ci_flags & (CPUF_BSP|CPUF_SP|CPUF_PRIMARY))
ci 376 arch/i386/i386/cpu.c cpu_boot_secondary(ci);
ci 383 arch/i386/i386/cpu.c struct cpu_info *ci;
ci 387 arch/i386/i386/cpu.c ci = cpu_info[i];
ci 388 arch/i386/i386/cpu.c if (ci == NULL)
ci 390 arch/i386/i386/cpu.c if (ci->ci_idle_pcb == NULL)
ci 392 arch/i386/i386/cpu.c if ((ci->ci_flags & CPUF_PRESENT) == 0)
ci 394 arch/i386/i386/cpu.c i386_init_pcb_tss_ldt(ci);
ci 399 arch/i386/i386/cpu.c cpu_boot_secondary(struct cpu_info *ci)
ci 407 arch/i386/i386/cpu.c printf("%s: starting", ci->ci_dev.dv_xname);
ci 412 arch/i386/i386/cpu.c pcb = ci->ci_idle_pcb;
ci 417 arch/i386/i386/cpu.c CPU_STARTUP(ci);
ci 422 arch/i386/i386/cpu.c for (i = 100000; (!(ci->ci_flags & CPUF_RUNNING)) && i > 0; i--) {
ci 425 arch/i386/i386/cpu.c if (!(ci->ci_flags & CPUF_RUNNING)) {
ci 426 arch/i386/i386/cpu.c printf("%s failed to become ready\n", ci->ci_dev.dv_xname);
ci 432 arch/i386/i386/cpu.c CPU_START_CLEANUP(ci);
ci 443 arch/i386/i386/cpu.c struct cpu_info *ci = (struct cpu_info *)v;
ci 450 arch/i386/i386/cpu.c gdt_init_cpu(ci);
ci 451 arch/i386/i386/cpu.c cpu_init_ldt(ci);
ci 452 arch/i386/i386/cpu.c npxinit(ci);
ci 456 arch/i386/i386/cpu.c cpu_init(ci);
ci 463 arch/i386/i386/cpu.c ci->ci_dev.dv_xname, ci->ci_cpuid);
ci 464 arch/i386/i386/cpu.c microuptime(&ci->ci_schedstate.spc_runtime);
ci 514 arch/i386/i386/cpu.c cpu_set_tss_gates(struct cpu_info *ci)
ci 518 arch/i386/i386/cpu.c ci->ci_doubleflt_stack = (char *)uvm_km_alloc(kernel_map, USPACE);
ci 519 arch/i386/i386/cpu.c cpu_init_tss(&ci->ci_doubleflt_tss, ci->ci_doubleflt_stack,
ci 521 arch/i386/i386/cpu.c setsegment(&sd, &ci->ci_doubleflt_tss, sizeof(struct i386tss) - 1,
ci 523 arch/i386/i386/cpu.c ci->ci_gdt[GTRAPTSS_SEL].sd = sd;
ci 535 arch/i386/i386/cpu.c ci->ci_ddbipi_stack = (char *)uvm_km_alloc(kernel_map, USPACE);
ci 536 arch/i386/i386/cpu.c cpu_init_tss(&ci->ci_ddbipi_tss, ci->ci_ddbipi_stack,
ci 539 arch/i386/i386/cpu.c setsegment(&sd, &ci->ci_ddbipi_tss, sizeof(struct i386tss) - 1,
ci 541 arch/i386/i386/cpu.c ci->ci_gdt[GIPITSS_SEL].sd = sd;
ci 551 arch/i386/i386/cpu.c mp_cpu_start(struct cpu_info *ci)
ci 582 arch/i386/i386/cpu.c if (ci->ci_flags & CPUF_AP) {
ci 583 arch/i386/i386/cpu.c if ((error = i386_ipi_init(ci->ci_apicid)) != 0)
ci 590 arch/i386/i386/cpu.c ci->ci_apicid, LAPIC_DLMODE_STARTUP)) != 0)
ci 595 arch/i386/i386/cpu.c ci->ci_apicid, LAPIC_DLMODE_STARTUP)) != 0)
ci 605 arch/i386/i386/cpu.c mp_cpu_start_cleanup(struct cpu_info *ci)
ci 138 arch/i386/i386/db_mp.c i386_ipi_db(struct cpu_info *ci)
ci 110 arch/i386/i386/gdt.c struct cpu_info *ci;
ci 115 arch/i386/i386/gdt.c CPU_INFO_FOREACH(cii, ci)
ci 116 arch/i386/i386/gdt.c if (ci->ci_gdt != NULL && ci->ci_gdt != gdt)
ci 117 arch/i386/i386/gdt.c ci->ci_gdt[sel].sd = *sd;
ci 129 arch/i386/i386/gdt.c struct cpu_info *ci = &cpu_info_primary;
ci 147 arch/i386/i386/gdt.c ci->ci_gdt = gdt;
ci 148 arch/i386/i386/gdt.c setsegment(&ci->ci_gdt[GCPU_SEL].sd, ci, sizeof(struct cpu_info)-1,
ci 151 arch/i386/i386/gdt.c gdt_init_cpu(ci);
ci 159 arch/i386/i386/gdt.c gdt_alloc_cpu(struct cpu_info *ci)
ci 164 arch/i386/i386/gdt.c ci->ci_gdt = (union descriptor *)uvm_km_valloc(kernel_map, max_len);
ci 165 arch/i386/i386/gdt.c uvm_map_pageable(kernel_map, (vaddr_t)ci->ci_gdt,
ci 166 arch/i386/i386/gdt.c (vaddr_t)ci->ci_gdt + min_len, FALSE, FALSE);
ci 167 arch/i386/i386/gdt.c bzero(ci->ci_gdt, min_len);
ci 168 arch/i386/i386/gdt.c bcopy(gdt, ci->ci_gdt, gdt_size * sizeof(union descriptor));
ci 169 arch/i386/i386/gdt.c setsegment(&ci->ci_gdt[GCPU_SEL].sd, ci, sizeof(struct cpu_info)-1,
ci 180 arch/i386/i386/gdt.c gdt_init_cpu(struct cpu_info *ci)
ci 184 arch/i386/i386/gdt.c setregion(®ion, ci->ci_gdt,
ci 197 arch/i386/i386/gdt.c struct cpu_info *ci;
ci 205 arch/i386/i386/gdt.c CPU_INFO_FOREACH(cii, ci) {
ci 206 arch/i386/i386/gdt.c for (va = (vaddr_t)(ci->ci_gdt) + old_len;
ci 207 arch/i386/i386/gdt.c va < (vaddr_t)(ci->ci_gdt) + new_len;
ci 97 arch/i386/i386/ipifuncs.c i386_ipi_halt(struct cpu_info *ci)
ci 101 arch/i386/i386/ipifuncs.c printf("%s: shutting down\n", ci->ci_dev.dv_xname);
ci 109 arch/i386/i386/ipifuncs.c i386_ipi_flush_fpu(struct cpu_info *ci)
ci 111 arch/i386/i386/ipifuncs.c npxsave_cpu(ci, 0);
ci 115 arch/i386/i386/ipifuncs.c i386_ipi_synch_fpu(struct cpu_info *ci)
ci 117 arch/i386/i386/ipifuncs.c npxsave_cpu(ci, 1);
ci 128 arch/i386/i386/ipifuncs.c i386_send_ipi(struct cpu_info *ci, int ipimask)
ci 132 arch/i386/i386/ipifuncs.c i386_atomic_setbits_l(&ci->ci_ipis, ipimask);
ci 135 arch/i386/i386/ipifuncs.c if (!(ci->ci_flags & CPUF_RUNNING))
ci 138 arch/i386/i386/ipifuncs.c ret = i386_ipi(LAPIC_IPI_VECTOR, ci->ci_cpuid, LAPIC_DLMODE_FIXED);
ci 141 arch/i386/i386/ipifuncs.c ipimask, curcpu()->ci_dev.dv_xname, ci->ci_dev.dv_xname);
ci 148 arch/i386/i386/ipifuncs.c i386_fast_ipi(struct cpu_info *ci, int ipi)
ci 150 arch/i386/i386/ipifuncs.c if (!(ci->ci_flags & CPUF_RUNNING))
ci 153 arch/i386/i386/ipifuncs.c return (i386_ipi(ipi, ci->ci_cpuid, LAPIC_DLMODE_FIXED));
ci 167 arch/i386/i386/ipifuncs.c struct cpu_info *ci, *self = curcpu();
ci 171 arch/i386/i386/ipifuncs.c CPU_INFO_FOREACH(cii, ci) {
ci 172 arch/i386/i386/ipifuncs.c if (ci == self)
ci 174 arch/i386/i386/ipifuncs.c if ((ci->ci_flags & CPUF_RUNNING) == 0)
ci 176 arch/i386/i386/ipifuncs.c i386_atomic_setbits_l(&ci->ci_ipis, ipimask);
ci 189 arch/i386/i386/ipifuncs.c struct cpu_info *ci = curcpu();
ci 193 arch/i386/i386/ipifuncs.c pending = i386_atomic_testset_ul(&ci->ci_ipis, 0);
ci 198 arch/i386/i386/ipifuncs.c (*ipifunc[bit])(ci);
ci 131 arch/i386/i386/lapic.c struct cpu_info *ci = curcpu();
ci 137 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "prelint", 0, 0,
ci 139 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "prelint", 1, 0,
ci 147 arch/i386/i386/lapic.c || mpi->cpu_id == ci->ci_apicid)) {
ci 162 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "timer", 0, 0,
ci 164 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "pcint", 0, 0,
ci 166 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "lint", 0, 0,
ci 168 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "lint", 1, 0,
ci 170 arch/i386/i386/lapic.c apic_format_redir(ci->ci_dev.dv_xname, "err", 0, 0,
ci 224 arch/i386/i386/lapic.c struct cpu_info *ci = curcpu();
ci 227 arch/i386/i386/lapic.c if (CPU_IS_PRIMARY(ci)) {
ci 228 arch/i386/i386/lapic.c ci->ci_tscbase = rdtsc();
ci 267 arch/i386/i386/lapic.c lapic_calibrate_timer(struct cpu_info *ci)
ci 276 arch/i386/i386/lapic.c printf("%s: calibrating local timer\n", ci->ci_dev.dv_xname);
ci 351 arch/i386/i386/lapic.c printf("%s: apic clock running at %s\n", ci->ci_dev.dv_xname, tbuf);
ci 432 arch/i386/i386/lapic.c i386_ipi_microset(struct cpu_info *ci)
ci 434 arch/i386/i386/lapic.c ci->ci_tscbase = rdtsc();
ci 445 arch/i386/i386/lapic.c struct cpu_info *ci = curcpu();
ci 451 arch/i386/i386/lapic.c tmp = rdtsc() - ci->ci_tscbase;
ci 502 arch/i386/i386/machdep.c i386_init_pcb_tss_ldt(struct cpu_info *ci)
ci 505 arch/i386/i386/machdep.c struct pcb *pcb = ci->ci_idle_pcb;
ci 514 arch/i386/i386/machdep.c pcb->pcb_ldt = ci->ci_ldt;
ci 516 arch/i386/i386/machdep.c ci->ci_idle_tss_sel = tss_alloc(pcb);
ci 1028 arch/i386/i386/machdep.c winchip_cpu_setup(struct cpu_info *ci)
ci 1032 arch/i386/i386/machdep.c switch ((ci->ci_signature >> 4) & 15) { /* model */
ci 1034 arch/i386/i386/machdep.c ci->ci_feature_flags &= ~CPUID_TSC;
ci 1037 arch/i386/i386/machdep.c printf("%s: TSC disabled\n", ci->ci_dev.dv_xname);
ci 1045 arch/i386/i386/machdep.c cyrix3_setperf_setup(struct cpu_info *ci)
ci 1049 arch/i386/i386/machdep.c est_init(ci->ci_dev.dv_xname, CPUVENDOR_VIA);
ci 1052 arch/i386/i386/machdep.c ci->ci_dev.dv_xname);
ci 1058 arch/i386/i386/machdep.c cyrix3_cpu_setup(struct cpu_info *ci)
ci 1061 arch/i386/i386/machdep.c int model = (ci->ci_signature >> 4) & 15;
ci 1062 arch/i386/i386/machdep.c int step = ci->ci_signature & 15;
ci 1073 arch/i386/i386/machdep.c cyrix3_get_bus_clock(ci);
ci 1118 arch/i386/i386/machdep.c printf("%s:", ci->ci_dev.dv_xname);
ci 1192 arch/i386/i386/machdep.c cyrix6x86_cpu_setup(struct cpu_info *ci)
ci 1196 arch/i386/i386/machdep.c switch ((ci->ci_signature >> 4) & 15) { /* model */
ci 1213 arch/i386/i386/machdep.c ci->ci_dev.dv_xname);
ci 1219 arch/i386/i386/machdep.c printf("%s: TSC disabled\n", ci->ci_dev.dv_xname);
ci 1225 arch/i386/i386/machdep.c natsem6x86_cpu_setup(struct cpu_info *ci)
ci 1229 arch/i386/i386/machdep.c int model = (ci->ci_signature >> 4) & 15;
ci 1235 arch/i386/i386/machdep.c printf("%s: TSC disabled\n", ci->ci_dev.dv_xname);
ci 1242 arch/i386/i386/machdep.c intel586_cpu_setup(struct cpu_info *ci)
ci 1248 arch/i386/i386/machdep.c ci->ci_dev.dv_xname);
ci 1255 arch/i386/i386/machdep.c amd_family5_setperf_setup(struct cpu_info *ci)
ci 1262 arch/i386/i386/machdep.c amd_family5_setup(struct cpu_info *ci)
ci 1264 arch/i386/i386/machdep.c int model = (ci->ci_signature >> 4) & 15;
ci 1292 arch/i386/i386/machdep.c amd_family6_setperf_setup(struct cpu_info *ci)
ci 1294 arch/i386/i386/machdep.c int family = (ci->ci_signature >> 8) & 15;
ci 1308 arch/i386/i386/machdep.c amd_family6_setup(struct cpu_info *ci)
ci 1338 arch/i386/i386/machdep.c struct cpu_info *ci = (struct cpu_info *) args;
ci 1347 arch/i386/i386/machdep.c ci->ci_sensor.value = max - MSR_THERM_STATUS_TEMP(msr);
ci 1349 arch/i386/i386/machdep.c ci->ci_sensor.value *= 1000000;
ci 1351 arch/i386/i386/machdep.c ci->ci_sensor.value += 273150000;
ci 1352 arch/i386/i386/machdep.c ci->ci_sensor.flags &= ~SENSOR_FINVALID;
ci 1354 arch/i386/i386/machdep.c ci->ci_sensor.value = 0;
ci 1355 arch/i386/i386/machdep.c ci->ci_sensor.flags |= SENSOR_FINVALID;
ci 1360 arch/i386/i386/machdep.c intel686_cpusensors_setup(struct cpu_info *ci)
ci 1373 arch/i386/i386/machdep.c strlcpy(ci->ci_sensordev.xname, ci->ci_dev.dv_xname,
ci 1374 arch/i386/i386/machdep.c sizeof(ci->ci_sensordev.xname));
ci 1375 arch/i386/i386/machdep.c ci->ci_sensor.type = SENSOR_TEMP;
ci 1376 arch/i386/i386/machdep.c sensor_task_register(ci, intelcore_update_sensor, 5);
ci 1377 arch/i386/i386/machdep.c sensor_attach(&ci->ci_sensordev, &ci->ci_sensor);
ci 1378 arch/i386/i386/machdep.c sensordev_install(&ci->ci_sensordev);
ci 1384 arch/i386/i386/machdep.c intel686_setperf_setup(struct cpu_info *ci)
ci 1386 arch/i386/i386/machdep.c int family = (ci->ci_signature >> 8) & 15;
ci 1387 arch/i386/i386/machdep.c int step = ci->ci_signature & 15;
ci 1391 arch/i386/i386/machdep.c est_init(ci->ci_dev.dv_xname, CPUVENDOR_INTEL);
ci 1394 arch/i386/i386/machdep.c ci->ci_dev.dv_xname);
ci 1402 arch/i386/i386/machdep.c intel686_common_cpu_setup(struct cpu_info *ci)
ci 1427 arch/i386/i386/machdep.c intel686_cpu_setup(struct cpu_info *ci)
ci 1429 arch/i386/i386/machdep.c int model = (ci->ci_signature >> 4) & 15;
ci 1430 arch/i386/i386/machdep.c int step = ci->ci_signature & 15;
ci 1434 arch/i386/i386/machdep.c p3_get_bus_clock(ci);
ci 1437 arch/i386/i386/machdep.c intel686_common_cpu_setup(ci);
ci 1444 arch/i386/i386/machdep.c ci->ci_feature_flags &= ~CPUID_SEP;
ci 1449 arch/i386/i386/machdep.c if ((model == 7) && (ci->ci_feature_flags & CPUID_SER)) {
ci 1455 arch/i386/i386/machdep.c ci->ci_dev.dv_xname);
ci 1456 arch/i386/i386/machdep.c ci->ci_feature_flags &= ~CPUID_SER;
ci 1457 arch/i386/i386/machdep.c ci->ci_level = 2;
ci 1467 arch/i386/i386/machdep.c intel686_p4_cpu_setup(struct cpu_info *ci)
ci 1470 arch/i386/i386/machdep.c p4_get_bus_clock(ci);
ci 1473 arch/i386/i386/machdep.c intel686_common_cpu_setup(ci);
ci 1476 arch/i386/i386/machdep.c p4_model = (ci->ci_signature >> 4) & 15;
ci 1482 arch/i386/i386/machdep.c tm86_cpu_setup(struct cpu_info *ci)
ci 1560 arch/i386/i386/machdep.c identifycpu(struct cpu_info *ci)
ci 1567 arch/i386/i386/machdep.c char *cpu_device = ci->ci_dev.dv_xname;
ci 1582 arch/i386/i386/machdep.c ci->cpu_setup = i386_nocpuid_cpus[cpu].cpu_setup;
ci 1587 arch/i386/i386/machdep.c modif = (ci->ci_signature >> 12) & 3;
ci 1588 arch/i386/i386/machdep.c family = (ci->ci_signature >> 8) & 15;
ci 1589 arch/i386/i386/machdep.c model = (ci->ci_signature >> 4) & 15;
ci 1590 arch/i386/i386/machdep.c step = ci->ci_signature & 15;
ci 1623 arch/i386/i386/machdep.c ci->cpu_setup = NULL;
ci 1666 arch/i386/i386/machdep.c ci->cpu_setup = cpup->cpu_family[i].cpu_setup;
ci 1706 arch/i386/i386/machdep.c if ((ci->ci_flags & CPUF_PRIMARY) == 0) {
ci 1725 arch/i386/i386/machdep.c if (ci->ci_feature_flags && (ci->ci_feature_flags & CPUID_TSC)) {
ci 1733 arch/i386/i386/machdep.c if ((ci->ci_flags & CPUF_PRIMARY) == 0) {
ci 1740 arch/i386/i386/machdep.c if ((ci->ci_flags & CPUF_PRIMARY) == 0) {
ci 1746 arch/i386/i386/machdep.c if ((ci->ci_flags & CPUF_PRIMARY) == 0) {
ci 1749 arch/i386/i386/machdep.c if (ci->ci_feature_flags) {
ci 1756 arch/i386/i386/machdep.c if (ci->ci_feature_flags &
ci 1779 arch/i386/i386/machdep.c if (ci->cpu_setup != NULL)
ci 1780 arch/i386/i386/machdep.c (ci->cpu_setup)(ci);
ci 1829 arch/i386/i386/machdep.c ci->cpu_class = class;
ci 1872 arch/i386/i386/machdep.c amd64_errata(ci);
ci 1898 arch/i386/i386/machdep.c cyrix3_get_bus_clock(struct cpu_info *ci)
ci 1922 arch/i386/i386/machdep.c p4_get_bus_clock(struct cpu_info *ci)
ci 1927 arch/i386/i386/machdep.c model = (ci->ci_signature >> 4) & 15;
ci 1941 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, model, bus);
ci 1962 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, model, bus);
ci 1969 arch/i386/i386/machdep.c p3_get_bus_clock(struct cpu_info *ci)
ci 1974 arch/i386/i386/machdep.c model = (ci->ci_signature >> 4) & 15;
ci 1991 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, bus);
ci 2020 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, bus);
ci 2046 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, bus);
ci 2052 arch/i386/i386/machdep.c ci->ci_dev.dv_xname, model);
ci 2800 arch/i386/i386/machdep.c cpu_default_ldt(struct cpu_info *ci)
ci 2802 arch/i386/i386/machdep.c ci->ci_ldt = ldt;
ci 2803 arch/i386/i386/machdep.c ci->ci_ldt_len = sizeof(ldt);
ci 2807 arch/i386/i386/machdep.c cpu_alloc_ldt(struct cpu_info *ci)
ci 2814 arch/i386/i386/machdep.c ci->ci_ldt = cpu_ldt;
ci 2815 arch/i386/i386/machdep.c ci->ci_ldt_len = len;
ci 2819 arch/i386/i386/machdep.c cpu_init_ldt(struct cpu_info *ci)
ci 2821 arch/i386/i386/machdep.c setsegment(&ci->ci_gdt[GLDT_SEL].sd, ci->ci_ldt, ci->ci_ldt_len - 1,
ci 3223 arch/i386/i386/machdep.c need_resched(struct cpu_info *ci)
ci 3227 arch/i386/i386/machdep.c ci->ci_want_resched = 1;
ci 3233 arch/i386/i386/machdep.c if ((p = ci->ci_curproc) != NULL)
ci 48 arch/i386/i386/mp_setperf.c struct cpu_info *ci;
ci 57 arch/i386/i386/mp_setperf.c CPU_INFO_FOREACH(cii, ci) {
ci 58 arch/i386/i386/mp_setperf.c if (ci->ci_setperf_state != CI_SETPERF_INTRANSIT) {
ci 59 arch/i386/i386/mp_setperf.c ci->ci_setperf_state =
ci 61 arch/i386/i386/mp_setperf.c i386_send_ipi(ci, I386_IPI_SETPERF);
ci 68 arch/i386/i386/mp_setperf.c CPU_INFO_FOREACH(cii, ci) {
ci 69 arch/i386/i386/mp_setperf.c if ((notready = (ci->ci_setperf_state
ci 86 arch/i386/i386/mp_setperf.c CPU_INFO_FOREACH(cii, ci) {
ci 87 arch/i386/i386/mp_setperf.c if ((notready = (ci->ci_setperf_state
ci 104 arch/i386/i386/mp_setperf.c i386_setperf_ipi(struct cpu_info *ci)
ci 107 arch/i386/i386/mp_setperf.c if (ci->ci_setperf_state == CI_SETPERF_SHOULDSTOP)
ci 108 arch/i386/i386/mp_setperf.c ci->ci_setperf_state = CI_SETPERF_INTRANSIT;
ci 115 arch/i386/i386/mp_setperf.c ci->ci_setperf_state = CI_SETPERF_DONE;
ci 119 arch/i386/i386/mp_setperf.c ci->ci_setperf_state = CI_SETPERF_READY;
ci 126 arch/i386/i386/mp_setperf.c struct cpu_info *ci;
ci 135 arch/i386/i386/mp_setperf.c CPU_INFO_FOREACH(cii, ci) {
ci 136 arch/i386/i386/mp_setperf.c ci->ci_setperf_state = CI_SETPERF_READY;
ci 2940 arch/i386/i386/pmap.c struct cpu_info *ci, *self = curcpu();
ci 2945 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) {
ci 2946 arch/i386/i386/pmap.c if (ci == self || !pmap_is_active(pm, ci->ci_cpuid) ||
ci 2947 arch/i386/i386/pmap.c !(ci->ci_flags & CPUF_RUNNING))
ci 2949 arch/i386/i386/pmap.c mask |= 1 << ci->ci_cpuid;
ci 2961 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) {
ci 2962 arch/i386/i386/pmap.c if ((mask & 1 << ci->ci_cpuid) == 0)
ci 2964 arch/i386/i386/pmap.c if (i386_fast_ipi(ci, LAPIC_IPI_INVLPG) != 0)
ci 2977 arch/i386/i386/pmap.c struct cpu_info *ci, *self = curcpu();
ci 2983 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) {
ci 2984 arch/i386/i386/pmap.c if (ci == self || !pmap_is_active(pm, ci->ci_cpuid) ||
ci 2985 arch/i386/i386/pmap.c !(ci->ci_flags & CPUF_RUNNING))
ci 2987 arch/i386/i386/pmap.c mask |= 1 << ci->ci_cpuid;
ci 3000 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) {
ci 3001 arch/i386/i386/pmap.c if ((mask & 1 << ci->ci_cpuid) == 0)
ci 3003 arch/i386/i386/pmap.c if (i386_fast_ipi(ci, LAPIC_IPI_INVLRANGE) != 0)
ci 3017 arch/i386/i386/pmap.c struct cpu_info *ci, *self = curcpu();
ci 3022 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) {
ci 3023 arch/i386/i386/pmap.c if (ci == self || !(ci->ci_flags & CPUF_RUNNING))
ci 3025 arch/i386/i386/pmap.c mask |= 1 << ci->ci_cpuid;
ci 3037 arch/i386/i386/pmap.c CPU_INFO_FOREACH(cii, ci) {
ci 3038 arch/i386/i386/pmap.c if ((mask & 1 << ci->ci_cpuid) == 0)
ci 3040 arch/i386/i386/pmap.c if (i386_fast_ipi(ci, LAPIC_IPI_INVLTLB) != 0)
ci 385 arch/i386/i386/powernow-k7.c struct cpu_info *ci;
ci 392 arch/i386/i386/powernow-k7.c ci = curcpu();
ci 410 arch/i386/i386/powernow-k7.c if (ci->ci_signature == AMD_ERRATA_A0_CPUSIG)
ci 426 arch/i386/i386/powernow-k7.c if (!k7pnow_states(cstate, ci->ci_signature, maxfid, startvid))
ci 436 arch/i386/i386/powernow-k7.c ci->ci_dev.dv_xname, techname, cpuspeed);
ci 435 arch/i386/i386/powernow-k8.c struct cpu_info * ci;
ci 439 arch/i386/i386/powernow-k8.c ci = curcpu();
ci 482 arch/i386/i386/powernow-k8.c if (!k8pnow_states(cstate, ci->ci_signature, maxfid, maxvid))
ci 487 arch/i386/i386/powernow-k8.c ci->ci_dev.dv_xname, techname, cpuspeed);
ci 177 arch/i386/include/cpu.h #define CPU_INFO_FOREACH(cii, ci) for (cii = 0, ci = cpu_info_list; \
ci 178 arch/i386/include/cpu.h ci != NULL; ci = ci->ci_next)
ci 180 arch/i386/include/cpu.h #define CPU_INFO_UNIT(ci) ((ci)->ci_dev.dv_unit)
ci 195 arch/i386/include/cpu.h struct cpu_info *ci;
ci 199 arch/i386/include/cpu.h "=r" (ci) : "m"
ci 201 arch/i386/include/cpu.h return ci;
ci 205 arch/i386/include/cpu.h #define CPU_IS_PRIMARY(ci) ((ci)->ci_flags & CPUF_PRIMARY)
ci 220 arch/i386/include/cpu.h #define CPU_IS_PRIMARY(ci) 1
ci 157 arch/i386/isa/npx.c npxdna_notset(struct cpu_info *ci)
ci 350 arch/i386/isa/npx.c npxinit(struct cpu_info *ci)
ci 357 arch/i386/isa/npx.c ci->ci_dev.dv_xname);
ci 422 arch/i386/isa/npx.c struct cpu_info *ci = curcpu();
ci 423 arch/i386/isa/npx.c struct proc *p = ci->ci_fpcurproc;
ci 430 arch/i386/isa/npx.c IPRINTF(("%s: fp intr\n", ci->ci_dev.dv_xname));
ci 446 arch/i386/isa/npx.c if (ci->ci_fpsaving)
ci 582 arch/i386/isa/npx.c npxdna_xmm(struct cpu_info *ci)
ci 587 arch/i386/isa/npx.c if (ci->ci_fpsaving) {
ci 595 arch/i386/isa/npx.c p = ci->ci_curproc;
ci 600 arch/i386/isa/npx.c IPRINTF(("%s: dna for %lx%s\n", ci->ci_dev.dv_xname, (u_long)p,
ci 611 arch/i386/isa/npx.c if (ci->ci_fpcurproc != NULL) {
ci 612 arch/i386/isa/npx.c IPRINTF(("%s: fp save %lx\n", ci->ci_dev.dv_xname,
ci 613 arch/i386/isa/npx.c (u_long)ci->ci_fpcurproc));
ci 614 arch/i386/isa/npx.c npxsave_cpu(ci, 1);
ci 617 arch/i386/isa/npx.c IPRINTF(("%s: fp init\n", ci->ci_dev.dv_xname));
ci 624 arch/i386/isa/npx.c IPRINTF(("%s: done saving\n", ci->ci_dev.dv_xname));
ci 625 arch/i386/isa/npx.c KDASSERT(ci->ci_fpcurproc == NULL);
ci 635 arch/i386/isa/npx.c ci->ci_fpcurproc = p;
ci 636 arch/i386/isa/npx.c p->p_addr->u_pcb.pcb_fpcpu = ci;
ci 660 arch/i386/isa/npx.c npxdna_s87(struct cpu_info *ci)
ci 667 arch/i386/isa/npx.c if (ci->ci_fpsaving) {
ci 674 arch/i386/isa/npx.c p = ci->ci_curproc;
ci 679 arch/i386/isa/npx.c IPRINTF(("%s: dna for %lx%s\n", ci->ci_dev.dv_xname, (u_long)p,
ci 687 arch/i386/isa/npx.c if (ci->ci_fpcurproc != NULL) {
ci 688 arch/i386/isa/npx.c IPRINTF(("%s: fp save %lx\n", ci->ci_dev.dv_xname,
ci 689 arch/i386/isa/npx.c (u_long)ci->ci_fpcurproc));
ci 690 arch/i386/isa/npx.c npxsave_cpu(ci, 1);
ci 693 arch/i386/isa/npx.c IPRINTF(("%s: fp init\n", ci->ci_dev.dv_xname));
ci 700 arch/i386/isa/npx.c IPRINTF(("%s: done saving\n", ci->ci_dev.dv_xname));
ci 701 arch/i386/isa/npx.c KDASSERT(ci->ci_fpcurproc == NULL);
ci 711 arch/i386/isa/npx.c ci->ci_fpcurproc = p;
ci 712 arch/i386/isa/npx.c p->p_addr->u_pcb.pcb_fpcpu = ci;
ci 748 arch/i386/isa/npx.c npxsave_cpu(struct cpu_info *ci, int save)
ci 753 arch/i386/isa/npx.c KDASSERT(ci == curcpu());
ci 755 arch/i386/isa/npx.c p = ci->ci_fpcurproc;
ci 759 arch/i386/isa/npx.c IPRINTF(("%s: fp cpu %s %lx\n", ci->ci_dev.dv_xname,
ci 764 arch/i386/isa/npx.c if (ci->ci_fpsaving != 0)
ci 781 arch/i386/isa/npx.c ci->ci_fpsaving = 1;
ci 783 arch/i386/isa/npx.c ci->ci_fpsaving = 0;
ci 797 arch/i386/isa/npx.c ci->ci_fpcurproc = NULL;
ci 807 arch/i386/isa/npx.c struct cpu_info *ci = curcpu();
ci 816 arch/i386/isa/npx.c IPRINTF(("%s: fp proc %s %lx\n", ci->ci_dev.dv_xname,
ci 820 arch/i386/isa/npx.c if (oci == ci) {
ci 822 arch/i386/isa/npx.c npxsave_cpu(ci, save);
ci 829 arch/i386/isa/npx.c IPRINTF(("%s: fp ipi to %s %s %lx\n", ci->ci_dev.dv_xname,
ci 843 arch/i386/isa/npx.c ci->ci_dev.dv_xname, oci->ci_dev.dv_xname);
ci 848 arch/i386/isa/npx.c KASSERT(ci->ci_fpcurproc == p);
ci 849 arch/i386/isa/npx.c npxsave_cpu(ci, save);
ci 180 arch/i386/pci/ichpcib.c struct cpu_info *ci;
ci 186 arch/i386/pci/ichpcib.c ci = curcpu();
ci 187 arch/i386/pci/ichpcib.c family = (ci->ci_signature >> 8) & 15;
ci 188 arch/i386/pci/ichpcib.c model = (ci->ci_signature >> 4) & 15;
ci 189 arch/i386/pci/ichpcib.c stepping = ci->ci_signature & 15;
ci 269 dev/ccd.c struct ccdcinfo *ci = NULL;
ci 307 dev/ccd.c ci = &cs->sc_cinfo[ix];
ci 308 dev/ccd.c ci->ci_vp = vp;
ci 315 dev/ccd.c MAXPATHLEN, &ci->ci_pathlen);
ci 323 dev/ccd.c ci->ci_path = malloc(ci->ci_pathlen, M_DEVBUF, M_WAITOK);
ci 324 dev/ccd.c bcopy(tmppath, ci->ci_path, ci->ci_pathlen);
ci 332 dev/ccd.c cs->sc_xname, ci->ci_path, error));
ci 333 dev/ccd.c free(ci->ci_path, M_DEVBUF);
ci 337 dev/ccd.c ci->ci_dev = va.va_rdev;
ci 347 dev/ccd.c cs->sc_xname, ci->ci_path, error));
ci 348 dev/ccd.c free(ci->ci_path, M_DEVBUF);
ci 361 dev/ccd.c cs->sc_xname, ci->ci_path));
ci 362 dev/ccd.c free(ci->ci_path, M_DEVBUF);
ci 376 dev/ccd.c ("%s: %s: size == 0\n", cs->sc_xname, ci->ci_path));
ci 377 dev/ccd.c free(ci->ci_path, M_DEVBUF);
ci 384 dev/ccd.c ci->ci_size = size;
ci 399 dev/ccd.c free(ci->ci_path, M_DEVBUF);
ci 414 dev/ccd.c free(ci->ci_path, M_DEVBUF);
ci 422 dev/ccd.c free(ci->ci_path, M_DEVBUF);
ci 434 dev/ccd.c for (ci = cs->sc_cinfo;
ci 435 dev/ccd.c ci < &cs->sc_cinfo[cs->sc_nccdisks]; ci++)
ci 436 dev/ccd.c ci->ci_size = minsize;
ci 468 dev/ccd.c struct ccdcinfo *ci, *smallci;
ci 523 dev/ccd.c for (ci = cs->sc_cinfo;
ci 524 dev/ccd.c ci < &cs->sc_cinfo[cs->sc_nccdisks]; ci++)
ci 525 dev/ccd.c if (ci->ci_size > size &&
ci 527 dev/ccd.c ci->ci_size < smallci->ci_size))
ci 528 dev/ccd.c smallci = ci;
ci 549 dev/ccd.c for (ci = cs->sc_cinfo;
ci 550 dev/ccd.c ci < &cs->sc_cinfo[cs->sc_nccunits]; ci++)
ci 551 dev/ccd.c if (ci->ci_size >= smallci->ci_size)
ci 552 dev/ccd.c ii->ii_index[ix++] = ci - cs->sc_cinfo;
ci 766 dev/ccd.c struct ccdcinfo *ci, *ci2 = NULL;
ci 788 dev/ccd.c for (ccdisk = 0, ci = &cs->sc_cinfo[ccdisk];
ci 789 dev/ccd.c cbn >= sblk + ci->ci_size;
ci 790 dev/ccd.c ccdisk++, ci = &cs->sc_cinfo[ccdisk])
ci 791 dev/ccd.c sblk += ci->ci_size;
ci 819 dev/ccd.c ci->ci_flags & CCIF_FAILED))
ci 823 dev/ccd.c ci = &cs->sc_cinfo[ccdisk];
ci 825 dev/ccd.c ccdisk, cbn, ci, ci2));
ci 830 dev/ccd.c cnt = dbtob(ci->ci_size - cbn);
ci 845 dev/ccd.c nbp->b_dev = ci->ci_dev; /* XXX */
ci 847 dev/ccd.c nbp->b_vp = ci->ci_vp;
ci 879 dev/ccd.c ci->ci_dev, ci-cs->sc_cinfo, cbp, bp->b_blkno,
ci 107 dev/i2o/iopsp.c struct i2o_param_hba_ctlr_info ci;
ci 123 dev/i2o/iopsp.c printf("iopsp_match: bustype = %d\n", param.ci.bustype);
ci 126 dev/i2o/iopsp.c return (param.ci.bustype == I2O_HBA_BUS_SCSI ||
ci 127 dev/i2o/iopsp.c param.ci.bustype == I2O_HBA_BUS_FCA);
ci 144 dev/i2o/iopsp.c struct i2o_param_hba_ctlr_info ci;
ci 171 dev/i2o/iopsp.c fcal = (param.p.ci.bustype == I2O_HBA_BUS_FCA); /* XXX */
ci 1888 dev/ic/aac.c u_int32_t pi, ci;
ci 1898 dev/ic/aac.c ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
ci 1905 dev/ic/aac.c if ((pi + 1) == ci) {
ci 1941 dev/ic/aac.c u_int32_t pi, ci;
ci 1948 dev/ic/aac.c ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
ci 1951 dev/ic/aac.c if (ci == pi) {
ci 1961 dev/ic/aac.c if (ci == pi + 1)
ci 1965 dev/ic/aac.c if (ci >= aac_qinfo[queue].size)
ci 1966 dev/ic/aac.c ci = 0;
ci 1969 dev/ic/aac.c *fib_size = (sc->aac_qentries[queue] + ci)->aq_fib_size;
ci 1980 dev/ic/aac.c fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr /
ci 1997 dev/ic/aac.c fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr;
ci 2018 dev/ic/aac.c sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX] = ci + 1;
ci 2035 dev/ic/aac.c u_int32_t pi, ci;
ci 2047 dev/ic/aac.c ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
ci 2054 dev/ic/aac.c if ((pi + 1) == ci) {
ci 443 dev/ic/pdq.c pdq_command_info_t * const ci = &pdq->pdq_command_info;
ci 452 dev/ic/pdq.c if (ci->ci_command_active || ci->ci_pending_commands == 0)
ci 459 dev/ic/pdq.c for (mask = 1 << ((int) op); (mask & ci->ci_pending_commands) == 0; mask >>= 1)
ci 479 dev/ic/pdq.c dbp->pdqdb_command_requests[ci->ci_request_producer].txd_seg_len = cmdlen;
ci 480 dev/ic/pdq.c PDQ_ADVANCE(ci->ci_request_producer, 1, PDQ_RING_MASK(dbp->pdqdb_command_requests));
ci 486 dev/ic/pdq.c dbp->pdqdb_command_responses[ci->ci_response_producer].rxd_seg_len_hi = cmdlen / 16;
ci 487 dev/ic/pdq.c PDQ_ADVANCE(ci->ci_response_producer, 1, PDQ_RING_MASK(dbp->pdqdb_command_responses));
ci 494 dev/ic/pdq.c PDQ_OS_MEMZERO(ci->ci_bufstart, cmdlen);
ci 495 dev/ic/pdq.c *(pdq_cmd_code_t *) ci->ci_bufstart = op;
ci 496 dev/ic/pdq.c ci->ci_pending_commands &= ~mask;
ci 503 dev/ic/pdq.c pdq_cmd_filter_set_t *filter_set = (pdq_cmd_filter_set_t *) ci->ci_bufstart;
ci 521 dev/ic/pdq.c pdq_cmd_addr_filter_set_t *addr_filter_set = (pdq_cmd_addr_filter_set_t *) ci->ci_bufstart;
ci 543 dev/ic/pdq.c ci->ci_command_active++;
ci 544 dev/ic/pdq.c PDQ_CSR_WRITE(csrs, csr_cmd_response_producer, ci->ci_response_producer | (ci->ci_response_completion << 8));
ci 545 dev/ic/pdq.c PDQ_CSR_WRITE(csrs, csr_cmd_request_producer, ci->ci_request_producer | (ci->ci_request_completion << 8));
ci 553 dev/ic/pdq.c pdq_command_info_t * const ci = &pdq->pdq_command_info;
ci 564 dev/ic/pdq.c if (cbp->pdqcb_command_response == ci->ci_response_completion)
ci 567 dev/ic/pdq.c PDQ_ASSERT(cbp->pdqcb_command_request != ci->ci_request_completion);
ci 569 dev/ic/pdq.c rspgen = (const pdq_response_generic_t *) ci->ci_bufstart;
ci 580 dev/ic/pdq.c PDQ_ADVANCE(ci->ci_request_completion, 1, PDQ_RING_MASK(dbp->pdqdb_command_requests));
ci 581 dev/ic/pdq.c PDQ_ADVANCE(ci->ci_response_completion, 1, PDQ_RING_MASK(dbp->pdqdb_command_responses));
ci 582 dev/ic/pdq.c ci->ci_command_active = 0;
ci 584 dev/ic/pdq.c if (ci->ci_pending_commands != 0) {
ci 588 dev/ic/pdq.c ci->ci_response_producer | (ci->ci_response_completion << 8));
ci 590 dev/ic/pdq.c ci->ci_request_producer | (ci->ci_request_completion << 8));
ci 437 dev/usb/uvisor.c uvisor_init(struct uvisor_softc *sc, struct uvisor_connection_info *ci,
ci 474 dev/usb/uvisor.c err = usbd_do_request_flags(sc->sc_udev, &req, ci,
ci 300 isofs/cd9660/cd9660_vnops.c struct cluster_info *ci = &ip->i_ci;
ci 314 isofs/cd9660/cd9660_vnops.c if (ci->ci_lastr + 1 == lbn) {
ci 334 isofs/cd9660/cd9660_vnops.c ci->ci_lastr = lbn;
ci 229 kern/kern_clock.c struct cpu_info *ci = curcpu();
ci 254 kern/kern_clock.c if (--ci->ci_schedstate.spc_rrticks <= 0)
ci 255 kern/kern_clock.c roundrobin(ci);
ci 261 kern/kern_clock.c if (CPU_IS_PRIMARY(ci) == 0)
ci 481 kern/kern_clock.c struct cpu_info *ci = curcpu();
ci 482 kern/kern_clock.c struct schedstate_percpu *spc = &ci->ci_schedstate;
ci 438 kern/kern_sysctl.c struct cpu_info *ci;
ci 443 kern/kern_sysctl.c CPU_INFO_FOREACH(cii, ci) {
ci 445 kern/kern_sysctl.c cp_time[i] += ci->ci_schedstate.spc_cp_time[i];
ci 1874 kern/kern_sysctl.c struct cpu_info *ci;
ci 1882 kern/kern_sysctl.c CPU_INFO_FOREACH(cii, ci) {
ci 1890 kern/kern_sysctl.c &ci->ci_schedstate.spc_cp_time,
ci 1891 kern/kern_sysctl.c sizeof(ci->ci_schedstate.spc_cp_time)));
ci 95 kern/sched_bsd.c roundrobin(struct cpu_info *ci)
ci 97 kern/sched_bsd.c struct schedstate_percpu *spc = &ci->ci_schedstate;
ci 480 kern/sched_bsd.c struct cpu_info *ci;
ci 504 kern/sched_bsd.c ci = (p->p_cpu != NULL) ? p->p_cpu : curcpu();
ci 505 kern/sched_bsd.c if (pri < ci->ci_schedstate.spc_curpriority)
ci 506 kern/sched_bsd.c need_resched(ci);
ci 62 kern/vfs_cluster.c cluster_write(struct buf *bp, struct cluster_info *ci, u_quad_t filesize)
ci 73 kern/vfs_cluster.c ci->ci_lasta = ci->ci_clen = ci->ci_cstart = ci->ci_lastw = 0;
ci 75 kern/vfs_cluster.c if (ci->ci_clen == 0 || lbn != ci->ci_lastw + 1 ||
ci 76 kern/vfs_cluster.c (bp->b_blkno != ci->ci_lasta + btodb(bp->b_bcount))) {
ci 78 kern/vfs_cluster.c if (ci->ci_clen != 0) {
ci 88 kern/vfs_cluster.c cursize = ci->ci_lastw - ci->ci_cstart + 1;
ci 90 kern/vfs_cluster.c lbn != ci->ci_lastw + 1 || ci->ci_clen <= cursize) {
ci 92 kern/vfs_cluster.c ci->ci_cstart, cursize, lbn);
ci 97 kern/vfs_cluster.c buflist = cluster_collectbufs(vp, ci, bp);
ci 109 kern/vfs_cluster.c ci->ci_cstart, cursize, lbn);
ci 118 kern/vfs_cluster.c ci->ci_lastw = lbn;
ci 119 kern/vfs_cluster.c ci->ci_lasta = bp->b_blkno;
ci 133 kern/vfs_cluster.c ci->ci_clen = 0;
ci 134 kern/vfs_cluster.c ci->ci_lasta = bp->b_blkno;
ci 135 kern/vfs_cluster.c ci->ci_cstart = lbn + 1;
ci 136 kern/vfs_cluster.c ci->ci_lastw = lbn;
ci 139 kern/vfs_cluster.c ci->ci_clen = maxclen;
ci 141 kern/vfs_cluster.c ci->ci_cstart = lbn + 1;
ci 144 kern/vfs_cluster.c ci->ci_cstart = lbn;
ci 147 kern/vfs_cluster.c } else if (lbn == ci->ci_cstart + ci->ci_clen) {
ci 151 kern/vfs_cluster.c cluster_wbuild(vp, bp, bp->b_bcount, ci->ci_cstart,
ci 152 kern/vfs_cluster.c ci->ci_clen + 1, lbn);
ci 153 kern/vfs_cluster.c ci->ci_clen = 0;
ci 154 kern/vfs_cluster.c ci->ci_cstart = lbn + 1;
ci 161 kern/vfs_cluster.c ci->ci_lastw = lbn;
ci 162 kern/vfs_cluster.c ci->ci_lasta = bp->b_blkno;
ci 225 kern/vfs_cluster.c cluster_collectbufs(struct vnode *vp, struct cluster_info *ci,
ci 232 kern/vfs_cluster.c len = ci->ci_lastw - ci->ci_cstart + 1;
ci 237 kern/vfs_cluster.c for (lbn = ci->ci_cstart, i = 0; i < len; lbn++, i++)