ci                235 arch/i386/i386/amd64errata.c amd64_errata_testmsr(struct cpu_info *ci, errata_t *e)
ci                239 arch/i386/i386/amd64errata.c 	(void)ci;
ci                250 arch/i386/i386/amd64errata.c amd64_errata_setmsr(struct cpu_info *ci, errata_t *e)
ci                254 arch/i386/i386/amd64errata.c 	(void)ci;
ci                274 arch/i386/i386/amd64errata.c amd64_errata(struct cpu_info *ci)
ci                320 arch/i386/i386/amd64errata.c 		    ci->ci_dev.dv_xname, e->e_num);
ci                327 arch/i386/i386/amd64errata.c 		rc = (e->e_act == NULL) ? 1 : (*e->e_act)(ci, e);
ci                340 arch/i386/i386/amd64errata.c 		    ci->ci_dev.dv_xname, e->e_num,
ci                351 arch/i386/i386/amd64errata.c 		printf("%s: AMD %s", ci->ci_dev.dv_xname,
ci                369 arch/i386/i386/amd64errata.c 		printf("%s: AMD %s", ci->ci_dev.dv_xname,
ci                187 arch/i386/i386/cpu.c 	struct cpu_info *ci = (struct cpu_info *)self;
ci                202 arch/i386/i386/cpu.c 		ci = &cpu_info_primary;
ci                203 arch/i386/i386/cpu.c 		bcopy(self, &ci->ci_dev, sizeof *self);
ci                213 arch/i386/i386/cpu.c 	cpu_info[cpunum] = ci;
ci                216 arch/i386/i386/cpu.c 	ci->ci_self = ci;
ci                217 arch/i386/i386/cpu.c 	ci->ci_apicid = caa->cpu_number;
ci                219 arch/i386/i386/cpu.c 	ci->ci_cpuid = ci->ci_apicid;
ci                221 arch/i386/i386/cpu.c 	ci->ci_cpuid = 0;	/* False for APs, so what, they're not used */
ci                223 arch/i386/i386/cpu.c 	ci->ci_signature = caa->cpu_signature;
ci                224 arch/i386/i386/cpu.c 	ci->ci_feature_flags = caa->feature_flags;
ci                225 arch/i386/i386/cpu.c 	ci->ci_func = caa->cpu_func;
ci                239 arch/i386/i386/cpu.c 		    ci->ci_dev.dv_xname);
ci                242 arch/i386/i386/cpu.c 	pcb = ci->ci_idle_pcb = (struct pcb *)kstack;
ci                254 arch/i386/i386/cpu.c 	cpu_default_ldt(ci);	/* Use the `global' ldt until one alloc'd */
ci                266 arch/i386/i386/cpu.c 		ci->ci_flags |= CPUF_PRESENT | CPUF_SP | CPUF_PRIMARY;
ci                267 arch/i386/i386/cpu.c 		identifycpu(ci);
ci                268 arch/i386/i386/cpu.c 		cpu_init(ci);
ci                273 arch/i386/i386/cpu.c 		ci->ci_flags |= CPUF_PRESENT | CPUF_BSP | CPUF_PRIMARY;
ci                274 arch/i386/i386/cpu.c 		identifycpu(ci);
ci                275 arch/i386/i386/cpu.c 		cpu_init(ci);
ci                282 arch/i386/i386/cpu.c 		lapic_calibrate_timer(ci);
ci                294 arch/i386/i386/cpu.c 		gdt_alloc_cpu(ci);
ci                295 arch/i386/i386/cpu.c 		cpu_alloc_ldt(ci);
ci                296 arch/i386/i386/cpu.c 		ci->ci_flags |= CPUF_PRESENT | CPUF_AP;
ci                297 arch/i386/i386/cpu.c 		identifycpu(ci);
ci                298 arch/i386/i386/cpu.c 		ci->ci_next = cpu_info_list->ci_next;
ci                299 arch/i386/i386/cpu.c 		cpu_info_list->ci_next = ci;
ci                308 arch/i386/i386/cpu.c 	if (ci->ci_apicid < IOAPIC_ID_MAX)
ci                309 arch/i386/i386/cpu.c 		ioapic_id_map &= ~(1 << ci->ci_apicid);
ci                313 arch/i386/i386/cpu.c 		    ci->ci_dev.dv_xname, kstack, USPACE);
ci                315 arch/i386/i386/cpu.c 		    ci->ci_dev.dv_xname, pcb, pcb->pcb_esp);
ci                328 arch/i386/i386/cpu.c cpu_init(struct cpu_info *ci)
ci                331 arch/i386/i386/cpu.c 	if (ci->cpu_setup != NULL)
ci                332 arch/i386/i386/cpu.c 		(*ci->cpu_setup)(ci);
ci                343 arch/i386/i386/cpu.c 	ci->ci_flags |= CPUF_RUNNING;
ci                363 arch/i386/i386/cpu.c 	struct cpu_info *ci;
ci                367 arch/i386/i386/cpu.c 		ci = cpu_info[i];
ci                368 arch/i386/i386/cpu.c 		if (ci == NULL)
ci                370 arch/i386/i386/cpu.c 		if (ci->ci_idle_pcb == NULL)
ci                372 arch/i386/i386/cpu.c 		if ((ci->ci_flags & CPUF_PRESENT) == 0)
ci                374 arch/i386/i386/cpu.c 		if (ci->ci_flags & (CPUF_BSP|CPUF_SP|CPUF_PRIMARY))
ci                376 arch/i386/i386/cpu.c 		cpu_boot_secondary(ci);
ci                383 arch/i386/i386/cpu.c 	struct cpu_info *ci;
ci                387 arch/i386/i386/cpu.c 		ci = cpu_info[i];
ci                388 arch/i386/i386/cpu.c 		if (ci == NULL)
ci                390 arch/i386/i386/cpu.c 		if (ci->ci_idle_pcb == NULL)
ci                392 arch/i386/i386/cpu.c 		if ((ci->ci_flags & CPUF_PRESENT) == 0)
ci                394 arch/i386/i386/cpu.c 		i386_init_pcb_tss_ldt(ci);
ci                399 arch/i386/i386/cpu.c cpu_boot_secondary(struct cpu_info *ci)
ci                407 arch/i386/i386/cpu.c 		printf("%s: starting", ci->ci_dev.dv_xname);
ci                412 arch/i386/i386/cpu.c 	pcb = ci->ci_idle_pcb;
ci                417 arch/i386/i386/cpu.c 	CPU_STARTUP(ci);
ci                422 arch/i386/i386/cpu.c 	for (i = 100000; (!(ci->ci_flags & CPUF_RUNNING)) && i > 0; i--) {
ci                425 arch/i386/i386/cpu.c 	if (!(ci->ci_flags & CPUF_RUNNING)) {
ci                426 arch/i386/i386/cpu.c 		printf("%s failed to become ready\n", ci->ci_dev.dv_xname);
ci                432 arch/i386/i386/cpu.c 	CPU_START_CLEANUP(ci);
ci                443 arch/i386/i386/cpu.c 	struct cpu_info *ci = (struct cpu_info *)v;
ci                450 arch/i386/i386/cpu.c 	gdt_init_cpu(ci);
ci                451 arch/i386/i386/cpu.c 	cpu_init_ldt(ci);
ci                452 arch/i386/i386/cpu.c 	npxinit(ci);
ci                456 arch/i386/i386/cpu.c 	cpu_init(ci);
ci                463 arch/i386/i386/cpu.c 		    ci->ci_dev.dv_xname, ci->ci_cpuid);
ci                464 arch/i386/i386/cpu.c 	microuptime(&ci->ci_schedstate.spc_runtime);
ci                514 arch/i386/i386/cpu.c cpu_set_tss_gates(struct cpu_info *ci)
ci                518 arch/i386/i386/cpu.c 	ci->ci_doubleflt_stack = (char *)uvm_km_alloc(kernel_map, USPACE);
ci                519 arch/i386/i386/cpu.c 	cpu_init_tss(&ci->ci_doubleflt_tss, ci->ci_doubleflt_stack,
ci                521 arch/i386/i386/cpu.c 	setsegment(&sd, &ci->ci_doubleflt_tss, sizeof(struct i386tss) - 1,
ci                523 arch/i386/i386/cpu.c 	ci->ci_gdt[GTRAPTSS_SEL].sd = sd;
ci                535 arch/i386/i386/cpu.c 	ci->ci_ddbipi_stack = (char *)uvm_km_alloc(kernel_map, USPACE);
ci                536 arch/i386/i386/cpu.c 	cpu_init_tss(&ci->ci_ddbipi_tss, ci->ci_ddbipi_stack,
ci                539 arch/i386/i386/cpu.c 	setsegment(&sd, &ci->ci_ddbipi_tss, sizeof(struct i386tss) - 1,
ci                541 arch/i386/i386/cpu.c 	ci->ci_gdt[GIPITSS_SEL].sd = sd;
ci                551 arch/i386/i386/cpu.c mp_cpu_start(struct cpu_info *ci)
ci                582 arch/i386/i386/cpu.c 	if (ci->ci_flags & CPUF_AP) {
ci                583 arch/i386/i386/cpu.c 		if ((error = i386_ipi_init(ci->ci_apicid)) != 0)
ci                590 arch/i386/i386/cpu.c 			    ci->ci_apicid, LAPIC_DLMODE_STARTUP)) != 0)
ci                595 arch/i386/i386/cpu.c 			    ci->ci_apicid, LAPIC_DLMODE_STARTUP)) != 0)
ci                605 arch/i386/i386/cpu.c mp_cpu_start_cleanup(struct cpu_info *ci)
ci                138 arch/i386/i386/db_mp.c i386_ipi_db(struct cpu_info *ci)
ci                110 arch/i386/i386/gdt.c 	struct cpu_info *ci;
ci                115 arch/i386/i386/gdt.c 	CPU_INFO_FOREACH(cii, ci)
ci                116 arch/i386/i386/gdt.c 		if (ci->ci_gdt != NULL && ci->ci_gdt != gdt)
ci                117 arch/i386/i386/gdt.c 			ci->ci_gdt[sel].sd = *sd;
ci                129 arch/i386/i386/gdt.c 	struct cpu_info *ci = &cpu_info_primary;
ci                147 arch/i386/i386/gdt.c 	ci->ci_gdt = gdt;
ci                148 arch/i386/i386/gdt.c 	setsegment(&ci->ci_gdt[GCPU_SEL].sd, ci, sizeof(struct cpu_info)-1,
ci                151 arch/i386/i386/gdt.c 	gdt_init_cpu(ci);
ci                159 arch/i386/i386/gdt.c gdt_alloc_cpu(struct cpu_info *ci)
ci                164 arch/i386/i386/gdt.c 	ci->ci_gdt = (union descriptor *)uvm_km_valloc(kernel_map, max_len);
ci                165 arch/i386/i386/gdt.c 	uvm_map_pageable(kernel_map, (vaddr_t)ci->ci_gdt,
ci                166 arch/i386/i386/gdt.c 	    (vaddr_t)ci->ci_gdt + min_len, FALSE, FALSE);
ci                167 arch/i386/i386/gdt.c 	bzero(ci->ci_gdt, min_len);
ci                168 arch/i386/i386/gdt.c 	bcopy(gdt, ci->ci_gdt, gdt_size * sizeof(union descriptor));
ci                169 arch/i386/i386/gdt.c 	setsegment(&ci->ci_gdt[GCPU_SEL].sd, ci, sizeof(struct cpu_info)-1,
ci                180 arch/i386/i386/gdt.c gdt_init_cpu(struct cpu_info *ci)
ci                184 arch/i386/i386/gdt.c 	setregion(&region, ci->ci_gdt,
ci                197 arch/i386/i386/gdt.c 	struct cpu_info *ci;
ci                205 arch/i386/i386/gdt.c 	CPU_INFO_FOREACH(cii, ci) {
ci                206 arch/i386/i386/gdt.c 		for (va = (vaddr_t)(ci->ci_gdt) + old_len;
ci                207 arch/i386/i386/gdt.c 		     va < (vaddr_t)(ci->ci_gdt) + new_len;
ci                 97 arch/i386/i386/ipifuncs.c i386_ipi_halt(struct cpu_info *ci)
ci                101 arch/i386/i386/ipifuncs.c 	printf("%s: shutting down\n", ci->ci_dev.dv_xname);
ci                109 arch/i386/i386/ipifuncs.c i386_ipi_flush_fpu(struct cpu_info *ci)
ci                111 arch/i386/i386/ipifuncs.c 	npxsave_cpu(ci, 0);
ci                115 arch/i386/i386/ipifuncs.c i386_ipi_synch_fpu(struct cpu_info *ci)
ci                117 arch/i386/i386/ipifuncs.c 	npxsave_cpu(ci, 1);
ci                128 arch/i386/i386/ipifuncs.c i386_send_ipi(struct cpu_info *ci, int ipimask)
ci                132 arch/i386/i386/ipifuncs.c 	i386_atomic_setbits_l(&ci->ci_ipis, ipimask);
ci                135 arch/i386/i386/ipifuncs.c 	if (!(ci->ci_flags & CPUF_RUNNING))
ci                138 arch/i386/i386/ipifuncs.c 	ret = i386_ipi(LAPIC_IPI_VECTOR, ci->ci_cpuid, LAPIC_DLMODE_FIXED);
ci                141 arch/i386/i386/ipifuncs.c 		    ipimask, curcpu()->ci_dev.dv_xname, ci->ci_dev.dv_xname);
ci                148 arch/i386/i386/ipifuncs.c i386_fast_ipi(struct cpu_info *ci, int ipi)
ci                150 arch/i386/i386/ipifuncs.c 	if (!(ci->ci_flags & CPUF_RUNNING))
ci                153 arch/i386/i386/ipifuncs.c 	return (i386_ipi(ipi, ci->ci_cpuid, LAPIC_DLMODE_FIXED));
ci                167 arch/i386/i386/ipifuncs.c 	struct cpu_info *ci, *self = curcpu();
ci                171 arch/i386/i386/ipifuncs.c 	CPU_INFO_FOREACH(cii, ci) {
ci                172 arch/i386/i386/ipifuncs.c 		if (ci == self)
ci                174 arch/i386/i386/ipifuncs.c 		if ((ci->ci_flags & CPUF_RUNNING) == 0)
ci                176 arch/i386/i386/ipifuncs.c 		i386_atomic_setbits_l(&ci->ci_ipis, ipimask);
ci                189 arch/i386/i386/ipifuncs.c 	struct cpu_info *ci = curcpu();
ci                193 arch/i386/i386/ipifuncs.c 	pending = i386_atomic_testset_ul(&ci->ci_ipis, 0);
ci                198 arch/i386/i386/ipifuncs.c 			(*ipifunc[bit])(ci);
ci                131 arch/i386/i386/lapic.c 	struct cpu_info *ci = curcpu();
ci                137 arch/i386/i386/lapic.c 		apic_format_redir(ci->ci_dev.dv_xname, "prelint", 0, 0,
ci                139 arch/i386/i386/lapic.c 		apic_format_redir(ci->ci_dev.dv_xname, "prelint", 1, 0,
ci                147 arch/i386/i386/lapic.c 					    || mpi->cpu_id == ci->ci_apicid)) {
ci                162 arch/i386/i386/lapic.c 		apic_format_redir(ci->ci_dev.dv_xname, "timer", 0, 0,
ci                164 arch/i386/i386/lapic.c 		apic_format_redir(ci->ci_dev.dv_xname, "pcint", 0, 0,
ci                166 arch/i386/i386/lapic.c 		apic_format_redir(ci->ci_dev.dv_xname, "lint", 0, 0,
ci                168 arch/i386/i386/lapic.c 		apic_format_redir(ci->ci_dev.dv_xname, "lint", 1, 0,
ci                170 arch/i386/i386/lapic.c 		apic_format_redir(ci->ci_dev.dv_xname, "err", 0, 0,
ci                224 arch/i386/i386/lapic.c 	struct cpu_info *ci = curcpu();
ci                227 arch/i386/i386/lapic.c 	if (CPU_IS_PRIMARY(ci)) {
ci                228 arch/i386/i386/lapic.c 		ci->ci_tscbase = rdtsc();
ci                267 arch/i386/i386/lapic.c lapic_calibrate_timer(struct cpu_info *ci)
ci                276 arch/i386/i386/lapic.c 		printf("%s: calibrating local timer\n", ci->ci_dev.dv_xname);
ci                351 arch/i386/i386/lapic.c 	printf("%s: apic clock running at %s\n", ci->ci_dev.dv_xname, tbuf);
ci                432 arch/i386/i386/lapic.c i386_ipi_microset(struct cpu_info *ci)
ci                434 arch/i386/i386/lapic.c 	ci->ci_tscbase = rdtsc();
ci                445 arch/i386/i386/lapic.c 	struct cpu_info *ci = curcpu();
ci                451 arch/i386/i386/lapic.c 	tmp = rdtsc() - ci->ci_tscbase;
ci                502 arch/i386/i386/machdep.c i386_init_pcb_tss_ldt(struct cpu_info *ci)
ci                505 arch/i386/i386/machdep.c 	struct pcb *pcb = ci->ci_idle_pcb;
ci                514 arch/i386/i386/machdep.c 	pcb->pcb_ldt = ci->ci_ldt;
ci                516 arch/i386/i386/machdep.c 	ci->ci_idle_tss_sel = tss_alloc(pcb);
ci               1028 arch/i386/i386/machdep.c winchip_cpu_setup(struct cpu_info *ci)
ci               1032 arch/i386/i386/machdep.c 	switch ((ci->ci_signature >> 4) & 15) { /* model */
ci               1034 arch/i386/i386/machdep.c 		ci->ci_feature_flags &= ~CPUID_TSC;
ci               1037 arch/i386/i386/machdep.c 		printf("%s: TSC disabled\n", ci->ci_dev.dv_xname);
ci               1045 arch/i386/i386/machdep.c cyrix3_setperf_setup(struct cpu_info *ci)
ci               1049 arch/i386/i386/machdep.c 			est_init(ci->ci_dev.dv_xname, CPUVENDOR_VIA);
ci               1052 arch/i386/i386/machdep.c 			    ci->ci_dev.dv_xname);
ci               1058 arch/i386/i386/machdep.c cyrix3_cpu_setup(struct cpu_info *ci)
ci               1061 arch/i386/i386/machdep.c 	int model = (ci->ci_signature >> 4) & 15;
ci               1062 arch/i386/i386/machdep.c 	int step = ci->ci_signature & 15;
ci               1073 arch/i386/i386/machdep.c 	cyrix3_get_bus_clock(ci);
ci               1118 arch/i386/i386/machdep.c 			printf("%s:", ci->ci_dev.dv_xname);
ci               1192 arch/i386/i386/machdep.c cyrix6x86_cpu_setup(struct cpu_info *ci)
ci               1196 arch/i386/i386/machdep.c 	switch ((ci->ci_signature >> 4) & 15) { /* model */
ci               1213 arch/i386/i386/machdep.c 		    ci->ci_dev.dv_xname);
ci               1219 arch/i386/i386/machdep.c 		printf("%s: TSC disabled\n", ci->ci_dev.dv_xname);
ci               1225 arch/i386/i386/machdep.c natsem6x86_cpu_setup(struct cpu_info *ci)
ci               1229 arch/i386/i386/machdep.c 	int model = (ci->ci_signature >> 4) & 15;
ci               1235 arch/i386/i386/machdep.c 		printf("%s: TSC disabled\n", ci->ci_dev.dv_xname);
ci               1242 arch/i386/i386/machdep.c intel586_cpu_setup(struct cpu_info *ci)
ci               1248 arch/i386/i386/machdep.c 		    ci->ci_dev.dv_xname);
ci               1255 arch/i386/i386/machdep.c amd_family5_setperf_setup(struct cpu_info *ci)
ci               1262 arch/i386/i386/machdep.c amd_family5_setup(struct cpu_info *ci)
ci               1264 arch/i386/i386/machdep.c 	int model = (ci->ci_signature >> 4) & 15;
ci               1292 arch/i386/i386/machdep.c amd_family6_setperf_setup(struct cpu_info *ci)
ci               1294 arch/i386/i386/machdep.c 	int family = (ci->ci_signature >> 8) & 15;
ci               1308 arch/i386/i386/machdep.c amd_family6_setup(struct cpu_info *ci)
ci               1338 arch/i386/i386/machdep.c 	struct cpu_info *ci = (struct cpu_info *) args;
ci               1347 arch/i386/i386/machdep.c 		ci->ci_sensor.value = max - MSR_THERM_STATUS_TEMP(msr);
ci               1349 arch/i386/i386/machdep.c 		ci->ci_sensor.value *= 1000000;
ci               1351 arch/i386/i386/machdep.c 		ci->ci_sensor.value += 273150000;
ci               1352 arch/i386/i386/machdep.c 		ci->ci_sensor.flags &= ~SENSOR_FINVALID;
ci               1354 arch/i386/i386/machdep.c 		ci->ci_sensor.value = 0;
ci               1355 arch/i386/i386/machdep.c 		ci->ci_sensor.flags |= SENSOR_FINVALID;
ci               1360 arch/i386/i386/machdep.c intel686_cpusensors_setup(struct cpu_info *ci)
ci               1373 arch/i386/i386/machdep.c 	strlcpy(ci->ci_sensordev.xname, ci->ci_dev.dv_xname,
ci               1374 arch/i386/i386/machdep.c 	    sizeof(ci->ci_sensordev.xname));
ci               1375 arch/i386/i386/machdep.c 	ci->ci_sensor.type = SENSOR_TEMP;
ci               1376 arch/i386/i386/machdep.c 	sensor_task_register(ci, intelcore_update_sensor, 5);
ci               1377 arch/i386/i386/machdep.c 	sensor_attach(&ci->ci_sensordev, &ci->ci_sensor);
ci               1378 arch/i386/i386/machdep.c 	sensordev_install(&ci->ci_sensordev);
ci               1384 arch/i386/i386/machdep.c intel686_setperf_setup(struct cpu_info *ci)
ci               1386 arch/i386/i386/machdep.c 	int family = (ci->ci_signature >> 8) & 15;
ci               1387 arch/i386/i386/machdep.c 	int step = ci->ci_signature & 15;
ci               1391 arch/i386/i386/machdep.c 			est_init(ci->ci_dev.dv_xname, CPUVENDOR_INTEL);
ci               1394 arch/i386/i386/machdep.c 			    ci->ci_dev.dv_xname);
ci               1402 arch/i386/i386/machdep.c intel686_common_cpu_setup(struct cpu_info *ci)
ci               1427 arch/i386/i386/machdep.c intel686_cpu_setup(struct cpu_info *ci)
ci               1429 arch/i386/i386/machdep.c 	int model = (ci->ci_signature >> 4) & 15;
ci               1430 arch/i386/i386/machdep.c 	int step = ci->ci_signature & 15;
ci               1434 arch/i386/i386/machdep.c 	p3_get_bus_clock(ci);
ci               1437 arch/i386/i386/machdep.c 	intel686_common_cpu_setup(ci);
ci               1444 arch/i386/i386/machdep.c 		ci->ci_feature_flags &= ~CPUID_SEP;
ci               1449 arch/i386/i386/machdep.c 	if ((model == 7) && (ci->ci_feature_flags & CPUID_SER)) {
ci               1455 arch/i386/i386/machdep.c 			 ci->ci_dev.dv_xname);
ci               1456 arch/i386/i386/machdep.c 		ci->ci_feature_flags &= ~CPUID_SER;
ci               1457 arch/i386/i386/machdep.c 		ci->ci_level = 2;
ci               1467 arch/i386/i386/machdep.c intel686_p4_cpu_setup(struct cpu_info *ci)
ci               1470 arch/i386/i386/machdep.c 	p4_get_bus_clock(ci);
ci               1473 arch/i386/i386/machdep.c 	intel686_common_cpu_setup(ci);
ci               1476 arch/i386/i386/machdep.c 	p4_model = (ci->ci_signature >> 4) & 15;
ci               1482 arch/i386/i386/machdep.c tm86_cpu_setup(struct cpu_info *ci)
ci               1560 arch/i386/i386/machdep.c identifycpu(struct cpu_info *ci)
ci               1567 arch/i386/i386/machdep.c 	char *cpu_device = ci->ci_dev.dv_xname;
ci               1582 arch/i386/i386/machdep.c 		ci->cpu_setup = i386_nocpuid_cpus[cpu].cpu_setup;
ci               1587 arch/i386/i386/machdep.c 		modif = (ci->ci_signature >> 12) & 3;
ci               1588 arch/i386/i386/machdep.c 		family = (ci->ci_signature >> 8) & 15;
ci               1589 arch/i386/i386/machdep.c 		model = (ci->ci_signature >> 4) & 15;
ci               1590 arch/i386/i386/machdep.c 		step = ci->ci_signature & 15;
ci               1623 arch/i386/i386/machdep.c 			ci->cpu_setup = NULL;
ci               1666 arch/i386/i386/machdep.c 			ci->cpu_setup = cpup->cpu_family[i].cpu_setup;
ci               1706 arch/i386/i386/machdep.c 	if ((ci->ci_flags & CPUF_PRIMARY) == 0) {
ci               1725 arch/i386/i386/machdep.c 	if (ci->ci_feature_flags && (ci->ci_feature_flags & CPUID_TSC)) {
ci               1733 arch/i386/i386/machdep.c 			if ((ci->ci_flags & CPUF_PRIMARY) == 0) {
ci               1740 arch/i386/i386/machdep.c 			if ((ci->ci_flags & CPUF_PRIMARY) == 0) {
ci               1746 arch/i386/i386/machdep.c 	if ((ci->ci_flags & CPUF_PRIMARY) == 0) {
ci               1749 arch/i386/i386/machdep.c 		if (ci->ci_feature_flags) {
ci               1756 arch/i386/i386/machdep.c 				if (ci->ci_feature_flags &
ci               1779 arch/i386/i386/machdep.c 	if (ci->cpu_setup != NULL)
ci               1780 arch/i386/i386/machdep.c 		(ci->cpu_setup)(ci);
ci               1829 arch/i386/i386/machdep.c 	ci->cpu_class = class;
ci               1872 arch/i386/i386/machdep.c 		amd64_errata(ci);
ci               1898 arch/i386/i386/machdep.c cyrix3_get_bus_clock(struct cpu_info *ci)
ci               1922 arch/i386/i386/machdep.c p4_get_bus_clock(struct cpu_info *ci)
ci               1927 arch/i386/i386/machdep.c 	model = (ci->ci_signature >> 4) & 15;
ci               1941 arch/i386/i386/machdep.c 			    ci->ci_dev.dv_xname, model, bus);
ci               1962 arch/i386/i386/machdep.c 			    ci->ci_dev.dv_xname, model, bus);
ci               1969 arch/i386/i386/machdep.c p3_get_bus_clock(struct cpu_info *ci)
ci               1974 arch/i386/i386/machdep.c 	model = (ci->ci_signature >> 4) & 15;
ci               1991 arch/i386/i386/machdep.c 			    ci->ci_dev.dv_xname, bus);
ci               2020 arch/i386/i386/machdep.c 			    ci->ci_dev.dv_xname, bus);
ci               2046 arch/i386/i386/machdep.c 			    ci->ci_dev.dv_xname, bus);
ci               2052 arch/i386/i386/machdep.c 		    ci->ci_dev.dv_xname, model);
ci               2800 arch/i386/i386/machdep.c cpu_default_ldt(struct cpu_info *ci)
ci               2802 arch/i386/i386/machdep.c 	ci->ci_ldt = ldt;
ci               2803 arch/i386/i386/machdep.c 	ci->ci_ldt_len = sizeof(ldt);
ci               2807 arch/i386/i386/machdep.c cpu_alloc_ldt(struct cpu_info *ci)
ci               2814 arch/i386/i386/machdep.c 	ci->ci_ldt = cpu_ldt;
ci               2815 arch/i386/i386/machdep.c 	ci->ci_ldt_len = len;
ci               2819 arch/i386/i386/machdep.c cpu_init_ldt(struct cpu_info *ci)
ci               2821 arch/i386/i386/machdep.c 	setsegment(&ci->ci_gdt[GLDT_SEL].sd, ci->ci_ldt, ci->ci_ldt_len - 1,
ci               3223 arch/i386/i386/machdep.c need_resched(struct cpu_info *ci)
ci               3227 arch/i386/i386/machdep.c 	ci->ci_want_resched = 1;
ci               3233 arch/i386/i386/machdep.c 	if ((p = ci->ci_curproc) != NULL)
ci                 48 arch/i386/i386/mp_setperf.c 	struct cpu_info *ci;
ci                 57 arch/i386/i386/mp_setperf.c 		CPU_INFO_FOREACH(cii, ci) {
ci                 58 arch/i386/i386/mp_setperf.c 			if (ci->ci_setperf_state != CI_SETPERF_INTRANSIT) {
ci                 59 arch/i386/i386/mp_setperf.c 				ci->ci_setperf_state =
ci                 61 arch/i386/i386/mp_setperf.c 				i386_send_ipi(ci, I386_IPI_SETPERF);
ci                 68 arch/i386/i386/mp_setperf.c 			CPU_INFO_FOREACH(cii, ci) {
ci                 69 arch/i386/i386/mp_setperf.c 				if ((notready = (ci->ci_setperf_state
ci                 86 arch/i386/i386/mp_setperf.c 			CPU_INFO_FOREACH(cii, ci) {
ci                 87 arch/i386/i386/mp_setperf.c 				if ((notready = (ci->ci_setperf_state
ci                104 arch/i386/i386/mp_setperf.c i386_setperf_ipi(struct cpu_info *ci)
ci                107 arch/i386/i386/mp_setperf.c 	if (ci->ci_setperf_state == CI_SETPERF_SHOULDSTOP)
ci                108 arch/i386/i386/mp_setperf.c 		ci->ci_setperf_state = CI_SETPERF_INTRANSIT;
ci                115 arch/i386/i386/mp_setperf.c 	ci->ci_setperf_state = CI_SETPERF_DONE;
ci                119 arch/i386/i386/mp_setperf.c 	ci->ci_setperf_state = CI_SETPERF_READY;
ci                126 arch/i386/i386/mp_setperf.c 	struct cpu_info *ci;
ci                135 arch/i386/i386/mp_setperf.c 	CPU_INFO_FOREACH(cii, ci) {
ci                136 arch/i386/i386/mp_setperf.c 		ci->ci_setperf_state = CI_SETPERF_READY;
ci               2940 arch/i386/i386/pmap.c 	struct cpu_info *ci, *self = curcpu();
ci               2945 arch/i386/i386/pmap.c 	CPU_INFO_FOREACH(cii, ci) {
ci               2946 arch/i386/i386/pmap.c 		if (ci == self || !pmap_is_active(pm, ci->ci_cpuid) ||
ci               2947 arch/i386/i386/pmap.c 		    !(ci->ci_flags & CPUF_RUNNING))
ci               2949 arch/i386/i386/pmap.c 		mask |= 1 << ci->ci_cpuid;
ci               2961 arch/i386/i386/pmap.c 		CPU_INFO_FOREACH(cii, ci) {
ci               2962 arch/i386/i386/pmap.c 			if ((mask & 1 << ci->ci_cpuid) == 0)
ci               2964 arch/i386/i386/pmap.c 			if (i386_fast_ipi(ci, LAPIC_IPI_INVLPG) != 0)
ci               2977 arch/i386/i386/pmap.c 	struct cpu_info *ci, *self = curcpu();
ci               2983 arch/i386/i386/pmap.c 	CPU_INFO_FOREACH(cii, ci) {
ci               2984 arch/i386/i386/pmap.c 		if (ci == self || !pmap_is_active(pm, ci->ci_cpuid) ||
ci               2985 arch/i386/i386/pmap.c 		    !(ci->ci_flags & CPUF_RUNNING))
ci               2987 arch/i386/i386/pmap.c 		mask |= 1 << ci->ci_cpuid;
ci               3000 arch/i386/i386/pmap.c 		CPU_INFO_FOREACH(cii, ci) {
ci               3001 arch/i386/i386/pmap.c 			if ((mask & 1 << ci->ci_cpuid) == 0)
ci               3003 arch/i386/i386/pmap.c 			if (i386_fast_ipi(ci, LAPIC_IPI_INVLRANGE) != 0)
ci               3017 arch/i386/i386/pmap.c 	struct cpu_info *ci, *self = curcpu();
ci               3022 arch/i386/i386/pmap.c 	CPU_INFO_FOREACH(cii, ci) {
ci               3023 arch/i386/i386/pmap.c 		if (ci == self || !(ci->ci_flags & CPUF_RUNNING))
ci               3025 arch/i386/i386/pmap.c 		mask |= 1 << ci->ci_cpuid;
ci               3037 arch/i386/i386/pmap.c 		CPU_INFO_FOREACH(cii, ci) {
ci               3038 arch/i386/i386/pmap.c 			if ((mask & 1 << ci->ci_cpuid) == 0)
ci               3040 arch/i386/i386/pmap.c 			if (i386_fast_ipi(ci, LAPIC_IPI_INVLTLB) != 0)
ci                385 arch/i386/i386/powernow-k7.c 	struct cpu_info *ci;
ci                392 arch/i386/i386/powernow-k7.c 	ci = curcpu();
ci                410 arch/i386/i386/powernow-k7.c 	if (ci->ci_signature == AMD_ERRATA_A0_CPUSIG)
ci                426 arch/i386/i386/powernow-k7.c 		if (!k7pnow_states(cstate, ci->ci_signature, maxfid, startvid))
ci                436 arch/i386/i386/powernow-k7.c 		    ci->ci_dev.dv_xname, techname, cpuspeed);
ci                435 arch/i386/i386/powernow-k8.c 	struct cpu_info * ci;
ci                439 arch/i386/i386/powernow-k8.c 	ci = curcpu();
ci                482 arch/i386/i386/powernow-k8.c 		if (!k8pnow_states(cstate, ci->ci_signature, maxfid, maxvid))
ci                487 arch/i386/i386/powernow-k8.c 		    ci->ci_dev.dv_xname, techname, cpuspeed);
ci                177 arch/i386/include/cpu.h #define	CPU_INFO_FOREACH(cii, ci)	for (cii = 0, ci = cpu_info_list; \
ci                178 arch/i386/include/cpu.h 					    ci != NULL; ci = ci->ci_next)
ci                180 arch/i386/include/cpu.h #define CPU_INFO_UNIT(ci)	((ci)->ci_dev.dv_unit)
ci                195 arch/i386/include/cpu.h 	struct cpu_info *ci;
ci                199 arch/i386/include/cpu.h 		"=r" (ci) : "m"
ci                201 arch/i386/include/cpu.h 	return ci;
ci                205 arch/i386/include/cpu.h #define CPU_IS_PRIMARY(ci)	((ci)->ci_flags & CPUF_PRIMARY)
ci                220 arch/i386/include/cpu.h #define CPU_IS_PRIMARY(ci)	1
ci                157 arch/i386/isa/npx.c npxdna_notset(struct cpu_info *ci)
ci                350 arch/i386/isa/npx.c npxinit(struct cpu_info *ci)
ci                357 arch/i386/isa/npx.c 		    ci->ci_dev.dv_xname);
ci                422 arch/i386/isa/npx.c 	struct cpu_info *ci = curcpu();
ci                423 arch/i386/isa/npx.c 	struct proc *p = ci->ci_fpcurproc;
ci                430 arch/i386/isa/npx.c 	IPRINTF(("%s: fp intr\n", ci->ci_dev.dv_xname));
ci                446 arch/i386/isa/npx.c 	if (ci->ci_fpsaving)
ci                582 arch/i386/isa/npx.c npxdna_xmm(struct cpu_info *ci)
ci                587 arch/i386/isa/npx.c 	if (ci->ci_fpsaving) {
ci                595 arch/i386/isa/npx.c 	p = ci->ci_curproc;
ci                600 arch/i386/isa/npx.c 	IPRINTF(("%s: dna for %lx%s\n", ci->ci_dev.dv_xname, (u_long)p,
ci                611 arch/i386/isa/npx.c 	if (ci->ci_fpcurproc != NULL) {
ci                612 arch/i386/isa/npx.c 		IPRINTF(("%s: fp save %lx\n", ci->ci_dev.dv_xname,
ci                613 arch/i386/isa/npx.c 		    (u_long)ci->ci_fpcurproc));
ci                614 arch/i386/isa/npx.c 		npxsave_cpu(ci, 1);
ci                617 arch/i386/isa/npx.c 		IPRINTF(("%s: fp init\n", ci->ci_dev.dv_xname));
ci                624 arch/i386/isa/npx.c 	IPRINTF(("%s: done saving\n", ci->ci_dev.dv_xname));
ci                625 arch/i386/isa/npx.c 	KDASSERT(ci->ci_fpcurproc == NULL);
ci                635 arch/i386/isa/npx.c 	ci->ci_fpcurproc = p;
ci                636 arch/i386/isa/npx.c 	p->p_addr->u_pcb.pcb_fpcpu = ci;
ci                660 arch/i386/isa/npx.c npxdna_s87(struct cpu_info *ci)
ci                667 arch/i386/isa/npx.c 	if (ci->ci_fpsaving) {
ci                674 arch/i386/isa/npx.c 	p = ci->ci_curproc;
ci                679 arch/i386/isa/npx.c 	IPRINTF(("%s: dna for %lx%s\n", ci->ci_dev.dv_xname, (u_long)p,
ci                687 arch/i386/isa/npx.c 	if (ci->ci_fpcurproc != NULL) {
ci                688 arch/i386/isa/npx.c 		IPRINTF(("%s: fp save %lx\n", ci->ci_dev.dv_xname,
ci                689 arch/i386/isa/npx.c 		    (u_long)ci->ci_fpcurproc));
ci                690 arch/i386/isa/npx.c 		npxsave_cpu(ci, 1);
ci                693 arch/i386/isa/npx.c 		IPRINTF(("%s: fp init\n", ci->ci_dev.dv_xname));
ci                700 arch/i386/isa/npx.c 	IPRINTF(("%s: done saving\n", ci->ci_dev.dv_xname));
ci                701 arch/i386/isa/npx.c 	KDASSERT(ci->ci_fpcurproc == NULL);
ci                711 arch/i386/isa/npx.c 	ci->ci_fpcurproc = p;
ci                712 arch/i386/isa/npx.c 	p->p_addr->u_pcb.pcb_fpcpu = ci;
ci                748 arch/i386/isa/npx.c npxsave_cpu(struct cpu_info *ci, int save)
ci                753 arch/i386/isa/npx.c 	KDASSERT(ci == curcpu());
ci                755 arch/i386/isa/npx.c 	p = ci->ci_fpcurproc;
ci                759 arch/i386/isa/npx.c 	IPRINTF(("%s: fp cpu %s %lx\n", ci->ci_dev.dv_xname,
ci                764 arch/i386/isa/npx.c 		if (ci->ci_fpsaving != 0)
ci                781 arch/i386/isa/npx.c 		ci->ci_fpsaving = 1;
ci                783 arch/i386/isa/npx.c 		ci->ci_fpsaving = 0;
ci                797 arch/i386/isa/npx.c 	ci->ci_fpcurproc = NULL;
ci                807 arch/i386/isa/npx.c 	struct cpu_info *ci = curcpu();
ci                816 arch/i386/isa/npx.c 	IPRINTF(("%s: fp proc %s %lx\n", ci->ci_dev.dv_xname,
ci                820 arch/i386/isa/npx.c 	if (oci == ci) {
ci                822 arch/i386/isa/npx.c 		npxsave_cpu(ci, save);
ci                829 arch/i386/isa/npx.c 		IPRINTF(("%s: fp ipi to %s %s %lx\n", ci->ci_dev.dv_xname,
ci                843 arch/i386/isa/npx.c 				    ci->ci_dev.dv_xname, oci->ci_dev.dv_xname);
ci                848 arch/i386/isa/npx.c 	KASSERT(ci->ci_fpcurproc == p);
ci                849 arch/i386/isa/npx.c 	npxsave_cpu(ci, save);
ci                180 arch/i386/pci/ichpcib.c 	struct cpu_info *ci;
ci                186 arch/i386/pci/ichpcib.c 	ci = curcpu();
ci                187 arch/i386/pci/ichpcib.c 	family = (ci->ci_signature >> 8) & 15;
ci                188 arch/i386/pci/ichpcib.c 	model = (ci->ci_signature >> 4) & 15;
ci                189 arch/i386/pci/ichpcib.c 	stepping = ci->ci_signature & 15;
ci                269 dev/ccd.c      	struct ccdcinfo *ci = NULL;
ci                307 dev/ccd.c      		ci = &cs->sc_cinfo[ix];
ci                308 dev/ccd.c      		ci->ci_vp = vp;
ci                315 dev/ccd.c      		    MAXPATHLEN, &ci->ci_pathlen);
ci                323 dev/ccd.c      		ci->ci_path = malloc(ci->ci_pathlen, M_DEVBUF, M_WAITOK);
ci                324 dev/ccd.c      		bcopy(tmppath, ci->ci_path, ci->ci_pathlen);
ci                332 dev/ccd.c      			    cs->sc_xname, ci->ci_path, error));
ci                333 dev/ccd.c      			free(ci->ci_path, M_DEVBUF);
ci                337 dev/ccd.c      		ci->ci_dev = va.va_rdev;
ci                347 dev/ccd.c      			    cs->sc_xname, ci->ci_path, error));
ci                348 dev/ccd.c      			free(ci->ci_path, M_DEVBUF);
ci                361 dev/ccd.c      			    cs->sc_xname, ci->ci_path));
ci                362 dev/ccd.c      			free(ci->ci_path, M_DEVBUF);
ci                376 dev/ccd.c      			    ("%s: %s: size == 0\n", cs->sc_xname, ci->ci_path));
ci                377 dev/ccd.c      			free(ci->ci_path, M_DEVBUF);
ci                384 dev/ccd.c      		ci->ci_size = size;
ci                399 dev/ccd.c      		free(ci->ci_path, M_DEVBUF);
ci                414 dev/ccd.c      			free(ci->ci_path, M_DEVBUF);
ci                422 dev/ccd.c      			free(ci->ci_path, M_DEVBUF);
ci                434 dev/ccd.c      		for (ci = cs->sc_cinfo;
ci                435 dev/ccd.c      		     ci < &cs->sc_cinfo[cs->sc_nccdisks]; ci++)
ci                436 dev/ccd.c      			ci->ci_size = minsize;
ci                468 dev/ccd.c      	struct ccdcinfo *ci, *smallci;
ci                523 dev/ccd.c      		for (ci = cs->sc_cinfo;
ci                524 dev/ccd.c      		    ci < &cs->sc_cinfo[cs->sc_nccdisks]; ci++)
ci                525 dev/ccd.c      			if (ci->ci_size > size &&
ci                527 dev/ccd.c      			    ci->ci_size < smallci->ci_size))
ci                528 dev/ccd.c      				smallci = ci;
ci                549 dev/ccd.c      		for (ci = cs->sc_cinfo;
ci                550 dev/ccd.c      		    ci < &cs->sc_cinfo[cs->sc_nccunits]; ci++)
ci                551 dev/ccd.c      			if (ci->ci_size >= smallci->ci_size)
ci                552 dev/ccd.c      				ii->ii_index[ix++] = ci - cs->sc_cinfo;
ci                766 dev/ccd.c      	struct ccdcinfo *ci, *ci2 = NULL;
ci                788 dev/ccd.c      		for (ccdisk = 0, ci = &cs->sc_cinfo[ccdisk];
ci                789 dev/ccd.c      		    cbn >= sblk + ci->ci_size;
ci                790 dev/ccd.c      		    ccdisk++, ci = &cs->sc_cinfo[ccdisk])
ci                791 dev/ccd.c      			sblk += ci->ci_size;
ci                819 dev/ccd.c      			      ci->ci_flags & CCIF_FAILED))
ci                823 dev/ccd.c      		ci = &cs->sc_cinfo[ccdisk];
ci                825 dev/ccd.c      		    ccdisk, cbn, ci, ci2));
ci                830 dev/ccd.c      		cnt = dbtob(ci->ci_size - cbn);
ci                845 dev/ccd.c      	nbp->b_dev = ci->ci_dev;		/* XXX */
ci                847 dev/ccd.c      	nbp->b_vp = ci->ci_vp;
ci                879 dev/ccd.c      	    ci->ci_dev, ci-cs->sc_cinfo, cbp, bp->b_blkno,
ci                107 dev/i2o/iopsp.c 		struct	i2o_param_hba_ctlr_info ci;
ci                123 dev/i2o/iopsp.c 	printf("iopsp_match: bustype = %d\n", param.ci.bustype);
ci                126 dev/i2o/iopsp.c 	return (param.ci.bustype == I2O_HBA_BUS_SCSI ||
ci                127 dev/i2o/iopsp.c 	    param.ci.bustype == I2O_HBA_BUS_FCA);
ci                144 dev/i2o/iopsp.c 			struct	i2o_param_hba_ctlr_info ci;
ci                171 dev/i2o/iopsp.c 	fcal = (param.p.ci.bustype == I2O_HBA_BUS_FCA);		/* XXX */
ci               1888 dev/ic/aac.c   	u_int32_t pi, ci;
ci               1898 dev/ic/aac.c   	ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
ci               1905 dev/ic/aac.c   	if ((pi + 1) == ci) {
ci               1941 dev/ic/aac.c   	u_int32_t pi, ci;
ci               1948 dev/ic/aac.c   	ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
ci               1951 dev/ic/aac.c   	if (ci == pi) {
ci               1961 dev/ic/aac.c   	if (ci == pi + 1)
ci               1965 dev/ic/aac.c   	if (ci >= aac_qinfo[queue].size)
ci               1966 dev/ic/aac.c   		ci = 0;
ci               1969 dev/ic/aac.c   	*fib_size = (sc->aac_qentries[queue] + ci)->aq_fib_size;
ci               1980 dev/ic/aac.c   		fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr /
ci               1997 dev/ic/aac.c   		fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr;
ci               2018 dev/ic/aac.c   	sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX] = ci + 1;
ci               2035 dev/ic/aac.c   	u_int32_t pi, ci;
ci               2047 dev/ic/aac.c   	ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
ci               2054 dev/ic/aac.c   	if ((pi + 1) == ci) {
ci                443 dev/ic/pdq.c       pdq_command_info_t * const ci = &pdq->pdq_command_info;
ci                452 dev/ic/pdq.c       if (ci->ci_command_active || ci->ci_pending_commands == 0)
ci                459 dev/ic/pdq.c       for (mask = 1 << ((int) op); (mask & ci->ci_pending_commands) == 0; mask >>= 1)
ci                479 dev/ic/pdq.c       dbp->pdqdb_command_requests[ci->ci_request_producer].txd_seg_len = cmdlen;
ci                480 dev/ic/pdq.c       PDQ_ADVANCE(ci->ci_request_producer, 1, PDQ_RING_MASK(dbp->pdqdb_command_requests));
ci                486 dev/ic/pdq.c       dbp->pdqdb_command_responses[ci->ci_response_producer].rxd_seg_len_hi = cmdlen / 16;
ci                487 dev/ic/pdq.c       PDQ_ADVANCE(ci->ci_response_producer, 1, PDQ_RING_MASK(dbp->pdqdb_command_responses));
ci                494 dev/ic/pdq.c       PDQ_OS_MEMZERO(ci->ci_bufstart, cmdlen);
ci                495 dev/ic/pdq.c       *(pdq_cmd_code_t *) ci->ci_bufstart = op;
ci                496 dev/ic/pdq.c       ci->ci_pending_commands &= ~mask;
ci                503 dev/ic/pdq.c   	    pdq_cmd_filter_set_t *filter_set = (pdq_cmd_filter_set_t *) ci->ci_bufstart;
ci                521 dev/ic/pdq.c   	    pdq_cmd_addr_filter_set_t *addr_filter_set = (pdq_cmd_addr_filter_set_t *) ci->ci_bufstart;
ci                543 dev/ic/pdq.c       ci->ci_command_active++;
ci                544 dev/ic/pdq.c       PDQ_CSR_WRITE(csrs, csr_cmd_response_producer, ci->ci_response_producer | (ci->ci_response_completion << 8));
ci                545 dev/ic/pdq.c       PDQ_CSR_WRITE(csrs, csr_cmd_request_producer, ci->ci_request_producer | (ci->ci_request_completion << 8));
ci                553 dev/ic/pdq.c       pdq_command_info_t * const ci = &pdq->pdq_command_info;
ci                564 dev/ic/pdq.c       if (cbp->pdqcb_command_response == ci->ci_response_completion)
ci                567 dev/ic/pdq.c       PDQ_ASSERT(cbp->pdqcb_command_request != ci->ci_request_completion);
ci                569 dev/ic/pdq.c       rspgen = (const pdq_response_generic_t *) ci->ci_bufstart;
ci                580 dev/ic/pdq.c       PDQ_ADVANCE(ci->ci_request_completion, 1, PDQ_RING_MASK(dbp->pdqdb_command_requests));
ci                581 dev/ic/pdq.c       PDQ_ADVANCE(ci->ci_response_completion, 1, PDQ_RING_MASK(dbp->pdqdb_command_responses));
ci                582 dev/ic/pdq.c       ci->ci_command_active = 0;
ci                584 dev/ic/pdq.c       if (ci->ci_pending_commands != 0) {
ci                588 dev/ic/pdq.c   		      ci->ci_response_producer | (ci->ci_response_completion << 8));
ci                590 dev/ic/pdq.c   		      ci->ci_request_producer | (ci->ci_request_completion << 8));
ci                437 dev/usb/uvisor.c uvisor_init(struct uvisor_softc *sc, struct uvisor_connection_info *ci,
ci                474 dev/usb/uvisor.c 		err = usbd_do_request_flags(sc->sc_udev, &req, ci,
ci                300 isofs/cd9660/cd9660_vnops.c 		struct cluster_info *ci = &ip->i_ci;
ci                314 isofs/cd9660/cd9660_vnops.c 		if (ci->ci_lastr + 1 == lbn) {
ci                334 isofs/cd9660/cd9660_vnops.c 		ci->ci_lastr = lbn;
ci                229 kern/kern_clock.c 	struct cpu_info *ci = curcpu();
ci                254 kern/kern_clock.c 	if (--ci->ci_schedstate.spc_rrticks <= 0)
ci                255 kern/kern_clock.c 		roundrobin(ci);
ci                261 kern/kern_clock.c 	if (CPU_IS_PRIMARY(ci) == 0)
ci                481 kern/kern_clock.c 	struct cpu_info *ci = curcpu();
ci                482 kern/kern_clock.c 	struct schedstate_percpu *spc = &ci->ci_schedstate;
ci                438 kern/kern_sysctl.c 		struct cpu_info *ci;
ci                443 kern/kern_sysctl.c 		CPU_INFO_FOREACH(cii, ci) {
ci                445 kern/kern_sysctl.c 				cp_time[i] += ci->ci_schedstate.spc_cp_time[i];
ci               1874 kern/kern_sysctl.c 	struct cpu_info *ci;
ci               1882 kern/kern_sysctl.c 	CPU_INFO_FOREACH(cii, ci) {
ci               1890 kern/kern_sysctl.c 	    &ci->ci_schedstate.spc_cp_time,
ci               1891 kern/kern_sysctl.c 	    sizeof(ci->ci_schedstate.spc_cp_time)));
ci                 95 kern/sched_bsd.c roundrobin(struct cpu_info *ci)
ci                 97 kern/sched_bsd.c 	struct schedstate_percpu *spc = &ci->ci_schedstate;
ci                480 kern/sched_bsd.c 	struct cpu_info *ci;
ci                504 kern/sched_bsd.c 	ci = (p->p_cpu != NULL) ? p->p_cpu : curcpu();
ci                505 kern/sched_bsd.c 	if (pri < ci->ci_schedstate.spc_curpriority)
ci                506 kern/sched_bsd.c 		need_resched(ci);
ci                 62 kern/vfs_cluster.c cluster_write(struct buf *bp, struct cluster_info *ci, u_quad_t filesize)
ci                 73 kern/vfs_cluster.c 		ci->ci_lasta = ci->ci_clen = ci->ci_cstart = ci->ci_lastw = 0;
ci                 75 kern/vfs_cluster.c 	if (ci->ci_clen == 0 || lbn != ci->ci_lastw + 1 ||
ci                 76 kern/vfs_cluster.c 	    (bp->b_blkno != ci->ci_lasta + btodb(bp->b_bcount))) {
ci                 78 kern/vfs_cluster.c 		if (ci->ci_clen != 0) {
ci                 88 kern/vfs_cluster.c 			cursize = ci->ci_lastw - ci->ci_cstart + 1;
ci                 90 kern/vfs_cluster.c 			    lbn != ci->ci_lastw + 1 || ci->ci_clen <= cursize) {
ci                 92 kern/vfs_cluster.c 				    ci->ci_cstart, cursize, lbn);
ci                 97 kern/vfs_cluster.c 				buflist = cluster_collectbufs(vp, ci, bp);
ci                109 kern/vfs_cluster.c 					    ci->ci_cstart, cursize, lbn);
ci                118 kern/vfs_cluster.c 					ci->ci_lastw = lbn;
ci                119 kern/vfs_cluster.c 					ci->ci_lasta = bp->b_blkno;
ci                133 kern/vfs_cluster.c 			ci->ci_clen = 0;
ci                134 kern/vfs_cluster.c 			ci->ci_lasta = bp->b_blkno;
ci                135 kern/vfs_cluster.c 			ci->ci_cstart = lbn + 1;
ci                136 kern/vfs_cluster.c 			ci->ci_lastw = lbn;
ci                139 kern/vfs_cluster.c 		ci->ci_clen = maxclen;
ci                141 kern/vfs_cluster.c 			ci->ci_cstart = lbn + 1;
ci                144 kern/vfs_cluster.c 			ci->ci_cstart = lbn;
ci                147 kern/vfs_cluster.c 	} else if (lbn == ci->ci_cstart + ci->ci_clen) {
ci                151 kern/vfs_cluster.c 		cluster_wbuild(vp, bp, bp->b_bcount, ci->ci_cstart,
ci                152 kern/vfs_cluster.c 		    ci->ci_clen + 1, lbn);
ci                153 kern/vfs_cluster.c 		ci->ci_clen = 0;
ci                154 kern/vfs_cluster.c 		ci->ci_cstart = lbn + 1;
ci                161 kern/vfs_cluster.c 	ci->ci_lastw = lbn;
ci                162 kern/vfs_cluster.c 	ci->ci_lasta = bp->b_blkno;
ci                225 kern/vfs_cluster.c cluster_collectbufs(struct vnode *vp, struct cluster_info *ci,
ci                232 kern/vfs_cluster.c 	len = ci->ci_lastw - ci->ci_cstart + 1;
ci                237 kern/vfs_cluster.c 	for (lbn = ci->ci_cstart, i = 0; i < len; lbn++, i++)