HelenOS sources

CPU                98 kernel/arch/amd64/src/cpu/cpu.c 	CPU->arch.tss = tss_p;
CPU                99 kernel/arch/amd64/src/cpu/cpu.c 	CPU->arch.tss->iomap_base = &CPU->arch.tss->iomap[0] -
CPU               100 kernel/arch/amd64/src/cpu/cpu.c 	    ((uint8_t *) CPU->arch.tss);
CPU               107 kernel/arch/amd64/src/cpu/cpu.c 	CPU->arch.vendor = VendorUnknown;
CPU               117 kernel/arch/amd64/src/cpu/cpu.c 			CPU->arch.vendor = VendorAMD;
CPU               126 kernel/arch/amd64/src/cpu/cpu.c 			CPU->arch.vendor = VendorIntel;
CPU               130 kernel/arch/amd64/src/cpu/cpu.c 		CPU->arch.family = (info.cpuid_eax >> 8) & 0xf;
CPU               131 kernel/arch/amd64/src/cpu/cpu.c 		CPU->arch.model = (info.cpuid_eax >> 4) & 0xf;
CPU               132 kernel/arch/amd64/src/cpu/cpu.c 		CPU->arch.stepping = (info.cpuid_eax >> 0) & 0xf;
CPU                70 kernel/arch/amd64/src/ddi/ddi.c 		    CPU->arch.tss->iomap);
CPU               113 kernel/arch/amd64/src/ddi/ddi.c 	CPU->arch.iomapver_copy = ver;
CPU               149 kernel/arch/amd64/src/drivers/i8254.c 	CPU->delay_loop_const =
CPU               157 kernel/arch/amd64/src/drivers/i8254.c 	CPU->frequency_mhz = (clk2 - clk1) >> SHIFT;
CPU               112 kernel/arch/amd64/src/interrupt.c 		if (CPU->arch.iomapver_copy != ver) {
CPU               179 kernel/arch/amd64/src/interrupt.c 		log(LF_ARCH, LVL_DEBUG, "cpu%u: unhandled IRQ %u", CPU->id,
CPU               198 kernel/arch/amd64/src/interrupt.c 	log(LF_ARCH, LVL_DEBUG, "cpu%u: PIC spurious interrupt %u", CPU->id,
CPU                57 kernel/arch/amd64/src/proc/scheduler.c 	CPU->arch.tss->rsp0 = (uintptr_t) &THREAD->kstack[STACK_SIZE];
CPU               469 kernel/arch/amd64/src/smp/apic.c 	if (CPU->arch.family >= 6)
CPU               516 kernel/arch/amd64/src/smp/apic.c 	assert(CPU->id < 8);
CPU               520 kernel/arch/amd64/src/smp/apic.c 	ldr.id = (uint8_t) (1 << CPU->id);
CPU               543 kernel/arch/amd64/src/smp/apic.c 	    CPU->id, l_apic_id());
CPU               205 kernel/arch/arm32/src/cpu/cpu.c 	arch_cpu_identify(&CPU->arch);
CPU               317 kernel/arch/arm32/src/cpu/cpu.c 	for (unsigned i = 0; i < CPU->arch.dcache_levels; ++i)
CPU               319 kernel/arch/arm32/src/cpu/cpu.c 		    CPU->arch.dcache[i].ways, CPU->arch.dcache[i].sets,
CPU               320 kernel/arch/arm32/src/cpu/cpu.c 		    CPU->arch.dcache[i].way_shift, CPU->arch.dcache[i].set_shift);
CPU               328 kernel/arch/arm32/src/cpu/cpu.c 		    CPU->arch.dcache[i].ways, CPU->arch.dcache[i].sets,
CPU               329 kernel/arch/arm32/src/cpu/cpu.c 		    CPU->arch.dcache[i].way_shift, CPU->arch.dcache[i].set_shift);
CPU               162 kernel/arch/arm32/src/mach/beagleboardxm/beagleboardxm.c 		    CPU->id, inum);
CPU               152 kernel/arch/arm32/src/mach/gta02/gta02.c 		    CPU->id, inum);
CPU               275 kernel/arch/arm32/src/mach/integratorcp/integratorcp.c 				    CPU->id, i);
CPU               164 kernel/arch/arm32/src/mach/raspberrypi/raspberrypi.c 		printf("cpu%d: spurious interrupt (inum=%d)\n", CPU->id, inum);
CPU               133 kernel/arch/arm64/src/arm64.c 	CPU->delay_loop_const = 1;
CPU                80 kernel/arch/arm64/src/cpu/cpu.c 	CPU->arch.implementer =
CPU                82 kernel/arch/arm64/src/cpu/cpu.c 	CPU->arch.variant = (midr & MIDR_VARIANT_MASK) >> MIDR_VARIANT_SHIFT;
CPU                83 kernel/arch/arm64/src/cpu/cpu.c 	CPU->arch.partnum = (midr & MIDR_PARTNUM_MASK) >> MIDR_PARTNUM_SHIFT;
CPU                84 kernel/arch/arm64/src/cpu/cpu.c 	CPU->arch.revision = (midr & MIDR_REVISION_MASK) >> MIDR_REVISION_SHIFT;
CPU                80 kernel/arch/arm64/src/mach/hikey960/hikey960.c 		printf("cpu%d: spurious interrupt (inum=%u)\n", CPU->id, inum);
CPU                98 kernel/arch/arm64/src/mach/virt/virt.c 		printf("cpu%d: spurious interrupt (inum=%u)\n", CPU->id, inum);
CPU                86 kernel/arch/ia32/src/cpu/cpu.c 	CPU->arch.tss = tss_p;
CPU                87 kernel/arch/ia32/src/cpu/cpu.c 	CPU->arch.tss->iomap_base = &CPU->arch.tss->iomap[0] - ((uint8_t *) CPU->arch.tss);
CPU                91 kernel/arch/ia32/src/cpu/cpu.c 	CPU->arch.fi.word = info.cpuid_edx;
CPU                93 kernel/arch/ia32/src/cpu/cpu.c 	if (CPU->arch.fi.bits.fxsr)
CPU                98 kernel/arch/ia32/src/cpu/cpu.c 	if (CPU->arch.fi.bits.sse) {
CPU               109 kernel/arch/ia32/src/cpu/cpu.c 	if (CPU->arch.fi.bits.sep) {
CPU               120 kernel/arch/ia32/src/cpu/cpu.c 	CPU->arch.vendor = VendorUnknown;
CPU               130 kernel/arch/ia32/src/cpu/cpu.c 			CPU->arch.vendor = VendorAMD;
CPU               138 kernel/arch/ia32/src/cpu/cpu.c 			CPU->arch.vendor = VendorIntel;
CPU               141 kernel/arch/ia32/src/cpu/cpu.c 		CPU->arch.family = (info.cpuid_eax >> 8) & 0x0fU;
CPU               142 kernel/arch/ia32/src/cpu/cpu.c 		CPU->arch.model = (info.cpuid_eax >> 4) & 0x0fU;
CPU               143 kernel/arch/ia32/src/cpu/cpu.c 		CPU->arch.stepping = (info.cpuid_eax >> 0) & 0x0fU;
CPU                70 kernel/arch/ia32/src/ddi/ddi.c 		    CPU->arch.tss->iomap);
CPU               112 kernel/arch/ia32/src/ddi/ddi.c 	CPU->arch.iomapver_copy = ver;
CPU               149 kernel/arch/ia32/src/drivers/i8254.c 	CPU->delay_loop_const =
CPU               157 kernel/arch/ia32/src/drivers/i8254.c 	CPU->frequency_mhz = (clk2 - clk1) >> SHIFT;
CPU               120 kernel/arch/ia32/src/interrupt.c 		if (CPU->arch.iomapver_copy != ver) {
CPU               200 kernel/arch/ia32/src/interrupt.c 		log(LF_ARCH, LVL_DEBUG, "cpu%u: unhandled IRQ %u", CPU->id,
CPU               219 kernel/arch/ia32/src/interrupt.c 	log(LF_ARCH, LVL_DEBUG, "cpu%u: PIC spurious interrupt %u", CPU->id,
CPU                63 kernel/arch/ia32/src/proc/scheduler.c 	if (CPU->arch.fi.bits.sep) {
CPU                70 kernel/arch/ia32/src/proc/scheduler.c 	CPU->arch.tss->esp0 = kstk;
CPU                71 kernel/arch/ia32/src/proc/scheduler.c 	CPU->arch.tss->ss0 = GDT_SELECTOR(KDATA_DES);
CPU               469 kernel/arch/ia32/src/smp/apic.c 	if (CPU->arch.family >= 6)
CPU               516 kernel/arch/ia32/src/smp/apic.c 	assert(CPU->id < 8);
CPU               520 kernel/arch/ia32/src/smp/apic.c 	ldr.id = (uint8_t) (1 << CPU->id);
CPU               543 kernel/arch/ia32/src/smp/apic.c 	    CPU->id, l_apic_id());
CPU                47 kernel/arch/ia64/src/cpu/cpu.c 	CPU->arch.cpuid0 = cpuid_read(0);
CPU                48 kernel/arch/ia64/src/cpu/cpu.c 	CPU->arch.cpuid1 = cpuid_read(1);
CPU                49 kernel/arch/ia64/src/cpu/cpu.c 	CPU->arch.cpuid3.value = cpuid_read(3);
CPU                57 kernel/arch/ia64/src/cpu/cpu.c 	memcpy(vendor, &CPU->arch.cpuid0, 8);
CPU                58 kernel/arch/ia64/src/cpu/cpu.c 	memcpy(vendor + 8, &CPU->arch.cpuid1, 8);
CPU                73 kernel/arch/ia64/src/cpu/cpu.c 	printf("cpu%d: %s (%s), archrev=%d, model=%d, revision=%d\n", CPU->id,
CPU                74 kernel/arch/ia64/src/cpu/cpu.c 	    family_str, vendor, CPU->arch.cpuid3.archrev,
CPU                75 kernel/arch/ia64/src/cpu/cpu.c 	    CPU->arch.cpuid3.model, CPU->arch.cpuid3.revision);
CPU               254 kernel/arch/ia64/src/interrupt.c 		printf("cpu%d: spurious interrupt\n", CPU->id);
CPU               100 kernel/arch/mips32/src/cpu/cpu.c 	CPU->arch.rev_num = cp0_prid_read() & 0xff;
CPU               101 kernel/arch/mips32/src/cpu/cpu.c 	CPU->arch.imp_num = (cp0_prid_read() >> 8) & 0xff;
CPU               195 kernel/arch/mips32/src/exception.c 				    CPU->id, i);
CPU                81 kernel/arch/mips32/src/mach/malta/malta.c 		    CPU->id, isa_irq);
CPU                92 kernel/arch/mips32/src/mach/malta/malta.c 		    CPU->id, isa_irq);
CPU                72 kernel/arch/mips32/src/mach/msim/msim.c 		    CPU->id, i);
CPU                50 kernel/arch/ppc32/src/cpu/cpu.c 	cpu_version(&CPU->arch);
CPU               141 kernel/arch/ppc32/src/interrupt.c 			    " (inum=%" PRIu8 ")", CPU->id, inum);
CPU                64 kernel/arch/sparc64/src/cpu/sun4u/cpu.c 		if (mid == CPU->arch.mid) {
CPU                83 kernel/arch/sparc64/src/cpu/sun4u/cpu.c 	CPU->arch.mid = read_mid();
CPU               112 kernel/arch/sparc64/src/cpu/sun4u/cpu.c 	CPU->arch.clock_frequency = clock_frequency;
CPU               119 kernel/arch/sparc64/src/cpu/sun4u/cpu.c 	CPU->arch.ver.value = ver_read();
CPU               149 kernel/arch/sparc64/src/cpu/sun4u/cpu.c 	switch (CPU->arch.ver.impl) {
CPU                53 kernel/arch/sparc64/src/cpu/sun4v/cpu.c 	CPU->arch.id = myid;
CPU                69 kernel/arch/sparc64/src/cpu/sun4v/cpu.c 			CPU->arch.clock_frequency = clock_frequency;
CPU                56 kernel/arch/sparc64/src/drivers/tick.c 	    CPU->arch.clock_frequency / HZ;
CPU                57 kernel/arch/sparc64/src/drivers/tick.c 	CPU->arch.next_tick_cmpr = compare.tick_cmpr;
CPU               116 kernel/arch/sparc64/src/drivers/tick.c 	drift = tick_counter_read() - CPU->arch.next_tick_cmpr;
CPU               117 kernel/arch/sparc64/src/drivers/tick.c 	while (drift > CPU->arch.clock_frequency / HZ) {
CPU               118 kernel/arch/sparc64/src/drivers/tick.c 		drift -= CPU->arch.clock_frequency / HZ;
CPU               121 kernel/arch/sparc64/src/drivers/tick.c 	CPU->arch.next_tick_cmpr = tick_counter_read() +
CPU               122 kernel/arch/sparc64/src/drivers/tick.c 	    (CPU->arch.clock_frequency / HZ) - drift;
CPU               123 kernel/arch/sparc64/src/drivers/tick.c 	tick_compare_write(CPU->arch.next_tick_cmpr);
CPU               168 kernel/arch/sparc64/src/smp/sun4u/ipi.c 		if (&cpus[i] == CPU)
CPU               103 kernel/arch/sparc64/src/smp/sun4u/smp.c 	if (CPU->arch.mid == mid)
CPU                65 kernel/arch/sparc64/src/smp/sun4v/ipi.c 	data[CPU->arch.id][0] = (uint64_t) func;
CPU                69 kernel/arch/sparc64/src/smp/sun4v/ipi.c 		ipi_cpu_list[CPU->arch.id][i] = cpu_list[i];
CPU                73 kernel/arch/sparc64/src/smp/sun4v/ipi.c 	    KA2PA(ipi_cpu_list[CPU->arch.id]), KA2PA(data[CPU->arch.id]));
CPU                86 kernel/arch/sparc64/src/smp/sun4v/ipi.c 	ipi_cpu_list[CPU->arch.id][0] = cpu_id;
CPU                87 kernel/arch/sparc64/src/smp/sun4v/ipi.c 	return ipi_brodcast_to(func, ipi_cpu_list[CPU->arch.id], 1);
CPU               113 kernel/arch/sparc64/src/smp/sun4v/ipi.c 		if (&cpus[i] == CPU) {
CPU               117 kernel/arch/sparc64/src/smp/sun4v/ipi.c 		ipi_cpu_list[CPU->id][idx] = (uint16_t) cpus[i].id;
CPU               121 kernel/arch/sparc64/src/smp/sun4v/ipi.c 	ipi_brodcast_to(func, ipi_cpu_list[CPU->arch.id], idx);
CPU               143 kernel/arch/sparc64/src/sun4u/sparc64.c 	CPU->delay_loop_const = 1;
CPU               155 kernel/arch/sparc64/src/sun4u/sparc64.c 	    CPU->arch.clock_frequency / 1000000;
CPU               141 kernel/arch/sparc64/src/sun4v/sparc64.c 	CPU->delay_loop_const = 1;
CPU               153 kernel/arch/sparc64/src/sun4v/sparc64.c 	    CPU->arch.clock_frequency / 1000000;
CPU               101 kernel/arch/sparc64/src/trap/sun4u/interrupt.c 		    PRIx64 ")", CPU->id, intrcv, data0);
CPU                86 kernel/arch/sparc64/src/trap/sun4v/interrupt.c 	    KA2PA(cpu_mondo_queues[CPU->id]),
CPU                89 kernel/arch/sparc64/src/trap/sun4v/interrupt.c 		    CPU->arch.id);
CPU               103 kernel/arch/sparc64/src/trap/sun4v/interrupt.c 		uint64_t data1 = cpu_mondo_queues[CPU->id][0];
CPU               113 kernel/arch/sparc64/src/trap/sun4v/interrupt.c 			    ", data = %" PRIx64 ".", CPU->arch.id, data1);
CPU                47 kernel/generic/include/cpu.h #define CPU_LOCAL            (&CPU->local)
CPU                59 kernel/generic/include/synch/spinlock.h 		    #pname, (value), CPU->id, __func__, __LINE__); \
CPU               108 kernel/generic/src/console/chardev.c 		if (CPU)
CPU               109 kernel/generic/src/console/chardev.c 			printf("cpu%u: ", CPU->id);
CPU               101 kernel/generic/src/cpu/cpu.c 	CPU = &cpus[config.cpu_active - 1];
CPU               103 kernel/generic/src/cpu/cpu.c 	CPU->active = true;
CPU               104 kernel/generic/src/cpu/cpu.c 	CPU->tlb_active = true;
CPU               108 kernel/generic/src/cpu/cpu.c 	CPU->idle_cycles = ATOMIC_TIME_INITIALIZER();
CPU               109 kernel/generic/src/cpu/cpu.c 	CPU->busy_cycles = ATOMIC_TIME_INITIALIZER();
CPU                53 kernel/generic/src/debug/panic.c 	if (CPU)
CPU                54 kernel/generic/src/debug/panic.c 		printf("on cpu%u ", CPU->id);
CPU               120 kernel/generic/src/interrupt/interrupt.c 	if (CPU && CPU_LOCAL->idle) {
CPU               122 kernel/generic/src/interrupt/interrupt.c 		atomic_time_increment(&CPU->idle_cycles, now - CPU_LOCAL->last_cycle);
CPU                74 kernel/generic/src/lib/halt.c 	if (CPU)
CPU                75 kernel/generic/src/lib/halt.c 		log(LF_OTHER, LVL_NOTE, "cpu%u: halted", CPU->id);
CPU               396 kernel/generic/src/mm/slab.c 	slab_magazine_t *cmag = cache->mag_cache[CPU->id].current;
CPU               397 kernel/generic/src/mm/slab.c 	slab_magazine_t *lastmag = cache->mag_cache[CPU->id].last;
CPU               399 kernel/generic/src/mm/slab.c 	assert(irq_spinlock_locked(&cache->mag_cache[CPU->id].lock));
CPU               406 kernel/generic/src/mm/slab.c 			cache->mag_cache[CPU->id].current = lastmag;
CPU               407 kernel/generic/src/mm/slab.c 			cache->mag_cache[CPU->id].last = cmag;
CPU               420 kernel/generic/src/mm/slab.c 	cache->mag_cache[CPU->id].last = cmag;
CPU               421 kernel/generic/src/mm/slab.c 	cache->mag_cache[CPU->id].current = newmag;
CPU               433 kernel/generic/src/mm/slab.c 	if (!CPU)
CPU               436 kernel/generic/src/mm/slab.c 	irq_spinlock_lock(&cache->mag_cache[CPU->id].lock, true);
CPU               440 kernel/generic/src/mm/slab.c 		irq_spinlock_unlock(&cache->mag_cache[CPU->id].lock, true);
CPU               445 kernel/generic/src/mm/slab.c 	irq_spinlock_unlock(&cache->mag_cache[CPU->id].lock, true);
CPU               463 kernel/generic/src/mm/slab.c 	slab_magazine_t *cmag = cache->mag_cache[CPU->id].current;
CPU               464 kernel/generic/src/mm/slab.c 	slab_magazine_t *lastmag = cache->mag_cache[CPU->id].last;
CPU               466 kernel/generic/src/mm/slab.c 	assert(irq_spinlock_locked(&cache->mag_cache[CPU->id].lock));
CPU               473 kernel/generic/src/mm/slab.c 			cache->mag_cache[CPU->id].last = cmag;
CPU               474 kernel/generic/src/mm/slab.c 			cache->mag_cache[CPU->id].current = lastmag;
CPU               500 kernel/generic/src/mm/slab.c 	cache->mag_cache[CPU->id].last = cmag;
CPU               501 kernel/generic/src/mm/slab.c 	cache->mag_cache[CPU->id].current = newmag;
CPU               513 kernel/generic/src/mm/slab.c 	if (!CPU)
CPU               516 kernel/generic/src/mm/slab.c 	irq_spinlock_lock(&cache->mag_cache[CPU->id].lock, true);
CPU               520 kernel/generic/src/mm/slab.c 		irq_spinlock_unlock(&cache->mag_cache[CPU->id].lock, true);
CPU               526 kernel/generic/src/mm/slab.c 	irq_spinlock_unlock(&cache->mag_cache[CPU->id].lock, true);
CPU                87 kernel/generic/src/mm/tlb.c 	CPU->tlb_active = false;
CPU                92 kernel/generic/src/mm/tlb.c 		if (i == CPU->id)
CPU               140 kernel/generic/src/mm/tlb.c 	CPU->tlb_active = true;
CPU               154 kernel/generic/src/mm/tlb.c 	assert(CPU);
CPU               156 kernel/generic/src/mm/tlb.c 	CPU->tlb_active = false;
CPU               160 kernel/generic/src/mm/tlb.c 	irq_spinlock_lock(&CPU->tlb_lock, false);
CPU               161 kernel/generic/src/mm/tlb.c 	assert(CPU->tlb_messages_count <= TLB_MESSAGE_QUEUE_LEN);
CPU               164 kernel/generic/src/mm/tlb.c 	for (i = 0; i < CPU->tlb_messages_count; i++) {
CPU               165 kernel/generic/src/mm/tlb.c 		tlb_invalidate_type_t type = CPU->tlb_messages[i].type;
CPU               166 kernel/generic/src/mm/tlb.c 		asid_t asid = CPU->tlb_messages[i].asid;
CPU               167 kernel/generic/src/mm/tlb.c 		uintptr_t page = CPU->tlb_messages[i].page;
CPU               168 kernel/generic/src/mm/tlb.c 		size_t count = CPU->tlb_messages[i].count;
CPU               190 kernel/generic/src/mm/tlb.c 	CPU->tlb_messages_count = 0;
CPU               191 kernel/generic/src/mm/tlb.c 	irq_spinlock_unlock(&CPU->tlb_lock, false);
CPU               192 kernel/generic/src/mm/tlb.c 	CPU->tlb_active = true;
CPU                76 kernel/generic/src/proc/scheduler.c 	irq_spinlock_lock(&CPU->fpu_lock, false);
CPU                79 kernel/generic/src/proc/scheduler.c 	thread_t *owner = atomic_load_explicit(&CPU->fpu_owner, memory_order_relaxed);
CPU                82 kernel/generic/src/proc/scheduler.c 		atomic_store_explicit(&CPU->fpu_owner, NULL, memory_order_relaxed);
CPU                85 kernel/generic/src/proc/scheduler.c 	irq_spinlock_unlock(&CPU->fpu_lock, false);
CPU                94 kernel/generic/src/proc/scheduler.c 	atomic_store_explicit(&CPU->fpu_owner, THREAD, memory_order_relaxed);
CPU               119 kernel/generic/src/proc/scheduler.c 	assert(CPU != NULL);
CPU               121 kernel/generic/src/proc/scheduler.c 	if (atomic_load(&CPU->nrdy) == 0)
CPU               125 kernel/generic/src/proc/scheduler.c 		irq_spinlock_lock(&(CPU->rq[i].lock), false);
CPU               126 kernel/generic/src/proc/scheduler.c 		if (CPU->rq[i].n == 0) {
CPU               130 kernel/generic/src/proc/scheduler.c 			irq_spinlock_unlock(&(CPU->rq[i].lock), false);
CPU               134 kernel/generic/src/proc/scheduler.c 		atomic_dec(&CPU->nrdy);
CPU               136 kernel/generic/src/proc/scheduler.c 		CPU->rq[i].n--;
CPU               142 kernel/generic/src/proc/scheduler.c 		    list_first(&CPU->rq[i].rq), thread_t, rq_link);
CPU               145 kernel/generic/src/proc/scheduler.c 		irq_spinlock_unlock(&(CPU->rq[i].lock), false);
CPU               166 kernel/generic/src/proc/scheduler.c 	assert(CPU != NULL);
CPU               242 kernel/generic/src/proc/scheduler.c 		irq_spinlock_lock(&CPU->rq[i].lock, false);
CPU               245 kernel/generic/src/proc/scheduler.c 		list_swap(&CPU->rq[i].rq, &list);
CPU               248 kernel/generic/src/proc/scheduler.c 		size_t tmpn = CPU->rq[i].n;
CPU               249 kernel/generic/src/proc/scheduler.c 		CPU->rq[i].n = n;
CPU               252 kernel/generic/src/proc/scheduler.c 		irq_spinlock_unlock(&CPU->rq[i].lock, false);
CPU               257 kernel/generic/src/proc/scheduler.c 		irq_spinlock_lock(&CPU->rq[start].lock, false);
CPU               258 kernel/generic/src/proc/scheduler.c 		list_concat(&CPU->rq[start].rq, &list);
CPU               259 kernel/generic/src/proc/scheduler.c 		CPU->rq[start].n += n;
CPU               260 kernel/generic/src/proc/scheduler.c 		irq_spinlock_unlock(&CPU->rq[start].lock, false);
CPU               285 kernel/generic/src/proc/scheduler.c 	thread_t *owner = atomic_load_explicit(&CPU->fpu_owner,
CPU               312 kernel/generic/src/proc/scheduler.c 	assert(atomic_get_unordered(&THREAD->cpu) == CPU);
CPU               326 kernel/generic/src/proc/scheduler.c 	    ", nrdy=%zu)", CPU->id, THREAD->tid, rq_index,
CPU               327 kernel/generic/src/proc/scheduler.c 	    THREAD->ticks, atomic_load(&CPU->nrdy));
CPU               387 kernel/generic/src/proc/scheduler.c 	assert(atomic_get_unordered(&thread->cpu) == CPU);
CPU               398 kernel/generic/src/proc/scheduler.c 	add_to_rq(thread, CPU, prio);
CPU               414 kernel/generic/src/proc/scheduler.c 		cpu = CPU;
CPU               415 kernel/generic/src/proc/scheduler.c 		atomic_set_unordered(&thread->cpu, CPU);
CPU               475 kernel/generic/src/proc/scheduler.c 	assert(CPU != NULL);
CPU               547 kernel/generic/src/proc/scheduler.c 	assert(CPU != NULL);
CPU               569 kernel/generic/src/proc/scheduler.c 		assert(CPU != NULL);
CPU               620 kernel/generic/src/proc/scheduler.c 	runq_t *new_rq = &CPU->rq[i];
CPU               647 kernel/generic/src/proc/scheduler.c 		atomic_set_unordered(&thread->cpu, CPU);
CPU               656 kernel/generic/src/proc/scheduler.c 		    "nrdy=%ld, avg=%ld", CPU->id, thread->tid,
CPU               657 kernel/generic/src/proc/scheduler.c 		    CPU->id, atomic_load(&CPU->nrdy),
CPU               673 kernel/generic/src/proc/scheduler.c 		atomic_inc(&CPU->nrdy);
CPU               710 kernel/generic/src/proc/scheduler.c 	rdy = atomic_load(&CPU->nrdy);
CPU               734 kernel/generic/src/proc/scheduler.c 			if (CPU == cpu)
CPU               745 kernel/generic/src/proc/scheduler.c 	if (atomic_load(&CPU->nrdy)) {
CPU                46 kernel/generic/src/synch/irq_spinlock.c #define CPU_OWNER ((CPU == NULL) ? (cpu_t *) UINTPTR_MAX : CPU)
CPU                86 kernel/generic/src/synch/mutex.c 			printf("cpu%u: looping on active mutex %p\n", CPU->id, mtx);
CPU                94 kernel/generic/src/synch/mutex.c 		printf("cpu%u: not deadlocked\n", CPU->id);
CPU               106 kernel/generic/src/synch/spinlock.c 			    "caller=%p (%s)\n", CPU->id, lock, lock->name,
CPU               120 kernel/generic/src/synch/spinlock.c 		printf("cpu%u: not deadlocked\n", CPU->id);
CPU               109 kernel/generic/src/time/clock.c 	if (CPU->id == 0) {
CPU               127 kernel/generic/src/time/clock.c 	atomic_time_increment(&CPU->busy_cycles, now - CPU_LOCAL->last_cycle);
CPU               156 kernel/generic/src/time/clock.c 	irq_spinlock_lock(&CPU->timeoutlock, false);
CPU               159 kernel/generic/src/time/clock.c 	while ((cur = list_first(&CPU->timeout_active_list)) != NULL) {
CPU               171 kernel/generic/src/time/clock.c 		irq_spinlock_unlock(&CPU->timeoutlock, false);
CPU               178 kernel/generic/src/time/clock.c 		irq_spinlock_lock(&CPU->timeoutlock, false);
CPU               181 kernel/generic/src/time/clock.c 	irq_spinlock_unlock(&CPU->timeoutlock, false);
CPU                60 kernel/generic/src/time/delay.c 	asm_delay_loop(usec * CPU->delay_loop_const);
CPU                56 kernel/generic/src/time/timeout.c 	irq_spinlock_initialize(&CPU->timeoutlock, "cpu.timeoutlock");
CPU                57 kernel/generic/src/time/timeout.c 	list_initialize(&CPU->timeout_active_list);
CPU                88 kernel/generic/src/time/timeout.c 		.cpu = CPU,
CPU                97 kernel/generic/src/time/timeout.c 	link_t *last = list_last(&CPU->timeout_active_list);
CPU                99 kernel/generic/src/time/timeout.c 		list_append(&timeout->link, &CPU->timeout_active_list);
CPU               101 kernel/generic/src/time/timeout.c 		for (link_t *cur = list_first(&CPU->timeout_active_list); cur != NULL;
CPU               102 kernel/generic/src/time/timeout.c 		    cur = list_next(cur, &CPU->timeout_active_list)) {
CPU               127 kernel/generic/src/time/timeout.c 	irq_spinlock_lock(&CPU->timeoutlock, true);
CPU               129 kernel/generic/src/time/timeout.c 	irq_spinlock_unlock(&CPU->timeoutlock, true);
CPU               135 kernel/generic/src/time/timeout.c 	irq_spinlock_lock(&CPU->timeoutlock, true);
CPU               137 kernel/generic/src/time/timeout.c 	irq_spinlock_unlock(&CPU->timeoutlock, true);
CPU                55 kernel/test/mm/falloc2.c 		    "Unable to allocate frames\n", THREAD->tid, CPU->id);
CPU                66 kernel/test/mm/falloc2.c 			    CPU->id, count, bytes);
CPU                79 kernel/test/mm/falloc2.c 			    "%u blocks allocated.\n", THREAD->tid, CPU->id,
CPU                82 kernel/test/mm/falloc2.c 			    "Deallocating ... \n", THREAD->tid, CPU->id);
CPU                89 kernel/test/mm/falloc2.c 						    THREAD->tid, CPU->id, ((char *) PA2KA(frames[i]))[k],
CPU                99 kernel/test/mm/falloc2.c 			    "Finished run.\n", THREAD->tid, CPU->id);
CPU               107 kernel/test/mm/falloc2.c 	    THREAD->tid, CPU->id);
CPU                66 kernel/test/synch/semaphore2.c 	TPRINTF("cpu%u, tid %" PRIu64 " down+ (%d)\n", CPU->id, THREAD->tid, to);
CPU                69 kernel/test/synch/semaphore2.c 		TPRINTF("cpu%u, tid %" PRIu64 " down!\n", CPU->id, THREAD->tid);
CPU                73 kernel/test/synch/semaphore2.c 	TPRINTF("cpu%u, tid %" PRIu64 " down=\n", CPU->id, THREAD->tid);
CPU                77 kernel/test/synch/semaphore2.c 	TPRINTF("cpu%u, tid %" PRIu64 " up\n", CPU->id, THREAD->tid);
HelenOS homepage, sources at GitHub