HelenOS sources
CPU 98 kernel/arch/amd64/src/cpu/cpu.c CPU->arch.tss = tss_p;
CPU 99 kernel/arch/amd64/src/cpu/cpu.c CPU->arch.tss->iomap_base = &CPU->arch.tss->iomap[0] -
CPU 100 kernel/arch/amd64/src/cpu/cpu.c ((uint8_t *) CPU->arch.tss);
CPU 107 kernel/arch/amd64/src/cpu/cpu.c CPU->arch.vendor = VendorUnknown;
CPU 117 kernel/arch/amd64/src/cpu/cpu.c CPU->arch.vendor = VendorAMD;
CPU 126 kernel/arch/amd64/src/cpu/cpu.c CPU->arch.vendor = VendorIntel;
CPU 130 kernel/arch/amd64/src/cpu/cpu.c CPU->arch.family = (info.cpuid_eax >> 8) & 0xf;
CPU 131 kernel/arch/amd64/src/cpu/cpu.c CPU->arch.model = (info.cpuid_eax >> 4) & 0xf;
CPU 132 kernel/arch/amd64/src/cpu/cpu.c CPU->arch.stepping = (info.cpuid_eax >> 0) & 0xf;
CPU 70 kernel/arch/amd64/src/ddi/ddi.c CPU->arch.tss->iomap);
CPU 113 kernel/arch/amd64/src/ddi/ddi.c CPU->arch.iomapver_copy = ver;
CPU 149 kernel/arch/amd64/src/drivers/i8254.c CPU->delay_loop_const =
CPU 157 kernel/arch/amd64/src/drivers/i8254.c CPU->frequency_mhz = (clk2 - clk1) >> SHIFT;
CPU 112 kernel/arch/amd64/src/interrupt.c if (CPU->arch.iomapver_copy != ver) {
CPU 179 kernel/arch/amd64/src/interrupt.c log(LF_ARCH, LVL_DEBUG, "cpu%u: unhandled IRQ %u", CPU->id,
CPU 198 kernel/arch/amd64/src/interrupt.c log(LF_ARCH, LVL_DEBUG, "cpu%u: PIC spurious interrupt %u", CPU->id,
CPU 57 kernel/arch/amd64/src/proc/scheduler.c CPU->arch.tss->rsp0 = (uintptr_t) &THREAD->kstack[STACK_SIZE];
CPU 469 kernel/arch/amd64/src/smp/apic.c if (CPU->arch.family >= 6)
CPU 516 kernel/arch/amd64/src/smp/apic.c assert(CPU->id < 8);
CPU 520 kernel/arch/amd64/src/smp/apic.c ldr.id = (uint8_t) (1 << CPU->id);
CPU 543 kernel/arch/amd64/src/smp/apic.c CPU->id, l_apic_id());
CPU 205 kernel/arch/arm32/src/cpu/cpu.c arch_cpu_identify(&CPU->arch);
CPU 317 kernel/arch/arm32/src/cpu/cpu.c for (unsigned i = 0; i < CPU->arch.dcache_levels; ++i)
CPU 319 kernel/arch/arm32/src/cpu/cpu.c CPU->arch.dcache[i].ways, CPU->arch.dcache[i].sets,
CPU 320 kernel/arch/arm32/src/cpu/cpu.c CPU->arch.dcache[i].way_shift, CPU->arch.dcache[i].set_shift);
CPU 328 kernel/arch/arm32/src/cpu/cpu.c CPU->arch.dcache[i].ways, CPU->arch.dcache[i].sets,
CPU 329 kernel/arch/arm32/src/cpu/cpu.c CPU->arch.dcache[i].way_shift, CPU->arch.dcache[i].set_shift);
CPU 162 kernel/arch/arm32/src/mach/beagleboardxm/beagleboardxm.c CPU->id, inum);
CPU 152 kernel/arch/arm32/src/mach/gta02/gta02.c CPU->id, inum);
CPU 275 kernel/arch/arm32/src/mach/integratorcp/integratorcp.c CPU->id, i);
CPU 164 kernel/arch/arm32/src/mach/raspberrypi/raspberrypi.c printf("cpu%d: spurious interrupt (inum=%d)\n", CPU->id, inum);
CPU 133 kernel/arch/arm64/src/arm64.c CPU->delay_loop_const = 1;
CPU 80 kernel/arch/arm64/src/cpu/cpu.c CPU->arch.implementer =
CPU 82 kernel/arch/arm64/src/cpu/cpu.c CPU->arch.variant = (midr & MIDR_VARIANT_MASK) >> MIDR_VARIANT_SHIFT;
CPU 83 kernel/arch/arm64/src/cpu/cpu.c CPU->arch.partnum = (midr & MIDR_PARTNUM_MASK) >> MIDR_PARTNUM_SHIFT;
CPU 84 kernel/arch/arm64/src/cpu/cpu.c CPU->arch.revision = (midr & MIDR_REVISION_MASK) >> MIDR_REVISION_SHIFT;
CPU 80 kernel/arch/arm64/src/mach/hikey960/hikey960.c printf("cpu%d: spurious interrupt (inum=%u)\n", CPU->id, inum);
CPU 98 kernel/arch/arm64/src/mach/virt/virt.c printf("cpu%d: spurious interrupt (inum=%u)\n", CPU->id, inum);
CPU 86 kernel/arch/ia32/src/cpu/cpu.c CPU->arch.tss = tss_p;
CPU 87 kernel/arch/ia32/src/cpu/cpu.c CPU->arch.tss->iomap_base = &CPU->arch.tss->iomap[0] - ((uint8_t *) CPU->arch.tss);
CPU 91 kernel/arch/ia32/src/cpu/cpu.c CPU->arch.fi.word = info.cpuid_edx;
CPU 93 kernel/arch/ia32/src/cpu/cpu.c if (CPU->arch.fi.bits.fxsr)
CPU 98 kernel/arch/ia32/src/cpu/cpu.c if (CPU->arch.fi.bits.sse) {
CPU 109 kernel/arch/ia32/src/cpu/cpu.c if (CPU->arch.fi.bits.sep) {
CPU 120 kernel/arch/ia32/src/cpu/cpu.c CPU->arch.vendor = VendorUnknown;
CPU 130 kernel/arch/ia32/src/cpu/cpu.c CPU->arch.vendor = VendorAMD;
CPU 138 kernel/arch/ia32/src/cpu/cpu.c CPU->arch.vendor = VendorIntel;
CPU 141 kernel/arch/ia32/src/cpu/cpu.c CPU->arch.family = (info.cpuid_eax >> 8) & 0x0fU;
CPU 142 kernel/arch/ia32/src/cpu/cpu.c CPU->arch.model = (info.cpuid_eax >> 4) & 0x0fU;
CPU 143 kernel/arch/ia32/src/cpu/cpu.c CPU->arch.stepping = (info.cpuid_eax >> 0) & 0x0fU;
CPU 70 kernel/arch/ia32/src/ddi/ddi.c CPU->arch.tss->iomap);
CPU 112 kernel/arch/ia32/src/ddi/ddi.c CPU->arch.iomapver_copy = ver;
CPU 149 kernel/arch/ia32/src/drivers/i8254.c CPU->delay_loop_const =
CPU 157 kernel/arch/ia32/src/drivers/i8254.c CPU->frequency_mhz = (clk2 - clk1) >> SHIFT;
CPU 120 kernel/arch/ia32/src/interrupt.c if (CPU->arch.iomapver_copy != ver) {
CPU 200 kernel/arch/ia32/src/interrupt.c log(LF_ARCH, LVL_DEBUG, "cpu%u: unhandled IRQ %u", CPU->id,
CPU 219 kernel/arch/ia32/src/interrupt.c log(LF_ARCH, LVL_DEBUG, "cpu%u: PIC spurious interrupt %u", CPU->id,
CPU 63 kernel/arch/ia32/src/proc/scheduler.c if (CPU->arch.fi.bits.sep) {
CPU 70 kernel/arch/ia32/src/proc/scheduler.c CPU->arch.tss->esp0 = kstk;
CPU 71 kernel/arch/ia32/src/proc/scheduler.c CPU->arch.tss->ss0 = GDT_SELECTOR(KDATA_DES);
CPU 469 kernel/arch/ia32/src/smp/apic.c if (CPU->arch.family >= 6)
CPU 516 kernel/arch/ia32/src/smp/apic.c assert(CPU->id < 8);
CPU 520 kernel/arch/ia32/src/smp/apic.c ldr.id = (uint8_t) (1 << CPU->id);
CPU 543 kernel/arch/ia32/src/smp/apic.c CPU->id, l_apic_id());
CPU 47 kernel/arch/ia64/src/cpu/cpu.c CPU->arch.cpuid0 = cpuid_read(0);
CPU 48 kernel/arch/ia64/src/cpu/cpu.c CPU->arch.cpuid1 = cpuid_read(1);
CPU 49 kernel/arch/ia64/src/cpu/cpu.c CPU->arch.cpuid3.value = cpuid_read(3);
CPU 57 kernel/arch/ia64/src/cpu/cpu.c memcpy(vendor, &CPU->arch.cpuid0, 8);
CPU 58 kernel/arch/ia64/src/cpu/cpu.c memcpy(vendor + 8, &CPU->arch.cpuid1, 8);
CPU 73 kernel/arch/ia64/src/cpu/cpu.c printf("cpu%d: %s (%s), archrev=%d, model=%d, revision=%d\n", CPU->id,
CPU 74 kernel/arch/ia64/src/cpu/cpu.c family_str, vendor, CPU->arch.cpuid3.archrev,
CPU 75 kernel/arch/ia64/src/cpu/cpu.c CPU->arch.cpuid3.model, CPU->arch.cpuid3.revision);
CPU 254 kernel/arch/ia64/src/interrupt.c printf("cpu%d: spurious interrupt\n", CPU->id);
CPU 100 kernel/arch/mips32/src/cpu/cpu.c CPU->arch.rev_num = cp0_prid_read() & 0xff;
CPU 101 kernel/arch/mips32/src/cpu/cpu.c CPU->arch.imp_num = (cp0_prid_read() >> 8) & 0xff;
CPU 195 kernel/arch/mips32/src/exception.c CPU->id, i);
CPU 81 kernel/arch/mips32/src/mach/malta/malta.c CPU->id, isa_irq);
CPU 92 kernel/arch/mips32/src/mach/malta/malta.c CPU->id, isa_irq);
CPU 72 kernel/arch/mips32/src/mach/msim/msim.c CPU->id, i);
CPU 50 kernel/arch/ppc32/src/cpu/cpu.c cpu_version(&CPU->arch);
CPU 141 kernel/arch/ppc32/src/interrupt.c " (inum=%" PRIu8 ")", CPU->id, inum);
CPU 64 kernel/arch/sparc64/src/cpu/sun4u/cpu.c if (mid == CPU->arch.mid) {
CPU 83 kernel/arch/sparc64/src/cpu/sun4u/cpu.c CPU->arch.mid = read_mid();
CPU 112 kernel/arch/sparc64/src/cpu/sun4u/cpu.c CPU->arch.clock_frequency = clock_frequency;
CPU 119 kernel/arch/sparc64/src/cpu/sun4u/cpu.c CPU->arch.ver.value = ver_read();
CPU 149 kernel/arch/sparc64/src/cpu/sun4u/cpu.c switch (CPU->arch.ver.impl) {
CPU 53 kernel/arch/sparc64/src/cpu/sun4v/cpu.c CPU->arch.id = myid;
CPU 69 kernel/arch/sparc64/src/cpu/sun4v/cpu.c CPU->arch.clock_frequency = clock_frequency;
CPU 56 kernel/arch/sparc64/src/drivers/tick.c CPU->arch.clock_frequency / HZ;
CPU 57 kernel/arch/sparc64/src/drivers/tick.c CPU->arch.next_tick_cmpr = compare.tick_cmpr;
CPU 116 kernel/arch/sparc64/src/drivers/tick.c drift = tick_counter_read() - CPU->arch.next_tick_cmpr;
CPU 117 kernel/arch/sparc64/src/drivers/tick.c while (drift > CPU->arch.clock_frequency / HZ) {
CPU 118 kernel/arch/sparc64/src/drivers/tick.c drift -= CPU->arch.clock_frequency / HZ;
CPU 121 kernel/arch/sparc64/src/drivers/tick.c CPU->arch.next_tick_cmpr = tick_counter_read() +
CPU 122 kernel/arch/sparc64/src/drivers/tick.c (CPU->arch.clock_frequency / HZ) - drift;
CPU 123 kernel/arch/sparc64/src/drivers/tick.c tick_compare_write(CPU->arch.next_tick_cmpr);
CPU 168 kernel/arch/sparc64/src/smp/sun4u/ipi.c if (&cpus[i] == CPU)
CPU 103 kernel/arch/sparc64/src/smp/sun4u/smp.c if (CPU->arch.mid == mid)
CPU 65 kernel/arch/sparc64/src/smp/sun4v/ipi.c data[CPU->arch.id][0] = (uint64_t) func;
CPU 69 kernel/arch/sparc64/src/smp/sun4v/ipi.c ipi_cpu_list[CPU->arch.id][i] = cpu_list[i];
CPU 73 kernel/arch/sparc64/src/smp/sun4v/ipi.c KA2PA(ipi_cpu_list[CPU->arch.id]), KA2PA(data[CPU->arch.id]));
CPU 86 kernel/arch/sparc64/src/smp/sun4v/ipi.c ipi_cpu_list[CPU->arch.id][0] = cpu_id;
CPU 87 kernel/arch/sparc64/src/smp/sun4v/ipi.c return ipi_brodcast_to(func, ipi_cpu_list[CPU->arch.id], 1);
CPU 113 kernel/arch/sparc64/src/smp/sun4v/ipi.c if (&cpus[i] == CPU) {
CPU 117 kernel/arch/sparc64/src/smp/sun4v/ipi.c ipi_cpu_list[CPU->id][idx] = (uint16_t) cpus[i].id;
CPU 121 kernel/arch/sparc64/src/smp/sun4v/ipi.c ipi_brodcast_to(func, ipi_cpu_list[CPU->arch.id], idx);
CPU 143 kernel/arch/sparc64/src/sun4u/sparc64.c CPU->delay_loop_const = 1;
CPU 155 kernel/arch/sparc64/src/sun4u/sparc64.c CPU->arch.clock_frequency / 1000000;
CPU 141 kernel/arch/sparc64/src/sun4v/sparc64.c CPU->delay_loop_const = 1;
CPU 153 kernel/arch/sparc64/src/sun4v/sparc64.c CPU->arch.clock_frequency / 1000000;
CPU 101 kernel/arch/sparc64/src/trap/sun4u/interrupt.c PRIx64 ")", CPU->id, intrcv, data0);
CPU 86 kernel/arch/sparc64/src/trap/sun4v/interrupt.c KA2PA(cpu_mondo_queues[CPU->id]),
CPU 89 kernel/arch/sparc64/src/trap/sun4v/interrupt.c CPU->arch.id);
CPU 103 kernel/arch/sparc64/src/trap/sun4v/interrupt.c uint64_t data1 = cpu_mondo_queues[CPU->id][0];
CPU 113 kernel/arch/sparc64/src/trap/sun4v/interrupt.c ", data = %" PRIx64 ".", CPU->arch.id, data1);
CPU 47 kernel/generic/include/cpu.h #define CPU_LOCAL (&CPU->local)
CPU 59 kernel/generic/include/synch/spinlock.h #pname, (value), CPU->id, __func__, __LINE__); \
CPU 108 kernel/generic/src/console/chardev.c if (CPU)
CPU 109 kernel/generic/src/console/chardev.c printf("cpu%u: ", CPU->id);
CPU 101 kernel/generic/src/cpu/cpu.c CPU = &cpus[config.cpu_active - 1];
CPU 103 kernel/generic/src/cpu/cpu.c CPU->active = true;
CPU 104 kernel/generic/src/cpu/cpu.c CPU->tlb_active = true;
CPU 108 kernel/generic/src/cpu/cpu.c CPU->idle_cycles = ATOMIC_TIME_INITIALIZER();
CPU 109 kernel/generic/src/cpu/cpu.c CPU->busy_cycles = ATOMIC_TIME_INITIALIZER();
CPU 53 kernel/generic/src/debug/panic.c if (CPU)
CPU 54 kernel/generic/src/debug/panic.c printf("on cpu%u ", CPU->id);
CPU 120 kernel/generic/src/interrupt/interrupt.c if (CPU && CPU_LOCAL->idle) {
CPU 122 kernel/generic/src/interrupt/interrupt.c atomic_time_increment(&CPU->idle_cycles, now - CPU_LOCAL->last_cycle);
CPU 74 kernel/generic/src/lib/halt.c if (CPU)
CPU 75 kernel/generic/src/lib/halt.c log(LF_OTHER, LVL_NOTE, "cpu%u: halted", CPU->id);
CPU 396 kernel/generic/src/mm/slab.c slab_magazine_t *cmag = cache->mag_cache[CPU->id].current;
CPU 397 kernel/generic/src/mm/slab.c slab_magazine_t *lastmag = cache->mag_cache[CPU->id].last;
CPU 399 kernel/generic/src/mm/slab.c assert(irq_spinlock_locked(&cache->mag_cache[CPU->id].lock));
CPU 406 kernel/generic/src/mm/slab.c cache->mag_cache[CPU->id].current = lastmag;
CPU 407 kernel/generic/src/mm/slab.c cache->mag_cache[CPU->id].last = cmag;
CPU 420 kernel/generic/src/mm/slab.c cache->mag_cache[CPU->id].last = cmag;
CPU 421 kernel/generic/src/mm/slab.c cache->mag_cache[CPU->id].current = newmag;
CPU 433 kernel/generic/src/mm/slab.c if (!CPU)
CPU 436 kernel/generic/src/mm/slab.c irq_spinlock_lock(&cache->mag_cache[CPU->id].lock, true);
CPU 440 kernel/generic/src/mm/slab.c irq_spinlock_unlock(&cache->mag_cache[CPU->id].lock, true);
CPU 445 kernel/generic/src/mm/slab.c irq_spinlock_unlock(&cache->mag_cache[CPU->id].lock, true);
CPU 463 kernel/generic/src/mm/slab.c slab_magazine_t *cmag = cache->mag_cache[CPU->id].current;
CPU 464 kernel/generic/src/mm/slab.c slab_magazine_t *lastmag = cache->mag_cache[CPU->id].last;
CPU 466 kernel/generic/src/mm/slab.c assert(irq_spinlock_locked(&cache->mag_cache[CPU->id].lock));
CPU 473 kernel/generic/src/mm/slab.c cache->mag_cache[CPU->id].last = cmag;
CPU 474 kernel/generic/src/mm/slab.c cache->mag_cache[CPU->id].current = lastmag;
CPU 500 kernel/generic/src/mm/slab.c cache->mag_cache[CPU->id].last = cmag;
CPU 501 kernel/generic/src/mm/slab.c cache->mag_cache[CPU->id].current = newmag;
CPU 513 kernel/generic/src/mm/slab.c if (!CPU)
CPU 516 kernel/generic/src/mm/slab.c irq_spinlock_lock(&cache->mag_cache[CPU->id].lock, true);
CPU 520 kernel/generic/src/mm/slab.c irq_spinlock_unlock(&cache->mag_cache[CPU->id].lock, true);
CPU 526 kernel/generic/src/mm/slab.c irq_spinlock_unlock(&cache->mag_cache[CPU->id].lock, true);
CPU 87 kernel/generic/src/mm/tlb.c CPU->tlb_active = false;
CPU 92 kernel/generic/src/mm/tlb.c if (i == CPU->id)
CPU 140 kernel/generic/src/mm/tlb.c CPU->tlb_active = true;
CPU 154 kernel/generic/src/mm/tlb.c assert(CPU);
CPU 156 kernel/generic/src/mm/tlb.c CPU->tlb_active = false;
CPU 160 kernel/generic/src/mm/tlb.c irq_spinlock_lock(&CPU->tlb_lock, false);
CPU 161 kernel/generic/src/mm/tlb.c assert(CPU->tlb_messages_count <= TLB_MESSAGE_QUEUE_LEN);
CPU 164 kernel/generic/src/mm/tlb.c for (i = 0; i < CPU->tlb_messages_count; i++) {
CPU 165 kernel/generic/src/mm/tlb.c tlb_invalidate_type_t type = CPU->tlb_messages[i].type;
CPU 166 kernel/generic/src/mm/tlb.c asid_t asid = CPU->tlb_messages[i].asid;
CPU 167 kernel/generic/src/mm/tlb.c uintptr_t page = CPU->tlb_messages[i].page;
CPU 168 kernel/generic/src/mm/tlb.c size_t count = CPU->tlb_messages[i].count;
CPU 190 kernel/generic/src/mm/tlb.c CPU->tlb_messages_count = 0;
CPU 191 kernel/generic/src/mm/tlb.c irq_spinlock_unlock(&CPU->tlb_lock, false);
CPU 192 kernel/generic/src/mm/tlb.c CPU->tlb_active = true;
CPU 76 kernel/generic/src/proc/scheduler.c irq_spinlock_lock(&CPU->fpu_lock, false);
CPU 79 kernel/generic/src/proc/scheduler.c thread_t *owner = atomic_load_explicit(&CPU->fpu_owner, memory_order_relaxed);
CPU 82 kernel/generic/src/proc/scheduler.c atomic_store_explicit(&CPU->fpu_owner, NULL, memory_order_relaxed);
CPU 85 kernel/generic/src/proc/scheduler.c irq_spinlock_unlock(&CPU->fpu_lock, false);
CPU 94 kernel/generic/src/proc/scheduler.c atomic_store_explicit(&CPU->fpu_owner, THREAD, memory_order_relaxed);
CPU 119 kernel/generic/src/proc/scheduler.c assert(CPU != NULL);
CPU 121 kernel/generic/src/proc/scheduler.c if (atomic_load(&CPU->nrdy) == 0)
CPU 125 kernel/generic/src/proc/scheduler.c irq_spinlock_lock(&(CPU->rq[i].lock), false);
CPU 126 kernel/generic/src/proc/scheduler.c if (CPU->rq[i].n == 0) {
CPU 130 kernel/generic/src/proc/scheduler.c irq_spinlock_unlock(&(CPU->rq[i].lock), false);
CPU 134 kernel/generic/src/proc/scheduler.c atomic_dec(&CPU->nrdy);
CPU 136 kernel/generic/src/proc/scheduler.c CPU->rq[i].n--;
CPU 142 kernel/generic/src/proc/scheduler.c list_first(&CPU->rq[i].rq), thread_t, rq_link);
CPU 145 kernel/generic/src/proc/scheduler.c irq_spinlock_unlock(&(CPU->rq[i].lock), false);
CPU 166 kernel/generic/src/proc/scheduler.c assert(CPU != NULL);
CPU 242 kernel/generic/src/proc/scheduler.c irq_spinlock_lock(&CPU->rq[i].lock, false);
CPU 245 kernel/generic/src/proc/scheduler.c list_swap(&CPU->rq[i].rq, &list);
CPU 248 kernel/generic/src/proc/scheduler.c size_t tmpn = CPU->rq[i].n;
CPU 249 kernel/generic/src/proc/scheduler.c CPU->rq[i].n = n;
CPU 252 kernel/generic/src/proc/scheduler.c irq_spinlock_unlock(&CPU->rq[i].lock, false);
CPU 257 kernel/generic/src/proc/scheduler.c irq_spinlock_lock(&CPU->rq[start].lock, false);
CPU 258 kernel/generic/src/proc/scheduler.c list_concat(&CPU->rq[start].rq, &list);
CPU 259 kernel/generic/src/proc/scheduler.c CPU->rq[start].n += n;
CPU 260 kernel/generic/src/proc/scheduler.c irq_spinlock_unlock(&CPU->rq[start].lock, false);
CPU 285 kernel/generic/src/proc/scheduler.c thread_t *owner = atomic_load_explicit(&CPU->fpu_owner,
CPU 312 kernel/generic/src/proc/scheduler.c assert(atomic_get_unordered(&THREAD->cpu) == CPU);
CPU 326 kernel/generic/src/proc/scheduler.c ", nrdy=%zu)", CPU->id, THREAD->tid, rq_index,
CPU 327 kernel/generic/src/proc/scheduler.c THREAD->ticks, atomic_load(&CPU->nrdy));
CPU 387 kernel/generic/src/proc/scheduler.c assert(atomic_get_unordered(&thread->cpu) == CPU);
CPU 398 kernel/generic/src/proc/scheduler.c add_to_rq(thread, CPU, prio);
CPU 414 kernel/generic/src/proc/scheduler.c cpu = CPU;
CPU 415 kernel/generic/src/proc/scheduler.c atomic_set_unordered(&thread->cpu, CPU);
CPU 475 kernel/generic/src/proc/scheduler.c assert(CPU != NULL);
CPU 547 kernel/generic/src/proc/scheduler.c assert(CPU != NULL);
CPU 569 kernel/generic/src/proc/scheduler.c assert(CPU != NULL);
CPU 620 kernel/generic/src/proc/scheduler.c runq_t *new_rq = &CPU->rq[i];
CPU 647 kernel/generic/src/proc/scheduler.c atomic_set_unordered(&thread->cpu, CPU);
CPU 656 kernel/generic/src/proc/scheduler.c "nrdy=%ld, avg=%ld", CPU->id, thread->tid,
CPU 657 kernel/generic/src/proc/scheduler.c CPU->id, atomic_load(&CPU->nrdy),
CPU 673 kernel/generic/src/proc/scheduler.c atomic_inc(&CPU->nrdy);
CPU 710 kernel/generic/src/proc/scheduler.c rdy = atomic_load(&CPU->nrdy);
CPU 734 kernel/generic/src/proc/scheduler.c if (CPU == cpu)
CPU 745 kernel/generic/src/proc/scheduler.c if (atomic_load(&CPU->nrdy)) {
CPU 46 kernel/generic/src/synch/irq_spinlock.c #define CPU_OWNER ((CPU == NULL) ? (cpu_t *) UINTPTR_MAX : CPU)
CPU 86 kernel/generic/src/synch/mutex.c printf("cpu%u: looping on active mutex %p\n", CPU->id, mtx);
CPU 94 kernel/generic/src/synch/mutex.c printf("cpu%u: not deadlocked\n", CPU->id);
CPU 106 kernel/generic/src/synch/spinlock.c "caller=%p (%s)\n", CPU->id, lock, lock->name,
CPU 120 kernel/generic/src/synch/spinlock.c printf("cpu%u: not deadlocked\n", CPU->id);
CPU 109 kernel/generic/src/time/clock.c if (CPU->id == 0) {
CPU 127 kernel/generic/src/time/clock.c atomic_time_increment(&CPU->busy_cycles, now - CPU_LOCAL->last_cycle);
CPU 156 kernel/generic/src/time/clock.c irq_spinlock_lock(&CPU->timeoutlock, false);
CPU 159 kernel/generic/src/time/clock.c while ((cur = list_first(&CPU->timeout_active_list)) != NULL) {
CPU 171 kernel/generic/src/time/clock.c irq_spinlock_unlock(&CPU->timeoutlock, false);
CPU 178 kernel/generic/src/time/clock.c irq_spinlock_lock(&CPU->timeoutlock, false);
CPU 181 kernel/generic/src/time/clock.c irq_spinlock_unlock(&CPU->timeoutlock, false);
CPU 60 kernel/generic/src/time/delay.c asm_delay_loop(usec * CPU->delay_loop_const);
CPU 56 kernel/generic/src/time/timeout.c irq_spinlock_initialize(&CPU->timeoutlock, "cpu.timeoutlock");
CPU 57 kernel/generic/src/time/timeout.c list_initialize(&CPU->timeout_active_list);
CPU 88 kernel/generic/src/time/timeout.c .cpu = CPU,
CPU 97 kernel/generic/src/time/timeout.c link_t *last = list_last(&CPU->timeout_active_list);
CPU 99 kernel/generic/src/time/timeout.c list_append(&timeout->link, &CPU->timeout_active_list);
CPU 101 kernel/generic/src/time/timeout.c for (link_t *cur = list_first(&CPU->timeout_active_list); cur != NULL;
CPU 102 kernel/generic/src/time/timeout.c cur = list_next(cur, &CPU->timeout_active_list)) {
CPU 127 kernel/generic/src/time/timeout.c irq_spinlock_lock(&CPU->timeoutlock, true);
CPU 129 kernel/generic/src/time/timeout.c irq_spinlock_unlock(&CPU->timeoutlock, true);
CPU 135 kernel/generic/src/time/timeout.c irq_spinlock_lock(&CPU->timeoutlock, true);
CPU 137 kernel/generic/src/time/timeout.c irq_spinlock_unlock(&CPU->timeoutlock, true);
CPU 55 kernel/test/mm/falloc2.c "Unable to allocate frames\n", THREAD->tid, CPU->id);
CPU 66 kernel/test/mm/falloc2.c CPU->id, count, bytes);
CPU 79 kernel/test/mm/falloc2.c "%u blocks allocated.\n", THREAD->tid, CPU->id,
CPU 82 kernel/test/mm/falloc2.c "Deallocating ... \n", THREAD->tid, CPU->id);
CPU 89 kernel/test/mm/falloc2.c THREAD->tid, CPU->id, ((char *) PA2KA(frames[i]))[k],
CPU 99 kernel/test/mm/falloc2.c "Finished run.\n", THREAD->tid, CPU->id);
CPU 107 kernel/test/mm/falloc2.c THREAD->tid, CPU->id);
CPU 66 kernel/test/synch/semaphore2.c TPRINTF("cpu%u, tid %" PRIu64 " down+ (%d)\n", CPU->id, THREAD->tid, to);
CPU 69 kernel/test/synch/semaphore2.c TPRINTF("cpu%u, tid %" PRIu64 " down!\n", CPU->id, THREAD->tid);
CPU 73 kernel/test/synch/semaphore2.c TPRINTF("cpu%u, tid %" PRIu64 " down=\n", CPU->id, THREAD->tid);
CPU 77 kernel/test/synch/semaphore2.c TPRINTF("cpu%u, tid %" PRIu64 " up\n", CPU->id, THREAD->tid);
HelenOS homepage, sources at GitHub