b258298c70
atomic_{get, set}64() are problematic on architectures without 64 bit compare and swap. Also, using sequential lock instead of atomic access ensures that any reads from cpu_ent::active_time won't require any writes to shared memory.
138 lines
2.8 KiB
C++
138 lines
2.8 KiB
C++
/*
|
|
* Copyright 2002-2006, Haiku Inc. All rights reserved.
|
|
* Distributed under the terms of the MIT License.
|
|
*
|
|
* Copyright 2002, Travis Geiselbrecht. All rights reserved.
|
|
* Distributed under the terms of the NewOS License.
|
|
*/
|
|
#ifndef _KERNEL_CPU_H
|
|
#define _KERNEL_CPU_H
|
|
|
|
|
|
#include <setjmp.h>
|
|
|
|
#include <int.h>
|
|
#include <smp.h>
|
|
#include <timer.h>
|
|
#include <arch/cpu.h>
|
|
|
|
#include <scheduler.h>
|
|
|
|
|
|
struct kernel_args;
|
|
|
|
namespace BKernel {
|
|
struct Thread;
|
|
}
|
|
|
|
using BKernel::Thread;
|
|
|
|
|
|
typedef enum cpu_topology_level {
|
|
CPU_TOPOLOGY_SMT,
|
|
CPU_TOPOLOGY_CORE,
|
|
CPU_TOPOLOGY_PACKAGE,
|
|
//
|
|
CPU_TOPOLOGY_LEVELS
|
|
} cpu_topology_level;
|
|
|
|
typedef struct cpu_topology_node {
|
|
cpu_topology_level level;
|
|
|
|
int id;
|
|
|
|
cpu_topology_node** children;
|
|
int children_count;
|
|
} cpu_topology_node;
|
|
|
|
|
|
/* CPU local data structure */
|
|
|
|
typedef struct cpu_ent {
|
|
int cpu_num;
|
|
|
|
// thread.c: used to force a reschedule at quantum expiration time
|
|
bool preempted;
|
|
timer quantum_timer;
|
|
|
|
// keeping track of CPU activity
|
|
seqlock active_time_lock;
|
|
bigtime_t active_time;
|
|
bigtime_t irq_time;
|
|
bigtime_t interrupt_time;
|
|
bigtime_t last_kernel_time;
|
|
bigtime_t last_user_time;
|
|
|
|
int32 ici_counter;
|
|
|
|
// used in the kernel debugger
|
|
addr_t fault_handler;
|
|
addr_t fault_handler_stack_pointer;
|
|
jmp_buf fault_jump_buffer;
|
|
|
|
Thread* running_thread;
|
|
Thread* previous_thread;
|
|
bool invoke_scheduler;
|
|
bool disabled;
|
|
|
|
// CPU topology information
|
|
int topology_id[CPU_TOPOLOGY_LEVELS];
|
|
int cache_id[CPU_MAX_CACHE_LEVEL];
|
|
|
|
// IRQs assigned to this CPU
|
|
struct list irqs;
|
|
spinlock irqs_lock;
|
|
|
|
// arch-specific stuff
|
|
arch_cpu_info arch;
|
|
} cpu_ent CACHE_LINE_ALIGN;
|
|
|
|
|
|
//extern cpu_ent gCPU[MAX_BOOT_CPUS];
|
|
extern cpu_ent gCPU[];
|
|
extern uint32 gCPUCacheLevelCount;
|
|
|
|
|
|
#ifdef __cplusplus
|
|
extern "C" {
|
|
#endif
|
|
|
|
status_t cpu_preboot_init_percpu(struct kernel_args *args, int curr_cpu);
|
|
status_t cpu_init(struct kernel_args *args);
|
|
status_t cpu_init_percpu(struct kernel_args *ka, int curr_cpu);
|
|
status_t cpu_init_post_vm(struct kernel_args *args);
|
|
status_t cpu_init_post_modules(struct kernel_args *args);
|
|
bigtime_t cpu_get_active_time(int32 cpu);
|
|
|
|
cpu_ent *get_cpu_struct(void);
|
|
extern inline cpu_ent *get_cpu_struct(void) { return &gCPU[smp_get_current_cpu()]; }
|
|
|
|
status_t cpu_build_topology_tree(void);
|
|
const cpu_topology_node* get_cpu_topology(void);
|
|
|
|
void cpu_set_scheduler_mode(enum scheduler_mode mode);
|
|
|
|
status_t increase_cpu_performance(int delta);
|
|
status_t decrease_cpu_performance(int delta);
|
|
|
|
void cpu_idle(void);
|
|
void cpu_wait(int32* variable, int32 test);
|
|
|
|
|
|
static inline void
|
|
cpu_pause(void)
|
|
{
|
|
arch_cpu_pause();
|
|
}
|
|
|
|
|
|
void _user_clear_caches(void *address, size_t length, uint32 flags);
|
|
bool _user_cpu_enabled(int32 cpu);
|
|
status_t _user_set_cpu_enabled(int32 cpu, bool enabled);
|
|
|
|
#ifdef __cplusplus
|
|
}
|
|
#endif
|
|
|
|
#endif /* _KERNEL_CPU_H */
|