Cleanup: Various comment and whitespace fixes.
Change-Id: I37c3e3346813efc595df651421b7e8ff4fbf3339 Reviewed-on: https://review.haiku-os.org/c/haiku/+/2845 Reviewed-by: waddlesplash <waddlesplash@gmail.com>
This commit is contained in:
parent
b3bd66961a
commit
2555f33549
@ -34,7 +34,7 @@
|
||||
// whole kernel address space is the top 512GB of the address space.
|
||||
#define KERNEL_BASE 0xffffff0000000000
|
||||
#define KERNEL_SIZE 0x10000000000
|
||||
#define KERNEL_TOP (KERNEL_BASE + (KERNEL_SIZE - 1))
|
||||
#define KERNEL_TOP (KERNEL_BASE + (KERNEL_SIZE - 1))
|
||||
#define KERNEL_LOAD_BASE 0xffffffff80000000
|
||||
|
||||
// Kernel physical memory map area.
|
||||
|
@ -154,12 +154,11 @@ extern void mutex_destroy(mutex* lock);
|
||||
extern void mutex_transfer_lock(mutex* lock, thread_id thread);
|
||||
extern status_t mutex_switch_lock(mutex* from, mutex* to);
|
||||
// Unlocks "from" and locks "to" such that unlocking and starting to wait
|
||||
// for the lock is atomically. I.e. if "from" guards the object "to" belongs
|
||||
// for the lock is atomic. I.e. if "from" guards the object "to" belongs
|
||||
// to, the operation is safe as long as "from" is held while destroying
|
||||
// "to".
|
||||
extern status_t mutex_switch_from_read_lock(rw_lock* from, mutex* to);
|
||||
// Like mutex_switch_lock(), just for a switching from a read-locked
|
||||
// rw_lock.
|
||||
// Like mutex_switch_lock(), just for switching from a read-locked rw_lock.
|
||||
|
||||
|
||||
// implementation private:
|
||||
|
@ -17,7 +17,7 @@
|
||||
(B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA | B_KERNEL_EXECUTE_AREA \
|
||||
| B_KERNEL_STACK_AREA)
|
||||
|
||||
// TODO: These aren't really a protection flags, but since the "protection"
|
||||
// TODO: These aren't really protection flags, but since the "protection"
|
||||
// field is the only flag field, we currently use it for this.
|
||||
// A cleaner approach would be appreciated - maybe just an official generic
|
||||
// flags region in the protection field.
|
||||
|
@ -27,7 +27,7 @@ struct X86VMTranslationMap : VMTranslationMap {
|
||||
|
||||
status_t Init(bool kernel);
|
||||
|
||||
virtual bool Lock() final;
|
||||
virtual bool Lock() final;
|
||||
virtual void Unlock() final;
|
||||
|
||||
virtual addr_t MappedSize() const final;
|
||||
|
@ -8,7 +8,7 @@
|
||||
*/
|
||||
|
||||
|
||||
/*! This is main - initializes the kernel and launches the Bootscript */
|
||||
/*! This is main - initializes the kernel and launches the launch_daemon */
|
||||
|
||||
|
||||
#include <string.h>
|
||||
|
@ -958,7 +958,7 @@ VMAnonymousCache::Merge(VMCache* _source)
|
||||
{
|
||||
VMAnonymousCache* source = dynamic_cast<VMAnonymousCache*>(_source);
|
||||
if (source == NULL) {
|
||||
panic("VMAnonymousCache::MergeStore(): merge with incompatible cache "
|
||||
panic("VMAnonymousCache::Merge(): merge with incompatible cache "
|
||||
"%p requested", _source);
|
||||
return;
|
||||
}
|
||||
|
@ -347,7 +347,7 @@ VMUserAddressSpace::UnreserveAddressRange(addr_t address, size_t size,
|
||||
// check to see if this address space has entered DELETE state
|
||||
if (fDeleting) {
|
||||
// okay, someone is trying to delete this address space now, so we can't
|
||||
// insert the area, so back out
|
||||
// remove the area, so back out
|
||||
return B_BAD_TEAM_ID;
|
||||
}
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user