core: replace getpagesize() with qemu_real_host_page_size
There are three page size in qemu: real host page size host page size target page size All of them have dedicate variable to represent. For the last two, we use the same form in the whole qemu project, while for the first one we use two forms: qemu_real_host_page_size and getpagesize(). qemu_real_host_page_size is defined to be a replacement of getpagesize(), so let it serve the role. [Note] Not fully tested for some arch or device. Signed-off-by: Wei Yang <richardw.yang@linux.intel.com> Message-Id: <20191013021145.16011-3-richardw.yang@linux.intel.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
This commit is contained in:
parent
5608956575
commit
038adc2f58
@ -52,7 +52,7 @@
|
||||
/* KVM uses PAGE_SIZE in its definition of KVM_COALESCED_MMIO_MAX. We
|
||||
* need to use the real host PAGE_SIZE, as that's what KVM will use.
|
||||
*/
|
||||
#define PAGE_SIZE getpagesize()
|
||||
#define PAGE_SIZE qemu_real_host_page_size
|
||||
|
||||
//#define DEBUG_KVM
|
||||
|
||||
@ -507,7 +507,7 @@ static int kvm_get_dirty_pages_log_range(MemoryRegionSection *section,
|
||||
{
|
||||
ram_addr_t start = section->offset_within_region +
|
||||
memory_region_get_ram_addr(section->mr);
|
||||
ram_addr_t pages = int128_get64(section->size) / getpagesize();
|
||||
ram_addr_t pages = int128_get64(section->size) / qemu_real_host_page_size;
|
||||
|
||||
cpu_physical_memory_set_dirty_lebitmap(bitmap, start, pages);
|
||||
return 0;
|
||||
@ -1841,7 +1841,7 @@ static int kvm_init(MachineState *ms)
|
||||
* even with KVM. TARGET_PAGE_SIZE is assumed to be the minimum
|
||||
* page size for the system though.
|
||||
*/
|
||||
assert(TARGET_PAGE_SIZE <= getpagesize());
|
||||
assert(TARGET_PAGE_SIZE <= qemu_real_host_page_size);
|
||||
|
||||
s->sigmask_len = 8;
|
||||
|
||||
|
@ -304,7 +304,7 @@ size_t host_memory_backend_pagesize(HostMemoryBackend *memdev)
|
||||
#else
|
||||
size_t host_memory_backend_pagesize(HostMemoryBackend *memdev)
|
||||
{
|
||||
return getpagesize();
|
||||
return qemu_real_host_page_size;
|
||||
}
|
||||
#endif
|
||||
|
||||
|
4
block.c
4
block.c
@ -106,7 +106,7 @@ size_t bdrv_opt_mem_align(BlockDriverState *bs)
|
||||
{
|
||||
if (!bs || !bs->drv) {
|
||||
/* page size or 4k (hdd sector size) should be on the safe side */
|
||||
return MAX(4096, getpagesize());
|
||||
return MAX(4096, qemu_real_host_page_size);
|
||||
}
|
||||
|
||||
return bs->bl.opt_mem_alignment;
|
||||
@ -116,7 +116,7 @@ size_t bdrv_min_mem_align(BlockDriverState *bs)
|
||||
{
|
||||
if (!bs || !bs->drv) {
|
||||
/* page size or 4k (hdd sector size) should be on the safe side */
|
||||
return MAX(4096, getpagesize());
|
||||
return MAX(4096, qemu_real_host_page_size);
|
||||
}
|
||||
|
||||
return bs->bl.min_mem_alignment;
|
||||
|
@ -327,7 +327,7 @@ static void raw_probe_alignment(BlockDriverState *bs, int fd, Error **errp)
|
||||
{
|
||||
BDRVRawState *s = bs->opaque;
|
||||
char *buf;
|
||||
size_t max_align = MAX(MAX_BLOCKSIZE, getpagesize());
|
||||
size_t max_align = MAX(MAX_BLOCKSIZE, qemu_real_host_page_size);
|
||||
size_t alignments[] = {1, 512, 1024, 2048, 4096};
|
||||
|
||||
/* For SCSI generic devices the alignment is not really used.
|
||||
@ -1136,13 +1136,14 @@ static void raw_refresh_limits(BlockDriverState *bs, Error **errp)
|
||||
|
||||
ret = sg_get_max_segments(s->fd);
|
||||
if (ret > 0) {
|
||||
bs->bl.max_transfer = MIN(bs->bl.max_transfer, ret * getpagesize());
|
||||
bs->bl.max_transfer = MIN(bs->bl.max_transfer,
|
||||
ret * qemu_real_host_page_size);
|
||||
}
|
||||
}
|
||||
|
||||
raw_probe_alignment(bs, s->fd, errp);
|
||||
bs->bl.min_mem_alignment = s->buf_align;
|
||||
bs->bl.opt_mem_alignment = MAX(s->buf_align, getpagesize());
|
||||
bs->bl.opt_mem_alignment = MAX(s->buf_align, qemu_real_host_page_size);
|
||||
}
|
||||
|
||||
static int check_for_dasd(int fd)
|
||||
@ -1705,7 +1706,7 @@ static int allocate_first_block(int fd, size_t max_size)
|
||||
size_t write_size = (max_size < MAX_BLOCKSIZE)
|
||||
? BDRV_SECTOR_SIZE
|
||||
: MAX_BLOCKSIZE;
|
||||
size_t max_align = MAX(MAX_BLOCKSIZE, getpagesize());
|
||||
size_t max_align = MAX(MAX_BLOCKSIZE, qemu_real_host_page_size);
|
||||
void *buf;
|
||||
ssize_t n;
|
||||
int ret;
|
||||
|
@ -160,7 +160,7 @@ void bdrv_refresh_limits(BlockDriverState *bs, Error **errp)
|
||||
bdrv_merge_limits(&bs->bl, &bs->file->bs->bl);
|
||||
} else {
|
||||
bs->bl.min_mem_alignment = 512;
|
||||
bs->bl.opt_mem_alignment = getpagesize();
|
||||
bs->bl.opt_mem_alignment = qemu_real_host_page_size;
|
||||
|
||||
/* Safe default since most protocols use readv()/writev()/etc */
|
||||
bs->bl.max_iov = IOV_MAX;
|
||||
|
@ -847,7 +847,7 @@ static int parallels_open(BlockDriverState *bs, QDict *options, int flags,
|
||||
}
|
||||
}
|
||||
|
||||
s->bat_dirty_block = 4 * getpagesize();
|
||||
s->bat_dirty_block = 4 * qemu_real_host_page_size;
|
||||
s->bat_dirty_bmap =
|
||||
bitmap_new(DIV_ROUND_UP(s->header_size, s->bat_dirty_block));
|
||||
|
||||
|
@ -74,7 +74,7 @@ static void qcow2_cache_table_release(Qcow2Cache *c, int i, int num_tables)
|
||||
/* Using MADV_DONTNEED to discard memory is a Linux-specific feature */
|
||||
#ifdef CONFIG_LINUX
|
||||
void *t = qcow2_cache_get_table_addr(c, i);
|
||||
int align = getpagesize();
|
||||
int align = qemu_real_host_page_size;
|
||||
size_t mem_size = (size_t) c->table_size * num_tables;
|
||||
size_t offset = QEMU_ALIGN_UP((uintptr_t) t, align) - (uintptr_t) t;
|
||||
size_t length = QEMU_ALIGN_DOWN(mem_size - offset, align);
|
||||
|
@ -52,7 +52,7 @@ struct udmabuf_create {
|
||||
static size_t
|
||||
udmabuf_get_size(struct vugbm_buffer *buf)
|
||||
{
|
||||
return ROUND_UP(buf->width * buf->height * 4, getpagesize());
|
||||
return ROUND_UP(buf->width * buf->height * 4, qemu_real_host_page_size);
|
||||
}
|
||||
|
||||
static bool
|
||||
|
6
exec.c
6
exec.c
@ -1756,11 +1756,11 @@ long qemu_maxrampagesize(void)
|
||||
#else
|
||||
long qemu_minrampagesize(void)
|
||||
{
|
||||
return getpagesize();
|
||||
return qemu_real_host_page_size;
|
||||
}
|
||||
long qemu_maxrampagesize(void)
|
||||
{
|
||||
return getpagesize();
|
||||
return qemu_real_host_page_size;
|
||||
}
|
||||
#endif
|
||||
|
||||
@ -2417,7 +2417,7 @@ RAMBlock *qemu_ram_alloc_internal(ram_addr_t size, ram_addr_t max_size,
|
||||
new_block->max_length = max_size;
|
||||
assert(max_size >= size);
|
||||
new_block->fd = -1;
|
||||
new_block->page_size = getpagesize();
|
||||
new_block->page_size = qemu_real_host_page_size;
|
||||
new_block->host = host;
|
||||
if (host) {
|
||||
new_block->flags |= RAM_PREALLOC;
|
||||
|
@ -25,7 +25,7 @@
|
||||
#include "migration/qemu-file-types.h"
|
||||
#include "trace.h"
|
||||
|
||||
#define FLIC_SAVE_INITIAL_SIZE getpagesize()
|
||||
#define FLIC_SAVE_INITIAL_SIZE qemu_real_host_page_size
|
||||
#define FLIC_FAILED (-1UL)
|
||||
#define FLIC_SAVEVM_VERSION 1
|
||||
|
||||
|
@ -439,7 +439,7 @@ static void ppc_core99_init(MachineState *machine)
|
||||
}
|
||||
|
||||
/* The NewWorld NVRAM is not located in the MacIO device */
|
||||
if (kvm_enabled() && getpagesize() > 4096) {
|
||||
if (kvm_enabled() && qemu_real_host_page_size > 4096) {
|
||||
/* We can't combine read-write and read-only in a single page, so
|
||||
move the NVRAM out of ROM again for KVM */
|
||||
nvram_addr = 0xFFE00000;
|
||||
|
@ -1942,7 +1942,7 @@ static void spapr_phb_realize(DeviceState *dev, Error **errp)
|
||||
* our memory slot is of page size granularity.
|
||||
*/
|
||||
if (kvm_enabled()) {
|
||||
msi_window_size = getpagesize();
|
||||
msi_window_size = qemu_real_host_page_size;
|
||||
}
|
||||
|
||||
memory_region_init_io(&sphb->msiwindow, OBJECT(sphb), &spapr_msi_ops, spapr,
|
||||
|
@ -601,7 +601,7 @@ static void pvrdma_realize(PCIDevice *pdev, Error **errp)
|
||||
rdma_info_report("Initializing device %s %x.%x", pdev->name,
|
||||
PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn));
|
||||
|
||||
if (TARGET_PAGE_SIZE != getpagesize()) {
|
||||
if (TARGET_PAGE_SIZE != qemu_real_host_page_size) {
|
||||
error_setg(errp, "Target page size must be the same as host page size");
|
||||
return;
|
||||
}
|
||||
|
@ -196,14 +196,15 @@ int vfio_spapr_create_window(VFIOContainer *container,
|
||||
* bits_per_level is a safe guess of how much we can allocate per level:
|
||||
* 8 is the current minimum for CONFIG_FORCE_MAX_ZONEORDER and MAX_ORDER
|
||||
* is usually bigger than that.
|
||||
* Below we look at getpagesize() as TCEs are allocated from system pages.
|
||||
* Below we look at qemu_real_host_page_size as TCEs are allocated from
|
||||
* system pages.
|
||||
*/
|
||||
bits_per_level = ctz64(getpagesize()) + 8;
|
||||
bits_per_level = ctz64(qemu_real_host_page_size) + 8;
|
||||
create.levels = bits_total / bits_per_level;
|
||||
if (bits_total % bits_per_level) {
|
||||
++create.levels;
|
||||
}
|
||||
max_levels = (64 - create.page_shift) / ctz64(getpagesize());
|
||||
max_levels = (64 - create.page_shift) / ctz64(qemu_real_host_page_size);
|
||||
for ( ; create.levels <= max_levels; ++create.levels) {
|
||||
ret = ioctl(container->fd, VFIO_IOMMU_SPAPR_TCE_CREATE, &create);
|
||||
if (!ret) {
|
||||
|
@ -373,7 +373,7 @@ static inline void cpu_physical_memory_set_dirty_lebitmap(unsigned long *bitmap,
|
||||
hwaddr addr;
|
||||
ram_addr_t ram_addr;
|
||||
unsigned long len = (pages + HOST_LONG_BITS - 1) / HOST_LONG_BITS;
|
||||
unsigned long hpratio = getpagesize() / TARGET_PAGE_SIZE;
|
||||
unsigned long hpratio = qemu_real_host_page_size / TARGET_PAGE_SIZE;
|
||||
unsigned long page = BIT_WORD(start >> TARGET_PAGE_BITS);
|
||||
|
||||
/* start address is aligned at the start of a word? */
|
||||
|
@ -423,9 +423,9 @@ void qemu_anon_ram_free(void *ptr, size_t size);
|
||||
# define QEMU_VMALLOC_ALIGN (256 * 4096)
|
||||
#elif defined(__linux__) && defined(__sparc__)
|
||||
#include <sys/shm.h>
|
||||
# define QEMU_VMALLOC_ALIGN MAX(getpagesize(), SHMLBA)
|
||||
# define QEMU_VMALLOC_ALIGN MAX(qemu_real_host_page_size, SHMLBA)
|
||||
#else
|
||||
# define QEMU_VMALLOC_ALIGN getpagesize()
|
||||
# define QEMU_VMALLOC_ALIGN qemu_real_host_page_size
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_POSIX
|
||||
|
@ -2284,7 +2284,7 @@ static struct rp_cmd_args {
|
||||
static void migrate_handle_rp_req_pages(MigrationState *ms, const char* rbname,
|
||||
ram_addr_t start, size_t len)
|
||||
{
|
||||
long our_host_ps = getpagesize();
|
||||
long our_host_ps = qemu_real_host_page_size;
|
||||
|
||||
trace_migrate_handle_rp_req_pages(rbname, start, len);
|
||||
|
||||
|
@ -308,7 +308,7 @@ static bool ufd_check_and_apply(int ufd, MigrationIncomingState *mis)
|
||||
return false;
|
||||
}
|
||||
|
||||
if (getpagesize() != ram_pagesize_summary()) {
|
||||
if (qemu_real_host_page_size != ram_pagesize_summary()) {
|
||||
bool have_hp = false;
|
||||
/* We've got a huge page */
|
||||
#ifdef UFFD_FEATURE_MISSING_HUGETLBFS
|
||||
@ -346,7 +346,7 @@ static int test_ramblock_postcopiable(RAMBlock *rb, void *opaque)
|
||||
*/
|
||||
bool postcopy_ram_supported_by_host(MigrationIncomingState *mis)
|
||||
{
|
||||
long pagesize = getpagesize();
|
||||
long pagesize = qemu_real_host_page_size;
|
||||
int ufd = -1;
|
||||
bool ret = false; /* Error unless we change it */
|
||||
void *testarea = NULL;
|
||||
|
@ -862,7 +862,7 @@ static uint64_t vtop(void *ptr, Error **errp)
|
||||
uint64_t pinfo;
|
||||
uint64_t ret = -1;
|
||||
uintptr_t addr = (uintptr_t) ptr;
|
||||
uintptr_t pagesize = getpagesize();
|
||||
uintptr_t pagesize = qemu_real_host_page_size;
|
||||
off_t offset = addr / pagesize * sizeof(pinfo);
|
||||
int fd;
|
||||
|
||||
|
@ -411,7 +411,7 @@ void kvm_check_mmu(PowerPCCPU *cpu, Error **errp)
|
||||
* will be a normal mapping, not a special hugepage one used
|
||||
* for RAM.
|
||||
*/
|
||||
if (getpagesize() < 0x10000) {
|
||||
if (qemu_real_host_page_size < 0x10000) {
|
||||
error_setg(errp,
|
||||
"KVM can't supply 64kiB CI pages, which guest expects");
|
||||
}
|
||||
|
@ -468,8 +468,8 @@ vubr_queue_set_started(VuDev *dev, int qidx, bool started)
|
||||
|
||||
if (started && vubr->notifier.fd >= 0) {
|
||||
vu_set_queue_host_notifier(dev, vq, vubr->notifier.fd,
|
||||
getpagesize(),
|
||||
qidx * getpagesize());
|
||||
qemu_real_host_page_size,
|
||||
qidx * qemu_real_host_page_size);
|
||||
}
|
||||
|
||||
if (qidx % 2 == 1) {
|
||||
@ -594,7 +594,7 @@ static void *notifier_thread(void *arg)
|
||||
{
|
||||
VuDev *dev = (VuDev *)arg;
|
||||
VubrDev *vubr = container_of(dev, VubrDev, vudev);
|
||||
int pagesize = getpagesize();
|
||||
int pagesize = qemu_real_host_page_size;
|
||||
int qidx;
|
||||
|
||||
while (true) {
|
||||
@ -630,7 +630,7 @@ vubr_host_notifier_setup(VubrDev *dev)
|
||||
void *addr;
|
||||
int fd;
|
||||
|
||||
length = getpagesize() * VHOST_USER_BRIDGE_MAX_QUEUES;
|
||||
length = qemu_real_host_page_size * VHOST_USER_BRIDGE_MAX_QUEUES;
|
||||
|
||||
fd = mkstemp(template);
|
||||
if (fd < 0) {
|
||||
|
@ -48,7 +48,7 @@ size_t qemu_fd_getpagesize(int fd)
|
||||
#endif
|
||||
#endif
|
||||
|
||||
return getpagesize();
|
||||
return qemu_real_host_page_size;
|
||||
}
|
||||
|
||||
size_t qemu_mempath_getpagesize(const char *mem_path)
|
||||
@ -79,7 +79,7 @@ size_t qemu_mempath_getpagesize(const char *mem_path)
|
||||
#endif
|
||||
#endif
|
||||
|
||||
return getpagesize();
|
||||
return qemu_real_host_page_size;
|
||||
}
|
||||
|
||||
void *qemu_ram_mmap(int fd,
|
||||
@ -114,7 +114,7 @@ void *qemu_ram_mmap(int fd,
|
||||
*/
|
||||
flags = MAP_PRIVATE;
|
||||
pagesize = qemu_fd_getpagesize(fd);
|
||||
if (fd == -1 || pagesize == getpagesize()) {
|
||||
if (fd == -1 || pagesize == qemu_real_host_page_size) {
|
||||
guardfd = -1;
|
||||
flags |= MAP_ANONYMOUS;
|
||||
} else {
|
||||
@ -123,7 +123,7 @@ void *qemu_ram_mmap(int fd,
|
||||
}
|
||||
#else
|
||||
guardfd = -1;
|
||||
pagesize = getpagesize();
|
||||
pagesize = qemu_real_host_page_size;
|
||||
flags = MAP_PRIVATE | MAP_ANONYMOUS;
|
||||
#endif
|
||||
|
||||
@ -205,7 +205,7 @@ void qemu_ram_munmap(int fd, void *ptr, size_t size)
|
||||
#if defined(__powerpc64__) && defined(__linux__)
|
||||
pagesize = qemu_fd_getpagesize(fd);
|
||||
#else
|
||||
pagesize = getpagesize();
|
||||
pagesize = qemu_real_host_page_size;
|
||||
#endif
|
||||
munmap(ptr, size + pagesize);
|
||||
}
|
||||
|
@ -617,7 +617,7 @@ void *qemu_alloc_stack(size_t *sz)
|
||||
#ifdef CONFIG_DEBUG_STACK_USAGE
|
||||
void *ptr2;
|
||||
#endif
|
||||
size_t pagesz = getpagesize();
|
||||
size_t pagesz = qemu_real_host_page_size;
|
||||
#ifdef _SC_THREAD_STACK_MIN
|
||||
/* avoid stacks smaller than _SC_THREAD_STACK_MIN */
|
||||
long min_stack_sz = sysconf(_SC_THREAD_STACK_MIN);
|
||||
@ -679,7 +679,7 @@ void qemu_free_stack(void *stack, size_t sz)
|
||||
unsigned int usage;
|
||||
void *ptr;
|
||||
|
||||
for (ptr = stack + getpagesize(); ptr < stack + sz;
|
||||
for (ptr = stack + qemu_real_host_page_size; ptr < stack + sz;
|
||||
ptr += sizeof(uint32_t)) {
|
||||
if (*(uint32_t *)ptr != 0xdeadbeaf) {
|
||||
break;
|
||||
|
@ -554,7 +554,7 @@ void os_mem_prealloc(int fd, char *area, size_t memory, int smp_cpus,
|
||||
Error **errp)
|
||||
{
|
||||
int i;
|
||||
size_t pagesize = getpagesize();
|
||||
size_t pagesize = qemu_real_host_page_size;
|
||||
|
||||
memory = (memory + pagesize - 1) & -pagesize;
|
||||
for (i = 0; i < memory / pagesize; i++) {
|
||||
|
@ -514,9 +514,9 @@ static IOVAMapping *qemu_vfio_add_mapping(QEMUVFIOState *s,
|
||||
IOVAMapping m = {.host = host, .size = size, .iova = iova};
|
||||
IOVAMapping *insert;
|
||||
|
||||
assert(QEMU_IS_ALIGNED(size, getpagesize()));
|
||||
assert(QEMU_IS_ALIGNED(s->low_water_mark, getpagesize()));
|
||||
assert(QEMU_IS_ALIGNED(s->high_water_mark, getpagesize()));
|
||||
assert(QEMU_IS_ALIGNED(size, qemu_real_host_page_size));
|
||||
assert(QEMU_IS_ALIGNED(s->low_water_mark, qemu_real_host_page_size));
|
||||
assert(QEMU_IS_ALIGNED(s->high_water_mark, qemu_real_host_page_size));
|
||||
trace_qemu_vfio_new_mapping(s, host, size, index, iova);
|
||||
|
||||
assert(index >= 0);
|
||||
@ -567,7 +567,7 @@ static void qemu_vfio_undo_mapping(QEMUVFIOState *s, IOVAMapping *mapping,
|
||||
|
||||
index = mapping - s->mappings;
|
||||
assert(mapping->size > 0);
|
||||
assert(QEMU_IS_ALIGNED(mapping->size, getpagesize()));
|
||||
assert(QEMU_IS_ALIGNED(mapping->size, qemu_real_host_page_size));
|
||||
assert(index >= 0 && index < s->nr_mappings);
|
||||
if (ioctl(s->container, VFIO_IOMMU_UNMAP_DMA, &unmap)) {
|
||||
error_setg(errp, "VFIO_UNMAP_DMA failed: %d", -errno);
|
||||
@ -613,8 +613,8 @@ int qemu_vfio_dma_map(QEMUVFIOState *s, void *host, size_t size,
|
||||
IOVAMapping *mapping;
|
||||
uint64_t iova0;
|
||||
|
||||
assert(QEMU_PTR_IS_ALIGNED(host, getpagesize()));
|
||||
assert(QEMU_IS_ALIGNED(size, getpagesize()));
|
||||
assert(QEMU_PTR_IS_ALIGNED(host, qemu_real_host_page_size));
|
||||
assert(QEMU_IS_ALIGNED(size, qemu_real_host_page_size));
|
||||
trace_qemu_vfio_dma_map(s, host, size, temporary, iova);
|
||||
qemu_mutex_lock(&s->lock);
|
||||
mapping = qemu_vfio_find_mapping(s, host, &index);
|
||||
|
Loading…
Reference in New Issue
Block a user