40c4a5533f
Implemented FIFO and LRU eviction policies. Now one of the three eviction policies can be chosen as an argument. On not specifying an argument, LRU is used by default. Signed-off-by: Mahmoud Mandour <ma.mandourr@gmail.com> Signed-off-by: Alex Bennée <alex.bennee@linaro.org> Message-Id: <20210623125458.450462-4-ma.mandourr@gmail.com> Message-Id: <20210709143005.1554-39-alex.bennee@linaro.org>
642 lines
17 KiB
C
642 lines
17 KiB
C
/*
|
|
* Copyright (C) 2021, Mahmoud Mandour <ma.mandourr@gmail.com>
|
|
*
|
|
* License: GNU GPL, version 2 or later.
|
|
* See the COPYING file in the top-level directory.
|
|
*/
|
|
|
|
#include <inttypes.h>
|
|
#include <stdio.h>
|
|
#include <glib.h>
|
|
|
|
#include <qemu-plugin.h>
|
|
|
|
QEMU_PLUGIN_EXPORT int qemu_plugin_version = QEMU_PLUGIN_VERSION;
|
|
|
|
static enum qemu_plugin_mem_rw rw = QEMU_PLUGIN_MEM_RW;
|
|
|
|
static GHashTable *miss_ht;
|
|
|
|
static GMutex mtx;
|
|
static GRand *rng;
|
|
|
|
static int limit;
|
|
static bool sys;
|
|
|
|
static uint64_t dmem_accesses;
|
|
static uint64_t dmisses;
|
|
|
|
static uint64_t imem_accesses;
|
|
static uint64_t imisses;
|
|
|
|
enum EvictionPolicy {
|
|
LRU,
|
|
FIFO,
|
|
RAND,
|
|
};
|
|
|
|
enum EvictionPolicy policy;
|
|
|
|
/*
|
|
* A CacheSet is a set of cache blocks. A memory block that maps to a set can be
|
|
* put in any of the blocks inside the set. The number of block per set is
|
|
* called the associativity (assoc).
|
|
*
|
|
* Each block contains the the stored tag and a valid bit. Since this is not
|
|
* a functional simulator, the data itself is not stored. We only identify
|
|
* whether a block is in the cache or not by searching for its tag.
|
|
*
|
|
* In order to search for memory data in the cache, the set identifier and tag
|
|
* are extracted from the address and the set is probed to see whether a tag
|
|
* match occur.
|
|
*
|
|
* An address is logically divided into three portions: The block offset,
|
|
* the set number, and the tag.
|
|
*
|
|
* The set number is used to identify the set in which the block may exist.
|
|
* The tag is compared against all the tags of a set to search for a match. If a
|
|
* match is found, then the access is a hit.
|
|
*
|
|
* The CacheSet also contains bookkeaping information about eviction details.
|
|
*/
|
|
|
|
typedef struct {
|
|
uint64_t tag;
|
|
bool valid;
|
|
} CacheBlock;
|
|
|
|
typedef struct {
|
|
CacheBlock *blocks;
|
|
uint64_t *lru_priorities;
|
|
uint64_t lru_gen_counter;
|
|
GQueue *fifo_queue;
|
|
} CacheSet;
|
|
|
|
typedef struct {
|
|
CacheSet *sets;
|
|
int num_sets;
|
|
int cachesize;
|
|
int assoc;
|
|
int blksize_shift;
|
|
uint64_t set_mask;
|
|
uint64_t tag_mask;
|
|
} Cache;
|
|
|
|
typedef struct {
|
|
char *disas_str;
|
|
const char *symbol;
|
|
uint64_t addr;
|
|
uint64_t dmisses;
|
|
uint64_t imisses;
|
|
} InsnData;
|
|
|
|
void (*update_hit)(Cache *cache, int set, int blk);
|
|
void (*update_miss)(Cache *cache, int set, int blk);
|
|
|
|
void (*metadata_init)(Cache *cache);
|
|
void (*metadata_destroy)(Cache *cache);
|
|
|
|
Cache *dcache, *icache;
|
|
|
|
static int pow_of_two(int num)
|
|
{
|
|
g_assert((num & (num - 1)) == 0);
|
|
int ret = 0;
|
|
while (num /= 2) {
|
|
ret++;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* LRU evection policy: For each set, a generation counter is maintained
|
|
* alongside a priority array.
|
|
*
|
|
* On each set access, the generation counter is incremented.
|
|
*
|
|
* On a cache hit: The hit-block is assigned the current generation counter,
|
|
* indicating that it is the most recently used block.
|
|
*
|
|
* On a cache miss: The block with the least priority is searched and replaced
|
|
* with the newly-cached block, of which the priority is set to the current
|
|
* generation number.
|
|
*/
|
|
|
|
static void lru_priorities_init(Cache *cache)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < cache->num_sets; i++) {
|
|
cache->sets[i].lru_priorities = g_new0(uint64_t, cache->assoc);
|
|
cache->sets[i].lru_gen_counter = 0;
|
|
}
|
|
}
|
|
|
|
static void lru_update_blk(Cache *cache, int set_idx, int blk_idx)
|
|
{
|
|
CacheSet *set = &cache->sets[set_idx];
|
|
set->lru_priorities[blk_idx] = cache->sets[set_idx].lru_gen_counter;
|
|
set->lru_gen_counter++;
|
|
}
|
|
|
|
static int lru_get_lru_block(Cache *cache, int set_idx)
|
|
{
|
|
int i, min_idx, min_priority;
|
|
|
|
min_priority = cache->sets[set_idx].lru_priorities[0];
|
|
min_idx = 0;
|
|
|
|
for (i = 1; i < cache->assoc; i++) {
|
|
if (cache->sets[set_idx].lru_priorities[i] < min_priority) {
|
|
min_priority = cache->sets[set_idx].lru_priorities[i];
|
|
min_idx = i;
|
|
}
|
|
}
|
|
return min_idx;
|
|
}
|
|
|
|
static void lru_priorities_destroy(Cache *cache)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < cache->num_sets; i++) {
|
|
g_free(cache->sets[i].lru_priorities);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* FIFO eviction policy: a FIFO queue is maintained for each CacheSet that
|
|
* stores accesses to the cache.
|
|
*
|
|
* On a compulsory miss: The block index is enqueued to the fifo_queue to
|
|
* indicate that it's the latest cached block.
|
|
*
|
|
* On a conflict miss: The first-in block is removed from the cache and the new
|
|
* block is put in its place and enqueued to the FIFO queue.
|
|
*/
|
|
|
|
static void fifo_init(Cache *cache)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < cache->num_sets; i++) {
|
|
cache->sets[i].fifo_queue = g_queue_new();
|
|
}
|
|
}
|
|
|
|
static int fifo_get_first_block(Cache *cache, int set)
|
|
{
|
|
GQueue *q = cache->sets[set].fifo_queue;
|
|
return GPOINTER_TO_INT(g_queue_pop_tail(q));
|
|
}
|
|
|
|
static void fifo_update_on_miss(Cache *cache, int set, int blk_idx)
|
|
{
|
|
GQueue *q = cache->sets[set].fifo_queue;
|
|
g_queue_push_head(q, GINT_TO_POINTER(blk_idx));
|
|
}
|
|
|
|
static void fifo_destroy(Cache *cache)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < cache->assoc; i++) {
|
|
g_queue_free(cache->sets[i].fifo_queue);
|
|
}
|
|
}
|
|
|
|
static inline uint64_t extract_tag(Cache *cache, uint64_t addr)
|
|
{
|
|
return addr & cache->tag_mask;
|
|
}
|
|
|
|
static inline uint64_t extract_set(Cache *cache, uint64_t addr)
|
|
{
|
|
return (addr & cache->set_mask) >> cache->blksize_shift;
|
|
}
|
|
|
|
static const char *cache_config_error(int blksize, int assoc, int cachesize)
|
|
{
|
|
if (cachesize % blksize != 0) {
|
|
return "cache size must be divisible by block size";
|
|
} else if (cachesize % (blksize * assoc) != 0) {
|
|
return "cache size must be divisible by set size (assoc * block size)";
|
|
} else {
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
static bool bad_cache_params(int blksize, int assoc, int cachesize)
|
|
{
|
|
return (cachesize % blksize) != 0 || (cachesize % (blksize * assoc) != 0);
|
|
}
|
|
|
|
static Cache *cache_init(int blksize, int assoc, int cachesize)
|
|
{
|
|
if (bad_cache_params(blksize, assoc, cachesize)) {
|
|
return NULL;
|
|
}
|
|
|
|
Cache *cache;
|
|
int i;
|
|
uint64_t blk_mask;
|
|
|
|
cache = g_new(Cache, 1);
|
|
cache->assoc = assoc;
|
|
cache->cachesize = cachesize;
|
|
cache->num_sets = cachesize / (blksize * assoc);
|
|
cache->sets = g_new(CacheSet, cache->num_sets);
|
|
cache->blksize_shift = pow_of_two(blksize);
|
|
|
|
for (i = 0; i < cache->num_sets; i++) {
|
|
cache->sets[i].blocks = g_new0(CacheBlock, assoc);
|
|
}
|
|
|
|
blk_mask = blksize - 1;
|
|
cache->set_mask = ((cache->num_sets - 1) << cache->blksize_shift);
|
|
cache->tag_mask = ~(cache->set_mask | blk_mask);
|
|
|
|
if (metadata_init) {
|
|
metadata_init(cache);
|
|
}
|
|
|
|
return cache;
|
|
}
|
|
|
|
static int get_invalid_block(Cache *cache, uint64_t set)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < cache->assoc; i++) {
|
|
if (!cache->sets[set].blocks[i].valid) {
|
|
return i;
|
|
}
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
static int get_replaced_block(Cache *cache, int set)
|
|
{
|
|
switch (policy) {
|
|
case RAND:
|
|
return g_rand_int_range(rng, 0, cache->assoc);
|
|
case LRU:
|
|
return lru_get_lru_block(cache, set);
|
|
case FIFO:
|
|
return fifo_get_first_block(cache, set);
|
|
default:
|
|
g_assert_not_reached();
|
|
}
|
|
}
|
|
|
|
static int in_cache(Cache *cache, uint64_t addr)
|
|
{
|
|
int i;
|
|
uint64_t tag, set;
|
|
|
|
tag = extract_tag(cache, addr);
|
|
set = extract_set(cache, addr);
|
|
|
|
for (i = 0; i < cache->assoc; i++) {
|
|
if (cache->sets[set].blocks[i].tag == tag &&
|
|
cache->sets[set].blocks[i].valid) {
|
|
return i;
|
|
}
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
/**
|
|
* access_cache(): Simulate a cache access
|
|
* @cache: The cache under simulation
|
|
* @addr: The address of the requested memory location
|
|
*
|
|
* Returns true if the requsted data is hit in the cache and false when missed.
|
|
* The cache is updated on miss for the next access.
|
|
*/
|
|
static bool access_cache(Cache *cache, uint64_t addr)
|
|
{
|
|
int hit_blk, replaced_blk;
|
|
uint64_t tag, set;
|
|
|
|
tag = extract_tag(cache, addr);
|
|
set = extract_set(cache, addr);
|
|
|
|
hit_blk = in_cache(cache, addr);
|
|
if (hit_blk != -1) {
|
|
if (update_hit) {
|
|
update_hit(cache, set, hit_blk);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
replaced_blk = get_invalid_block(cache, set);
|
|
|
|
if (replaced_blk == -1) {
|
|
replaced_blk = get_replaced_block(cache, set);
|
|
}
|
|
|
|
if (update_miss) {
|
|
update_miss(cache, set, replaced_blk);
|
|
}
|
|
|
|
cache->sets[set].blocks[replaced_blk].tag = tag;
|
|
cache->sets[set].blocks[replaced_blk].valid = true;
|
|
|
|
return false;
|
|
}
|
|
|
|
static void vcpu_mem_access(unsigned int vcpu_index, qemu_plugin_meminfo_t info,
|
|
uint64_t vaddr, void *userdata)
|
|
{
|
|
uint64_t effective_addr;
|
|
struct qemu_plugin_hwaddr *hwaddr;
|
|
InsnData *insn;
|
|
|
|
g_mutex_lock(&mtx);
|
|
hwaddr = qemu_plugin_get_hwaddr(info, vaddr);
|
|
if (hwaddr && qemu_plugin_hwaddr_is_io(hwaddr)) {
|
|
g_mutex_unlock(&mtx);
|
|
return;
|
|
}
|
|
|
|
effective_addr = hwaddr ? qemu_plugin_hwaddr_phys_addr(hwaddr) : vaddr;
|
|
|
|
if (!access_cache(dcache, effective_addr)) {
|
|
insn = (InsnData *) userdata;
|
|
insn->dmisses++;
|
|
dmisses++;
|
|
}
|
|
dmem_accesses++;
|
|
g_mutex_unlock(&mtx);
|
|
}
|
|
|
|
static void vcpu_insn_exec(unsigned int vcpu_index, void *userdata)
|
|
{
|
|
uint64_t insn_addr;
|
|
InsnData *insn;
|
|
|
|
g_mutex_lock(&mtx);
|
|
insn_addr = ((InsnData *) userdata)->addr;
|
|
|
|
if (!access_cache(icache, insn_addr)) {
|
|
insn = (InsnData *) userdata;
|
|
insn->imisses++;
|
|
imisses++;
|
|
}
|
|
imem_accesses++;
|
|
g_mutex_unlock(&mtx);
|
|
}
|
|
|
|
static void vcpu_tb_trans(qemu_plugin_id_t id, struct qemu_plugin_tb *tb)
|
|
{
|
|
size_t n_insns;
|
|
size_t i;
|
|
InsnData *data;
|
|
|
|
n_insns = qemu_plugin_tb_n_insns(tb);
|
|
for (i = 0; i < n_insns; i++) {
|
|
struct qemu_plugin_insn *insn = qemu_plugin_tb_get_insn(tb, i);
|
|
uint64_t effective_addr;
|
|
|
|
if (sys) {
|
|
effective_addr = (uint64_t) qemu_plugin_insn_haddr(insn);
|
|
} else {
|
|
effective_addr = (uint64_t) qemu_plugin_insn_vaddr(insn);
|
|
}
|
|
|
|
/*
|
|
* Instructions might get translated multiple times, we do not create
|
|
* new entries for those instructions. Instead, we fetch the same
|
|
* entry from the hash table and register it for the callback again.
|
|
*/
|
|
g_mutex_lock(&mtx);
|
|
data = g_hash_table_lookup(miss_ht, GUINT_TO_POINTER(effective_addr));
|
|
if (data == NULL) {
|
|
data = g_new0(InsnData, 1);
|
|
data->disas_str = qemu_plugin_insn_disas(insn);
|
|
data->symbol = qemu_plugin_insn_symbol(insn);
|
|
data->addr = effective_addr;
|
|
g_hash_table_insert(miss_ht, GUINT_TO_POINTER(effective_addr),
|
|
(gpointer) data);
|
|
}
|
|
g_mutex_unlock(&mtx);
|
|
|
|
qemu_plugin_register_vcpu_mem_cb(insn, vcpu_mem_access,
|
|
QEMU_PLUGIN_CB_NO_REGS,
|
|
rw, data);
|
|
|
|
qemu_plugin_register_vcpu_insn_exec_cb(insn, vcpu_insn_exec,
|
|
QEMU_PLUGIN_CB_NO_REGS, data);
|
|
}
|
|
}
|
|
|
|
static void insn_free(gpointer data)
|
|
{
|
|
InsnData *insn = (InsnData *) data;
|
|
g_free(insn->disas_str);
|
|
g_free(insn);
|
|
}
|
|
|
|
static void cache_free(Cache *cache)
|
|
{
|
|
for (int i = 0; i < cache->num_sets; i++) {
|
|
g_free(cache->sets[i].blocks);
|
|
}
|
|
|
|
if (metadata_destroy) {
|
|
metadata_destroy(cache);
|
|
}
|
|
|
|
g_free(cache->sets);
|
|
g_free(cache);
|
|
}
|
|
|
|
static int dcmp(gconstpointer a, gconstpointer b)
|
|
{
|
|
InsnData *insn_a = (InsnData *) a;
|
|
InsnData *insn_b = (InsnData *) b;
|
|
|
|
return insn_a->dmisses < insn_b->dmisses ? 1 : -1;
|
|
}
|
|
|
|
static int icmp(gconstpointer a, gconstpointer b)
|
|
{
|
|
InsnData *insn_a = (InsnData *) a;
|
|
InsnData *insn_b = (InsnData *) b;
|
|
|
|
return insn_a->imisses < insn_b->imisses ? 1 : -1;
|
|
}
|
|
|
|
static void log_stats()
|
|
{
|
|
g_autoptr(GString) rep = g_string_new("");
|
|
g_string_append_printf(rep,
|
|
"Data accesses: %lu, Misses: %lu\nMiss rate: %lf%%\n\n",
|
|
dmem_accesses,
|
|
dmisses,
|
|
((double) dmisses / (double) dmem_accesses) * 100.0);
|
|
|
|
g_string_append_printf(rep,
|
|
"Instruction accesses: %lu, Misses: %lu\nMiss rate: %lf%%\n\n",
|
|
imem_accesses,
|
|
imisses,
|
|
((double) imisses / (double) imem_accesses) * 100.0);
|
|
|
|
qemu_plugin_outs(rep->str);
|
|
}
|
|
|
|
static void log_top_insns()
|
|
{
|
|
int i;
|
|
GList *curr, *miss_insns;
|
|
InsnData *insn;
|
|
|
|
miss_insns = g_hash_table_get_values(miss_ht);
|
|
miss_insns = g_list_sort(miss_insns, dcmp);
|
|
g_autoptr(GString) rep = g_string_new("");
|
|
g_string_append_printf(rep, "%s", "address, data misses, instruction\n");
|
|
|
|
for (curr = miss_insns, i = 0; curr && i < limit; i++, curr = curr->next) {
|
|
insn = (InsnData *) curr->data;
|
|
g_string_append_printf(rep, "0x%" PRIx64, insn->addr);
|
|
if (insn->symbol) {
|
|
g_string_append_printf(rep, " (%s)", insn->symbol);
|
|
}
|
|
g_string_append_printf(rep, ", %ld, %s\n", insn->dmisses,
|
|
insn->disas_str);
|
|
}
|
|
|
|
miss_insns = g_list_sort(miss_insns, icmp);
|
|
g_string_append_printf(rep, "%s", "\naddress, fetch misses, instruction\n");
|
|
|
|
for (curr = miss_insns, i = 0; curr && i < limit; i++, curr = curr->next) {
|
|
insn = (InsnData *) curr->data;
|
|
g_string_append_printf(rep, "0x%" PRIx64, insn->addr);
|
|
if (insn->symbol) {
|
|
g_string_append_printf(rep, " (%s)", insn->symbol);
|
|
}
|
|
g_string_append_printf(rep, ", %ld, %s\n", insn->imisses,
|
|
insn->disas_str);
|
|
}
|
|
|
|
qemu_plugin_outs(rep->str);
|
|
g_list_free(miss_insns);
|
|
}
|
|
|
|
static void plugin_exit(qemu_plugin_id_t id, void *p)
|
|
{
|
|
log_stats();
|
|
log_top_insns();
|
|
|
|
cache_free(dcache);
|
|
cache_free(icache);
|
|
|
|
g_hash_table_destroy(miss_ht);
|
|
}
|
|
|
|
static void policy_init()
|
|
{
|
|
switch (policy) {
|
|
case LRU:
|
|
update_hit = lru_update_blk;
|
|
update_miss = lru_update_blk;
|
|
metadata_init = lru_priorities_init;
|
|
metadata_destroy = lru_priorities_destroy;
|
|
break;
|
|
case FIFO:
|
|
update_miss = fifo_update_on_miss;
|
|
metadata_init = fifo_init;
|
|
metadata_destroy = fifo_destroy;
|
|
break;
|
|
case RAND:
|
|
rng = g_rand_new();
|
|
break;
|
|
default:
|
|
g_assert_not_reached();
|
|
}
|
|
}
|
|
|
|
QEMU_PLUGIN_EXPORT
|
|
int qemu_plugin_install(qemu_plugin_id_t id, const qemu_info_t *info,
|
|
int argc, char **argv)
|
|
{
|
|
int i;
|
|
int iassoc, iblksize, icachesize;
|
|
int dassoc, dblksize, dcachesize;
|
|
|
|
limit = 32;
|
|
sys = info->system_emulation;
|
|
|
|
dassoc = 8;
|
|
dblksize = 64;
|
|
dcachesize = dblksize * dassoc * 32;
|
|
|
|
iassoc = 8;
|
|
iblksize = 64;
|
|
icachesize = iblksize * iassoc * 32;
|
|
|
|
policy = LRU;
|
|
|
|
for (i = 0; i < argc; i++) {
|
|
char *opt = argv[i];
|
|
if (g_str_has_prefix(opt, "iblksize=")) {
|
|
iblksize = g_ascii_strtoll(opt + 9, NULL, 10);
|
|
} else if (g_str_has_prefix(opt, "iassoc=")) {
|
|
iassoc = g_ascii_strtoll(opt + 7, NULL, 10);
|
|
} else if (g_str_has_prefix(opt, "icachesize=")) {
|
|
icachesize = g_ascii_strtoll(opt + 11, NULL, 10);
|
|
} else if (g_str_has_prefix(opt, "dblksize=")) {
|
|
dblksize = g_ascii_strtoll(opt + 9, NULL, 10);
|
|
} else if (g_str_has_prefix(opt, "dassoc=")) {
|
|
dassoc = g_ascii_strtoll(opt + 7, NULL, 10);
|
|
} else if (g_str_has_prefix(opt, "dcachesize=")) {
|
|
dcachesize = g_ascii_strtoll(opt + 11, NULL, 10);
|
|
} else if (g_str_has_prefix(opt, "limit=")) {
|
|
limit = g_ascii_strtoll(opt + 6, NULL, 10);
|
|
} else if (g_str_has_prefix(opt, "evict=")) {
|
|
gchar *p = opt + 6;
|
|
if (g_strcmp0(p, "rand") == 0) {
|
|
policy = RAND;
|
|
} else if (g_strcmp0(p, "lru") == 0) {
|
|
policy = LRU;
|
|
} else if (g_strcmp0(p, "fifo") == 0) {
|
|
policy = FIFO;
|
|
} else {
|
|
fprintf(stderr, "invalid eviction policy: %s\n", opt);
|
|
return -1;
|
|
}
|
|
} else {
|
|
fprintf(stderr, "option parsing failed: %s\n", opt);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
policy_init();
|
|
|
|
dcache = cache_init(dblksize, dassoc, dcachesize);
|
|
if (!dcache) {
|
|
const char *err = cache_config_error(dblksize, dassoc, dcachesize);
|
|
fprintf(stderr, "dcache cannot be constructed from given parameters\n");
|
|
fprintf(stderr, "%s\n", err);
|
|
return -1;
|
|
}
|
|
|
|
icache = cache_init(iblksize, iassoc, icachesize);
|
|
if (!icache) {
|
|
const char *err = cache_config_error(iblksize, iassoc, icachesize);
|
|
fprintf(stderr, "icache cannot be constructed from given parameters\n");
|
|
fprintf(stderr, "%s\n", err);
|
|
return -1;
|
|
}
|
|
|
|
qemu_plugin_register_vcpu_tb_trans_cb(id, vcpu_tb_trans);
|
|
qemu_plugin_register_atexit_cb(id, plugin_exit, NULL);
|
|
|
|
miss_ht = g_hash_table_new_full(NULL, g_direct_equal, NULL, insn_free);
|
|
|
|
return 0;
|
|
}
|