migration: code clean up
Use 'QemuMutex comp_done_lock' and 'QemuCond comp_done_cond' instead of 'QemuMutex *comp_done_lock' and 'QemuCond comp_done_cond'. To keep consistent with 'QemuMutex decomp_done_lock' and 'QemuCond comp_done_cond'. Signed-off-by: Liang Li <liang.z.li@intel.com> Message-Id: <1462433579-13691-10-git-send-email-liang.z.li@intel.com> Signed-off-by: Amit Shah <amit.shah@redhat.com>
This commit is contained in:
parent
33d151f418
commit
0d9f9a5c52
@ -280,8 +280,8 @@ static QemuThread *compress_threads;
|
|||||||
* one of the compression threads has finished the compression.
|
* one of the compression threads has finished the compression.
|
||||||
* comp_done_lock is used to co-work with comp_done_cond.
|
* comp_done_lock is used to co-work with comp_done_cond.
|
||||||
*/
|
*/
|
||||||
static QemuMutex *comp_done_lock;
|
static QemuMutex comp_done_lock;
|
||||||
static QemuCond *comp_done_cond;
|
static QemuCond comp_done_cond;
|
||||||
/* The empty QEMUFileOps will be used by file in CompressParam */
|
/* The empty QEMUFileOps will be used by file in CompressParam */
|
||||||
static const QEMUFileOps empty_ops = { };
|
static const QEMUFileOps empty_ops = { };
|
||||||
|
|
||||||
@ -310,10 +310,10 @@ static void *do_data_compress(void *opaque)
|
|||||||
|
|
||||||
do_compress_ram_page(param->file, block, offset);
|
do_compress_ram_page(param->file, block, offset);
|
||||||
|
|
||||||
qemu_mutex_lock(comp_done_lock);
|
qemu_mutex_lock(&comp_done_lock);
|
||||||
param->done = true;
|
param->done = true;
|
||||||
qemu_cond_signal(comp_done_cond);
|
qemu_cond_signal(&comp_done_cond);
|
||||||
qemu_mutex_unlock(comp_done_lock);
|
qemu_mutex_unlock(&comp_done_lock);
|
||||||
|
|
||||||
qemu_mutex_lock(¶m->mutex);
|
qemu_mutex_lock(¶m->mutex);
|
||||||
} else {
|
} else {
|
||||||
@ -353,16 +353,12 @@ void migrate_compress_threads_join(void)
|
|||||||
qemu_mutex_destroy(&comp_param[i].mutex);
|
qemu_mutex_destroy(&comp_param[i].mutex);
|
||||||
qemu_cond_destroy(&comp_param[i].cond);
|
qemu_cond_destroy(&comp_param[i].cond);
|
||||||
}
|
}
|
||||||
qemu_mutex_destroy(comp_done_lock);
|
qemu_mutex_destroy(&comp_done_lock);
|
||||||
qemu_cond_destroy(comp_done_cond);
|
qemu_cond_destroy(&comp_done_cond);
|
||||||
g_free(compress_threads);
|
g_free(compress_threads);
|
||||||
g_free(comp_param);
|
g_free(comp_param);
|
||||||
g_free(comp_done_cond);
|
|
||||||
g_free(comp_done_lock);
|
|
||||||
compress_threads = NULL;
|
compress_threads = NULL;
|
||||||
comp_param = NULL;
|
comp_param = NULL;
|
||||||
comp_done_cond = NULL;
|
|
||||||
comp_done_lock = NULL;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void migrate_compress_threads_create(void)
|
void migrate_compress_threads_create(void)
|
||||||
@ -376,10 +372,8 @@ void migrate_compress_threads_create(void)
|
|||||||
thread_count = migrate_compress_threads();
|
thread_count = migrate_compress_threads();
|
||||||
compress_threads = g_new0(QemuThread, thread_count);
|
compress_threads = g_new0(QemuThread, thread_count);
|
||||||
comp_param = g_new0(CompressParam, thread_count);
|
comp_param = g_new0(CompressParam, thread_count);
|
||||||
comp_done_cond = g_new0(QemuCond, 1);
|
qemu_cond_init(&comp_done_cond);
|
||||||
comp_done_lock = g_new0(QemuMutex, 1);
|
qemu_mutex_init(&comp_done_lock);
|
||||||
qemu_cond_init(comp_done_cond);
|
|
||||||
qemu_mutex_init(comp_done_lock);
|
|
||||||
for (i = 0; i < thread_count; i++) {
|
for (i = 0; i < thread_count; i++) {
|
||||||
/* com_param[i].file is just used as a dummy buffer to save data, set
|
/* com_param[i].file is just used as a dummy buffer to save data, set
|
||||||
* it's ops to empty.
|
* it's ops to empty.
|
||||||
@ -840,13 +834,13 @@ static void flush_compressed_data(QEMUFile *f)
|
|||||||
}
|
}
|
||||||
thread_count = migrate_compress_threads();
|
thread_count = migrate_compress_threads();
|
||||||
|
|
||||||
qemu_mutex_lock(comp_done_lock);
|
qemu_mutex_lock(&comp_done_lock);
|
||||||
for (idx = 0; idx < thread_count; idx++) {
|
for (idx = 0; idx < thread_count; idx++) {
|
||||||
while (!comp_param[idx].done) {
|
while (!comp_param[idx].done) {
|
||||||
qemu_cond_wait(comp_done_cond, comp_done_lock);
|
qemu_cond_wait(&comp_done_cond, &comp_done_lock);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
qemu_mutex_unlock(comp_done_lock);
|
qemu_mutex_unlock(&comp_done_lock);
|
||||||
|
|
||||||
for (idx = 0; idx < thread_count; idx++) {
|
for (idx = 0; idx < thread_count; idx++) {
|
||||||
qemu_mutex_lock(&comp_param[idx].mutex);
|
qemu_mutex_lock(&comp_param[idx].mutex);
|
||||||
@ -872,7 +866,7 @@ static int compress_page_with_multi_thread(QEMUFile *f, RAMBlock *block,
|
|||||||
int idx, thread_count, bytes_xmit = -1, pages = -1;
|
int idx, thread_count, bytes_xmit = -1, pages = -1;
|
||||||
|
|
||||||
thread_count = migrate_compress_threads();
|
thread_count = migrate_compress_threads();
|
||||||
qemu_mutex_lock(comp_done_lock);
|
qemu_mutex_lock(&comp_done_lock);
|
||||||
while (true) {
|
while (true) {
|
||||||
for (idx = 0; idx < thread_count; idx++) {
|
for (idx = 0; idx < thread_count; idx++) {
|
||||||
if (comp_param[idx].done) {
|
if (comp_param[idx].done) {
|
||||||
@ -891,10 +885,10 @@ static int compress_page_with_multi_thread(QEMUFile *f, RAMBlock *block,
|
|||||||
if (pages > 0) {
|
if (pages > 0) {
|
||||||
break;
|
break;
|
||||||
} else {
|
} else {
|
||||||
qemu_cond_wait(comp_done_cond, comp_done_lock);
|
qemu_cond_wait(&comp_done_cond, &comp_done_lock);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
qemu_mutex_unlock(comp_done_lock);
|
qemu_mutex_unlock(&comp_done_lock);
|
||||||
|
|
||||||
return pages;
|
return pages;
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user