tests: introduce a framework for testing migration performance
This introduces a moderately general purpose framework for testing performance of migration. The initial guest workload is provided by the included 'stress' program, which is configured to spawn one thread per guest CPU and run a maximally memory intensive workload. It will loop over GB of memory, xor'ing each byte with data from a 4k array of random bytes. This ensures heavy read and write load across all of guest memory to stress the migration performance. While running the 'stress' program will record how long it takes to xor each GB of memory and print this data for later reporting. The test engine will spawn a pair of QEMU processes, either on the same host, or with the target on a remote host via ssh, using the host kernel and a custom initrd built with 'stress' as the /init binary. Kernel command line args are set to ensure a fast kernel boot time (< 1 second) between launching QEMU and the stress program starting execution. None the less, the test engine will initially wait N seconds for the guest workload to stablize, before starting the migration operation. When migration is running, the engine will use pause, post-copy, autoconverge, xbzrle compression and multithread compression features, as well as downtime & bandwidth tuning to encourage completion. If migration completes, the test engine will wait N seconds again for the guest workooad to stablize on the target host. If migration does not complete after a preset number of iterations, it will be aborted. While the QEMU process is running on the source host, the test engine will sample the host CPU usage of QEMU as a whole, and each vCPU thread. While migration is running, it will record all the stats reported by 'query-migration'. Finally, it will capture the output of the stress program running in the guest. All the data produced from a single test execution is recorded in a structured JSON file. A separate program is then able to create interactive charts using the "plotly" python + javascript libraries, showing the characteristics of the migration. The data output provides visualization of the effect on guest vCPU workloads from the migration process, the corresponding vCPU utilization on the host, and the overall CPU hit from QEMU on the host. This is correlated from statistics from the migration process, such as downtime, vCPU throttling and iteration number. While the tests can be run individually with arbitrary parameters, there is also a facility for producing batch reports for a number of pre-defined scenarios / comparisons, in order to be able to get standardized results across different hardware configurations (eg TCP vs RDMA, or comparing different VCPU counts / memory sizes, etc). To use this, first you must build the initrd image $ make tests/migration/initrd-stress.img To run a a one-shot test with all default parameters $ ./tests/migration/guestperf.py > result.json This has many command line args for varying its behaviour. For example, to increase the RAM size and CPU count and bind it to specific host NUMA nodes $ ./tests/migration/guestperf.py \ --mem 4 --cpus 2 \ --src-mem-bind 0 --src-cpu-bind 0,1 \ --dst-mem-bind 1 --dst-cpu-bind 2,3 \ > result.json Using mem + cpu binding is strongly recommended on NUMA machines, otherwise the guest performance results will vary wildly between runs of the test due to lucky/unlucky NUMA placement, making sensible data analysis impossible. To make it run across separate hosts: $ ./tests/migration/guestperf.py \ --dst-host somehostname > result.json To request that post-copy is enabled, with switchover after 5 iterations $ ./tests/migration/guestperf.py \ --post-copy --post-copy-iters 5 > result.json Once a result.json file is created, a graph of the data can be generated, showing guest workload performance per thread and the migration iteration points: $ ./tests/migration/guestperf-plot.py --output result.html \ --migration-iters --split-guest-cpu result.json To further include host vCPU utilization and overall QEMU utilization $ ./tests/migration/guestperf-plot.py --output result.html \ --migration-iters --split-guest-cpu \ --qemu-cpu --vcpu-cpu result.json NB, the 'guestperf-plot.py' command requires that you have the plotly python library installed. eg you must do $ pip install --user plotly Viewing the result.html file requires that you have the plotly.min.js file in the same directory as the HTML output. This js file is installed as part of the plotly python library, so can be found in $HOME/.local/lib/python2.7/site-packages/plotly/offline/plotly.min.js The guestperf-plot.py program can accept multiple json files to plot, enabling results from different configurations to be compared. Finally, to run the entire standardized set of comparisons $ ./tests/migration/guestperf-batch.py \ --dst-host somehost \ --mem 4 --cpus 2 \ --src-mem-bind 0 --src-cpu-bind 0,1 \ --dst-mem-bind 1 --dst-cpu-bind 2,3 --output tcp-somehost-4gb-2cpu will store JSON files from all scenarios in the directory named tcp-somehost-4gb-2cpu Signed-off-by: Daniel P. Berrange <berrange@redhat.com> Message-Id: <1469020993-29426-7-git-send-email-berrange@redhat.com> Signed-off-by: Amit Shah <amit.shah@redhat.com>
This commit is contained in:
parent
168ae6c24b
commit
409437e16d
2
configure
vendored
2
configure
vendored
@ -3121,6 +3121,7 @@ else
|
||||
if test "$found" = "no"; then
|
||||
LIBS="$pthread_lib $LIBS"
|
||||
fi
|
||||
PTHREAD_LIB="$pthread_lib"
|
||||
break
|
||||
fi
|
||||
done
|
||||
@ -5537,6 +5538,7 @@ echo "LDFLAGS_NOPIE=$LDFLAGS_NOPIE" >> $config_host_mak
|
||||
echo "LD_REL_FLAGS=$LD_REL_FLAGS" >> $config_host_mak
|
||||
echo "LIBS+=$LIBS" >> $config_host_mak
|
||||
echo "LIBS_TOOLS+=$libs_tools" >> $config_host_mak
|
||||
echo "PTHREAD_LIB=$PTHREAD_LIB" >> $config_host_mak
|
||||
echo "EXESUF=$EXESUF" >> $config_host_mak
|
||||
echo "DSOSUF=$DSOSUF" >> $config_host_mak
|
||||
echo "LDFLAGS_SHARED=$LDFLAGS_SHARED" >> $config_host_mak
|
||||
|
@ -627,6 +627,18 @@ tests/test-filter-redirector$(EXESUF): tests/test-filter-redirector.o $(qtest-ob
|
||||
tests/ivshmem-test$(EXESUF): tests/ivshmem-test.o contrib/ivshmem-server/ivshmem-server.o $(libqos-pc-obj-y)
|
||||
tests/vhost-user-bridge$(EXESUF): tests/vhost-user-bridge.o
|
||||
|
||||
tests/migration/stress$(EXESUF): tests/migration/stress.o
|
||||
$(call quiet-command, $(LINKPROG) -static -O3 $(PTHREAD_LIB) -o $@ $< ," LINK $(TARGET_DIR)$@")
|
||||
|
||||
INITRD_WORK_DIR=tests/migration/initrd
|
||||
|
||||
tests/migration/initrd-stress.img: tests/migration/stress$(EXESUF)
|
||||
mkdir -p $(INITRD_WORK_DIR)
|
||||
cp $< $(INITRD_WORK_DIR)/init
|
||||
(cd $(INITRD_WORK_DIR) && (find | cpio --quiet -o -H newc | gzip -9)) > $@
|
||||
rm $(INITRD_WORK_DIR)/init
|
||||
rmdir $(INITRD_WORK_DIR)
|
||||
|
||||
ifeq ($(CONFIG_POSIX),y)
|
||||
LIBS += -lutil
|
||||
endif
|
||||
|
2
tests/migration/.gitignore
vendored
Normal file
2
tests/migration/.gitignore
vendored
Normal file
@ -0,0 +1,2 @@
|
||||
initrd-stress.img
|
||||
stress
|
26
tests/migration/guestperf-batch.py
Executable file
26
tests/migration/guestperf-batch.py
Executable file
@ -0,0 +1,26 @@
|
||||
#!/usr/bin/python
|
||||
#
|
||||
# Migration test batch comparison invokation
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
import sys
|
||||
|
||||
from guestperf.shell import BatchShell
|
||||
|
||||
shell = BatchShell()
|
||||
sys.exit(shell.run(sys.argv[1:]))
|
26
tests/migration/guestperf-plot.py
Executable file
26
tests/migration/guestperf-plot.py
Executable file
@ -0,0 +1,26 @@
|
||||
#!/usr/bin/python
|
||||
#
|
||||
# Migration test graph plotting command
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
import sys
|
||||
|
||||
from guestperf.shell import PlotShell
|
||||
|
||||
shell = PlotShell()
|
||||
sys.exit(shell.run(sys.argv[1:]))
|
27
tests/migration/guestperf.py
Executable file
27
tests/migration/guestperf.py
Executable file
@ -0,0 +1,27 @@
|
||||
#!/usr/bin/python
|
||||
#
|
||||
# Migration test direct invokation command
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
|
||||
import sys
|
||||
|
||||
from guestperf.shell import Shell
|
||||
|
||||
shell = Shell()
|
||||
sys.exit(shell.run(sys.argv[1:]))
|
0
tests/migration/guestperf/__init__.py
Normal file
0
tests/migration/guestperf/__init__.py
Normal file
124
tests/migration/guestperf/comparison.py
Normal file
124
tests/migration/guestperf/comparison.py
Normal file
@ -0,0 +1,124 @@
|
||||
#
|
||||
# Migration test scenario comparison mapping
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
from guestperf.scenario import Scenario
|
||||
|
||||
class Comparison(object):
|
||||
def __init__(self, name, scenarios):
|
||||
self._name = name
|
||||
self._scenarios = scenarios
|
||||
|
||||
COMPARISONS = [
|
||||
# Looking at effect of pausing guest during migration
|
||||
# at various stages of iteration over RAM
|
||||
Comparison("pause-iters", scenarios = [
|
||||
Scenario("pause-iters-0",
|
||||
pause=True, pause_iters=0),
|
||||
Scenario("pause-iters-1",
|
||||
pause=True, pause_iters=1),
|
||||
Scenario("pause-iters-5",
|
||||
pause=True, pause_iters=5),
|
||||
Scenario("pause-iters-20",
|
||||
pause=True, pause_iters=20),
|
||||
]),
|
||||
|
||||
|
||||
# Looking at use of post-copy in relation to bandwidth
|
||||
# available for migration
|
||||
Comparison("post-copy-bandwidth", scenarios = [
|
||||
Scenario("post-copy-bw-100mbs",
|
||||
post_copy=True, bandwidth=12),
|
||||
Scenario("post-copy-bw-300mbs",
|
||||
post_copy=True, bandwidth=37),
|
||||
Scenario("post-copy-bw-1gbs",
|
||||
post_copy=True, bandwidth=125),
|
||||
Scenario("post-copy-bw-10gbs",
|
||||
post_copy=True, bandwidth=1250),
|
||||
Scenario("post-copy-bw-100gbs",
|
||||
post_copy=True, bandwidth=12500),
|
||||
]),
|
||||
|
||||
|
||||
# Looking at effect of starting post-copy at different
|
||||
# stages of the migration
|
||||
Comparison("post-copy-iters", scenarios = [
|
||||
Scenario("post-copy-iters-0",
|
||||
post_copy=True, post_copy_iters=0),
|
||||
Scenario("post-copy-iters-1",
|
||||
post_copy=True, post_copy_iters=1),
|
||||
Scenario("post-copy-iters-5",
|
||||
post_copy=True, post_copy_iters=5),
|
||||
Scenario("post-copy-iters-20",
|
||||
post_copy=True, post_copy_iters=20),
|
||||
]),
|
||||
|
||||
|
||||
# Looking at effect of auto-converge with different
|
||||
# throttling percentage step rates
|
||||
Comparison("auto-converge-iters", scenarios = [
|
||||
Scenario("auto-converge-step-5",
|
||||
auto_converge=True, auto_converge_step=5),
|
||||
Scenario("auto-converge-step-10",
|
||||
auto_converge=True, auto_converge_step=10),
|
||||
Scenario("auto-converge-step-20",
|
||||
auto_converge=True, auto_converge_step=20),
|
||||
]),
|
||||
|
||||
|
||||
# Looking at use of auto-converge in relation to bandwidth
|
||||
# available for migration
|
||||
Comparison("auto-converge-bandwidth", scenarios = [
|
||||
Scenario("auto-converge-bw-100mbs",
|
||||
auto_converge=True, bandwidth=12),
|
||||
Scenario("auto-converge-bw-300mbs",
|
||||
auto_converge=True, bandwidth=37),
|
||||
Scenario("auto-converge-bw-1gbs",
|
||||
auto_converge=True, bandwidth=125),
|
||||
Scenario("auto-converge-bw-10gbs",
|
||||
auto_converge=True, bandwidth=1250),
|
||||
Scenario("auto-converge-bw-100gbs",
|
||||
auto_converge=True, bandwidth=12500),
|
||||
]),
|
||||
|
||||
|
||||
# Looking at effect of multi-thread compression with
|
||||
# varying numbers of threads
|
||||
Comparison("compr-mt", scenarios = [
|
||||
Scenario("compr-mt-threads-1",
|
||||
compression_mt=True, compression_mt_threads=1),
|
||||
Scenario("compr-mt-threads-2",
|
||||
compression_mt=True, compression_mt_threads=2),
|
||||
Scenario("compr-mt-threads-4",
|
||||
compression_mt=True, compression_mt_threads=4),
|
||||
]),
|
||||
|
||||
|
||||
# Looking at effect of xbzrle compression with varying
|
||||
# cache sizes
|
||||
Comparison("compr-xbzrle", scenarios = [
|
||||
Scenario("compr-xbzrle-cache-5",
|
||||
compression_xbzrle=True, compression_xbzrle_cache=5),
|
||||
Scenario("compr-xbzrle-cache-10",
|
||||
compression_xbzrle=True, compression_xbzrle_cache=10),
|
||||
Scenario("compr-xbzrle-cache-20",
|
||||
compression_xbzrle=True, compression_xbzrle_cache=10),
|
||||
Scenario("compr-xbzrle-cache-50",
|
||||
compression_xbzrle=True, compression_xbzrle_cache=50),
|
||||
]),
|
||||
]
|
439
tests/migration/guestperf/engine.py
Normal file
439
tests/migration/guestperf/engine.py
Normal file
@ -0,0 +1,439 @@
|
||||
#
|
||||
# Migration test main engine
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
|
||||
import os
|
||||
import re
|
||||
import sys
|
||||
import time
|
||||
|
||||
sys.path.append(os.path.join(os.path.dirname(__file__), '..', '..', '..', 'scripts'))
|
||||
import qemu
|
||||
import qmp.qmp
|
||||
from guestperf.progress import Progress, ProgressStats
|
||||
from guestperf.report import Report
|
||||
from guestperf.timings import TimingRecord, Timings
|
||||
|
||||
|
||||
class Engine(object):
|
||||
|
||||
def __init__(self, binary, dst_host, kernel, initrd, transport="tcp",
|
||||
sleep=15, verbose=False, debug=False):
|
||||
|
||||
self._binary = binary # Path to QEMU binary
|
||||
self._dst_host = dst_host # Hostname of target host
|
||||
self._kernel = kernel # Path to kernel image
|
||||
self._initrd = initrd # Path to stress initrd
|
||||
self._transport = transport # 'unix' or 'tcp' or 'rdma'
|
||||
self._sleep = sleep
|
||||
self._verbose = verbose
|
||||
self._debug = debug
|
||||
|
||||
if debug:
|
||||
self._verbose = debug
|
||||
|
||||
def _vcpu_timing(self, pid, tid_list):
|
||||
records = []
|
||||
now = time.time()
|
||||
|
||||
jiffies_per_sec = os.sysconf(os.sysconf_names['SC_CLK_TCK'])
|
||||
for tid in tid_list:
|
||||
statfile = "/proc/%d/task/%d/stat" % (pid, tid)
|
||||
with open(statfile, "r") as fh:
|
||||
stat = fh.readline()
|
||||
fields = stat.split(" ")
|
||||
stime = int(fields[13])
|
||||
utime = int(fields[14])
|
||||
records.append(TimingRecord(tid, now, 1000 * (stime + utime) / jiffies_per_sec))
|
||||
return records
|
||||
|
||||
def _cpu_timing(self, pid):
|
||||
records = []
|
||||
now = time.time()
|
||||
|
||||
jiffies_per_sec = os.sysconf(os.sysconf_names['SC_CLK_TCK'])
|
||||
statfile = "/proc/%d/stat" % pid
|
||||
with open(statfile, "r") as fh:
|
||||
stat = fh.readline()
|
||||
fields = stat.split(" ")
|
||||
stime = int(fields[13])
|
||||
utime = int(fields[14])
|
||||
return TimingRecord(pid, now, 1000 * (stime + utime) / jiffies_per_sec)
|
||||
|
||||
def _migrate_progress(self, vm):
|
||||
info = vm.command("query-migrate")
|
||||
|
||||
if "ram" not in info:
|
||||
info["ram"] = {}
|
||||
|
||||
return Progress(
|
||||
info.get("status", "active"),
|
||||
ProgressStats(
|
||||
info["ram"].get("transferred", 0),
|
||||
info["ram"].get("remaining", 0),
|
||||
info["ram"].get("total", 0),
|
||||
info["ram"].get("duplicate", 0),
|
||||
info["ram"].get("skipped", 0),
|
||||
info["ram"].get("normal", 0),
|
||||
info["ram"].get("normal-bytes", 0),
|
||||
info["ram"].get("dirty-pages-rate", 0),
|
||||
info["ram"].get("mbps", 0),
|
||||
info["ram"].get("dirty-sync-count", 0)
|
||||
),
|
||||
time.time(),
|
||||
info.get("total-time", 0),
|
||||
info.get("downtime", 0),
|
||||
info.get("expected-downtime", 0),
|
||||
info.get("setup-time", 0),
|
||||
info.get("x-cpu-throttle-percentage", 0),
|
||||
)
|
||||
|
||||
def _migrate(self, hardware, scenario, src, dst, connect_uri):
|
||||
src_qemu_time = []
|
||||
src_vcpu_time = []
|
||||
src_pid = src.get_pid()
|
||||
|
||||
vcpus = src.command("query-cpus")
|
||||
src_threads = []
|
||||
for vcpu in vcpus:
|
||||
src_threads.append(vcpu["thread_id"])
|
||||
|
||||
# XXX how to get dst timings on remote host ?
|
||||
|
||||
if self._verbose:
|
||||
print "Sleeping %d seconds for initial guest workload run" % self._sleep
|
||||
sleep_secs = self._sleep
|
||||
while sleep_secs > 1:
|
||||
src_qemu_time.append(self._cpu_timing(src_pid))
|
||||
src_vcpu_time.extend(self._vcpu_timing(src_pid, src_threads))
|
||||
time.sleep(1)
|
||||
sleep_secs -= 1
|
||||
|
||||
if self._verbose:
|
||||
print "Starting migration"
|
||||
if scenario._auto_converge:
|
||||
resp = src.command("migrate-set-capabilities",
|
||||
capabilities = [
|
||||
{ "capability": "auto-converge",
|
||||
"state": True }
|
||||
])
|
||||
resp = src.command("migrate-set-parameters",
|
||||
x_cpu_throttle_increment=scenario._auto_converge_step)
|
||||
|
||||
if scenario._post_copy:
|
||||
resp = src.command("migrate-set-capabilities",
|
||||
capabilities = [
|
||||
{ "capability": "postcopy-ram",
|
||||
"state": True }
|
||||
])
|
||||
resp = dst.command("migrate-set-capabilities",
|
||||
capabilities = [
|
||||
{ "capability": "postcopy-ram",
|
||||
"state": True }
|
||||
])
|
||||
|
||||
resp = src.command("migrate_set_speed",
|
||||
value=scenario._bandwidth * 1024 * 1024)
|
||||
|
||||
resp = src.command("migrate_set_downtime",
|
||||
value=scenario._downtime / 1024.0)
|
||||
|
||||
if scenario._compression_mt:
|
||||
resp = src.command("migrate-set-capabilities",
|
||||
capabilities = [
|
||||
{ "capability": "compress",
|
||||
"state": True }
|
||||
])
|
||||
resp = src.command("migrate-set-parameters",
|
||||
compress_threads=scenario._compression_mt_threads)
|
||||
resp = dst.command("migrate-set-capabilities",
|
||||
capabilities = [
|
||||
{ "capability": "compress",
|
||||
"state": True }
|
||||
])
|
||||
resp = dst.command("migrate-set-parameters",
|
||||
decompress_threads=scenario._compression_mt_threads)
|
||||
|
||||
if scenario._compression_xbzrle:
|
||||
resp = src.command("migrate-set-capabilities",
|
||||
capabilities = [
|
||||
{ "capability": "xbzrle",
|
||||
"state": True }
|
||||
])
|
||||
resp = dst.command("migrate-set-capabilities",
|
||||
capabilities = [
|
||||
{ "capability": "xbzrle",
|
||||
"state": True }
|
||||
])
|
||||
resp = src.command("migrate-set-cache-size",
|
||||
value=(hardware._mem * 1024 * 1024 * 1024 / 100 *
|
||||
scenario._compression_xbzrle_cache))
|
||||
|
||||
resp = src.command("migrate", uri=connect_uri)
|
||||
|
||||
post_copy = False
|
||||
paused = False
|
||||
|
||||
progress_history = []
|
||||
|
||||
start = time.time()
|
||||
loop = 0
|
||||
while True:
|
||||
loop = loop + 1
|
||||
time.sleep(0.05)
|
||||
|
||||
progress = self._migrate_progress(src)
|
||||
if (loop % 20) == 0:
|
||||
src_qemu_time.append(self._cpu_timing(src_pid))
|
||||
src_vcpu_time.extend(self._vcpu_timing(src_pid, src_threads))
|
||||
|
||||
if (len(progress_history) == 0 or
|
||||
(progress_history[-1]._ram._iterations <
|
||||
progress._ram._iterations)):
|
||||
progress_history.append(progress)
|
||||
|
||||
if progress._status in ("completed", "failed", "cancelled"):
|
||||
if progress._status == "completed" and paused:
|
||||
dst.command("cont")
|
||||
if progress_history[-1] != progress:
|
||||
progress_history.append(progress)
|
||||
|
||||
if progress._status == "completed":
|
||||
if self._verbose:
|
||||
print "Sleeping %d seconds for final guest workload run" % self._sleep
|
||||
sleep_secs = self._sleep
|
||||
while sleep_secs > 1:
|
||||
time.sleep(1)
|
||||
src_qemu_time.append(self._cpu_timing(src_pid))
|
||||
src_vcpu_time.extend(self._vcpu_timing(src_pid, src_threads))
|
||||
sleep_secs -= 1
|
||||
|
||||
return [progress_history, src_qemu_time, src_vcpu_time]
|
||||
|
||||
if self._verbose and (loop % 20) == 0:
|
||||
print "Iter %d: remain %5dMB of %5dMB (total %5dMB @ %5dMb/sec)" % (
|
||||
progress._ram._iterations,
|
||||
progress._ram._remaining_bytes / (1024 * 1024),
|
||||
progress._ram._total_bytes / (1024 * 1024),
|
||||
progress._ram._transferred_bytes / (1024 * 1024),
|
||||
progress._ram._transfer_rate_mbs,
|
||||
)
|
||||
|
||||
if progress._ram._iterations > scenario._max_iters:
|
||||
if self._verbose:
|
||||
print "No completion after %d iterations over RAM" % scenario._max_iters
|
||||
src.command("migrate_cancel")
|
||||
continue
|
||||
|
||||
if time.time() > (start + scenario._max_time):
|
||||
if self._verbose:
|
||||
print "No completion after %d seconds" % scenario._max_time
|
||||
src.command("migrate_cancel")
|
||||
continue
|
||||
|
||||
if (scenario._post_copy and
|
||||
progress._ram._iterations >= scenario._post_copy_iters and
|
||||
not post_copy):
|
||||
if self._verbose:
|
||||
print "Switching to post-copy after %d iterations" % scenario._post_copy_iters
|
||||
resp = src.command("migrate-start-postcopy")
|
||||
post_copy = True
|
||||
|
||||
if (scenario._pause and
|
||||
progress._ram._iterations >= scenario._pause_iters and
|
||||
not paused):
|
||||
if self._verbose:
|
||||
print "Pausing VM after %d iterations" % scenario._pause_iters
|
||||
resp = src.command("stop")
|
||||
paused = True
|
||||
|
||||
def _get_common_args(self, hardware, tunnelled=False):
|
||||
args = [
|
||||
"noapic",
|
||||
"edd=off",
|
||||
"printk.time=1",
|
||||
"noreplace-smp",
|
||||
"cgroup_disable=memory",
|
||||
"pci=noearly",
|
||||
"console=ttyS0",
|
||||
]
|
||||
if self._debug:
|
||||
args.append("debug")
|
||||
else:
|
||||
args.append("quiet")
|
||||
|
||||
args.append("ramsize=%s" % hardware._mem)
|
||||
|
||||
cmdline = " ".join(args)
|
||||
if tunnelled:
|
||||
cmdline = "'" + cmdline + "'"
|
||||
|
||||
argv = [
|
||||
"-machine", "accel=kvm",
|
||||
"-cpu", "host",
|
||||
"-kernel", self._kernel,
|
||||
"-initrd", self._initrd,
|
||||
"-append", cmdline,
|
||||
"-chardev", "stdio,id=cdev0",
|
||||
"-device", "isa-serial,chardev=cdev0",
|
||||
"-m", str((hardware._mem * 1024) + 512),
|
||||
"-smp", str(hardware._cpus),
|
||||
]
|
||||
|
||||
if self._debug:
|
||||
argv.extend(["-device", "sga"])
|
||||
|
||||
if hardware._prealloc_pages:
|
||||
argv_source += ["-mem-path", "/dev/shm",
|
||||
"-mem-prealloc"]
|
||||
if hardware._locked_pages:
|
||||
argv_source += ["-realtime", "mlock=on"]
|
||||
if hardware._huge_pages:
|
||||
pass
|
||||
|
||||
return argv
|
||||
|
||||
def _get_src_args(self, hardware):
|
||||
return self._get_common_args(hardware)
|
||||
|
||||
def _get_dst_args(self, hardware, uri):
|
||||
tunnelled = False
|
||||
if self._dst_host != "localhost":
|
||||
tunnelled = True
|
||||
argv = self._get_common_args(hardware, tunnelled)
|
||||
return argv + ["-incoming", uri]
|
||||
|
||||
@staticmethod
|
||||
def _get_common_wrapper(cpu_bind, mem_bind):
|
||||
wrapper = []
|
||||
if len(cpu_bind) > 0 or len(mem_bind) > 0:
|
||||
wrapper.append("numactl")
|
||||
if cpu_bind:
|
||||
wrapper.append("--physcpubind=%s" % ",".join(cpu_bind))
|
||||
if mem_bind:
|
||||
wrapper.append("--membind=%s" % ",".join(mem_bind))
|
||||
|
||||
return wrapper
|
||||
|
||||
def _get_src_wrapper(self, hardware):
|
||||
return self._get_common_wrapper(hardware._src_cpu_bind, hardware._src_mem_bind)
|
||||
|
||||
def _get_dst_wrapper(self, hardware):
|
||||
wrapper = self._get_common_wrapper(hardware._dst_cpu_bind, hardware._dst_mem_bind)
|
||||
if self._dst_host != "localhost":
|
||||
return ["ssh",
|
||||
"-R", "9001:localhost:9001",
|
||||
self._dst_host] + wrapper
|
||||
else:
|
||||
return wrapper
|
||||
|
||||
def _get_timings(self, vm):
|
||||
log = vm.get_log()
|
||||
if not log:
|
||||
return []
|
||||
if self._debug:
|
||||
print log
|
||||
|
||||
regex = r"[^\s]+\s\((\d+)\):\sINFO:\s(\d+)ms\scopied\s\d+\sGB\sin\s(\d+)ms"
|
||||
matcher = re.compile(regex)
|
||||
records = []
|
||||
for line in log.split("\n"):
|
||||
match = matcher.match(line)
|
||||
if match:
|
||||
records.append(TimingRecord(int(match.group(1)),
|
||||
int(match.group(2)) / 1000.0,
|
||||
int(match.group(3))))
|
||||
return records
|
||||
|
||||
def run(self, hardware, scenario, result_dir=os.getcwd()):
|
||||
abs_result_dir = os.path.join(result_dir, scenario._name)
|
||||
|
||||
if self._transport == "tcp":
|
||||
uri = "tcp:%s:9000" % self._dst_host
|
||||
elif self._transport == "rdma":
|
||||
uri = "rdma:%s:9000" % self._dst_host
|
||||
elif self._transport == "unix":
|
||||
if self._dst_host != "localhost":
|
||||
raise Exception("Running use unix migration transport for non-local host")
|
||||
uri = "unix:/var/tmp/qemu-migrate-%d.migrate" % os.getpid()
|
||||
try:
|
||||
os.remove(uri[5:])
|
||||
os.remove(monaddr)
|
||||
except:
|
||||
pass
|
||||
|
||||
if self._dst_host != "localhost":
|
||||
dstmonaddr = ("localhost", 9001)
|
||||
else:
|
||||
dstmonaddr = "/var/tmp/qemu-dst-%d-monitor.sock" % os.getpid()
|
||||
srcmonaddr = "/var/tmp/qemu-src-%d-monitor.sock" % os.getpid()
|
||||
|
||||
src = qemu.QEMUMachine(self._binary,
|
||||
args=self._get_src_args(hardware),
|
||||
wrapper=self._get_src_wrapper(hardware),
|
||||
name="qemu-src-%d" % os.getpid(),
|
||||
monitor_address=srcmonaddr,
|
||||
debug=self._debug)
|
||||
|
||||
dst = qemu.QEMUMachine(self._binary,
|
||||
args=self._get_dst_args(hardware, uri),
|
||||
wrapper=self._get_dst_wrapper(hardware),
|
||||
name="qemu-dst-%d" % os.getpid(),
|
||||
monitor_address=dstmonaddr,
|
||||
debug=self._debug)
|
||||
|
||||
try:
|
||||
src.launch()
|
||||
dst.launch()
|
||||
|
||||
ret = self._migrate(hardware, scenario, src, dst, uri)
|
||||
progress_history = ret[0]
|
||||
qemu_timings = ret[1]
|
||||
vcpu_timings = ret[2]
|
||||
if uri[0:5] == "unix:":
|
||||
os.remove(uri[5:])
|
||||
if self._verbose:
|
||||
print "Finished migration"
|
||||
|
||||
src.shutdown()
|
||||
dst.shutdown()
|
||||
|
||||
return Report(hardware, scenario, progress_history,
|
||||
Timings(self._get_timings(src) + self._get_timings(dst)),
|
||||
Timings(qemu_timings),
|
||||
Timings(vcpu_timings),
|
||||
self._binary, self._dst_host, self._kernel,
|
||||
self._initrd, self._transport, self._sleep)
|
||||
except Exception as e:
|
||||
if self._debug:
|
||||
print "Failed: %s" % str(e)
|
||||
try:
|
||||
src.shutdown()
|
||||
except:
|
||||
pass
|
||||
try:
|
||||
dst.shutdown()
|
||||
except:
|
||||
pass
|
||||
|
||||
if self._debug:
|
||||
print src.get_log()
|
||||
print dst.get_log()
|
||||
raise
|
||||
|
62
tests/migration/guestperf/hardware.py
Normal file
62
tests/migration/guestperf/hardware.py
Normal file
@ -0,0 +1,62 @@
|
||||
#
|
||||
# Migration test hardware configuration description
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
|
||||
class Hardware(object):
|
||||
def __init__(self, cpus=1, mem=1,
|
||||
src_cpu_bind=None, src_mem_bind=None,
|
||||
dst_cpu_bind=None, dst_mem_bind=None,
|
||||
prealloc_pages = False,
|
||||
huge_pages=False, locked_pages=False):
|
||||
self._cpus = cpus
|
||||
self._mem = mem # GiB
|
||||
self._src_mem_bind = src_mem_bind # List of NUMA nodes
|
||||
self._src_cpu_bind = src_cpu_bind # List of pCPUs
|
||||
self._dst_mem_bind = dst_mem_bind # List of NUMA nodes
|
||||
self._dst_cpu_bind = dst_cpu_bind # List of pCPUs
|
||||
self._prealloc_pages = prealloc_pages
|
||||
self._huge_pages = huge_pages
|
||||
self._locked_pages = locked_pages
|
||||
|
||||
|
||||
def serialize(self):
|
||||
return {
|
||||
"cpus": self._cpus,
|
||||
"mem": self._mem,
|
||||
"src_mem_bind": self._src_mem_bind,
|
||||
"dst_mem_bind": self._dst_mem_bind,
|
||||
"src_cpu_bind": self._src_cpu_bind,
|
||||
"dst_cpu_bind": self._dst_cpu_bind,
|
||||
"prealloc_pages": self._prealloc_pages,
|
||||
"huge_pages": self._huge_pages,
|
||||
"locked_pages": self._locked_pages,
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def deserialize(cls, data):
|
||||
return cls(
|
||||
data["cpus"],
|
||||
data["mem"],
|
||||
data["src_cpu_bind"],
|
||||
data["src_mem_bind"],
|
||||
data["dst_cpu_bind"],
|
||||
data["dst_mem_bind"],
|
||||
data["prealloc_pages"],
|
||||
data["huge_pages"],
|
||||
data["locked_pages"])
|
623
tests/migration/guestperf/plot.py
Normal file
623
tests/migration/guestperf/plot.py
Normal file
@ -0,0 +1,623 @@
|
||||
#
|
||||
# Migration test graph plotting
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
import sys
|
||||
|
||||
|
||||
class Plot(object):
|
||||
|
||||
# Generated using
|
||||
# http://tools.medialab.sciences-po.fr/iwanthue/
|
||||
COLORS = ["#CD54D0",
|
||||
"#79D94C",
|
||||
"#7470CD",
|
||||
"#D2D251",
|
||||
"#863D79",
|
||||
"#76DDA6",
|
||||
"#D4467B",
|
||||
"#61923D",
|
||||
"#CB9CCA",
|
||||
"#D98F36",
|
||||
"#8CC8DA",
|
||||
"#CE4831",
|
||||
"#5E7693",
|
||||
"#9B803F",
|
||||
"#412F4C",
|
||||
"#CECBA6",
|
||||
"#6D3229",
|
||||
"#598B73",
|
||||
"#C8827C",
|
||||
"#394427"]
|
||||
|
||||
def __init__(self,
|
||||
reports,
|
||||
migration_iters,
|
||||
total_guest_cpu,
|
||||
split_guest_cpu,
|
||||
qemu_cpu,
|
||||
vcpu_cpu):
|
||||
|
||||
self._reports = reports
|
||||
self._migration_iters = migration_iters
|
||||
self._total_guest_cpu = total_guest_cpu
|
||||
self._split_guest_cpu = split_guest_cpu
|
||||
self._qemu_cpu = qemu_cpu
|
||||
self._vcpu_cpu = vcpu_cpu
|
||||
self._color_idx = 0
|
||||
|
||||
def _next_color(self):
|
||||
color = self.COLORS[self._color_idx]
|
||||
self._color_idx += 1
|
||||
if self._color_idx >= len(self.COLORS):
|
||||
self._color_idx = 0
|
||||
return color
|
||||
|
||||
def _get_progress_label(self, progress):
|
||||
if progress:
|
||||
return "\n\n" + "\n".join(
|
||||
["Status: %s" % progress._status,
|
||||
"Iteration: %d" % progress._ram._iterations,
|
||||
"Throttle: %02d%%" % progress._throttle_pcent,
|
||||
"Dirty rate: %dMB/s" % (progress._ram._dirty_rate_pps * 4 / 1024.0)])
|
||||
else:
|
||||
return "\n\n" + "\n".join(
|
||||
["Status: %s" % "none",
|
||||
"Iteration: %d" % 0])
|
||||
|
||||
def _find_start_time(self, report):
|
||||
startqemu = report._qemu_timings._records[0]._timestamp
|
||||
startguest = report._guest_timings._records[0]._timestamp
|
||||
if startqemu < startguest:
|
||||
return startqemu
|
||||
else:
|
||||
return stasrtguest
|
||||
|
||||
def _get_guest_max_value(self, report):
|
||||
maxvalue = 0
|
||||
for record in report._guest_timings._records:
|
||||
if record._value > maxvalue:
|
||||
maxvalue = record._value
|
||||
return maxvalue
|
||||
|
||||
def _get_qemu_max_value(self, report):
|
||||
maxvalue = 0
|
||||
oldvalue = None
|
||||
oldtime = None
|
||||
for record in report._qemu_timings._records:
|
||||
if oldvalue is not None:
|
||||
cpudelta = (record._value - oldvalue) / 1000.0
|
||||
timedelta = record._timestamp - oldtime
|
||||
if timedelta == 0:
|
||||
continue
|
||||
util = cpudelta / timedelta * 100.0
|
||||
else:
|
||||
util = 0
|
||||
oldvalue = record._value
|
||||
oldtime = record._timestamp
|
||||
|
||||
if util > maxvalue:
|
||||
maxvalue = util
|
||||
return maxvalue
|
||||
|
||||
def _get_total_guest_cpu_graph(self, report, starttime):
|
||||
xaxis = []
|
||||
yaxis = []
|
||||
labels = []
|
||||
progress_idx = -1
|
||||
for record in report._guest_timings._records:
|
||||
while ((progress_idx + 1) < len(report._progress_history) and
|
||||
report._progress_history[progress_idx + 1]._now < record._timestamp):
|
||||
progress_idx = progress_idx + 1
|
||||
|
||||
if progress_idx >= 0:
|
||||
progress = report._progress_history[progress_idx]
|
||||
else:
|
||||
progress = None
|
||||
|
||||
xaxis.append(record._timestamp - starttime)
|
||||
yaxis.append(record._value)
|
||||
labels.append(self._get_progress_label(progress))
|
||||
|
||||
from plotly import graph_objs as go
|
||||
return go.Scatter(x=xaxis,
|
||||
y=yaxis,
|
||||
name="Guest PIDs: %s" % report._scenario._name,
|
||||
mode='lines',
|
||||
line={
|
||||
"dash": "solid",
|
||||
"color": self._next_color(),
|
||||
"shape": "linear",
|
||||
"width": 1
|
||||
},
|
||||
text=labels)
|
||||
|
||||
def _get_split_guest_cpu_graphs(self, report, starttime):
|
||||
threads = {}
|
||||
for record in report._guest_timings._records:
|
||||
if record._tid in threads:
|
||||
continue
|
||||
threads[record._tid] = {
|
||||
"xaxis": [],
|
||||
"yaxis": [],
|
||||
"labels": [],
|
||||
}
|
||||
|
||||
progress_idx = -1
|
||||
for record in report._guest_timings._records:
|
||||
while ((progress_idx + 1) < len(report._progress_history) and
|
||||
report._progress_history[progress_idx + 1]._now < record._timestamp):
|
||||
progress_idx = progress_idx + 1
|
||||
|
||||
if progress_idx >= 0:
|
||||
progress = report._progress_history[progress_idx]
|
||||
else:
|
||||
progress = None
|
||||
|
||||
threads[record._tid]["xaxis"].append(record._timestamp - starttime)
|
||||
threads[record._tid]["yaxis"].append(record._value)
|
||||
threads[record._tid]["labels"].append(self._get_progress_label(progress))
|
||||
|
||||
|
||||
graphs = []
|
||||
from plotly import graph_objs as go
|
||||
for tid in threads.keys():
|
||||
graphs.append(
|
||||
go.Scatter(x=threads[tid]["xaxis"],
|
||||
y=threads[tid]["yaxis"],
|
||||
name="PID %s: %s" % (tid, report._scenario._name),
|
||||
mode="lines",
|
||||
line={
|
||||
"dash": "solid",
|
||||
"color": self._next_color(),
|
||||
"shape": "linear",
|
||||
"width": 1
|
||||
},
|
||||
text=threads[tid]["labels"]))
|
||||
return graphs
|
||||
|
||||
def _get_migration_iters_graph(self, report, starttime):
|
||||
xaxis = []
|
||||
yaxis = []
|
||||
labels = []
|
||||
for progress in report._progress_history:
|
||||
xaxis.append(progress._now - starttime)
|
||||
yaxis.append(0)
|
||||
labels.append(self._get_progress_label(progress))
|
||||
|
||||
from plotly import graph_objs as go
|
||||
return go.Scatter(x=xaxis,
|
||||
y=yaxis,
|
||||
text=labels,
|
||||
name="Migration iterations",
|
||||
mode="markers",
|
||||
marker={
|
||||
"color": self._next_color(),
|
||||
"symbol": "star",
|
||||
"size": 5
|
||||
})
|
||||
|
||||
def _get_qemu_cpu_graph(self, report, starttime):
|
||||
xaxis = []
|
||||
yaxis = []
|
||||
labels = []
|
||||
progress_idx = -1
|
||||
|
||||
first = report._qemu_timings._records[0]
|
||||
abstimestamps = [first._timestamp]
|
||||
absvalues = [first._value]
|
||||
|
||||
for record in report._qemu_timings._records[1:]:
|
||||
while ((progress_idx + 1) < len(report._progress_history) and
|
||||
report._progress_history[progress_idx + 1]._now < record._timestamp):
|
||||
progress_idx = progress_idx + 1
|
||||
|
||||
if progress_idx >= 0:
|
||||
progress = report._progress_history[progress_idx]
|
||||
else:
|
||||
progress = None
|
||||
|
||||
oldvalue = absvalues[-1]
|
||||
oldtime = abstimestamps[-1]
|
||||
|
||||
cpudelta = (record._value - oldvalue) / 1000.0
|
||||
timedelta = record._timestamp - oldtime
|
||||
if timedelta == 0:
|
||||
continue
|
||||
util = cpudelta / timedelta * 100.0
|
||||
|
||||
abstimestamps.append(record._timestamp)
|
||||
absvalues.append(record._value)
|
||||
|
||||
xaxis.append(record._timestamp - starttime)
|
||||
yaxis.append(util)
|
||||
labels.append(self._get_progress_label(progress))
|
||||
|
||||
from plotly import graph_objs as go
|
||||
return go.Scatter(x=xaxis,
|
||||
y=yaxis,
|
||||
yaxis="y2",
|
||||
name="QEMU: %s" % report._scenario._name,
|
||||
mode='lines',
|
||||
line={
|
||||
"dash": "solid",
|
||||
"color": self._next_color(),
|
||||
"shape": "linear",
|
||||
"width": 1
|
||||
},
|
||||
text=labels)
|
||||
|
||||
def _get_vcpu_cpu_graphs(self, report, starttime):
|
||||
threads = {}
|
||||
for record in report._vcpu_timings._records:
|
||||
if record._tid in threads:
|
||||
continue
|
||||
threads[record._tid] = {
|
||||
"xaxis": [],
|
||||
"yaxis": [],
|
||||
"labels": [],
|
||||
"absvalue": [record._value],
|
||||
"abstime": [record._timestamp],
|
||||
}
|
||||
|
||||
progress_idx = -1
|
||||
for record in report._vcpu_timings._records:
|
||||
while ((progress_idx + 1) < len(report._progress_history) and
|
||||
report._progress_history[progress_idx + 1]._now < record._timestamp):
|
||||
progress_idx = progress_idx + 1
|
||||
|
||||
if progress_idx >= 0:
|
||||
progress = report._progress_history[progress_idx]
|
||||
else:
|
||||
progress = None
|
||||
|
||||
oldvalue = threads[record._tid]["absvalue"][-1]
|
||||
oldtime = threads[record._tid]["abstime"][-1]
|
||||
|
||||
cpudelta = (record._value - oldvalue) / 1000.0
|
||||
timedelta = record._timestamp - oldtime
|
||||
if timedelta == 0:
|
||||
continue
|
||||
util = cpudelta / timedelta * 100.0
|
||||
if util > 100:
|
||||
util = 100
|
||||
|
||||
threads[record._tid]["absvalue"].append(record._value)
|
||||
threads[record._tid]["abstime"].append(record._timestamp)
|
||||
|
||||
threads[record._tid]["xaxis"].append(record._timestamp - starttime)
|
||||
threads[record._tid]["yaxis"].append(util)
|
||||
threads[record._tid]["labels"].append(self._get_progress_label(progress))
|
||||
|
||||
|
||||
graphs = []
|
||||
from plotly import graph_objs as go
|
||||
for tid in threads.keys():
|
||||
graphs.append(
|
||||
go.Scatter(x=threads[tid]["xaxis"],
|
||||
y=threads[tid]["yaxis"],
|
||||
yaxis="y2",
|
||||
name="VCPU %s: %s" % (tid, report._scenario._name),
|
||||
mode="lines",
|
||||
line={
|
||||
"dash": "solid",
|
||||
"color": self._next_color(),
|
||||
"shape": "linear",
|
||||
"width": 1
|
||||
},
|
||||
text=threads[tid]["labels"]))
|
||||
return graphs
|
||||
|
||||
def _generate_chart_report(self, report):
|
||||
graphs = []
|
||||
starttime = self._find_start_time(report)
|
||||
if self._total_guest_cpu:
|
||||
graphs.append(self._get_total_guest_cpu_graph(report, starttime))
|
||||
if self._split_guest_cpu:
|
||||
graphs.extend(self._get_split_guest_cpu_graphs(report, starttime))
|
||||
if self._qemu_cpu:
|
||||
graphs.append(self._get_qemu_cpu_graph(report, starttime))
|
||||
if self._vcpu_cpu:
|
||||
graphs.extend(self._get_vcpu_cpu_graphs(report, starttime))
|
||||
if self._migration_iters:
|
||||
graphs.append(self._get_migration_iters_graph(report, starttime))
|
||||
return graphs
|
||||
|
||||
def _generate_annotation(self, starttime, progress):
|
||||
return {
|
||||
"text": progress._status,
|
||||
"x": progress._now - starttime,
|
||||
"y": 10,
|
||||
}
|
||||
|
||||
def _generate_annotations(self, report):
|
||||
starttime = self._find_start_time(report)
|
||||
annotations = {}
|
||||
started = False
|
||||
for progress in report._progress_history:
|
||||
if progress._status == "setup":
|
||||
continue
|
||||
if progress._status not in annotations:
|
||||
annotations[progress._status] = self._generate_annotation(starttime, progress)
|
||||
|
||||
return annotations.values()
|
||||
|
||||
def _generate_chart(self):
|
||||
from plotly.offline import plot
|
||||
from plotly import graph_objs as go
|
||||
|
||||
graphs = []
|
||||
yaxismax = 0
|
||||
yaxismax2 = 0
|
||||
for report in self._reports:
|
||||
graphs.extend(self._generate_chart_report(report))
|
||||
|
||||
maxvalue = self._get_guest_max_value(report)
|
||||
if maxvalue > yaxismax:
|
||||
yaxismax = maxvalue
|
||||
|
||||
maxvalue = self._get_qemu_max_value(report)
|
||||
if maxvalue > yaxismax2:
|
||||
yaxismax2 = maxvalue
|
||||
|
||||
yaxismax += 100
|
||||
if not self._qemu_cpu:
|
||||
yaxismax2 = 110
|
||||
yaxismax2 += 10
|
||||
|
||||
annotations = []
|
||||
if self._migration_iters:
|
||||
for report in self._reports:
|
||||
annotations.extend(self._generate_annotations(report))
|
||||
|
||||
layout = go.Layout(title="Migration comparison",
|
||||
xaxis={
|
||||
"title": "Wallclock time (secs)",
|
||||
"showgrid": False,
|
||||
},
|
||||
yaxis={
|
||||
"title": "Memory update speed (ms/GB)",
|
||||
"showgrid": False,
|
||||
"range": [0, yaxismax],
|
||||
},
|
||||
yaxis2={
|
||||
"title": "Hostutilization (%)",
|
||||
"overlaying": "y",
|
||||
"side": "right",
|
||||
"range": [0, yaxismax2],
|
||||
"showgrid": False,
|
||||
},
|
||||
annotations=annotations)
|
||||
|
||||
figure = go.Figure(data=graphs, layout=layout)
|
||||
|
||||
return plot(figure,
|
||||
show_link=False,
|
||||
include_plotlyjs=False,
|
||||
output_type="div")
|
||||
|
||||
|
||||
def _generate_report(self):
|
||||
pieces = []
|
||||
for report in self._reports:
|
||||
pieces.append("""
|
||||
<h3>Report %s</h3>
|
||||
<table>
|
||||
""" % report._scenario._name)
|
||||
|
||||
pieces.append("""
|
||||
<tr class="subhead">
|
||||
<th colspan="2">Test config</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Emulator:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Kernel:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Ramdisk:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Transport:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Host:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
""" % (report._binary, report._kernel,
|
||||
report._initrd, report._transport, report._dst_host))
|
||||
|
||||
hardware = report._hardware
|
||||
pieces.append("""
|
||||
<tr class="subhead">
|
||||
<th colspan="2">Hardware config</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>CPUs:</th>
|
||||
<td>%d</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>RAM:</th>
|
||||
<td>%d GB</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Source CPU bind:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Source RAM bind:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Dest CPU bind:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Dest RAM bind:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Preallocate RAM:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Locked RAM:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Huge pages:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
""" % (hardware._cpus, hardware._mem,
|
||||
",".join(hardware._src_cpu_bind),
|
||||
",".join(hardware._src_mem_bind),
|
||||
",".join(hardware._dst_cpu_bind),
|
||||
",".join(hardware._dst_mem_bind),
|
||||
"yes" if hardware._prealloc_pages else "no",
|
||||
"yes" if hardware._locked_pages else "no",
|
||||
"yes" if hardware._huge_pages else "no"))
|
||||
|
||||
scenario = report._scenario
|
||||
pieces.append("""
|
||||
<tr class="subhead">
|
||||
<th colspan="2">Scenario config</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Max downtime:</th>
|
||||
<td>%d milli-sec</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Max bandwidth:</th>
|
||||
<td>%d MB/sec</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Max iters:</th>
|
||||
<td>%d</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Max time:</th>
|
||||
<td>%d secs</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Pause:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Pause iters:</th>
|
||||
<td>%d</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Post-copy:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Post-copy iters:</th>
|
||||
<td>%d</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Auto-converge:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>Auto-converge iters:</th>
|
||||
<td>%d</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>MT compression:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>MT compression threads:</th>
|
||||
<td>%d</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>XBZRLE compression:</th>
|
||||
<td>%s</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<th>XBZRLE compression cache:</th>
|
||||
<td>%d%% of RAM</td>
|
||||
</tr>
|
||||
""" % (scenario._downtime, scenario._bandwidth,
|
||||
scenario._max_iters, scenario._max_time,
|
||||
"yes" if scenario._pause else "no", scenario._pause_iters,
|
||||
"yes" if scenario._post_copy else "no", scenario._post_copy_iters,
|
||||
"yes" if scenario._auto_converge else "no", scenario._auto_converge_step,
|
||||
"yes" if scenario._compression_mt else "no", scenario._compression_mt_threads,
|
||||
"yes" if scenario._compression_xbzrle else "no", scenario._compression_xbzrle_cache))
|
||||
|
||||
pieces.append("""
|
||||
</table>
|
||||
""")
|
||||
|
||||
return "\n".join(pieces)
|
||||
|
||||
def _generate_style(self):
|
||||
return """
|
||||
#report table tr th {
|
||||
text-align: right;
|
||||
}
|
||||
#report table tr td {
|
||||
text-align: left;
|
||||
}
|
||||
#report table tr.subhead th {
|
||||
background: rgb(192, 192, 192);
|
||||
text-align: center;
|
||||
}
|
||||
|
||||
"""
|
||||
|
||||
def generate_html(self, fh):
|
||||
print >>fh, """<html>
|
||||
<head>
|
||||
<script type="text/javascript" src="plotly.min.js">
|
||||
</script>
|
||||
<style type="text/css">
|
||||
%s
|
||||
</style>
|
||||
<title>Migration report</title>
|
||||
</head>
|
||||
<body>
|
||||
<h1>Migration report</h1>
|
||||
<h2>Chart summary</h2>
|
||||
<div id="chart">
|
||||
""" % self._generate_style()
|
||||
print >>fh, self._generate_chart()
|
||||
print >>fh, """
|
||||
</div>
|
||||
<h2>Report details</h2>
|
||||
<div id="report">
|
||||
"""
|
||||
print >>fh, self._generate_report()
|
||||
print >>fh, """
|
||||
</div>
|
||||
</body>
|
||||
</html>
|
||||
"""
|
||||
|
||||
def generate(self, filename):
|
||||
if filename is None:
|
||||
self.generate_html(sys.stdout)
|
||||
else:
|
||||
with open(filename, "w") as fh:
|
||||
self.generate_html(fh)
|
117
tests/migration/guestperf/progress.py
Normal file
117
tests/migration/guestperf/progress.py
Normal file
@ -0,0 +1,117 @@
|
||||
#
|
||||
# Migration test migration operation progress
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
|
||||
class ProgressStats(object):
|
||||
|
||||
def __init__(self,
|
||||
transferred_bytes,
|
||||
remaining_bytes,
|
||||
total_bytes,
|
||||
duplicate_pages,
|
||||
skipped_pages,
|
||||
normal_pages,
|
||||
normal_bytes,
|
||||
dirty_rate_pps,
|
||||
transfer_rate_mbs,
|
||||
iterations):
|
||||
self._transferred_bytes = transferred_bytes
|
||||
self._remaining_bytes = remaining_bytes
|
||||
self._total_bytes = total_bytes
|
||||
self._duplicate_pages = duplicate_pages
|
||||
self._skipped_pages = skipped_pages
|
||||
self._normal_pages = normal_pages
|
||||
self._normal_bytes = normal_bytes
|
||||
self._dirty_rate_pps = dirty_rate_pps
|
||||
self._transfer_rate_mbs = transfer_rate_mbs
|
||||
self._iterations = iterations
|
||||
|
||||
def serialize(self):
|
||||
return {
|
||||
"transferred_bytes": self._transferred_bytes,
|
||||
"remaining_bytes": self._remaining_bytes,
|
||||
"total_bytes": self._total_bytes,
|
||||
"duplicate_pages": self._duplicate_pages,
|
||||
"skipped_pages": self._skipped_pages,
|
||||
"normal_pages": self._normal_pages,
|
||||
"normal_bytes": self._normal_bytes,
|
||||
"dirty_rate_pps": self._dirty_rate_pps,
|
||||
"transfer_rate_mbs": self._transfer_rate_mbs,
|
||||
"iterations": self._iterations,
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def deserialize(cls, data):
|
||||
return cls(
|
||||
data["transferred_bytes"],
|
||||
data["remaining_bytes"],
|
||||
data["total_bytes"],
|
||||
data["duplicate_pages"],
|
||||
data["skipped_pages"],
|
||||
data["normal_pages"],
|
||||
data["normal_bytes"],
|
||||
data["dirty_rate_pps"],
|
||||
data["transfer_rate_mbs"],
|
||||
data["iterations"])
|
||||
|
||||
|
||||
class Progress(object):
|
||||
|
||||
def __init__(self,
|
||||
status,
|
||||
ram,
|
||||
now,
|
||||
duration,
|
||||
downtime,
|
||||
downtime_expected,
|
||||
setup_time,
|
||||
throttle_pcent):
|
||||
|
||||
self._status = status
|
||||
self._ram = ram
|
||||
self._now = now
|
||||
self._duration = duration
|
||||
self._downtime = downtime
|
||||
self._downtime_expected = downtime_expected
|
||||
self._setup_time = setup_time
|
||||
self._throttle_pcent = throttle_pcent
|
||||
|
||||
def serialize(self):
|
||||
return {
|
||||
"status": self._status,
|
||||
"ram": self._ram.serialize(),
|
||||
"now": self._now,
|
||||
"duration": self._duration,
|
||||
"downtime": self._downtime,
|
||||
"downtime_expected": self._downtime_expected,
|
||||
"setup_time": self._setup_time,
|
||||
"throttle_pcent": self._throttle_pcent,
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def deserialize(cls, data):
|
||||
return cls(
|
||||
data["status"],
|
||||
ProgressStats.deserialize(data["ram"]),
|
||||
data["now"],
|
||||
data["duration"],
|
||||
data["downtime"],
|
||||
data["downtime_expected"],
|
||||
data["setup_time"],
|
||||
data["throttle_pcent"])
|
98
tests/migration/guestperf/report.py
Normal file
98
tests/migration/guestperf/report.py
Normal file
@ -0,0 +1,98 @@
|
||||
#
|
||||
# Migration test output result reporting
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
import json
|
||||
|
||||
from guestperf.hardware import Hardware
|
||||
from guestperf.scenario import Scenario
|
||||
from guestperf.progress import Progress
|
||||
from guestperf.timings import Timings
|
||||
|
||||
class Report(object):
|
||||
|
||||
def __init__(self,
|
||||
hardware,
|
||||
scenario,
|
||||
progress_history,
|
||||
guest_timings,
|
||||
qemu_timings,
|
||||
vcpu_timings,
|
||||
binary,
|
||||
dst_host,
|
||||
kernel,
|
||||
initrd,
|
||||
transport,
|
||||
sleep):
|
||||
|
||||
self._hardware = hardware
|
||||
self._scenario = scenario
|
||||
self._progress_history = progress_history
|
||||
self._guest_timings = guest_timings
|
||||
self._qemu_timings = qemu_timings
|
||||
self._vcpu_timings = vcpu_timings
|
||||
self._binary = binary
|
||||
self._dst_host = dst_host
|
||||
self._kernel = kernel
|
||||
self._initrd = initrd
|
||||
self._transport = transport
|
||||
self._sleep = sleep
|
||||
|
||||
def serialize(self):
|
||||
return {
|
||||
"hardware": self._hardware.serialize(),
|
||||
"scenario": self._scenario.serialize(),
|
||||
"progress_history": [progress.serialize() for progress in self._progress_history],
|
||||
"guest_timings": self._guest_timings.serialize(),
|
||||
"qemu_timings": self._qemu_timings.serialize(),
|
||||
"vcpu_timings": self._vcpu_timings.serialize(),
|
||||
"binary": self._binary,
|
||||
"dst_host": self._dst_host,
|
||||
"kernel": self._kernel,
|
||||
"initrd": self._initrd,
|
||||
"transport": self._transport,
|
||||
"sleep": self._sleep,
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def deserialize(cls, data):
|
||||
return cls(
|
||||
Hardware.deserialize(data["hardware"]),
|
||||
Scenario.deserialize(data["scenario"]),
|
||||
[Progress.deserialize(record) for record in data["progress_history"]],
|
||||
Timings.deserialize(data["guest_timings"]),
|
||||
Timings.deserialize(data["qemu_timings"]),
|
||||
Timings.deserialize(data["vcpu_timings"]),
|
||||
data["binary"],
|
||||
data["dst_host"],
|
||||
data["kernel"],
|
||||
data["initrd"],
|
||||
data["transport"],
|
||||
data["sleep"])
|
||||
|
||||
def to_json(self):
|
||||
return json.dumps(self.serialize(), indent=4)
|
||||
|
||||
@classmethod
|
||||
def from_json(cls, data):
|
||||
return cls.deserialize(json.loads(data))
|
||||
|
||||
@classmethod
|
||||
def from_json_file(cls, filename):
|
||||
with open(filename, "r") as fh:
|
||||
return cls.deserialize(json.load(fh))
|
95
tests/migration/guestperf/scenario.py
Normal file
95
tests/migration/guestperf/scenario.py
Normal file
@ -0,0 +1,95 @@
|
||||
#
|
||||
# Migration test scenario parameter description
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
|
||||
class Scenario(object):
|
||||
|
||||
def __init__(self, name,
|
||||
downtime=500,
|
||||
bandwidth=125000, # 1000 gig-e, effectively unlimited
|
||||
max_iters=30,
|
||||
max_time=300,
|
||||
pause=False, pause_iters=5,
|
||||
post_copy=False, post_copy_iters=5,
|
||||
auto_converge=False, auto_converge_step=10,
|
||||
compression_mt=False, compression_mt_threads=1,
|
||||
compression_xbzrle=False, compression_xbzrle_cache=10):
|
||||
|
||||
self._name = name
|
||||
|
||||
# General migration tunables
|
||||
self._downtime = downtime # milliseconds
|
||||
self._bandwidth = bandwidth # MiB per second
|
||||
self._max_iters = max_iters
|
||||
self._max_time = max_time # seconds
|
||||
|
||||
|
||||
# Strategies for ensuring completion
|
||||
self._pause = pause
|
||||
self._pause_iters = pause_iters
|
||||
|
||||
self._post_copy = post_copy
|
||||
self._post_copy_iters = post_copy_iters
|
||||
|
||||
self._auto_converge = auto_converge
|
||||
self._auto_converge_step = auto_converge_step # percentage CPU time
|
||||
|
||||
self._compression_mt = compression_mt
|
||||
self._compression_mt_threads = compression_mt_threads
|
||||
|
||||
self._compression_xbzrle = compression_xbzrle
|
||||
self._compression_xbzrle_cache = compression_xbzrle_cache # percentage of guest RAM
|
||||
|
||||
def serialize(self):
|
||||
return {
|
||||
"name": self._name,
|
||||
"downtime": self._downtime,
|
||||
"bandwidth": self._bandwidth,
|
||||
"max_iters": self._max_iters,
|
||||
"max_time": self._max_time,
|
||||
"pause": self._pause,
|
||||
"pause_iters": self._pause_iters,
|
||||
"post_copy": self._post_copy,
|
||||
"post_copy_iters": self._post_copy_iters,
|
||||
"auto_converge": self._auto_converge,
|
||||
"auto_converge_step": self._auto_converge_step,
|
||||
"compression_mt": self._compression_mt,
|
||||
"compression_mt_threads": self._compression_mt_threads,
|
||||
"compression_xbzrle": self._compression_xbzrle,
|
||||
"compression_xbzrle_cache": self._compression_xbzrle_cache,
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def deserialize(cls, data):
|
||||
return cls(
|
||||
data["name"],
|
||||
data["downtime"],
|
||||
data["bandwidth"],
|
||||
data["max_iters"],
|
||||
data["max_time"],
|
||||
data["pause"],
|
||||
data["pause_iters"],
|
||||
data["post_copy"],
|
||||
data["post_copy_iters"],
|
||||
data["auto_converge"],
|
||||
data["auto_converge_step"],
|
||||
data["compression_mt"],
|
||||
data["compression_mt_threads"],
|
||||
data["compression_xbzrle"],
|
||||
data["compression_xbzrle_cache"])
|
255
tests/migration/guestperf/shell.py
Normal file
255
tests/migration/guestperf/shell.py
Normal file
@ -0,0 +1,255 @@
|
||||
#
|
||||
# Migration test command line shell integration
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
|
||||
import argparse
|
||||
import fnmatch
|
||||
import os
|
||||
import os.path
|
||||
import platform
|
||||
import sys
|
||||
|
||||
from guestperf.hardware import Hardware
|
||||
from guestperf.engine import Engine
|
||||
from guestperf.scenario import Scenario
|
||||
from guestperf.comparison import COMPARISONS
|
||||
from guestperf.plot import Plot
|
||||
from guestperf.report import Report
|
||||
|
||||
|
||||
class BaseShell(object):
|
||||
|
||||
def __init__(self):
|
||||
parser = argparse.ArgumentParser(description="Migration Test Tool")
|
||||
|
||||
# Test args
|
||||
parser.add_argument("--debug", dest="debug", default=False, action="store_true")
|
||||
parser.add_argument("--verbose", dest="verbose", default=False, action="store_true")
|
||||
parser.add_argument("--sleep", dest="sleep", default=15, type=int)
|
||||
parser.add_argument("--binary", dest="binary", default="/usr/bin/qemu-system-x86_64")
|
||||
parser.add_argument("--dst-host", dest="dst_host", default="localhost")
|
||||
parser.add_argument("--kernel", dest="kernel", default="/boot/vmlinuz-%s" % platform.release())
|
||||
parser.add_argument("--initrd", dest="initrd", default="tests/migration/initrd-stress.img")
|
||||
parser.add_argument("--transport", dest="transport", default="unix")
|
||||
|
||||
|
||||
# Hardware args
|
||||
parser.add_argument("--cpus", dest="cpus", default=1, type=int)
|
||||
parser.add_argument("--mem", dest="mem", default=1, type=int)
|
||||
parser.add_argument("--src-cpu-bind", dest="src_cpu_bind", default="")
|
||||
parser.add_argument("--src-mem-bind", dest="src_mem_bind", default="")
|
||||
parser.add_argument("--dst-cpu-bind", dest="dst_cpu_bind", default="")
|
||||
parser.add_argument("--dst-mem-bind", dest="dst_mem_bind", default="")
|
||||
parser.add_argument("--prealloc-pages", dest="prealloc_pages", default=False)
|
||||
parser.add_argument("--huge-pages", dest="huge_pages", default=False)
|
||||
parser.add_argument("--locked-pages", dest="locked_pages", default=False)
|
||||
|
||||
self._parser = parser
|
||||
|
||||
def get_engine(self, args):
|
||||
return Engine(binary=args.binary,
|
||||
dst_host=args.dst_host,
|
||||
kernel=args.kernel,
|
||||
initrd=args.initrd,
|
||||
transport=args.transport,
|
||||
sleep=args.sleep,
|
||||
debug=args.debug,
|
||||
verbose=args.verbose)
|
||||
|
||||
def get_hardware(self, args):
|
||||
def split_map(value):
|
||||
if value == "":
|
||||
return []
|
||||
return value.split(",")
|
||||
|
||||
return Hardware(cpus=args.cpus,
|
||||
mem=args.mem,
|
||||
|
||||
src_cpu_bind=split_map(args.src_cpu_bind),
|
||||
src_mem_bind=split_map(args.src_mem_bind),
|
||||
dst_cpu_bind=split_map(args.dst_cpu_bind),
|
||||
dst_mem_bind=split_map(args.dst_mem_bind),
|
||||
|
||||
locked_pages=args.locked_pages,
|
||||
huge_pages=args.huge_pages,
|
||||
prealloc_pages=args.prealloc_pages)
|
||||
|
||||
|
||||
class Shell(BaseShell):
|
||||
|
||||
def __init__(self):
|
||||
super(Shell, self).__init__()
|
||||
|
||||
parser = self._parser
|
||||
|
||||
parser.add_argument("--output", dest="output", default=None)
|
||||
|
||||
# Scenario args
|
||||
parser.add_argument("--max-iters", dest="max_iters", default=30, type=int)
|
||||
parser.add_argument("--max-time", dest="max_time", default=300, type=int)
|
||||
parser.add_argument("--bandwidth", dest="bandwidth", default=125000, type=int)
|
||||
parser.add_argument("--downtime", dest="downtime", default=500, type=int)
|
||||
|
||||
parser.add_argument("--pause", dest="pause", default=False, action="store_true")
|
||||
parser.add_argument("--pause-iters", dest="pause_iters", default=5, type=int)
|
||||
|
||||
parser.add_argument("--post-copy", dest="post_copy", default=False, action="store_true")
|
||||
parser.add_argument("--post-copy-iters", dest="post_copy_iters", default=5, type=int)
|
||||
|
||||
parser.add_argument("--auto-converge", dest="auto_converge", default=False, action="store_true")
|
||||
parser.add_argument("--auto-converge-step", dest="auto_converge_step", default=10, type=int)
|
||||
|
||||
parser.add_argument("--compression-mt", dest="compression_mt", default=False, action="store_true")
|
||||
parser.add_argument("--compression-mt-threads", dest="compression_mt_threads", default=1, type=int)
|
||||
|
||||
parser.add_argument("--compression-xbzrle", dest="compression_xbzrle", default=False, action="store_true")
|
||||
parser.add_argument("--compression-xbzrle-cache", dest="compression_xbzrle_cache", default=10, type=int)
|
||||
|
||||
def get_scenario(self, args):
|
||||
return Scenario(name="perfreport",
|
||||
downtime=args.downtime,
|
||||
bandwidth=args.bandwidth,
|
||||
max_iters=args.max_iters,
|
||||
max_time=args.max_time,
|
||||
|
||||
pause=args.pause,
|
||||
pause_iters=args.pause_iters,
|
||||
|
||||
post_copy=args.post_copy,
|
||||
post_copy_iters=args.post_copy_iters,
|
||||
|
||||
auto_converge=args.auto_converge,
|
||||
auto_converge_step=args.auto_converge_step,
|
||||
|
||||
compression_mt=args.compression_mt,
|
||||
compression_mt_threads=args.compression_mt_threads,
|
||||
|
||||
compression_xbzrle=args.compression_xbzrle,
|
||||
compression_xbzrle_cache=args.compression_xbzrle_cache)
|
||||
|
||||
def run(self, argv):
|
||||
args = self._parser.parse_args(argv)
|
||||
|
||||
engine = self.get_engine(args)
|
||||
hardware = self.get_hardware(args)
|
||||
scenario = self.get_scenario(args)
|
||||
|
||||
try:
|
||||
report = engine.run(hardware, scenario)
|
||||
if args.output is None:
|
||||
print report.to_json()
|
||||
else:
|
||||
with open(args.output, "w") as fh:
|
||||
print >>fh, report.to_json()
|
||||
return 0
|
||||
except Exception as e:
|
||||
print >>sys.stderr, "Error: %s" % str(e)
|
||||
if args.debug:
|
||||
raise
|
||||
return 1
|
||||
|
||||
|
||||
class BatchShell(BaseShell):
|
||||
|
||||
def __init__(self):
|
||||
super(BatchShell, self).__init__()
|
||||
|
||||
parser = self._parser
|
||||
|
||||
parser.add_argument("--filter", dest="filter", default="*")
|
||||
parser.add_argument("--output", dest="output", default=os.getcwd())
|
||||
|
||||
def run(self, argv):
|
||||
args = self._parser.parse_args(argv)
|
||||
|
||||
engine = self.get_engine(args)
|
||||
hardware = self.get_hardware(args)
|
||||
|
||||
try:
|
||||
for comparison in COMPARISONS:
|
||||
compdir = os.path.join(args.output, comparison._name)
|
||||
for scenario in comparison._scenarios:
|
||||
name = os.path.join(comparison._name, scenario._name)
|
||||
if not fnmatch.fnmatch(name, args.filter):
|
||||
if args.verbose:
|
||||
print "Skipping %s" % name
|
||||
continue
|
||||
|
||||
if args.verbose:
|
||||
print "Running %s" % name
|
||||
|
||||
dirname = os.path.join(args.output, comparison._name)
|
||||
filename = os.path.join(dirname, scenario._name + ".json")
|
||||
if not os.path.exists(dirname):
|
||||
os.makedirs(dirname)
|
||||
report = engine.run(hardware, scenario)
|
||||
with open(filename, "w") as fh:
|
||||
print >>fh, report.to_json()
|
||||
except Exception as e:
|
||||
print >>sys.stderr, "Error: %s" % str(e)
|
||||
if args.debug:
|
||||
raise
|
||||
|
||||
|
||||
class PlotShell(object):
|
||||
|
||||
def __init__(self):
|
||||
super(PlotShell, self).__init__()
|
||||
|
||||
self._parser = argparse.ArgumentParser(description="Migration Test Tool")
|
||||
|
||||
self._parser.add_argument("--output", dest="output", default=None)
|
||||
|
||||
self._parser.add_argument("--debug", dest="debug", default=False, action="store_true")
|
||||
self._parser.add_argument("--verbose", dest="verbose", default=False, action="store_true")
|
||||
|
||||
self._parser.add_argument("--migration-iters", dest="migration_iters", default=False, action="store_true")
|
||||
self._parser.add_argument("--total-guest-cpu", dest="total_guest_cpu", default=False, action="store_true")
|
||||
self._parser.add_argument("--split-guest-cpu", dest="split_guest_cpu", default=False, action="store_true")
|
||||
self._parser.add_argument("--qemu-cpu", dest="qemu_cpu", default=False, action="store_true")
|
||||
self._parser.add_argument("--vcpu-cpu", dest="vcpu_cpu", default=False, action="store_true")
|
||||
|
||||
self._parser.add_argument("reports", nargs='*')
|
||||
|
||||
def run(self, argv):
|
||||
args = self._parser.parse_args(argv)
|
||||
|
||||
if len(args.reports) == 0:
|
||||
print >>sys.stderr, "At least one report required"
|
||||
return 1
|
||||
|
||||
if not (args.qemu_cpu or
|
||||
args.vcpu_cpu or
|
||||
args.total_guest_cpu or
|
||||
args.split_guest_cpu):
|
||||
print >>sys.stderr, "At least one chart type is required"
|
||||
return 1
|
||||
|
||||
reports = []
|
||||
for report in args.reports:
|
||||
reports.append(Report.from_json_file(report))
|
||||
|
||||
plot = Plot(reports,
|
||||
args.migration_iters,
|
||||
args.total_guest_cpu,
|
||||
args.split_guest_cpu,
|
||||
args.qemu_cpu,
|
||||
args.vcpu_cpu)
|
||||
|
||||
plot.generate(args.output)
|
55
tests/migration/guestperf/timings.py
Normal file
55
tests/migration/guestperf/timings.py
Normal file
@ -0,0 +1,55 @@
|
||||
#
|
||||
# Migration test timing records
|
||||
#
|
||||
# Copyright (c) 2016 Red Hat, Inc.
|
||||
#
|
||||
# This library is free software; you can redistribute it and/or
|
||||
# modify it under the terms of the GNU Lesser General Public
|
||||
# License as published by the Free Software Foundation; either
|
||||
# version 2 of the License, or (at your option) any later version.
|
||||
#
|
||||
# This library is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
# Lesser General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU Lesser General Public
|
||||
# License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
|
||||
class TimingRecord(object):
|
||||
|
||||
def __init__(self, tid, timestamp, value):
|
||||
|
||||
self._tid = tid
|
||||
self._timestamp = timestamp
|
||||
self._value = value
|
||||
|
||||
def serialize(self):
|
||||
return {
|
||||
"tid": self._tid,
|
||||
"timestamp": self._timestamp,
|
||||
"value": self._value
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def deserialize(cls, data):
|
||||
return cls(
|
||||
data["tid"],
|
||||
data["timestamp"],
|
||||
data["value"])
|
||||
|
||||
|
||||
class Timings(object):
|
||||
|
||||
def __init__(self, records):
|
||||
|
||||
self._records = records
|
||||
|
||||
def serialize(self):
|
||||
return [record.serialize() for record in self._records]
|
||||
|
||||
@classmethod
|
||||
def deserialize(cls, data):
|
||||
return Timings([TimingRecord.deserialize(record) for record in data])
|
367
tests/migration/stress.c
Normal file
367
tests/migration/stress.c
Normal file
@ -0,0 +1,367 @@
|
||||
/*
|
||||
* Migration stress workload
|
||||
*
|
||||
* Copyright (c) 2016 Red Hat, Inc.
|
||||
*
|
||||
* This library is free software; you can redistribute it and/or
|
||||
* modify it under the terms of the GNU Lesser General Public
|
||||
* License as published by the Free Software Foundation; either
|
||||
* version 2 of the License, or (at your option) any later version.
|
||||
*
|
||||
* This library is distributed in the hope that it will be useful,
|
||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
* Lesser General Public License for more details.
|
||||
*
|
||||
* You should have received a copy of the GNU Lesser General Public
|
||||
* License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include <stdio.h>
|
||||
#include <getopt.h>
|
||||
#include <string.h>
|
||||
#include <stdlib.h>
|
||||
#include <errno.h>
|
||||
#include <unistd.h>
|
||||
#include <sys/reboot.h>
|
||||
#include <sys/syscall.h>
|
||||
#include <linux/random.h>
|
||||
#include <sys/time.h>
|
||||
#include <pthread.h>
|
||||
#include <fcntl.h>
|
||||
#include <sys/mount.h>
|
||||
#include <sys/stat.h>
|
||||
#include <sys/mman.h>
|
||||
|
||||
const char *argv0;
|
||||
|
||||
#define PAGE_SIZE 4096
|
||||
|
||||
static int gettid(void)
|
||||
{
|
||||
return syscall(SYS_gettid);
|
||||
}
|
||||
|
||||
static __attribute__((noreturn)) void exit_failure(void)
|
||||
{
|
||||
if (getpid() == 1) {
|
||||
sync();
|
||||
reboot(RB_POWER_OFF);
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot reboot: %s\n",
|
||||
argv0, gettid(), strerror(errno));
|
||||
abort();
|
||||
} else {
|
||||
exit(1);
|
||||
}
|
||||
}
|
||||
|
||||
static __attribute__((noreturn)) void exit_success(void)
|
||||
{
|
||||
if (getpid() == 1) {
|
||||
sync();
|
||||
reboot(RB_POWER_OFF);
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot reboot: %s\n",
|
||||
argv0, gettid(), strerror(errno));
|
||||
abort();
|
||||
} else {
|
||||
exit(0);
|
||||
}
|
||||
}
|
||||
|
||||
static int get_command_arg_str(const char *name,
|
||||
char **val)
|
||||
{
|
||||
static char line[1024];
|
||||
FILE *fp = fopen("/proc/cmdline", "r");
|
||||
char *start, *end;
|
||||
|
||||
if (fp == NULL) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot open /proc/cmdline: %s\n",
|
||||
argv0, gettid(), strerror(errno));
|
||||
return -1;
|
||||
}
|
||||
|
||||
if (!fgets(line, sizeof line, fp)) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot read /proc/cmdline: %s\n",
|
||||
argv0, gettid(), strerror(errno));
|
||||
fclose(fp);
|
||||
return -1;
|
||||
}
|
||||
fclose(fp);
|
||||
|
||||
start = strstr(line, name);
|
||||
if (!start)
|
||||
return 0;
|
||||
|
||||
start += strlen(name);
|
||||
|
||||
if (*start != '=') {
|
||||
fprintf(stderr, "%s (%05d): ERROR: no value provided for '%s' in /proc/cmdline\n",
|
||||
argv0, gettid(), name);
|
||||
}
|
||||
start++;
|
||||
|
||||
end = strstr(start, " ");
|
||||
if (!end)
|
||||
end = strstr(start, "\n");
|
||||
|
||||
if (end == start) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: no value provided for '%s' in /proc/cmdline\n",
|
||||
argv0, gettid(), name);
|
||||
return -1;
|
||||
}
|
||||
|
||||
if (end)
|
||||
*val = strndup(start, end - start);
|
||||
else
|
||||
*val = strdup(start);
|
||||
return 1;
|
||||
}
|
||||
|
||||
|
||||
static int get_command_arg_ull(const char *name,
|
||||
unsigned long long *val)
|
||||
{
|
||||
char *valstr;
|
||||
char *end;
|
||||
|
||||
int ret = get_command_arg_str(name, &valstr);
|
||||
if (ret <= 0)
|
||||
return ret;
|
||||
|
||||
errno = 0;
|
||||
*val = strtoll(valstr, &end, 10);
|
||||
if (errno || *end) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot parse %s value %s\n",
|
||||
argv0, gettid(), name, valstr);
|
||||
free(valstr);
|
||||
return -1;
|
||||
}
|
||||
free(valstr);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
static int random_bytes(char *buf, size_t len)
|
||||
{
|
||||
int fd;
|
||||
|
||||
fd = open("/dev/urandom", O_RDONLY);
|
||||
if (fd < 0) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot open /dev/urandom: %s\n",
|
||||
argv0, gettid(), strerror(errno));
|
||||
return -1;
|
||||
}
|
||||
|
||||
if (read(fd, buf, len) != len) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot read /dev/urandom: %s\n",
|
||||
argv0, gettid(), strerror(errno));
|
||||
close(fd);
|
||||
return -1;
|
||||
}
|
||||
|
||||
close(fd);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
static unsigned long long now(void)
|
||||
{
|
||||
struct timeval tv;
|
||||
|
||||
gettimeofday(&tv, NULL);
|
||||
|
||||
return (tv.tv_sec * 1000ull) + (tv.tv_usec / 1000ull);
|
||||
}
|
||||
|
||||
static int stressone(unsigned long long ramsizeMB)
|
||||
{
|
||||
size_t pagesPerMB = 1024 * 1024 / PAGE_SIZE;
|
||||
char *ram = malloc(ramsizeMB * 1024 * 1024);
|
||||
char *ramptr;
|
||||
size_t i, j, k;
|
||||
char *data = malloc(PAGE_SIZE);
|
||||
char *dataptr;
|
||||
size_t nMB = 0;
|
||||
unsigned long long before, after;
|
||||
|
||||
if (!ram) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot allocate %llu MB of RAM: %s\n",
|
||||
argv0, gettid(), ramsizeMB, strerror(errno));
|
||||
return -1;
|
||||
}
|
||||
if (!data) {
|
||||
fprintf(stderr, "%s (%d): ERROR: cannot allocate %d bytes of RAM: %s\n",
|
||||
argv0, gettid(), PAGE_SIZE, strerror(errno));
|
||||
free(ram);
|
||||
return -1;
|
||||
}
|
||||
|
||||
/* We don't care about initial state, but we do want
|
||||
* to fault it all into RAM, otherwise the first iter
|
||||
* of the loop below will be quite slow. We cna't use
|
||||
* 0x0 as the byte as gcc optimizes that away into a
|
||||
* calloc instead :-) */
|
||||
memset(ram, 0xfe, ramsizeMB * 1024 * 1024);
|
||||
|
||||
if (random_bytes(data, PAGE_SIZE) < 0) {
|
||||
free(ram);
|
||||
free(data);
|
||||
return -1;
|
||||
}
|
||||
|
||||
before = now();
|
||||
|
||||
while (1) {
|
||||
|
||||
ramptr = ram;
|
||||
for (i = 0; i < ramsizeMB; i++, nMB++) {
|
||||
for (j = 0; j < pagesPerMB; j++) {
|
||||
dataptr = data;
|
||||
for (k = 0; k < PAGE_SIZE; k += sizeof(long long)) {
|
||||
ramptr += sizeof(long long);
|
||||
dataptr += sizeof(long long);
|
||||
*(unsigned long long *)ramptr ^= *(unsigned long long *)dataptr;
|
||||
}
|
||||
}
|
||||
|
||||
if (nMB == 1024) {
|
||||
after = now();
|
||||
fprintf(stderr, "%s (%05d): INFO: %06llums copied 1 GB in %05llums\n",
|
||||
argv0, gettid(), after, after - before);
|
||||
before = now();
|
||||
nMB = 0;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
free(data);
|
||||
free(ram);
|
||||
}
|
||||
|
||||
|
||||
static void *stressthread(void *arg)
|
||||
{
|
||||
unsigned long long ramsizeMB = *(unsigned long long *)arg;
|
||||
|
||||
stressone(ramsizeMB);
|
||||
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static int stress(unsigned long long ramsizeGB, int ncpus)
|
||||
{
|
||||
size_t i;
|
||||
unsigned long long ramsizeMB = ramsizeGB * 1024 / ncpus;
|
||||
ncpus--;
|
||||
|
||||
for (i = 0; i < ncpus; i++) {
|
||||
pthread_t thr;
|
||||
pthread_create(&thr, NULL,
|
||||
stressthread, &ramsizeMB);
|
||||
}
|
||||
|
||||
stressone(ramsizeMB);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
static int mount_misc(const char *fstype, const char *dir)
|
||||
{
|
||||
if (mkdir(dir, 0755) < 0 && errno != EEXIST) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot create %s: %s\n",
|
||||
argv0, gettid(), dir, strerror(errno));
|
||||
return -1;
|
||||
}
|
||||
|
||||
if (mount("none", dir, fstype, 0, NULL) < 0) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: cannot mount %s: %s\n",
|
||||
argv0, gettid(), dir, strerror(errno));
|
||||
return -1;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int mount_all(void)
|
||||
{
|
||||
if (mount_misc("proc", "/proc") < 0 ||
|
||||
mount_misc("sysfs", "/sys") < 0 ||
|
||||
mount_misc("tmpfs", "/dev") < 0)
|
||||
return -1;
|
||||
|
||||
mknod("/dev/urandom", 0777 | S_IFCHR, makedev(1, 9));
|
||||
mknod("/dev/random", 0777 | S_IFCHR, makedev(1, 8));
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
int main(int argc, char **argv)
|
||||
{
|
||||
unsigned long long ramsizeGB = 1;
|
||||
char *end;
|
||||
int ch;
|
||||
int opt_ind = 0;
|
||||
const char *sopt = "hr:c:";
|
||||
struct option lopt[] = {
|
||||
{ "help", no_argument, NULL, 'h' },
|
||||
{ "ramsize", required_argument, NULL, 'r' },
|
||||
{ "cpus", required_argument, NULL, 'c' },
|
||||
{ NULL, 0, NULL, 0 }
|
||||
};
|
||||
int ret;
|
||||
int ncpus = 0;
|
||||
|
||||
argv0 = argv[0];
|
||||
|
||||
while ((ch = getopt_long(argc, argv, sopt, lopt, &opt_ind)) != -1) {
|
||||
switch (ch) {
|
||||
case 'r':
|
||||
errno = 0;
|
||||
ramsizeGB = strtoll(optarg, &end, 10);
|
||||
if (errno != 0 || *end) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: Cannot parse RAM size %s\n",
|
||||
argv0, gettid(), optarg);
|
||||
exit_failure();
|
||||
}
|
||||
break;
|
||||
|
||||
case 'c':
|
||||
errno = 0;
|
||||
ncpus = strtoll(optarg, &end, 10);
|
||||
if (errno != 0 || *end) {
|
||||
fprintf(stderr, "%s (%05d): ERROR: Cannot parse CPU count %s\n",
|
||||
argv0, gettid(), optarg);
|
||||
exit_failure();
|
||||
}
|
||||
break;
|
||||
|
||||
case '?':
|
||||
case 'h':
|
||||
fprintf(stderr, "%s: [--help][--ramsize GB][--cpus N]\n", argv0);
|
||||
exit_failure();
|
||||
}
|
||||
}
|
||||
|
||||
if (getpid() == 1) {
|
||||
if (mount_all() < 0)
|
||||
exit_failure();
|
||||
|
||||
ret = get_command_arg_ull("ramsize", &ramsizeGB);
|
||||
if (ret < 0)
|
||||
exit_failure();
|
||||
}
|
||||
|
||||
if (ncpus == 0)
|
||||
ncpus = sysconf(_SC_NPROCESSORS_ONLN);
|
||||
|
||||
fprintf(stdout, "%s (%05d): INFO: RAM %llu GiB across %d CPUs\n",
|
||||
argv0, gettid(), ramsizeGB, ncpus);
|
||||
|
||||
if (stress(ramsizeGB, ncpus) < 0)
|
||||
exit_failure();
|
||||
|
||||
exit_success();
|
||||
}
|
Loading…
Reference in New Issue
Block a user