bfd12c92cc
make check-avocado AVOCADO_TESTS=tests/avocado/tuxrun_baselines.py: \ TuxRunBaselineTest:test_riscv64_rv32 Signed-off-by: LIU Zhiwei <zhiwei_liu@linux.alibaba.com> Suggested-by: Daniel Henrique Barboza <dbarboza@ventanamicro.com> Acked-by: Alistair Francis <alistair.francis@wdc.com> Reviewed-by: Daniel Henrique Barboza <dbarboza@ventanamicro.com> Message-ID: <20240919055048.562-9-zhiwei_liu@linux.alibaba.com> Signed-off-by: Alistair Francis <alistair.francis@wdc.com>
241 lines
8.2 KiB
Python
241 lines
8.2 KiB
Python
# Functional test that boots known good tuxboot images the same way
|
|
# that tuxrun (www.tuxrun.org) does. This tool is used by things like
|
|
# the LKFT project to run regression tests on kernels.
|
|
#
|
|
# Copyright (c) 2023 Linaro Ltd.
|
|
#
|
|
# Author:
|
|
# Alex Bennée <alex.bennee@linaro.org>
|
|
#
|
|
# SPDX-License-Identifier: GPL-2.0-or-later
|
|
|
|
import os
|
|
import time
|
|
import tempfile
|
|
|
|
from avocado import skip, skipUnless
|
|
from avocado_qemu import QemuSystemTest
|
|
from avocado_qemu import exec_command, exec_command_and_wait_for_pattern
|
|
from avocado_qemu import wait_for_console_pattern
|
|
from avocado.utils import process
|
|
from avocado.utils.path import find_command
|
|
|
|
class TuxRunBaselineTest(QemuSystemTest):
|
|
"""
|
|
:avocado: tags=accel:tcg
|
|
"""
|
|
|
|
KERNEL_COMMON_COMMAND_LINE = 'printk.time=0'
|
|
# Tests are ~10-40s, allow for --debug/--enable-gcov overhead
|
|
timeout = 100
|
|
|
|
def get_tag(self, tagname, default=None):
|
|
"""
|
|
Get the metadata tag or return the default.
|
|
"""
|
|
utag = self._get_unique_tag_val(tagname)
|
|
print(f"{tagname}/{default} -> {utag}")
|
|
if utag:
|
|
return utag
|
|
|
|
return default
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
|
|
# We need zstd for all the tuxrun tests
|
|
# See https://github.com/avocado-framework/avocado/issues/5609
|
|
zstd = find_command('zstd', False)
|
|
if zstd is False:
|
|
self.cancel('Could not find "zstd", which is required to '
|
|
'decompress rootfs')
|
|
self.zstd = zstd
|
|
|
|
# Process the TuxRun specific tags, most machines work with
|
|
# reasonable defaults but we sometimes need to tweak the
|
|
# config. To avoid open coding everything we store all these
|
|
# details in the metadata for each test.
|
|
|
|
# The tuxboot tag matches the root directory
|
|
self.tuxboot = self.get_tag('tuxboot')
|
|
|
|
# Most Linux's use ttyS0 for their serial port
|
|
self.console = self.get_tag('console', "ttyS0")
|
|
|
|
# Does the machine shutdown QEMU nicely on "halt"
|
|
self.shutdown = self.get_tag('shutdown')
|
|
|
|
# The name of the kernel Image file
|
|
self.image = self.get_tag('image', "Image")
|
|
|
|
self.root = self.get_tag('root', "vda")
|
|
|
|
# Occasionally we need extra devices to hook things up
|
|
self.extradev = self.get_tag('extradev')
|
|
|
|
self.qemu_img = super().get_qemu_img()
|
|
|
|
def wait_for_console_pattern(self, success_message, vm=None):
|
|
wait_for_console_pattern(self, success_message,
|
|
failure_message='Kernel panic - not syncing',
|
|
vm=vm)
|
|
|
|
def fetch_tuxrun_assets(self, csums=None, dt=None):
|
|
"""
|
|
Fetch the TuxBoot assets. They are stored in a standard way so we
|
|
use the per-test tags to fetch details.
|
|
"""
|
|
base_url = f"https://storage.tuxboot.com/20230331/{self.tuxboot}/"
|
|
|
|
# empty hash if we weren't passed one
|
|
csums = {} if csums is None else csums
|
|
ksum = csums.get(self.image, None)
|
|
isum = csums.get("rootfs.ext4.zst", None)
|
|
|
|
kernel_image = self.fetch_asset(base_url + self.image,
|
|
asset_hash = ksum,
|
|
algorithm = "sha256")
|
|
disk_image_zst = self.fetch_asset(base_url + "rootfs.ext4.zst",
|
|
asset_hash = isum,
|
|
algorithm = "sha256")
|
|
|
|
cmd = f"{self.zstd} -d {disk_image_zst} -o {self.workdir}/rootfs.ext4"
|
|
process.run(cmd)
|
|
|
|
if dt:
|
|
dsum = csums.get(dt, None)
|
|
dtb = self.fetch_asset(base_url + dt,
|
|
asset_hash = dsum,
|
|
algorithm = "sha256")
|
|
else:
|
|
dtb = None
|
|
|
|
return (kernel_image, self.workdir + "/rootfs.ext4", dtb)
|
|
|
|
def prepare_run(self, kernel, disk, drive, dtb=None, console_index=0):
|
|
"""
|
|
Setup to run and add the common parameters to the system
|
|
"""
|
|
self.vm.set_console(console_index=console_index)
|
|
|
|
# all block devices are raw ext4's
|
|
blockdev = "driver=raw,file.driver=file," \
|
|
+ f"file.filename={disk},node-name=hd0"
|
|
|
|
kcmd_line = self.KERNEL_COMMON_COMMAND_LINE
|
|
kcmd_line += f" root=/dev/{self.root}"
|
|
kcmd_line += f" console={self.console}"
|
|
|
|
self.vm.add_args('-kernel', kernel,
|
|
'-append', kcmd_line,
|
|
'-blockdev', blockdev)
|
|
|
|
# Sometimes we need extra devices attached
|
|
if self.extradev:
|
|
self.vm.add_args('-device', self.extradev)
|
|
|
|
self.vm.add_args('-device',
|
|
f"{drive},drive=hd0")
|
|
|
|
# Some machines need an explicit DTB
|
|
if dtb:
|
|
self.vm.add_args('-dtb', dtb)
|
|
|
|
def run_tuxtest_tests(self, haltmsg):
|
|
"""
|
|
Wait for the system to boot up, wait for the login prompt and
|
|
then do a few things on the console. Trigger a shutdown and
|
|
wait to exit cleanly.
|
|
"""
|
|
self.wait_for_console_pattern("Welcome to TuxTest")
|
|
time.sleep(0.2)
|
|
exec_command(self, 'root')
|
|
time.sleep(0.2)
|
|
exec_command(self, 'cat /proc/interrupts')
|
|
time.sleep(0.1)
|
|
exec_command(self, 'cat /proc/self/maps')
|
|
time.sleep(0.1)
|
|
exec_command(self, 'uname -a')
|
|
time.sleep(0.1)
|
|
exec_command_and_wait_for_pattern(self, 'halt', haltmsg)
|
|
|
|
# Wait for VM to shut down gracefully if it can
|
|
if self.shutdown == "nowait":
|
|
self.vm.shutdown()
|
|
else:
|
|
self.vm.wait()
|
|
|
|
def common_tuxrun(self,
|
|
csums=None,
|
|
dt=None,
|
|
drive="virtio-blk-device",
|
|
haltmsg="reboot: System halted",
|
|
console_index=0):
|
|
"""
|
|
Common path for LKFT tests. Unless we need to do something
|
|
special with the command line we can process most things using
|
|
the tag metadata.
|
|
"""
|
|
(kernel, disk, dtb) = self.fetch_tuxrun_assets(csums, dt)
|
|
|
|
self.prepare_run(kernel, disk, drive, dtb, console_index)
|
|
self.vm.launch()
|
|
self.run_tuxtest_tests(haltmsg)
|
|
|
|
|
|
#
|
|
# The tests themselves. The configuration is derived from how
|
|
# tuxrun invokes qemu (with minor tweaks like using -blockdev
|
|
# consistently). The tuxrun equivalent is something like:
|
|
#
|
|
# tuxrun --device qemu-{ARCH} \
|
|
# --kernel https://storage.tuxboot.com/{TUXBOOT}/{IMAGE}
|
|
#
|
|
|
|
def test_arm64(self):
|
|
"""
|
|
:avocado: tags=arch:aarch64
|
|
:avocado: tags=cpu:cortex-a57
|
|
:avocado: tags=machine:virt
|
|
:avocado: tags=tuxboot:arm64
|
|
:avocado: tags=console:ttyAMA0
|
|
:avocado: tags=shutdown:nowait
|
|
"""
|
|
sums = {"Image" :
|
|
"ce95a7101a5fecebe0fe630deee6bd97b32ba41bc8754090e9ad8961ea8674c7",
|
|
"rootfs.ext4.zst" :
|
|
"bbd5ed4b9c7d3f4ca19ba71a323a843c6b585e880115df3b7765769dbd9dd061"}
|
|
self.common_tuxrun(csums=sums)
|
|
|
|
def test_arm64be(self):
|
|
"""
|
|
:avocado: tags=arch:aarch64
|
|
:avocado: tags=cpu:cortex-a57
|
|
:avocado: tags=endian:big
|
|
:avocado: tags=machine:virt
|
|
:avocado: tags=tuxboot:arm64be
|
|
:avocado: tags=console:ttyAMA0
|
|
:avocado: tags=shutdown:nowait
|
|
"""
|
|
sums = { "Image" :
|
|
"e0df4425eb2cd9ea9a283e808037f805641c65d8fcecc8f6407d8f4f339561b4",
|
|
"rootfs.ext4.zst" :
|
|
"e6ffd8813c8a335bc15728f2835f90539c84be7f8f5f691a8b01451b47fb4bd7"}
|
|
self.common_tuxrun(csums=sums)
|
|
|
|
def test_riscv64_rv32(self):
|
|
"""
|
|
:avocado: tags=arch:riscv64
|
|
:avocado: tags=machine:virt
|
|
:avocado: tags=tuxboot:riscv32
|
|
:avocado: tags=cpu:rv32
|
|
"""
|
|
sums = { "Image" :
|
|
"89599407d7334de629a40e7ad6503c73670359eb5f5ae9d686353a3d6deccbd5",
|
|
"fw_jump.elf" :
|
|
"f2ef28a0b77826f79d085d3e4aa686f1159b315eff9099a37046b18936676985",
|
|
"rootfs.ext4.zst" :
|
|
"7168d296d0283238ea73cd5a775b3dd608e55e04c7b92b76ecce31bb13108cba" }
|
|
|
|
self.common_tuxrun(csums=sums)
|