[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: Problems with MIPS Malta SSH tests in make check-acceptance
From: |
Cleber Rosa |
Subject: |
Re: Problems with MIPS Malta SSH tests in make check-acceptance |
Date: |
Thu, 19 Sep 2019 12:56:32 -0400 |
User-agent: |
Mutt/1.12.1 (2019-06-15) |
On Wed, Sep 18, 2019 at 09:14:58PM -0400, Cleber Rosa wrote:
> On Wed, Sep 18, 2019 at 05:16:54PM +1000, David Gibson wrote:
> > Hi,
> >
> > I'm finding make check-acceptance is currently useless for me as a
> > pre-pull test, because a bunch of the tests are not at all reliable.
> > There are a bunch which I'm still investigating, but for now I'm
> > looking at the MIPS Malta SSH tests.
> >
> > There seem to be at least two problems here. First, the test includes
> > a download of a pretty big guest disk image. This can easily exhaust
> > the 2m30 timeout on its own.
> >
>
> You're correct that successes and failures on those tests depend
> largely on bandwith. On a shared environment I used for tests
> the download of those images take roughly 400 seconds, resulting
> in failures. On my own machine, around 60, and the tests pass.
>
> There's a conceptual and conflicting problem in that the environment
> for tests to run should be prepared beforehand. The conflicting
> solutions can be:
>
> * extensive bootstrapping of the test execution environment, such
> as the installation of guests from ISOs or installation trees, or
> the download of "default" images wether the tests will use it or
> not (this is what Avocado-VT does/requires)
>
> * keeping test assets in the tree (Avocado allows this if you have
> a your_test.py.data/ directory), but it's not practical for large
> files or files that can't or shouldn't be redistributed
>
> > Even without the timeout, it makes the test really slow, even on
> > repeated runs. Is there some way we can make the image download part
> > of "building" the tests rather than actually running the testsuite, so
> > that a) the test themselves go faster and b) we don't include the
> > download in the test timeout - obviously the download speed is hugely
> > dependent on factors that aren't really related to what we're testing
> > here.
> >
>
> On Avocado version 72.0 we attempted to minimize the isse by
> implementing a "vmimage" command. So, if you expect to use Fedora 30
> aarch64 images, you could run before your tests:
>
> $ avocado vmimage get --distro fedora --distro-version 30 --arch aarch64
>
> And to list the images on your cache:
>
> $ avocado vmimage list
>
> Unfortunately, this test doesn't use the vmimage API. Actually that
> is fine because not all test assets map nicely to the vmimage goal,
> and should keep using the more generic (and lower level) fetch_asset().
>
> We're now working on various "asset fetcher" improvements that should
> allow us to check/cache all assets before a test is executed. Also,
> we're adding a mode in which the "fetch_asset()" API will default to
> cancel (aka SKIP) a test if the asset could not be downloaded.
>
> If you're interested in the card we're using to track that new feature:
>
>
> https://trello.com/c/T3SC1sZs/1521-implement-fetch-assets-command-line-parameter
>
> Another possibility that we've prototyped, and we'll be working on
> further, is to make a specific part of the "test" code execution
> (really a pre-test phase) to be executed without a timeout and even be
> tried a number of times before bailing out and skipping the test.
>
> > In the meantime, I tried hacking it by just increasing the timeout to
> > 10m. That got several of the tests working for me, but one still
> > failed. Specifically 'LinuxSSH.test_mips_malta32eb_kernel3_2_0' still
> > timed out for me, but now after booting the guest, rather than during
> > the image download. Looking at the avocado log file I'm seeing a
> > bunch of soft lockup messages from the guest console, AFAICT. So it
> > looks like we have a real bug here, which I suspect has been
> > overlooked precisely because the download problems mean this test
> > isn't reliable.
> >
>
> I've schedulled a 100 executions of `make check-acceptance` builds, with
> the linux_ssh_mips_malta.py tests having a 1500 seconds timeout. The
> very first execution already brought interesting results:
>
> ...
> (15/39)
> /home/cleber/src/qemu/tests/acceptance/linux_ssh_mips_malta.py:LinuxSSH.test_mips_malta32eb_kernel3_2_0:
> PASS (198.38 s)
> (16/39)
> /home/cleber/src/qemu/tests/acceptance/linux_ssh_mips_malta.py:LinuxSSH.test_mips_malta64el_kernel3_2_0:
> FAIL: Failure message found in console: Oops (22.83 s)
>
> I'll let you know about my full results. This should also serve as a
> starting point to a discussion about the reliability of other tests,
> as you mentioned before.
Out of the 100 executions on a ppc64le host, the results that contain
failures and errors:
15-/home/cleber/src/qemu/tests/acceptance/linux_ssh_mips_malta.py:LinuxSSH.test_mips_malta32eb_kernel3_2_0
- PASS: 92
- INTERRUPTED: 4
- FAIL: 4
16-/home/cleber/src/qemu/tests/acceptance/linux_ssh_mips_malta.py:LinuxSSH.test_mips_malta64el_kernel3_2_0
- PASS: 95
- FAIL: 5
FAIL means that self.fail() was called, which means 'Oops' was found
in the console. INTERRUPTED here means that the test timeout kicked
in, and I can back David's statements about soft lockups.
Let me know if anyone wants access to the full logs/results.
- Cleber.
>
> In my experience, and backed by the executions on Travis, most tests
> have been really stable on x86_64 hosts. Last week I've worked in
> ppc64 and aarch64 hosts, and posted a number of patches addressing
> the failures I found. I'll compile a list of the posted patches and
> their status.
>
> Thanks for reporting those issues.
> - Cleber.
>
> > Any thoughts on how to improve the situation?
> >
> > --
> > David Gibson | I'll have my music baroque, and my
> > code
> > david AT gibson.dropbear.id.au | minimalist, thank you. NOT _the_
> > _other_
> > | _way_ _around_!
> > http://www.ozlabs.org/~dgibson
>
>