From Fedora Project Wiki

Revision as of 15:55, 4 July 2016 by Adamwill (talk | contribs) (Relval bot-created Base validation results page for 25 Rawhide 20160704.n.0)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

No longer current
The compose for which this page contains results is no longer the current one. This page contains the results for the current compose.

This page records Base validation testing test results for the Fedora 25 20160704.n.0 nightly compose.

Which tests to run

Test coverage page
This page provides information about test coverage for the tests on this page across all the composes for the current release: it can help you see which test cases most need to be run.

Tests with a Milestone of Basic, Beta or Final are the most important. Optional tests are less so, but still useful to run. The milestone indicates that for that milestone release or a later one to be approved for release, the test must have been run against the release candidate build (so at Beta, all Basic and Beta tests must have been run, for instance). However, it is important to run the tests for all milestones as early and often as possible. Please refer to the test coverage page linked above and try to find and run tests which have not been run at all, or not run recently, for the current release.

How to test

1. Download one or more media for testing:

Image armhfp i386 x86_64
Everything boot Download Download
Workstation live Download Download
Workstation boot Download Download
Server dvd Download Download
Server boot Download Download
Atomic boot Download
Cloud_Base qcow2 Download
Cloud boot Download Download
Cloud_Base raw-xz Download
KDE live Download Download
Xfce live Download Download
SoaS live Download Download
Mate live Download Download
Cinnamon live Download Download
Security live Download Download
Robotics live Download Download
Scientific_KDE live Download Download
Docker_Base docker Download Download


2. Perform one or more of the test cases and add your results to the table below

  • You can submit results by editing the page directly, or by using relval, with the relval report-results command. It provides a simple text interface for reporting test results.

3. If a test fails, file a bug report. You may propose the bug as a release blocker or freeze exception bug for the appropriate release - see blocker bug process and freeze exception bug process.

Some tests must be run against particular Products or images - for example, the #Default boot and install tests. If no particular product or image is specified either in this page or the test case page, you can use any appropriate image. For example, you can run most of the #General Tests with the Workstation live image, or either of the Server install images.

If you notice a problem during a test which does not constitute a complete failure of the test, you should still file a bug report, but it may not be appropriate to propose it as a release blocker bug. Use your judgment in deciding this, with reference to the Fedora_Release_Criteria, which these tests are intended to verify. If you are unsure, err on the side of proposing the bug as a blocker.

Don't install updates
Don't install updates before performing any of the tests, as when you are testing pre-releases, available updates are not part of the proposed released package set.
Virtual machine testing
In most cases, testing in a virtual machine is OK. You can test UEFI (including SecureBoot) in VMs as well.

Results summary page

The Test Results:Fedora 25 Rawhide 20160704.n.0 Summary page contains the results from this page and all the other validation pages for the same compose listed together to provide an overview.

Add, Modify or Remove a Test Case

  1. Please request review for your changes by publishing your test case for review to the test mailing list and/or the appropriate working group mailing list (e.g. server, cloud, or desktop).
  2. Once reviewed, make your changes to any current documents that use the template (e.g. Test_Results:Current_Base_Test).
  3. Lastly, update Template:Base_test_matrix with the same changes.

Key

See the table below for a sample format for test results. All test results are posted using the result template.


Test Result Explanation Code Entered
none
Untested - This test has not been run, and is available for anyone to contribute feedback. {{result|none}}
Pass pass robatino
Passed - The test has been run and the tester determine the test met the expected results {{result|pass|robatino}}
Inprogress inprogress adamwill
Inprogress - An inprogress result is often used for tests that take a long time to execute. Inprogress results should be temporary and change to pass, fail or warn. {{result|inprogress|adamwill}}
Fail fail jlaska [1] [2]
Failed - Indicates a failed test. A link to a bug must be provided. See Template:Result for details on providing bug information.
  1. RHBZ #XYZ
  2. RHBZ #ZXY
{{result|fail|jlaska|XYZ|ZXY}}
Warning warn rhe
[1]
Warning - This test completed and met the expected results of the test, but other issues were encountered during testing that warrant attention.
  1. Brief description about the warning status
{{result|warn|rhe}} <ref>Brief description about the warning status</ref>
Pass pass hongqing
Warning warn kparal
Multiple results - More people can easily provide results to a single test case. {{result|pass|hongqing}} {{result|warn|kparal}}
Fail fail pboy [1] [2]
Failed - Same issue with LVM again
  1. RHBZ #2246871
  2. RHBZ #'2244305
{{result|fail|pboy|2246871|2244305}}
Pass pass previous <build> run
Result from previous test run - This test result is directly moved from the test run of previous <build>. {{result|pass|previous <build> run}}
Unsupported - An unsupported test or configuration. No testing is required.

Test Matrix

Milestone Test Case Workstation Server KDE ARM Cloud
Alpha QA:Testcase_base_initial_setup
none
none
none
Alpha QA:Testcase_base_startup
none
none
none
none
Alpha QA:Testcase_base_system_logging
none
none
none
none
none
Alpha QA:Testcase_base_update_cli
none
none
none
none
none
Final QA:Testcase_Services_start
none
none
none
none
none
Final QA:Testcase_base_selinux
none
none
none
none
none
Final QA:Testcase_base_service_manipulation
none
none
none
none
none