From Fedora Project Wiki

m (Remove more absolute links)
Line 17: Line 17:
* {{result|pass|Revise Fedora Install test plan to ensure adequate test coverage exists for failure scenarios listed above}}
* {{result|pass|Revise Fedora Install test plan to ensure adequate test coverage exists for failure scenarios listed above}}
* {{result|pass|Select a small, but representative, subset of test cases from the install test plan to automate}}The following test cases are selected:
* {{result|pass|Select a small, but representative, subset of test cases from the install test plan to automate}}The following test cases are selected:
** Rawhide Acceptance Test Plan [[https://fedoraproject.org/wiki/QA:Rawhide_Acceptance_Test_Plan]]
** [[QA:Rawhide_Acceptance_Test_Plan]]
** DVD.iso Installation
** DVD.iso Installation
** Boot.iso/Netinst.iso Installation
** Boot.iso/Netinst.iso Installation
Line 24: Line 24:
** system with basic video driver
** system with basic video driver
** Rescue installed system
** Rescue installed system
** Memory test [[https://fedoraproject.org/wiki/QA:Testcase_Memtest86]]
** [[QA:Testcase_Memtest86]]
* {{result|pass|Create python scripts to prepare KVM-based virtual environments for testing, initiate kickstart installs, and validate results}}
* {{result|pass|Create python scripts to prepare KVM-based virtual environments for testing, initiate kickstart installs, and validate results}}
** {{result|pass|Check the virtualization environment system sanity}}
** {{result|pass|Check the virtualization environment system sanity}}
Line 30: Line 30:
** {{result|pass|Print out test results}}
** {{result|pass|Print out test results}}
* {{result|pass| Create python scripts to parse differenet parameters}}
* {{result|pass| Create python scripts to parse differenet parameters}}
** {{result|pass|parse different repository: cdrom, http,ftp(anonymout, non anonymus), nfs, nfsiso, hard drive[[https://fedoraproject.org/wiki/QA:Testcase_Ftp_Repository]][[https://fedoraproject.org/wiki/QA:Testcase_Nfs_Repository]]}}
** {{result|pass|parse different repository: cdrom, http,ftp(anonymout, non anonymus), nfs, nfsiso, hard drive, [[QA:Testcase_Ftp_Repository]], [[QA:Testcase_Nfs_Repository]]}}
** {{result|pass|parse different kickstart delivery: http, file, hard drive, nfs [[https://fedoraproject.org/wiki/QA/TestCases/KickstartKsHttpServerKsCfg]] [[https://fedoraproject.org/wiki/QA/TestCases/KickstartKsFilePathKsCfg]][[https://fedoraproject.org/wiki/QA/TestCases/KickstartKsHdDevicePathKsCfg]][[https://fedoraproject.org/wiki/QA/TestCases/KickstartKsNfsServerPathKsCfg]] }}
** {{result|pass|parse different kickstart delivery: http, file, hard drive, nfs [[QA/TestCases/KickstartKsHttpServerKsCfg]], [[QA/TestCases/KickstartKsFilePathKsCfg]], [[QA/TestCases/KickstartKsHdDevicePathKsCfg]], [[QA/TestCases/KickstartKsNfsServerPathKsCfg]]}}
* Investigate methods for leveraging GUI automation to aid in automating applicable test cases
* Investigate methods for leveraging GUI automation to aid in automating applicable test cases
** {{result|pass| Open Virt Viewer of the guest}}
** {{result|pass| Open Virt Viewer of the guest}}

Revision as of 19:14, 12 May 2011

This page provides a high-level roadmap for implementing the Is_anaconda_broken_proposal project. More detailed tasks can be found in autoqa TRAC roadmap. We follow these steps to define the methods by which we initiate testing


First, in order to provide a consistent and documented test approach, the existing Fedora Install test plan [1] will be revisited. The test plan will be adjusted to ensure proper test coverage for the failure scenarios listed above. Existing test cases will be reviewed for accuracy. New test cases will be created using the Template:QA/Test_Case template. Finally, the test plan will be adjusted to match the improved Fedora Release Criteria [2]. This includes adjusting the test case priority to match milestone criteria.

Next, in order to reduce the setup/execution time, improve efficiency and to provide test results on a more consistent basis, a subset of test cases will be chosen for automation. Tests will be written in python and will be developed and executed on a system supporting KVM virtualization. Test scripts will be responsible for preparing a virtual install environment, initiating a kickstart install and validating the results. Once an initial batch of tests exist, they will be formally integrated into the AutoQA project.

Last, a method will be developed for collecting test results into a single test result matrix. Results may be posted to the wiki directly, or a custom turbogears application may be needed to display results [3]. The results will be easily accessible for testers and the installer development team.


The project will be divided into several phases.

Phase#1 - proof of concept

  • Pass pass Revise Fedora Install test plan to ensure adequate test coverage exists for failure scenarios listed above
  • Pass pass Select a small, but representative, subset of test cases from the install test plan to automate
    The following test cases are selected:
  • Pass pass Create python scripts to prepare KVM-based virtual environments for testing, initiate kickstart installs, and validate results
    • Pass pass Check the virtualization environment system sanity
    • Pass pass virt-install tree and ISO media.
    • Pass pass Print out test results
  • Pass pass Create python scripts to parse differenet parameters
  • Investigate methods for leveraging GUI automation to aid in automating applicable test cases
    • Pass pass Open Virt Viewer of the guest
    • Pass pass Input text into the GUI
    • Close the Virt Viewer of the guest

Phase#2 - implementation

Implement the selected test cases

Automate remainder of test cases from the install test plan

Implement the general test cases

Create kick start database to cover test cases that can be covered with kick start

Phase#3 - integration

  • Identify test event triggers which will be used to automatically initiate testing
  • Create appropriate control files and test wrappers to allow for scheduling tests through AutoQA (see Writing_AutoQA_Tests)
  • Develop or update AutoQA test event hooks to accommodate new test events (see Writing_AutoQA_Hooks)
  • Implement initial test result dashboard intended to eventually replace the wiki test matrix. The dashboard will also support FAS user test result submission. This will likely rely on