Executing a test case means either executing the specified test steps one by one on a specific version of the system under test in a selected test environment (only for test cases with at least one step) or importing the associated test result files.
The Run Test Case page shows all test cases of the selected project in a table.
The table shows the following values:
ID |
The automatically assigned ID. |
Additional Information |
A tooltip appears when the cursor is placed over the icon shown here. |
Revision |
The revision of the test case. |
Name |
The name of the test case. |
Execution |
Manual oder automated. |
Traceability |
The requirement assigned to this test case. |
Steps |
The number of associated test steps. |
Results |
The number of test case results. |
Action |
The executable actions. |
The following actions can be performed in the action column:
The following operations can be performed in the line above the table on the right:
All operations are described in detail in Section 5.2.3.1, “Overview Page”.
Manual or Automated Test Execution | |
---|---|
If a test case has at least one test step, it can be executed manually. Alternatively, test result files can be imported for all test cases (see Section 8.5, “Import Test Results”. for more information. |
Note | |
---|---|
Only test cases with the status Draft or Approved can be executed. |
Clicking the icon displays a dialog with information about the test case ( Figure 8.3).
Before a test case can be executed, the test environment the test case is run in and the system under test itself have to be selected. In addition, the user-defined properties of the test run can be set here (for more information, see Section 5.2.3.2.4, “User Defined Properties”).
Clicking the
button opens a second browser window where the test execution takes place. Make sure that pop-ups are allowed in your browser or add an exception for Klaros-Testmanagement. By default, a view with step-by-step instructions will be shown, but this can be changed to a tabular view during the execution.The step-by-step instructions screen shows the Action, Expected Result, Precondition and Postcondition values of the current test step.
Depending on the results of the step, the user can click the following buttons:
Passed |
Mark the test step as successfully completed. |
Error |
Mark the test step result as Error. |
Failure |
Mark the test step result as Failure. |
Inconclusive |
Mark the test step result as Inconclusive. |
Skip |
Skip the current test step without changing the result of the test case. |
Skip All |
Skip all remaining test steps without changing the result of the test case. |
Edit |
Edit this test step result. |
Review |
Request a review for this test case. |
Create Issue |
Create an issue related to this test case in an issue management system. |
Link Issue |
Link an existing issue from an issue management system to this test case. |
Back |
Go back to the last test step to repeat it or to edit it. |
The same process is applied for each test step of the test case.
As soon as all steps have a result or have been skipped, the execution can be finished via a dialog.
Clicking the Figure 8.7).
button, causes Klaros-Testmanagement to show the test run result of the test case (The tabular step instructions screen shows the Description, Expected Result, Precondition, Postcondition and result of all steps, the default result being Unknown.
The following actions can be performed at each step:
Skip |
Skip the current test step without changing the result of the test case. |
Back |
Go back to the last test step to repeat it again or to edit it. |
Review |
Request a review for this test case. |
Create Issue |
Create an issue related to this test case in an issue management system. |
Link Issue |
Link an existing issue from an issue management system to this test case. |
Pause execution |
Pauses the test execution and stops the time measurement. |
Skip all remaining steps |
Finish the test and mark all remaining steps as skipped. |
What is the difference between a failure and an error? | |
---|---|
A failure is a discrepancy between a computed, observed, or measured value or condition and the true, specified, or theoretically correct value or condition. An error is the inability of the system to perform the test correctly. |
For each error, failure or inconclusive result the following dialog will be shown: ( Figure 8.6).
Tip | |
---|---|
The Description and Summary fields appear in all comment dialogs. |
After the last step result of a test case has been entered, a dialog is displayed asking whether the execution should be finished.
Clicking the Figure 8.7).
button displays the result of the test run for this test case (In the Test Run - Overview view, a summary and description can be entered for the test case result as well as for each individual test step result. ( Figure 8.7).
Note | |
---|---|
If an issue management system is configured for the active project, the and buttons are activated. |
During the execution of a test case, issues can be created for this test case in an external issue management system by clicking the button. Issues already created in an issue management system can be linked to the test case by clicking the button.
For detailed information on creating and linking issues, see Section 9.6.6, “Issue Details (Creating a new Issue)” and Section 9.6.7, “Link Issue”.