Progress report

Reporting takes place in accordance with the reporting structure described in the test plan. The progress report contains data on the most recent reporting period and cumulative data on the entire test process. 

Besides figures, the report should also provide textual explanation and advice on the results, progress, risks and any problem areas. The latter is inclined to be forgotten in reports that are generated from test-management tools. It should be realised that explanation and advice are very important in the provision of quick and reliable insight into the figures. It is the most important product of testing. While the explanation can and should be given verbally, it most definitely should be contained in the written report. This forces the test manager to think carefully, as well as making the advice stronger, reaching a wider audience and helping with the process evaluation in retrospect. 

In more detail

Progress report versus final report 

Although the terms 'interim report' or 'progress report' may suggest that these are less important than the final report, in fact the opposite is true. The progress report supplies early information and advice, with which the recipients (such as client, project manager and others) can often make timely adjustments for keeping the total process on the right track. The final report is more a retrospective evaluation that mainly benefits subsequent test processes and projects. 

In outline, a progress report has the following content (based on the BDTM method with the four aspects of Result, Risks, Time and Costs). In practice, the list of contents may follow a different sequence; subjects may be combined, or even omitted. It depends on the report's target group. 

1Status of the test object (BDTM: Result)
 1.1Status per characteristic/object part 
 1.2Status of test goals 
 1.3Trends and recommendations 
2Product risk and strategy adjustment (BDTM: Risks)
3Progress of the test process (BDTM: Time and Costs)
 3.1Progress (in hours and data) of activities or products over the recent period
 3.2Activities in the coming period
 3.3Hours lost
 3.4rends and recommendations
4Problem areas/points of discussion (all the BDTM aspects)
5Agreements
6Quality of the test process (optional, all the BDTM aspects)
 6.1Effectiveness 
 6.2Efficiency 
 6.3Verifiability 

These subjects are further explained below.

Status of the test object (Result) - 1

Status per characteristic/object part - 1.1

It is shown per characteristic/object part:

  • The status of the tests (not started, planned, specification, execution, retest X, completed), optionally with the progress percentage, e.g. the progress of the execution is estimated at 60% 
  • Overview of numbers of defects (sorted by status and severity, optionally also by other aspects, such as cause)
  • If test products (such as test cases or test scripts) are seen emphatically as results, they can also be included in the overview, with an indication of whether a start has been made on the product and whether it is ready. 

The closer the end of the test period approaches, the more attention is paid in the progress report to the consequences of open defects. In the beginning, it is less useful to include this in the report, since it is expected that the defects will be solved. But the consequences should always be included in the defect report itself. 

  • Defects that remain open and their impact
  • Defects not solved (known errors), and their impact.
Status of test goals - 1.2

Based on the above, the status per test goal (user requirement, business process, critical success factor, etc.) is reported. Sometimes a test goal can be directly linked to a number of characteristics/object parts and to the test status related to them; sometimes the status per characteristic/object part is not sufficiently usable and the test manager still has to determine the test status per test goal. The risk tables from the Product Risk Analysis make the link possible. 

Trends and recommendations  - 1.3

Relevant trends and related recommendations can be reported here.

In more detail

Below are some overviews that will reveal whether certain trends are taking place:

  • The number of open defects per week will indicate whether the testing can tail off or if a backlog is building up
  • The relationship between numbers of defects and test cases per subsystem provides an indication of whether extra testing on that part will deliver many more defects 
  • The number of found defects and number of solved (including retested) defects within a certain period says something about the stability of the system 
  • Status of the defect versus who should carry out the following step in the handling of it. This shows up where any bottleneck lies. For example, where all the complex faults are allocated to that one experienced developer, with the result that a backlog of unsolved defects is created 
  • Cause of defects (requirements, design, code, test environment, wrong installation/ operation, wrong test case) versus subsystem. Provides insight into the concentrations of specific mistakes 
  • Number of defects versus tables (with data warehousing). This tells us what the error-sensitive system parts are.

In order to give the trend significance for the stakeholders, it is advisable to use graphics, making the trend visible. This is not as easy as it seems. It is difficult to produce a clear and legible graphic. A few tips (quoted freely from [Tufte, 2001]: 

  1. Make the data and the message the centre piece
  2. Maximise the data/ink ratio (i.e. leave out all the symbols, lines and colours that don't add anything)
  3. Remove redundancies
  4. Review and amend.

Product risk and strategy adjustment (Risks)

In this part of the report, the stakeholders are given insight into the degree to which the coverage of the various product risks has changed, as well as into any process risks. 

In the test plan strategy, it is determined whether and to what degree product risks will be covered by testing. During the test process, aberrations may occur: the estimate of the risk appears different and/or the test coverage requires adjusting. The adjustments over the reporting period, with associated consequences, are reported in this part. In this, the translation is made into the test goals: what kind of impact will the changed risks have on the attainment of these goals? 

Progress of the test process (Time and Costs) -  3

Regarding the progress of the test process, the points below are significant.

Progress over the latest period  - 3.1

At the level of phases and/or main products, the following could be reported:

  • Number of planned hours
  • Number of hours spent so far
  • Number of hours expected still to be spent
  • Percentage completed
  • Dates: planned/expected/actual start date; planned/expected/actual end date.

Products could be the test plan, test scripts, test-execution files and reports.

If the test manager is responsible for the budget, he will also include in the progress report information on completing the test process within budget. 

Activities in the coming period  - 3.2

Here, the activities to be carried out in the coming period are reported.

Hours lost  - 3.3

This refers to non-productive hours of the testers. If the test process environment does not meet certain preconditions, this will result in inefficiency and loss of hours. Examples are a non-functioning test infrastructure, much or lengthy test-obstructing defects or lack of support. Hours lost, and the causes, are reported here. 

Trends and recommendations - 3.4

As with trends in the status of the test object, trends and recommendations in connection with the progress of the testing should also be reported. The central question here is whether the agreed milestones are (or appear to be) feasible.

In more detail

One of the trends that can be watched is the average time required for the reworking of a defect. If this increases, it is possibly a signal that the volume of the backlog of work is increasing sharply. The percentage of wrongly reworked defects can also be observed. 

Problem areas/discussion points (all the BDTM aspects) - 4

In this section of the report, the test manager points out any problem areas or points for discussion that jeopardize completion of the test assignment within the set limits of time and costs. For example: 

  • The test object being delivered later than planned
  • The quality of the test basis being less than expected
  • The test environment not being available on the agreed date
  • Test-obstructing defects present in the test environment or test object.

Besides the various problem areas, their consequences and possible measures are shown. Here, too, the test manager makes the translation into the test goals. 

Agreements - 5

This part shows the agreements made in the current period between the test team and other parties that are relevant to the recipients of the report. 

Quality of the test process (optional) - 6

If required, this part of the report can include information on the quality of the test process. The following questions play a part here: 

  • Are the significant defects being found (as early as possible)? (Effectiveness)
  • How economical is the test process with time and resources? (Efficiency)
  • Is the test process working as agreed? (Verifiability)

The three quality aspects of the test process.

The three quality aspects of the test process

In more detail

A point of focus here is the general problem with metrics: how to draw the right conclusions from the figures; how to avoid comparing apples with oranges. See also [Link {13.xml}Chapter 13] "Metrics".

Effectiveness - 6.1

In more detail

The difficulty with the question of whether the testing is effective, is that this can usually only be established in retrospect. The effectiveness issue can be split into two parts: 

  • Is there a good strategy in place?
  • Is the testing being carried out in accordance with this strategy?

There are various indicators that can be included in the report:

  • The percentage of found defects in the test level / the number of defects present or an approximation thereof; the number of defects can be approximated by, for example, the number of defects still being found during the first 3 months of production 
  • The percentage of found defects in a test level that should reasonably have been
  • found in a preceding test (30% of the defects found in the acceptance test concern programming defects; these should actually already have been found in the development tests and system test) 
  • Degree of testing coverage; the more thorough the test, the more defects will be found
  • The percentage of mistakes (= test faults).
Efficiency - 6.2

In more detail

The following are possible indicators of this:

  • The number of defects found per test hour
  • Estimating prevented damage in relation to the test costs (through finding faults)
  • Number of specified or executed test cases per hour
  • Number of reviewed pages per hour.

By comparing these figures with an established standard, a picture is created of the efficiency of the test process.

Verifiability  - 6.3

In more detail

This aspect is difficult to communicate through indicators. What the test manager can say in the report about this is whether and how in the latest period it was verified that the test team was working as agreed. The verification can focus on the test 

products or the processes and can be based on the planned quality measures, or on monitoring, or on a random check at the overall level. The test manager should make a good risk estimate as regards what checking would be useful. In particular, the test levels that are placed with inexperienced test managers or that have been outsourced are eligible for verification. 

Below is an example of a dashboard, enabling the most important information to be seen at a glance.

An example of a dashboard

Later in the report, these points are worked out in detail in overviews with notes. Examples of overviews (without notes):

Quality of test object – defects

DefectsOpenTo be retestedClosedTOTAL
Obstructing  44
Severe111214
Disruptive3124964
Cosmetic1573759
TOTAL1920102141

Quality of test object – subsystem x causes

 RequirementsDesignSoftwareInfrastructureTestTOTAL
Subsystem 135186234
Subsystem 2 1162423
Subsystem 36143014165
Total system      
TOTAL92064227122

Progress

 TimeHours
Milestone/ActivityPlanExpectedRealisedEstimate (A)Spent (B)To be spent (C)Difference (A-B-C)
Planning       
- Test planApr 1 Apr 1605406
Preparation       
- Testability reviewApr 8Apr 12 40464-10
Specification       
- Test unit 1May 2May 2 12060600