SW 테스트 실전

(Case-study) Test Result reporting is a core part of the art of testing!

톰n제리 2026. 2. 10. 18:05

In a recent brief testing consulting, I noticed a recurring trend: many testing activities were being treated as mere formalities.

 

Some were skipped entirely, while others were finished with just a couple of perfunctory lines.

 

So I try to write about Test Result Reporting. Here are my thoughts.


Main Title : Approaching Test Reporting at three types of view.

When I think about my old test result reports, I realized they generally fall into three categories based on the answers to two questions:

 

a) Did the test proceed smoothly and according to plan?

b) Is there a specific message or concern I need to highlight?

 

Depending on the answers, you can choose one of the following approaches:

[1] The Fact-Oriented Report

Use this when things went as expected.

  • Content: Overview, objectives, methodology, and coverage.
  • Comparative Data: Distinctive features compared to previous similar tests (e.g., defect rates relative to lines of code or test cases, average days to resolve defects).

[2] The Issue-Oriented Report (Highlighting Obstacles)

Use this when the testing process faced significant hurdles.

  • Content: High volume of unexecuted TCs, difficulties in environment isolation or data acquisition, a high rate of "Not a Bug" resolutions, etc.

[3] The Message-Oriented Report (Advocacy Focus)

Use this when you need to drive a specific point home.

  • Content: Evidence of low initial quality (e.g., comparison of first-run pass rates) to justify future resource needs or process changes.

[0] Common Core Elements

Regardless of the lens, every report should include these basics (based on my actual project captures):

  • Test Overview: Reuse the test plan content, but specifically note any deviations.
    • Example: Ended testing after Round 1 (of a planned 2-round cycle) due to specific issues.
  • Test Schedule: Compare planned vs. actual dates and explain any variances.
  • Test Environment: If you had to use the Dev environment because the QA environment wasn't ready, document it.
    • Tip: This serves as critical evidence if code management issues arise later.
  • Test Results: Display standard metrics (Execution rate, Pass rate, Defect resolution rate). If you closed the test conditionally (under 100%), state the rationale.
    • Example: Legal review for Terms of Service was delayed. Mentioning this provides an "excuse" or grounds for re-verification if issues occur post-launch.
  • Defect Details: Resolution rates. If not 100%, break down the "Deferred," "Not a Bug," "Known Issue," or "Cancelled" counts.

Note: I prefer to separate Residual Risks and Conclusions/Recommendations into a final "Sign-off" document if QA is responsible for the go/no-go decision, keeping the "Result Report" strictly to objective facts.

 

A Sample of Test Resut report.

더보기

A Sample of Test Resut report.

5.4 Test Results

Test Execution Overview

The test was originally planned in two phases; however, the scope was adjusted as follows:

  • Phase 1 (Execution & Defect Verification):
    • Full execution of planned test cases and verification of individual bug fixes.
  • Phase 2 (Final Regression & Verification) - [POSTPONED]:
    • Reason: Originally intended for a full re-test under "Code Freeze" conditions. However, due to ongoing discussions regarding new features, upcoming major milestones, and pending decisions on the final launch date/scope, Phase 2 has been put on hold.

Test Schedule

  • Phase 1 Execution: 2023-05-02 (Tue) ~ 2023-05-12 (Fri) [2 Weeks]
  • Defect Verification: 2023-05-15 (Mon) ~ 2023-06-16 (Fri) [5 Weeks]
  • Phase 2 Execution: Postponed

Test Environment

  • Phase 1: Conducted in the Development Environment(http://dev.xxx.com) due to the unavailability of a dedicated QA environment at the time of testing.

Test Metrics (Overall)

- (link) the link that can check the whole test results.

  • Test Execution Rate: 100.0% (377/377 Test Cases)
    • Note: 10 cases were "Blocked" due to dependencies on defects from previous stages.
  • Test Pass Rate: 98.3% (373/377 Test Cases)
    • Fail (4 cases): Related to Terms & Policies (Privacy Policy, Terms of Use, AI Utilization, Notification Settings). These were marked as "Fail" because typos and undefined content were not finalized within the schedule.
  • Defect Resolution Rate: 100.0% (95/95 Defects)
    • Includes: 27 Cancelled, 27 Deferred.
    • [Link to Jira Dashboard]

(Reference) Initial Test Results (Before Fixes)

  • Execution Rate: 97.4% (373/383 TCs)
  • Pass Rate: 68.4% (255/373 TCs)

Defect Status Breakdown

  • Resolution Rate: 100.0% (98/98)
  • Detailed Jira List: [Link] | Dashboard: [Link]

1. Deferred Defects (27 cases)

Issues classified as low frequency/impact or those where the fix direction is currently undecided.

  • [EA-266] Requires verification of 4 Terms of Service items (placeholders like [date], [email address] remain).
  • [EA-265] Automatic redirect to login page upon session expiry, even on public pages (Home, Discover).
  • [EA-260] Multi-user boards: If one user deletes an image, others can still attempt comments/reactions despite server errors (data is not saved).
  • [EA-255] Need for a dedicated error page when the server returns 5xx responses.
  • [EA-240] Generation Page: Input values are cleared when switching between FullSentence and WordBlock tabs.

2. Cancelled Defects (27 cases)

Issues identified as "Not a Bug," unreproducible, or cancelled due to spec changes.

  • [EA-219] Direct input in the top WordBlock is disabled (as per intended design).
  • [EA-125] (API) Image deletion: Returns 204 instead of 404 when attempting to delete a non-existent image (confirmed as intended behavior).

Deep Dive: Specialized Content

[1] Fact-Heavy: Demonstrating Maturity

If the project was "business as usual," use these metrics to show that development and testing maturity are stable:

  • Defect Rate per TC: A downward trend proves quality activities are working. If it’s higher, investigate if it's due to a larger volume of changes.
  • Defect Rate per LOC (Lines of Code): A more objective measure of dev quality. (Though complexity should be noted as a variable).
  • Initial Pass Rate: Comparing the success rate of the first run against previous versions helps identify early-stage quality trends.

[2] Issue-Heavy: Highlighting Risk

If you "powered through" a messy process, make sure those risks are visible:

  • Blocked TCs: Explicitly list which areas couldn't be tested and why. These are "potential landmines."
  • Conditional Exit: If you stopped testing despite failures, document the stakeholders' agreement and the reasoning.
  • Environmental/Data Integrity: If the data was limited or the environment was "unclean," note that the results may not fully reflect real-world performance.

[3] Message-Heavy: Communicating Quality Concerns

When the initial build quality is poor (a common complaint I hear from QA leads), use data to speak for you:

  • Initial Pass Rate: A disastrous first-run pass rate is hard evidence that the build was handed over prematurely, causing high overhead for re-testing.
  • Defect Severity Ratio: A high percentage of "Critical/Major" bugs usually blocks the path to finding smaller, more nuanced defects.
  • Defect Density by Module: Identify "hotspots" or specific functional areas that are prone to failure. (Focusing on functional areas is often more professional than targeting individual developers).

Conclusion

Summarizing and sharing test results is not a mundane, clerical task. It is the "Endgame" of Test Engineering. It is where data meets strategy. Let’s strive to turn our reports into the "Art of Testing."