aboutsummaryrefslogtreecommitdiffstats
path: root/lib/test_server/doc/src/why_test_chapter.xml
blob: 3d0e8271b19d4900c1acbec5386ce6311c0484b0 (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
<?xml version="1.0" encoding="utf-8" ?>
<!DOCTYPE chapter SYSTEM "chapter.dtd">

<chapter>
  <header>
    <copyright>
      <year>2002</year><year>2013</year>
      <holder>Ericsson AB. All Rights Reserved.</holder>
    </copyright>
    <legalnotice>
      Licensed under the Apache License, Version 2.0 (the "License");
      you may not use this file except in compliance with the License.
      You may obtain a copy of the License at
 
          http://www.apache.org/licenses/LICENSE-2.0

      Unless required by applicable law or agreed to in writing, software
      distributed under the License is distributed on an "AS IS" BASIS,
      WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
      See the License for the specific language governing permissions and
      limitations under the License.
    
    </legalnotice>

    <title>Why Test</title>
    <prepared>Siri Hansen</prepared>
    <docno></docno>
    <date></date>
    <rev></rev>
  </header>

  <section>
    <title>Goals</title>
    <p>It's not possible to prove that a program is correct by
      testing. On the contrary, it has been formally proven that it is
      impossible to prove programs in general by testing. Theoretical
      program proofs or plain examination of code may be viable options
      for those that wish to certify that a program is correct. The test
      server, as it is based on testing, cannot be used for
      certification. Its intended use is instead to (cost effectively)
      <em>find bugs</em>. A successful test suite is one that reveals a
      bug. If a test suite results in Ok, then we know very little that
      we didn't know before.
      </p>
  </section>

  <section>
    <title>What to test?</title>
    <p>There are many kinds of test suites. Some concentrate on
      calling every function in the interface to some module or
      server. Some other do the same, but uses all kinds of illegal
      parameters, and verifies that the server stays alive and rejects
      the requests with reasonable error codes. Some test suites
      simulate an application (typically consisting of a few modules of
      an application), some try to do tricky requests in general, some
      test suites even test internal functions.
      </p>
    <p>Another interesting category of test suites are the ones that
      check that fixed bugs don't reoccur. When a bugfix is introduced,
      a test case that checks for that specific bug should be written
      and submitted to the affected test suite(s).
      </p>
    <p>Aim for finding bugs. Write whatever test that has the highest
      probability of finding a bug, now or in the future. Concentrate
      more on the critical parts. Bugs in critical subsystems are a lot
      more expensive than others.
      </p>
    <p>Aim for functionality testing rather than implementation
      details. Implementation details change quite often, and the test
      suites should be long lived. Often implementation details differ
      on different platforms and versions. If implementation details
      have to be tested, try to factor them out into separate test
      cases. Later on these test cases may be rewritten, or just
      skipped.
      </p>
    <p>Also, aim for testing everything once, no less, no more. It's
      not effective having every test case fail just because one
      function in the interface changed.
      </p>
  </section>

  <section>
    <title>How much to test</title>
    <p>There is a unix shell script that counts the number of non
      commented words (lines and characters too) of source code in each
      application's test directory and divides with the number of such
      source words in the src directory. This is a measure of how much
      test code there is.
      </p>
    <p>There has been much debate over how much test code, compared to
      production code, should be written in a project. More test code
      finds more bugs, but test code needs to be maintained just like
      the production code, and it's expensive to write it in the first
      place. In several articles from relatively mature software
      organizations that I have read, the amount of test code has been
      about the same as the production code.  </p>
    <p>In OTP, at the time of
      writing, few applications come even close to this, some have no
      test code at all.
      </p>

    <section>
      <title>Full coverage</title>
      <p>It is possible to cover compile the modules being tested
        before running the test suites. Doing so displays which branches
        of the code that are tested by the test suite, and which are
        not. Many use this as a measure of a good test suite. When every
        single line of source code is covered once by the test suite,
        the test suite is finished.
        </p>
      <p>A coverage of 100% still proves nothing, though. It doesn't
        mean that the code is error free, that everything is tested. For
        instance, if a function contains a division, it has to be
        executed at least twice. Once with parameters that cause
        division by zero, and once with other parameters.
        </p>
      <p>High degree of coverage is good of course, it means that no
        major parts of the code has been left untested. It's another
        question whether it's cost effective. You're only likely to find
        50% more bugs when going from 67% to 100% coverage, but the work
        (cost) is maybe 200% as large, or more, because reaching all of
        those obscure branches is usually complicated.
        </p>
      <p>Again, the reason for testing with the test server is to find
        bugs, not to create certificates of valid code. Maximizing the
        number of found bugs per hour probably means not going for 100%
        coverage. For some module the optimum may be 70%, for some other
        maybe 250%. 100% shouldn't be a goal in itself.</p>
    </section>

    <section>
      <title>User interface testing</title>
      <p>It is very difficult to do sensible testing of user
        interfaces, especially the graphic ones. The test server has
        some support for capturing the text I/O that goes to the user,
        but none for graphics. There are several tools on the market
        that help with this.</p>
    </section>
  </section>
</chapter>