1# Copyright (C) 2004, 2005, 2006 Nathaniel Smith
2# Copyright (C) 2006, 2007 Holger Hans Peter Freyther
3#
4# Redistribution and use in source and binary forms, with or without
5# modification, are permitted provided that the following conditions
6# are met:
7#
8# 1.  Redistributions of source code must retain the above copyright
9#     notice, this list of conditions and the following disclaimer.
10# 2.  Redistributions in binary form must reproduce the above copyright
11#     notice, this list of conditions and the following disclaimer in the
12#     documentation and/or other materials provided with the distribution.
13# 3.  Neither the name of Apple Computer, Inc. ("Apple") nor the names of
14#     its contributors may be used to endorse or promote products derived
15#     from this software without specific prior written permission.
16#
17# THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY
18# EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
19# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
20# DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY
21# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
22# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
23# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
24# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26# THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27
28import csv
29import time
30import os.path
31import shutil
32
33def analyze_coverage(possible_gcov_files, source_files, runid, data_dir, base):
34
35    if not os.path.exists(data_dir):
36        os.makedirs(data_dir)
37
38    output = open(os.path.join(data_dir, runid + ".csv"), "w")
39    w = csv.writer(output)
40    # First row: id and time
41    w.writerow([runid, time.time()])
42
43    results = scan_gcov_files(possible_gcov_files, source_files)
44    annotated_dir = os.path.join(data_dir, runid + ".annotated")
45    if os.path.exists(annotated_dir):
46        shutil.rmtree(annotated_dir)
47
48    keys = results.keys()
49    keys.sort()
50    for path in keys:
51        (total, covered, annotated_data) = results[path]
52        path = path[path.find(base)+len(base):]
53        # Rest of the rows: filename, total_lines, covered_lines
54        w.writerow([path, total, covered])
55
56        if path[:1] == "/":
57            path = path[1:]
58        annotated_path = os.path.join(annotated_dir, path)
59        try:
60            os.makedirs(os.path.dirname(annotated_path))
61        except OSError:
62            pass
63        a = open(annotated_path, "w")
64        a.write(annotated_data)
65        a.close()
66
67
68# zecke's rewrite
69STATE_NOT_CODE = -1
70STATE_NOT_SEEN = -2
71STATE_TEST_CODE = -3
72
73def find_gcov(f, possible_gcovs):
74    """
75    Find .gcov files that could be of interest for us
76    """
77    try:
78        return possible_gcovs[f]
79    except:
80        return []
81
82
83def parse_source_file(file):
84    """
85    Parse one source file and return a list of lines
86    """
87    f_source_list = []
88    init_state = STATE_NOT_SEEN
89    in_test_code = False
90    nesting = 0
91
92    for line in open(file, "r"):
93        code = line.split(":", 2)[-1]
94        if not in_test_code and code.startswith("#ifdef BUILD_UNIT_TESTS"):
95            in_test_code = 1
96        if in_test_code and code.startswith("#if"):
97            nesting += 1
98        if in_test_code and code.startswith("#endif"):
99            nesting -= 1
100            if not nesting:
101                in_test_code = True
102        if in_test_code:
103            init_state = STATE_TEST_CODE
104        else:
105            init_state = STATE_NOT_SEEN
106        f_source_list.append([init_state, line.split(":", 1)[1]])
107
108    return f_source_list
109
110# Runner-up, 3rd annual "write Python that looks like Perl" competition,
111# Well, not really.  It doesn't even use regexps.
112# He is right so I'm cleaning it up (zecke)
113def scan_gcov_files(possible_gcov_files, source_files):
114    """Takes a list of gcov filenames and a list of source filenames.
115
116    The gcov files should have names of the form foo.o##foo.cc.gcov, as
117    created by 'gcov -l'.
118
119    Returns a dict mapping source filenames to tuples
120      (total_lines, tested_lines, gcov_annotated_source)
121    which are a number, a number, and a very long string, respectively.
122
123    The fun bit is that we merge .gcov output generated by different object
124    files; this way we can provide accurate information for header files and
125    for monotone's current unit test system."""
126    results = {}
127    for f in source_files:
128        possible_gcovs = find_gcov(f, possible_gcov_files)
129        base_name = os.path.splitext(os.path.basename(f))[0]
130        if len(possible_gcovs) == 0:
131            print "No gcov files found for: '%s' but it was compiled" % f
132            continue
133
134            (garbage,extension) = os.path.splitext(f)
135            if extension  in [".cc", ".c", ".moc", ".cpp", ".cxx", ".m", ".mm"]:
136                lines = open(f, "r").readlines()
137                results[f] = (len(lines), 0, "".join(lines))
138            continue
139        elif len(possible_gcovs) > 1:
140            print "More than one gcov file for %s %d" % (f,len(possible_gcovs))
141        base_gcov_lines = parse_source_file(possible_gcovs[0])
142
143        # Now we will try hard to merge the results with others
144        # Our requirement is that we have the same amount of lines as
145        # as the original file
146        for cov_file in possible_gcovs:
147            lines = open(cov_file, "r").readlines()
148
149            # e.g. with phonon we have visualisation.h and we can not know
150            # which header file (foldername) it is refering to. This is a gcov
151            # limitation and i have no workaround yet. We just hope we will pick
152            # the right header file...
153            if len(lines) != len(base_gcov_lines):
154                print "Error Base: %s and Target: %s have different amount of lines" % (possible_gcovs[0],cov_file)
155                continue
156
157            # now do the merging of the file. If it has the same basename
158            # and the same number of lines things might work out
159            # In the future take a look at the header of the file
160            i = 0
161            for line in lines:
162                accumulator = base_gcov_lines[i]
163                if accumulator[0] != STATE_TEST_CODE:
164                    info = line.split(":", 1)[0]
165                    if info.endswith("-"):
166                        if accumulator[0] == STATE_NOT_SEEN:
167                            accumulator[0] = STATE_NOT_CODE
168                    else:
169                        if info.endswith("#"):
170                            num = 0
171                        else:
172                            num = int(info)
173                        if accumulator[0] in (STATE_NOT_SEEN, STATE_NOT_CODE):
174                            accumulator[0] = 0
175                        accumulator[0] += num
176                i += 1
177
178        # post processing of ths file
179        (total_lines, total_covered) = (0, 0)
180        annotated_lines = []
181        for state, line in base_gcov_lines:
182            if state == STATE_NOT_SEEN:
183                desc = "?????"
184            elif state == STATE_TEST_CODE:
185                desc = "+"
186            elif state == STATE_NOT_CODE:
187                desc = "-"
188            elif state == 0:
189                desc = "#####"
190                total_lines += 1
191            else:
192                desc = str(state)
193                total_lines += 1
194                total_covered += 1
195            annotated_lines.append(":".join([desc.rjust(9), line]))
196        results[f] = (total_lines, total_covered, "".join(annotated_lines))
197    return results
198
199
200
201    return results
202