aboutsummaryrefslogtreecommitdiffstats
path: root/contrib/tools/cython/Cython/Compiler/Pipeline.py
blob: 91196417da8747d581b60f371ee7659f60ca4138 (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
from __future__ import absolute_import 
 
import itertools 
from time import time 
 
from . import Errors 
from . import DebugFlags 
from . import Options 
from .Errors import CompileError, InternalError, AbortError 
from . import Naming 
 
# 
# Really small pipeline stages 
# 
def dumptree(t): 
    # For quick debugging in pipelines 
    print(t.dump())
    return t 
 
def abort_on_errors(node): 
    # Stop the pipeline if there are any errors. 
    if Errors.num_errors != 0: 
        raise AbortError("pipeline break") 
    return node 
 
def parse_stage_factory(context): 
    def parse(compsrc): 
        source_desc = compsrc.source_desc 
        full_module_name = compsrc.full_module_name 
        initial_pos = (source_desc, 1, 0) 
        saved_cimport_from_pyx, Options.cimport_from_pyx = Options.cimport_from_pyx, False 
        scope = context.find_module(full_module_name, pos = initial_pos, need_pxd = 0)
        Options.cimport_from_pyx = saved_cimport_from_pyx 
        tree = context.parse(source_desc, scope, pxd = 0, full_module_name = full_module_name) 
        tree.compilation_source = compsrc 
        tree.scope = scope 
        tree.is_pxd = False 
        return tree 
    return parse 
 
def parse_pxd_stage_factory(context, scope, module_name): 
    def parse(source_desc): 
        tree = context.parse(source_desc, scope, pxd=True, 
                             full_module_name=module_name) 
        tree.scope = scope 
        tree.is_pxd = True 
        return tree 
    return parse 
 
def generate_pyx_code_stage_factory(options, result): 
    def generate_pyx_code_stage(module_node): 
        module_node.process_implementation(options, result) 
        result.compilation_source = module_node.compilation_source 
        return result 
    return generate_pyx_code_stage 
 

def inject_pxd_code_stage_factory(context): 
    def inject_pxd_code_stage(module_node): 
        for name, (statlistnode, scope) in context.pxds.items():
            module_node.merge_in(statlistnode, scope) 
        return module_node 
    return inject_pxd_code_stage 
 

def use_utility_code_definitions(scope, target, seen=None): 
    if seen is None: 
        seen = set() 
 
    for entry in scope.entries.values():
        if entry in seen: 
            continue 
 
        seen.add(entry) 
        if entry.used and entry.utility_code_definition: 
            target.use_utility_code(entry.utility_code_definition) 
            for required_utility in entry.utility_code_definition.requires: 
                target.use_utility_code(required_utility) 
        elif entry.as_module: 
            use_utility_code_definitions(entry.as_module, target, seen) 
 

def sort_utility_codes(utilcodes):
    ranks = {}
    def get_rank(utilcode):
        if utilcode not in ranks:
            ranks[utilcode] = 0  # prevent infinite recursion on circular dependencies
            original_order = len(ranks)
            ranks[utilcode] = 1 + min([get_rank(dep) for dep in utilcode.requires or ()] or [-1]) + original_order * 1e-8
        return ranks[utilcode]
    for utilcode in utilcodes:
        get_rank(utilcode)
    return [utilcode for utilcode, _ in sorted(ranks.items(), key=lambda kv: kv[1])]


def normalize_deps(utilcodes):
    deps = {}
    for utilcode in utilcodes:
        deps[utilcode] = utilcode

    def unify_dep(dep):
        if dep in deps:
            return deps[dep]
        else:
            deps[dep] = dep
            return dep

    for utilcode in utilcodes:
        utilcode.requires = [unify_dep(dep) for dep in utilcode.requires or ()]


def inject_utility_code_stage_factory(context): 
    def inject_utility_code_stage(module_node): 
        module_node.prepare_utility_code()
        use_utility_code_definitions(context.cython_scope, module_node.scope) 
        module_node.scope.utility_code_list = sort_utility_codes(module_node.scope.utility_code_list)
        normalize_deps(module_node.scope.utility_code_list)
        added = [] 
        # Note: the list might be extended inside the loop (if some utility code 
        # pulls in other utility code, explicitly or implicitly) 
        for utilcode in module_node.scope.utility_code_list: 
            if utilcode in added:
                continue
            added.append(utilcode) 
            if utilcode.requires: 
                for dep in utilcode.requires: 
                    if dep not in added and dep not in module_node.scope.utility_code_list:
                        module_node.scope.utility_code_list.append(dep) 
            tree = utilcode.get_tree(cython_scope=context.cython_scope)
            if tree: 
                module_node.merge_in(tree.body, tree.scope, merge_scope=True) 
        return module_node 
    return inject_utility_code_stage 
 
 
# 
# Pipeline factories 
# 
 
def create_pipeline(context, mode, exclude_classes=()): 
    assert mode in ('pyx', 'py', 'pxd') 
    from .Visitor import PrintTree 
    from .ParseTreeTransforms import WithTransform, NormalizeTree, PostParse, PxdPostParse 
    from .ParseTreeTransforms import ForwardDeclareTypes, InjectGilHandling, AnalyseDeclarationsTransform
    from .ParseTreeTransforms import AnalyseExpressionsTransform, FindInvalidUseOfFusedTypes 
    from .ParseTreeTransforms import CreateClosureClasses, MarkClosureVisitor, DecoratorTransform 
    from .ParseTreeTransforms import TrackNumpyAttributes, InterpretCompilerDirectives, TransformBuiltinMethods
    from .ParseTreeTransforms import ExpandInplaceOperators, ParallelRangeTransform 
    from .ParseTreeTransforms import CalculateQualifiedNamesTransform 
    from .TypeInference import MarkParallelAssignments, MarkOverflowingArithmetic 
    from .ParseTreeTransforms import AdjustDefByDirectives, AlignFunctionDefinitions 
    from .ParseTreeTransforms import RemoveUnreachableCode, GilCheck 
    from .FlowControl import ControlFlowAnalysis 
    from .AnalysedTreeTransforms import AutoTestDictTransform 
    from .AutoDocTransforms import EmbedSignature 
    from .Optimize import FlattenInListTransform, SwitchTransform, IterationTransform 
    from .Optimize import EarlyReplaceBuiltinCalls, OptimizeBuiltinCalls 
    from .Optimize import InlineDefNodeCalls 
    from .Optimize import ConstantFolding, FinalOptimizePhase 
    from .Optimize import DropRefcountingTransform 
    from .Optimize import ConsolidateOverflowCheck 
    from .Buffer import IntroduceBufferAuxiliaryVars 
    from .ModuleNode import check_c_declarations, check_c_declarations_pxd 
 
 
    if mode == 'pxd': 
        _check_c_declarations = check_c_declarations_pxd 
        _specific_post_parse = PxdPostParse(context) 
    else: 
        _check_c_declarations = check_c_declarations 
        _specific_post_parse = None 
 
    if mode == 'py': 
        _align_function_definitions = AlignFunctionDefinitions(context) 
    else: 
        _align_function_definitions = None 
 
    # NOTE: This is the "common" parts of the pipeline, which is also 
    # code in pxd files. So it will be run multiple times in a 
    # compilation stage. 
    stages = [ 
        NormalizeTree(context), 
        PostParse(context), 
        _specific_post_parse, 
        TrackNumpyAttributes(),
        InterpretCompilerDirectives(context, context.compiler_directives), 
        ParallelRangeTransform(context), 
        AdjustDefByDirectives(context), 
        WithTransform(context),
        MarkClosureVisitor(context), 
        _align_function_definitions, 
        RemoveUnreachableCode(context), 
        ConstantFolding(), 
        FlattenInListTransform(), 
        DecoratorTransform(context), 
        ForwardDeclareTypes(context), 
        InjectGilHandling(),
        AnalyseDeclarationsTransform(context), 
        AutoTestDictTransform(context), 
        EmbedSignature(context), 
        EarlyReplaceBuiltinCalls(context),  ## Necessary? 
        TransformBuiltinMethods(context),
        MarkParallelAssignments(context), 
        ControlFlowAnalysis(context), 
        RemoveUnreachableCode(context), 
        # MarkParallelAssignments(context), 
        MarkOverflowingArithmetic(context), 
        IntroduceBufferAuxiliaryVars(context), 
        _check_c_declarations, 
        InlineDefNodeCalls(context), 
        AnalyseExpressionsTransform(context), 
        FindInvalidUseOfFusedTypes(context), 
        ExpandInplaceOperators(context), 
        IterationTransform(context),
        SwitchTransform(context),
        OptimizeBuiltinCalls(context),  ## Necessary? 
        CreateClosureClasses(context),  ## After all lookups and type inference 
        CalculateQualifiedNamesTransform(context), 
        ConsolidateOverflowCheck(context), 
        DropRefcountingTransform(), 
        FinalOptimizePhase(context), 
        GilCheck(), 
        ] 
    filtered_stages = [] 
    for s in stages: 
        if s.__class__ not in exclude_classes: 
            filtered_stages.append(s) 
    return filtered_stages 
 
def create_pyx_pipeline(context, options, result, py=False, exclude_classes=()): 
    if py: 
        mode = 'py' 
    else: 
        mode = 'pyx' 
    test_support = [] 
    if options.evaluate_tree_assertions: 
        from ..TestUtils import TreeAssertVisitor 
        test_support.append(TreeAssertVisitor()) 
 
    if options.gdb_debug: 
        from ..Debugger import DebugWriter # requires Py2.5+ 
        from .ParseTreeTransforms import DebugTransform 
        context.gdb_debug_outputwriter = DebugWriter.CythonDebugWriter( 
            options.output_dir) 
        debug_transform = [DebugTransform(context, options, result)] 
    else: 
        debug_transform = [] 
 
    return list(itertools.chain( 
        [parse_stage_factory(context)], 
        create_pipeline(context, mode, exclude_classes=exclude_classes), 
        test_support, 
        [inject_pxd_code_stage_factory(context), 
         inject_utility_code_stage_factory(context), 
         abort_on_errors], 
        debug_transform, 
        [generate_pyx_code_stage_factory(options, result)])) 
 
def create_pxd_pipeline(context, scope, module_name): 
    from .CodeGeneration import ExtractPxdCode 
 
    # The pxd pipeline ends up with a CCodeWriter containing the 
    # code of the pxd, as well as a pxd scope. 
    return [ 
        parse_pxd_stage_factory(context, scope, module_name) 
        ] + create_pipeline(context, 'pxd') + [ 
        ExtractPxdCode() 
        ] 
 
def create_py_pipeline(context, options, result): 
    return create_pyx_pipeline(context, options, result, py=True) 
 
def create_pyx_as_pxd_pipeline(context, result): 
    from .ParseTreeTransforms import AlignFunctionDefinitions, \ 
        MarkClosureVisitor, WithTransform, AnalyseDeclarationsTransform 
    from .Optimize import ConstantFolding, FlattenInListTransform 
    from .Nodes import StatListNode 
    pipeline = [] 
    pyx_pipeline = create_pyx_pipeline(context, context.options, result, 
                                       exclude_classes=[ 
                                           AlignFunctionDefinitions, 
                                           MarkClosureVisitor, 
                                           ConstantFolding, 
                                           FlattenInListTransform, 
                                           WithTransform 
                                           ]) 
    for stage in pyx_pipeline: 
        pipeline.append(stage) 
        if isinstance(stage, AnalyseDeclarationsTransform): 
            # This is the last stage we need. 
            break 
    def fake_pxd(root): 
        for entry in root.scope.entries.values(): 
            if not entry.in_cinclude: 
                entry.defined_in_pxd = 1 
                if entry.name == entry.cname and entry.visibility != 'extern': 
                    # Always mangle non-extern cimported entries. 
                    entry.cname = entry.scope.mangle(Naming.func_prefix, entry.name) 
        return StatListNode(root.pos, stats=[]), root.scope 
    pipeline.append(fake_pxd) 
    return pipeline 
 
def insert_into_pipeline(pipeline, transform, before=None, after=None): 
    """ 
    Insert a new transform into the pipeline after or before an instance of 
    the given class. e.g. 
 
        pipeline = insert_into_pipeline(pipeline, transform, 
                                        after=AnalyseDeclarationsTransform) 
    """ 
    assert before or after 
 
    cls = before or after 
    for i, t in enumerate(pipeline): 
        if isinstance(t, cls): 
            break 
 
    if after: 
        i += 1 
 
    return pipeline[:i] + [transform] + pipeline[i:] 
 
# 
# Running a pipeline 
# 
 
_pipeline_entry_points = {}


def run_pipeline(pipeline, source, printtree=True): 
    from .Visitor import PrintTree 
    exec_ns = globals().copy() if DebugFlags.debug_verbose_pipeline else None
 
    def run(phase, data):
        return phase(data)

    error = None 
    data = source 
    try: 
        try: 
            for phase in pipeline: 
                if phase is not None: 
                    if not printtree and isinstance(phase, PrintTree):
                        continue
                    if DebugFlags.debug_verbose_pipeline: 
                        t = time() 
                        print("Entering pipeline phase %r" % phase)
                        # create a new wrapper for each step to show the name in profiles
                        phase_name = getattr(phase, '__name__', type(phase).__name__)
                        try:
                            run = _pipeline_entry_points[phase_name]
                        except KeyError:
                            exec("def %s(phase, data): return phase(data)" % phase_name, exec_ns)
                            run = _pipeline_entry_points[phase_name] = exec_ns[phase_name]
                    data = run(phase, data)
                    if DebugFlags.debug_verbose_pipeline: 
                        print("    %.3f seconds" % (time() - t))
        except CompileError as err:
            # err is set 
            Errors.report_error(err, use_stack=False)
            error = err 
    except InternalError as err:
        # Only raise if there was not an earlier error 
        if Errors.num_errors == 0: 
            raise 
        error = err 
    except AbortError as err:
        error = err 
    return (error, data)