ALPyNA: Acceleration of Loops in Python for Novel Architectures
We present ALPyNA, an automatic loop parallelization framework for Python, which analyzes data dependences within nested loops and dynamically generates CUDA kernels for GPU execution. The ALPyNA system applies classical dependence analysis techniques to discover and exploit potential parallelism. The skeletal structure of the dependence graph is determined statically; this is combined with type and bounds information discovered at runtime, to auto-generate high-performance kernels for offload to GPU. We demonstrate speedups of up to 1000x relative to the native CPython interpreter across four array-intensive numerical Python benchmarks. Performance improvement is related to iteration domain sizes and the complexity of the dependence graph. Nevertheless, this approach promises to bring the benefits of manycore parallelism to end-user developers.
Sat 22 JunDisplayed time zone: Tijuana, Baja California change
16:00 - 17:30
|ALPyNA: Acceleration of Loops in Python for Novel Architectures|
|Code Generation in Linnea (extended abstract)|
|High-Level Synthesis of Functional Patterns with Lift|