control dependences, recurrences, parallelism, control height reduction, back-substitution, blocked back-substitution, software pipeline, loop optimization The performance of applications executing on processors with instruction level parallelism is often limited by control and data dependences. Performance bottlenecks caused by dependences can frequently be eliminated through transformations which reduce the height of critical paths through the program. While height reduction techniques are not always helpful, their utility can be demonstrated in an increasingly broad range of important situations. This report focuses on the height reduction of control recurrences within loops with data dependent exits. Loops with data dependent exits are ...
To exploit parallelism in Fortran code, this dissertation consists of a study of the following three...
Many loop nests in scientific codes contain a parallelizable outer loop but have an inner loop for w...
The trend in high-performance microprocessor design is toward increasing computational power on the ...
Reduction recognition and optimization are crucial techniques in parallelizing compilers. They are u...
Developing efficient programs for many of the current parallel computers is not easy due to the arch...
Architectures with instruction level parallelism such as VLIW and superscalar processors provide par...
The challenge of exploiting high degrees of instruction-level parallelism is often hampered by frequ...
Modern computers will increasingly rely on parallelism to achieve high computation rates. Techniques...
T here is an insatiable demand for computers ofever-increasing performance. Old applicationsare appl...
Control divergence poses many problems in parallelizing loops. While predicated execution is commonl...
Over the past 20 years, increases in processor speed have dramatically outstripped performance incre...
Journal PaperCurrent microprocessors incorporate techniques to exploit instruction-level parallelism...
Parallel programming involves finding the potential parallelism in an application, choos-ing an algo...
Advanced computer architectures rely mainly on compiler optimizations for parallelization, vectoriza...
Thesis (Ph. D.)--University of Rochester. Dept. of Computer Science, 2012.Speculative parallelizatio...
To exploit parallelism in Fortran code, this dissertation consists of a study of the following three...
Many loop nests in scientific codes contain a parallelizable outer loop but have an inner loop for w...
The trend in high-performance microprocessor design is toward increasing computational power on the ...
Reduction recognition and optimization are crucial techniques in parallelizing compilers. They are u...
Developing efficient programs for many of the current parallel computers is not easy due to the arch...
Architectures with instruction level parallelism such as VLIW and superscalar processors provide par...
The challenge of exploiting high degrees of instruction-level parallelism is often hampered by frequ...
Modern computers will increasingly rely on parallelism to achieve high computation rates. Techniques...
T here is an insatiable demand for computers ofever-increasing performance. Old applicationsare appl...
Control divergence poses many problems in parallelizing loops. While predicated execution is commonl...
Over the past 20 years, increases in processor speed have dramatically outstripped performance incre...
Journal PaperCurrent microprocessors incorporate techniques to exploit instruction-level parallelism...
Parallel programming involves finding the potential parallelism in an application, choos-ing an algo...
Advanced computer architectures rely mainly on compiler optimizations for parallelization, vectoriza...
Thesis (Ph. D.)--University of Rochester. Dept. of Computer Science, 2012.Speculative parallelizatio...
To exploit parallelism in Fortran code, this dissertation consists of a study of the following three...
Many loop nests in scientific codes contain a parallelizable outer loop but have an inner loop for w...
The trend in high-performance microprocessor design is toward increasing computational power on the ...